DiGS (Version 3.1) Setup Guide

Size: px
Start display at page:

Download "DiGS (Version 3.1) Setup Guide"

Transcription

1 DiGS (Version 3.1) Setup Guide George Beckett, Daragh Byrne, Radosław Ostrowski, James Perry, Eilidh Grant 10th May

2 Contents Preface 3 1 An Overview of DiGS A typical DiGS-powered infrastructure Useful terms to know A note on the history of DiGS Overview of Setup Process Prerequisites Globus Toolkit X.509 Digital Certificates Other software dependencies The DiGS Package Account Running on 32-bit and 64-bit platforms Types of installation User Management and Access Control Policy The Ant-based build system The Setup Process, Details of Client Installation Installing the Software (pre-compiled binaries) Installing the Software (building from source) Installing the Software (building from source from CVS) Creating the configuration files Configuration of User Environment GridFTP-based Storage Element Setup Creating a DiGS Package Account Authorising Control Node to access Storage Element Setting up User Access to Storage Element Setting up Storage Directories SRM Storage Element Setup Control Node Installation Creating a DiGS Package Account Installing the software Creating the Control Node Configuration Files Setting up X.509 certificate for Control Thread Setting up User Access to the Control Node Starting the Control Thread The Metadata Catalogue Installing the software Setting up the Servlet container Deploying the XML database exist Database Administration

3 3.6 Configuring a Backup Node Installing the software Mirroring the File Catalogue Starting the DiGS Backup Thread

4 Preface This document describes how to set up a computer to become part of a DiGS 1 -powered data grid. It lists the steps necessary to install and configure the DiGS software itself, identifies dependencies and prerequisites, and explains how to solve common problems that may be experienced during setup. Finally, it describes how to test the software installation in order to verify that it is correct and complete. Note that this document applies to DiGS Version DiGS is an abbreviation for Distributed Grid Storage. DiGS was formerly known as QCDgrid 4

5 Chapter 1 An Overview of DiGS DiGS is a a data grid application a distributed system that supports the management, sharing, publication and preservation of collections of data. It is specifically designed to handle large volumes of scientific data on dispersed, online (that is, disk-based) facilities, and furnishes semantic data management through the binding of application-specific metadata to datafiles. The key features of DiGS are: multi-terabyte, distributed storage. automatic and transparent replication of data across multiple sites. continuous data validation and consistency checking. support for bulk data transport operations. data provenance with application-specific metadata. simple and intuitive client tools. interface between managed storage and external compute resources. From the perspective of the user, a DiGS system looks like a local file store containing datasets (and possibly associated metadata); although in reality it is split across multiple, distributed machines that may be geographically dispersed. To hide this complexity, each dataset (or file) in the system is attributed with a unique and persistent identified called a Logical Filename (LFN). DiGS can be used to store arbitrary, file-based data. However, its development is motivatived by datadriven computing, focusing in particular on semantic-based data access. With this in mind, DiGS organises datasets into collections called experiments. Each experiment has an associated, scientific commentary captured in XML 1 which describes the experiment as a whole. An experiment consists of one or more datasets (or files), which are individually described by separate XML documents that capture the scientific meaning of the particular datasets. There is a one-to-one correspondence between datasets and the metadata documents that describe them. This connection is captured in a (configurable) text node of the XML description that contains the Logical Filename of the corresponding dataset. There is a one-to-many relationship between an experiment and the datasets it contains. This connection is captured by a text node in the XML description of the datafile, which contains the name of the experiment. This name is matched to a text node in the experiment description, which also contains the name. 1 An open standard for describing data, provided by the W3C. 1

6 1.1 A typical DiGS-powered infrastructure The component systems within a typical DiGS-powered infrustructure are illustrated in Figure 1.1. The system contains three types of node: Client, Storage Element, and Control/Backup Node, as described below. Figure 1.1: Overview of the typical DiGS architecture showing flow of data and control between clients, storage elements, control node and the various catalogue services. Client A user accesses the data grid from a Client. This can be any computer on which: the DiGS client tools are installed, plus the user has a user account and appropriate credentials. From a Client, a user can: Confirm that the data grid is operational. List and retrieve files from the data grid. Query the Metadata Catalogue for scientific metadata. Submit new data/metadata to the data grid. Submit a compute job to a remote compute resource. All of these interactions are supported by a suite of command line tools called the DiGS Command Line Interface (CLI). In addition, a subset of the above functionality is furnished by a graphical client called the DiGS Browser. More information about the client tools is provided in the DiGS User Guide [1]. 2

7 Storage Element A Storage Element provides disk-based storage space to the data grid for holding copies of user data files. The data files held on a storage element are assumed to be immediately available (a property often referred to as on-line when compared to tape-based storage space which is referred to as off-line). User file transfers to and from Storage Elements are normally initiated from a Client. Therefore, a user need never access the Storage Element directly. Other file transfers are also implemented by the Control Node as part of its data replication and validation functionality. Control Node The Control Node hosts a persistent agent (commonly referred to as the Control Thread) that continuously tests and validates the integrity and availability of the contents of the data grid. It performs a number of checks, including: checks that Storage Elements are responding. checks if new data is awaiting insertion into the data grid. checks that sufficient copies of each file exist on the data grid, and performs a replication operation if a file is detected with insufficient copies. compares the state of a Storage Element s file systems with the record that it maintains locally in a file catalogue. checks that each copy of a file is consistent with its peers that is, all replicas of a file are identical. computes the free capacity that remains in the data grid and takes appropriate action if the capacity is deemed to be too low. The Control Node is autonomous and typically orchestrates management of the data grid without intervention from a user/administrator. The Control Node is a critical component of the DiGS system architecture. If the Control Node becomes unavailable, then the system switches to a backup control node (called the Backup Node) that provides read-only access to the data grid. Backup Node The Backup Node provides a subset of the functionality of the Control Node, permitting users to retrieve data from the grid, in the event of a failure of the Control Node. Specifically, the Backup Node hosts a copy of the file catalogue allowing a Client to locate and retrieve data from the Storage Elements. The Backup Node is intended as a temporary substitute for the Control Node: it cannot add new data to the grid, nor can it replicate or validate existing data on the grid. 1.2 Useful terms to know The terms, below, are commonly used to describe the functions of a DiGS-powered data grid. A familiarity with these terms is important, since they are used widely within this document. Data Grid is a term used to describe a distributed file system (spanning multiple institutional facilities) that supports the management, sharing, publication and preservation of collections of files. Logical Filename (LFN) is a unique and persistent identifier for a dataset that is stored on a DiGSpowered data grid. 3

8 File Catalogue is a specialised database that maintains the mapping between each LFN and the physical locations of the actual copies of the data file. The DiGS File Catalogue also maintains basic filebased metadata, recording the size, the MD5 checksum, the owner, and the access permissions associated with each dataset. Metadata Catalogue is a database that stores scientific metadata pertaining to the experiments and datasets stored on the data grid. The DiGS Metadata Catalogue is XML-based. 1.3 A note on the history of DiGS The development of DiGS has been motivated by data management requirements in the computational particle physics community. The first version of DiGS (that is Version 1) was actually called QCDgrid a name that reflects the background of the application. One aim in the name change and the more recent development path has been to generalise the application space of the software. However, to maintain some backward compatiblity and to reduce the scope for introducing new bugs, the name QCDgrid is still present in many of the configuration files and underlying technologies. Subsequent sub-versions of DiGS will continue to reduce the number of references to the original name. 4

9 Chapter 2 Overview of Setup Process 2.1 Prerequisites Before one can begin to install the DiGS software onto a system, one needs to confirm that various prerequisites are satisfied. These prerequisites are described, in turn, below Globus Toolkit DiGS is a grid application that utilises services and functionality from a middleware called the Globus Toolkit. Each system on a DiGS-powered data grid requires this toolkit to be installed. DiGS Version 3.1 requires Globus Version 2.4 or higher. However, Version 4.0 or higher is recommended, since this includes significant bug fixes and security enhancements. Please refer to [2] for detailed information on how to set up a Globus Toolkit installation for use with DiGS X.509 Digital Certificates Each server (that is, Storage Element and Control Node), on a DiGS-powered data grid, needs to have an X.509 host certificate (also known as a server certificate) from a suitable certification authority (CA), such as the UK e-science CA ( Each user of the system will also need an X.509 user certificate to allow them to authenticate to, and be attributed with privileges on, the data grid Other software dependencies There are several other software dependencies that must be fulfilled before one installs DiGS software, as follows: DiGS client tools require access to a Java Virtual Machine. The developers recommend Sun Java Version 1.4 or higher. In addition, if one intends to build DiGS software from source, a Java Development Kit is required for example, Sun Java SDK Version 1.4. If one intends to build DiGS software from source, then the Ant build system (Version 1.6 or higher) is required. In addition to Ant, the build script requires access to the GNU Make utility (Version 3.8 or higher) for building various C-based components. The DiGS Metadata Catalogue requires access to an XMLDB-compliant database. As distributed, the DiGS software is set up to use the exist database [5] (Version 1.0 or higher). 5

10 If a VOMS proxy is required in order to access any of the storage elements on the grid, the glite VOMS client tools should be installed to enable users to run the voms-proxy-init command. If the software is to be built from source and SRM [3] support is required, the gsoap toolkit and the gsoap GSI plugin should first be installed The DiGS Package Account For a multi-user system, it is strongly recommended that the software is installed and configured using a DiGS Package Account a Unix account created specifically to manage the DiGS software installation. DiGS imposes no restrictions on the name of the account (over and above restrictions of the Operating System), though the developers recommend using the name digs 1. For a Storage Element, in addition to owning 2 the DiGS software, the package account also owns all data that is held on the Storage Element. For a single-user Client node, one may choose to install DiGS using the target user s account. However, this type of installation is not discussed further in this document Running on 32-bit and 64-bit platforms All versions of DiGS, since and including Version are compatible with both 64-bit and 32-bit platforms. 2.2 Types of installation To reflect the different types of node that are present on a DiGS-powered data grid, there are several different installation types, which are described in turn below: Client sets up command-line (and possibly graphical) client tools that allow a user to interact with a data grid for example to upload/download files and perform basic data manipulation. See Section 3.1 for more information about Client node installation. Storage Element DiGS software is not required to be installed on a storage element. However, some configuration is required. See Section 3.2 for more information on how to set up the hosting node as a storage element for the data grid. Control Node sets up the hosting node as a Control Node for the data grid. The Control Node installation includes the client toolset. See Section 3.4 for more information about Control Node installation. Backup Node sets up the hosting node as a Backup Node, which will automatically assume a limited Control Node role, should the Control Node become unavailable. See Section 3.6 for more information about Backup Node installation. For each of these installation types, one may choose to complete the set up process using a pre-compiled binary package (provided that one is available for the target architecture) or building the various tools and services from source. 1 Historically, the package account has been called qcdgrid. 2 Here ownership refers to Unix file system permissions. 6

11 2.2.1 User Management and Access Control Policy DiGS employs a group-based access control policy, intended to allow groups of collaborating users to share data, without making it accessible to users of the data grid as a whole (unless they choose so). Two types of group are defined by DiGS: privileged and guest. Each user must be a member of either a privileged group or a guest group, though (a restriction in Version 3.1 of DiGS is that) they may only be a member of one such group. Members of a privileged group may: access (usually view) data that belongs to the group; upload new datasets to the data grid (to be owned by the group); and view public data that belongs to other groups. Members of a guest group may only read public data. Each dataset, on the data grid, belongs to a privileged group and is classified as either: private data which may only be accessed by members of the corresponding privileged group. public data which may be read by any user of the data grid. The implementation of this access control mechanism is described in more detail in Chapter The Ant-based build system As of DiGS Version 2.0.1, the process of building DiGS tools and services from source has been greatly simplified, thanks to the use of an Ant-based build system. This Ant-based build system is encapsulated in an Ant build script that is included in the top-level directory of the DiGS source code bundle (or may be downloaded from the software homepage [6]). The build script defines the following targets: checkout Checks out source code from web-based repository. make-client Compiles client tools. make-client-from-cvs Checks out and compiles client tools. make-storagenode Compiles Storage Element tools and services (Version 2.x only). make-storagenode-from-cvs Checks out and compiles Storage Element tools and services (Version 2.x only). make-server Compiles Control/Backup Node tools and services. make-server-from-cvs Checks out and compiles Control/Backup Node tools and services. dist-source Generates a source distribution. dist-ildg-browser Generates a stand-alone DiGS Browser. dist-client Generates a client distribution. dist-storagenode Generates storage node distribution (Version 2.x only). dist-server Generates Control Node distribution. clean-dist Deletes distribution and browser directories. To invoke one of the build targets, run a command of the form: > ant <build target> However, before one can run the build script, one needs to check some system-specific properties defined within build.xml as follows: The make property should be set to the full path to the GNU Make utility on the target system. 7

12 In addition, if one plans to check out the DiGS software from the project CVS repository, one also needs to check several other properties: The cvs.username property should be set to the username with which one accesses the software CVS repository [6]. One is asked to enter their password during the build process, as required. The cvs.tag property should be set to the CVS tag name for the required version of DiGS. Typically, the tag name has the form DIGSx_y_z, where x, y, and z are the version, major issue, and minor issue number, respectively. For example, the tag name for DiGS Version is DIGS2_0_2. The build script is designed to perform all intermediate steps, as required, when one invokes one of the dist-* targets. For example, ant dist-client both compiles it and packages it up. The output from a build process can be substantial. One may save a copy of the output to file, for diagnostics, with a command of the following form: > ant <build target> tee build.log When building DiGS 3.1 including the SRM support, you may encounter compilation errors related to gsoap if the version of the gsoap toolkit on your system is not compatible with the one used to produce the automatically generated SOAP source files included with DiGS. This can be remedied by regenerating the files using your version of gsoap: > cd StorageElementInterface/src > wsdl2h -c -o srm2_2.h srm.v2.2.wsdl > soapcpp2 srm2_2.h 8

13 Chapter 3 The Setup Process, Details of The setup process for each of the installation types (client, storage element, and control/backup node) is described, in turn, in this section. The instructions below assume that one has created a package account (see Section 2.1.4) and that one runs the various shell commands as the package account user. The instructions below also assume that the Globus environment has been configured within the command shell. This is usually straightforward, though is dependent on the Globus installation approach that has been followed 1 : If Globus has been installed using Virtual Data Toolkit (VDT), then one needs to initialise the $VDT_LOCATION environment variable. For example: export VDT_LOCATION=/opt/vdt and then source the VDT setup script: source $VDT_LOCATION/setup.sh If Globus has been installed as a separate toolkit, then one needs to initialise the $GLOBUS_LOCATION environment variable. For example: export GLOBUS_LOCATION=/opt/globus and then source the Globus setup script: $GLOBUS_LOCATION/etc/globus-user-env.sh 3.1 Client Installation The client-only installation is the simplest to complete and is common to all installation types. It consists of the following steps: 1. Install the DiGS software (from either a pre-compiled binary package or from source code). 2. Create static configuration files that identify the particular characteristics of the target data grid. 3. Set up the user environment for correct operation of the client tools. These three steps are discussed, in turn, below. 1 The initialisation examples assume a Bourne-style shell, though an equivalent setup is available for the C-shell. 9

14 3.1.1 Installing the Software (pre-compiled binaries) Firstly, one should check that a suitable binary distribution is available for the target architecture by visiting the project website [6]. Next, one should decide on where to set up the DiGS binaries and configuration files the home directory of the DiGS Package Account (for example, /home/digs) is a suitable and commonly used location for a multi-user installation. Then, one should download the binary distribution (which is distributed as a compressed TAR ball) and unpack it into the chosen location. For example, as follows: > cd ~digs > tar -zxvf digs-3.1-linux-client.tar.gz Note that the above command sequence is assumed to be completed as the DiGS Package Account user (digs). By default, the bundle unpacks into a directory named digs-2.x.y, which we refer to as the DiGS installation directory from hereon. Note: For a multi-user installation, all of the files should be owned by the DiGS Package Account, should be world-readable and (for binaries) world-executable. One should ensure that an appropriate file mode creation mask (for example, UMASK environment) is configured before unpacking the TAR ball Installing the Software (building from source) Firstly, one should decide where to install the DiGS binaries and configuration files the home directory of the DiGS Package Account (for example, /home/digs) is a suitable and commonly used location for a multi-user installation. One can download a pre-packaged source code bundle from the project website [6], and then unpack it into the chosen location, as follows: > cd ~digs > tar -zxvf <path_to_bundle>/digs-<x.y.z>-source.tar.gz > cd digs-<x.y.z>-source At this point one should have a file called build.xml and a directory called digs-3.x.y. Having obtained a suitable source code bundle, one can run the command: > ant make-client tee build.log to build the client tools. At this point, one should have a directory, named digs-3.x.y, which contains the software installation Installing the Software (building from source from CVS) Firstly, one should decide where to install the DiGS binaries and configuration files the home directory of the DiGS Package Account (for example, /home/digs) is a suitable and commonly used location for a multi-user installation. 10

15 Next, one should download the latest version of the Ant build script build.xml (for example, from the project website [6]) to the target location and make any required configuration changes (see Section 2.2.2). Thirdly, one should obtain a source code bundle. To checkout the code one must have access to the project CVS repository (hosted at [6]), then one can check out a specific version of the source and build the client tools using the make-client-from-cvs build target: > ant make-client-from-cvs tee build.log At this point you should have two directories, named: fromcvs, which contains the source code; and digs-3.x.y, which contains the software installation Creating the configuration files Having built the source code (or unpacked the DiGS client binaries), there are a number of configuration files that need to be created: nodes.conf Configuration file that contains information to allow DiGS client tools to contact the Control Node and Backup Node. nodeprefs.conf Configuration file that enumerates Storage Elements that are available on the data grid, in order of preference. browser.properties Configuration file that provides information to allow DiGS client tools to locate the Metadata Catalogue. Specific details of the particular values that are required for a particular data grid should be obtained from the data grid administrator. What follows, below, is a description of the parameters that must, or can, be specified in each configuration file. Main configuration file The main configuration file nodes.conf, which should be created in the DiGS installation directory (for example, /home/digs/digs-3.1), contains the information that the client tools need in order to locate the Control Node (and Backup Node). Specifically, the file should contain the fully qualified domain name (FQDN) of the Control Node, the path to the DiGS software on that node, and the fully qualified domain name of the server hosting the Metadata Catalogue. It may also (optionally) contain the fully qualified domain name of the Backup Node, which is contacted if the Control Node is unreachable. The format for nodes.conf is, as follows: node = <FQDN of Control Node> path = <path to DiGS software on Control Node> mdc = <FQDN of Metadata Cataloge node, without port (assumed 8080)> backup_node = <FQDN of Backup Node> backup_path = <path to DiGS software on Backup Node> As noted above, the backup_node and backup_path entries are optional: all other entries are mandatory. If you are unsure of the location for the Control Node and Backup Node, then you should contact your data grid administrator. At the time of writing, the Control Node and Metadata Catalogue for UKQCD Grid are running on ukqcdcontrol.epcc.ed.ac.uk at University of Edinburgh, and the backup node is running on pytier2.swan.ac.uk at University of Swansea. The nodes.conf file is, as follows: 11

16 node = ukqcdcontrol.epcc.ed.ac.uk path = /home/qcdgrid/qcdgrid mdc = ukqcdcontrol.epcc.ed.ac.uk backup_node = pytier2.swan.ac.uk backup_path = /home/qcdgrid/qcdgrid For the Control Node (and Backup Node), the package account needs to have write access to this configuration file. For example, > chmod o+rw nodes.conf Storage Element preferences The Storage Element preferences file, nodeprefs.conf, lists the Storage Elements that are defined for the data grid, in order of preference (one per line). The file should be located in the DiGS Installation Directory (for example, /home/digs/digs-3.1), and those Storage Elements that are geographically closest should normally be at the top of the list for reasons of efficiency. For example: local_node.ed.ac.uk remote1.liv.ac.uk remote2.liv.ac.uk far-away_node.swan.ac.uk For an up-to-date list of Storage Elements for a particular data grid, one should contact the data grid administrator. Browser configuration The Metadata Catalogue properties file, browser.properties must also be created, and should be located in the java/ildg_browser subdirectory in the DiGS installation directory (for example, /home/digs/digs-3.1/java/ildg_browser). The file contains the authentication information for the database, as follows: xmldb.user.name=<username for Metadata Catalogue> xmldb.user.password=<password for Metadata Catalogue> This properties file is also used by the DiGS Browser, which is an application-specific client (see Section 3.5 for more information). Additional properties may be available for your particular data grid. For example, the UKQCD data grid has the following form: qcdgrid.browser.mode=<either UKQCD for local or ILDG for guest> queryrunner.maxresults=0 services.file.location=<location of ILDG Services definition> xmldb.user.name=<ukqcd or ILDG username> xmldb.user.password=<ukqcd or ILDG password> Configuration of User Environment In addition to the Globus environment requirements (see discussion at beginning of this chapter), a user needs to have certain environment variables configured in order that the DiGS software functions correctly. These environment variables are described, in turn, below. Since the environment needs to be initialised each time a shell is created, one might consider folding the command sequence into the normal shell initialisation process. The DiGS-specific requirements for a user s environment are, as follows: 12

17 The environment variable DIGS_HOME must be initialised to the DiGS installation directory. For example: export DIGS_HOME=/home/digs/digs-3.1 The DiGS executables must be in the user s path: export PATH=$PATH:$DIGS_HOME Note: Simply giving a full path when executing a DiGS commands in not sufficient for most commands. The DiGS client library (which, by default, is in DiGS installation directory) must be in the library path. For example: export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$DIGS_HOME The Java Runtime Environment must be in the path. This is usually the case by default, and can be checked by a successful invocation of: > java -version java version "1.5.0_09" Java(TM) 2 Runtime Environment, Standard Edition (build 1.5.0_09-b03) Java HotSpot(TM) Server VM (build 1.5.0_09-b03, mixed mode) The BROWSER_HOME environment variable must be assigned to the location of the DiGS Browser code this is usually $DIGS_HOME/java/ildg_browser. For example: export BROWSER_HOME=${DIGS_HOME}/java/ildg-browser The exist libraries 2 (otherwise known as JARs) must be in the class path. The exist libraries are distributed with both source and binary DiGS distributions in the exist_libraries sub-directory. For example: export EXIST_LIB_DIR=${DIGS_HOME}/exist_libraries for i in ${EXIST_LIB_DIR}/*.jar ; do export CLASSPATH=${i}:${CLASSPATH} done In addition to the above environment variables, the following additional entries may be required for specific circumstances: The assignment: export LD_ASSUME_KERNEL=2.2.5 is required on some Linux-based systems, due to a threading bug in early versions of the Globus Toolkit. Specifically, the assignment is necessary if using Globus Toolkit Version 2.4.x or Version 3.x, with Linux kernel Version or newer the bug does not affect older Linux kernels, and it has been fixed properly in Globus 4.0. On some operating systems, the Globus command globus-hostname does not return the fully qualified domain name of the system on which it is runnning (for example, digs-host.ed.ac.uk), as intended. One can run globus-hostname to check if this is the case. If the command returns either the first element of the hostname (continuing with our example, this would be digs-host) or localhost, then one needs to define the additional environment variable: 2 This assumes that exist is the underlying database for the Metadata Catalogue. 13

18 export GLOBUS_HOSTNAME=<fqdn.of.machine> The environment variable DIGS_GROUP may need to be set to the DiGS group to which the user belongs: export DIGS_GROUP=<users group> One only needs to do this, if instructed to by their data grid administrator. 3. Once the environment is set up as described above, the DiGS software is ready for use. One may wish to create a script to automate this environment setup, so that it can quickly be actioned in any particular shell instance. A template script, called example_setup.sh, is included in the distribution, which may be edited accordingly it may simply require the appropriate value of DIGS_HOME to be specified at beginning of the script. At this point, one may wish to run the DiGS command digs-test-setup to confirm that everything is in order. See the DiGS User Guide [1] for more information on this command. 3.2 GridFTP-based Storage Element Setup As of DiGS Version 3.0, the setup of a storage element has been simplified and it is no longer necessary to install DiGS software onto the resource. The instructions below should provide sufficient information for a competent Unix user to complete the process of setting up a GridFTP-based storage element. The process consists of the following steps: Create a DiGS package account this step, which is optional for a Client Node installation, is mandatory in this case. Map the Control Node s host certificate subject to the DiGS package account, so that the Control Thread process can interact with the Storage Element. Set up Unix accounts and groups reflecting the make-up of the user base. Set up a directory structure to be used for data storage. Note: Once a Storage Element bas been configured, one must contact the data grid administrator to have it added to the list of Storage Elements for the grid Creating a DiGS Package Account The use of a DiGS package account (see Section 2.1.4) is mandatory during the setup of a Storage Element. For security reasons, all datafiles stored on a DiGS Storage Element are owned by the DiGS package account. It is recommended that a new user account (for example, named digs) be created for this purpose, instead of using an existing user account or the root account Authorising Control Node to access Storage Element Having created a DiGS package account, one is ready to add the Control Node s host certificate to the authorisation list for over-grid-access to the DiGS package account on the Storage Element. This is achieved by adding an entry to the machine s grid-mapfile, usually: /etc/grid-security/grid-mapfile 3 For more information regarding the requirement for this environment variable, see the NeSCForge Problem Report #1374 [6] 14

19 a step that requires root permission 4. Specifically, one should add a line to the grid-mapfile, associating the Control Node s host certificate to the local DiGS Package Account. For example, as follows: "/C=UK/O=eScience/OU=Edinburgh/L=NeSC/CN=ukqdcontrol.epcc.ed.ac.uk/Em ail=<admin>@epcc.ed.ac.uk" digs This entry must be added as a single, continous line and not as shown above (where it has been wrapped for printing) Setting up User Access to Storage Element The group-based access control policy, which is described in Section 2.2.1, is implemented on each GridFTP-based Storage Element using Unix groups and accounts. On every Storage Element, each DiGS group must be instantiated as a Unix group and each DiGS user must be mapped to a Unix account that belongs to the corresponding Unix group. There does not need to be a one-to-one mapping between data grid users and Unix accounts. It is more typical to either: Create a single Unix account for each group, and map all users (in the corresponding DiGS group) to this Unix account. This approach works well provided that the Storage Element is not also used as a grid compute resource implying that a DiGS user may need to maintain time-limited though stateful data on the resource. Use a technology such as the EDG Pooled Accounts mechanism to create a set of generic, timeleased Unix accounts for each group, with each user being temporarily mapped to a free pooled account whenever they access the Storage Element resource. For more information about the EDG Pooled Accounts mechanism, check out the EDG VO files package, which is available (for example) from the project website [6]. Having set up the group structure on a storage element, one must ensure that the DiGS Package Account is added to each of the groups. This is to ensure that the DiGS Package Account can manage each groups data, as actioned by the Control Thread process. The mechanism used to manage the mapping user accounts to Unix equivalents is outside of the scope of the DiGS software. By default, a Globus installation is configured to utilise the grid-mapfile mechanism to map each user s certificate credentials to a local Unix account: this is a suitable mechanism for describing the user mappings as described above. For small-scale data grid installations with small (and stable) user bases, it is possible to administer groups and users manually. However, the DiGS team recommend that in order to both reduce the administrative load and the scope for errors one should invest effort to set up a more automated process: for example, based on Virtual Organisation Management System. A selection of tools to assist with the automation of user management is provided in the EDG Files package on the project website [6] Setting up Storage Directories The final step in setting up a Storage Element is to create a place (a Storage Directory) in which user data can be stored. The actual target storage can be located anywhere on the file system, though must be owned by and writable by the DiGS package account. 4 Note: If the target data grid uses VOMS for user management, then the host certificate mapping should be added to the local static grid-mapfile (usually in /etc/grid-security/grid-mapfile-local), instead of to the standard grid-mapfile. One should contact the data grid administrator for more information about this step. 15

20 For example, if the Storage Element has a local RAID system that is mounted on the file system as /raid, one may wish to create a target folder for data grid storage (that is, a Storage Directory), called /raid/digs-data. Within the Storage Directory, one must create a sub-directory called data. This directory structure could be implemented by running the following sequence of commands (as root): > cd /raid > mkdir digs-data > chown <package account> digs-data > cd /digs-data > mkdir data > chown <package account> data The data directory should support read access for each of the Unix groups that corresponds to a DiGS privileged group. This can be achieved using: > chmod a+rx data or (in a more controlled manner) using access control lists (ACLs), on systems that support this. An optional Inbox can also be created. This is a folder in which new data may be temporarily stored, in the first stage of being added to the data grid. At least one of the storage elements must have an Inbox. This directory, which must be located in the data directory, should be called NEW. It should be owned by DiGS Package Account and must be writable by members of each of the Unix groups that corresponds to a DiGS privileged group. Continuing with our example, this could be achieved as follows (again, running as root user): > cd /raid/digs-data/data > mkdir NEW > chown <package account> NEW > chmod a+rwx NEW Note: As above, one may configured group-based write permissions using ACLs, on systems that support this. Having created a suitable data storage directory (and possibly an Inbox), one needs to inform the Data Grid Administrator of the details, so that the Control Node can be configured as described in Section Configuring multiple Storage Directories It is possible to have multiple data directories on a single Storage Element. For example, this could allow multiple mass storage devices to be accessed from a single Storage Element. The process for adding further data directories is similar to that described above, with two differences of note: Symbolic links to each data directory should be created in the same directory as the first data directory, named data for the first entry, data1 for the second entry, then data2, and so on. For example, this can be done with the command: > cd /raid/digs-data > ln -s /raid2/digs-data data1 > chown <package account> data1 > chmod a+rx data1 DiGS automatically recognises that multiple Storage Directories are available and aims to share grid datasets between the different directories in order to balance free space on the associated disk units. 16

21 3.3 SRM Storage Element Setup As of DiGS version 3.1, storage elements that expose an Storage Resource Manager (SRM) [3] interface are supported. Any SRM v2 compliant implementation should work with DiGS, however at this stage DPM [4] has been most thoroughly tested. The details of setting up an SRM server for use with DiGS will vary depending on the implementation used, and are beyond the scope of this document. At a minimum, the following two steps will be necessary: Set up authorisation to allow access from DiGS users and the control thread. This is typically done by sourcing user information from a VOMS server or similar. Create the directory structure required by DiGS. This may be as simple as one single data directory and an optional inbox directory. The data directory should be writable by the control thread s identity and readable by all users. The inbox directory (if present) should be owned by the control thread certificate but writable by all users with permission to add data to the grid. As with GridFTP storage elements, details of the SRM storage elements must be added to the control node configuration (see next section). 3.4 Control Node Installation As described in Chapter 1, the Control Node is an essential component of a DiGS-powered system, hosting a daemon process, the Control Thread, that maintains the integrity of the data grid and hosts a database, the File Catalogue, describing the layout of date in the data grid. The setup process for the Control Node consists of the following steps: 1. Create a DiGS package account this step, which is optional for a Client Node installation, is mandatory for the Control Node. 2. Install the DiGS software (from either a pre-compiled binary package or from source code). 3. Set up the Control Node configuration files. 4. Set up an X.509 certificate to allow the Control Thread to authenticate to other nodes in data grid. 5. Set up user access to the Control Node. 6. Start the Control Thread. The Control Node can also host the Metadata Catalogue, though this can be installed elsewhere. The setup process for the Metadata Catalogue is described in Section Creating a DiGS Package Account The use of a DiGS package account (see Section 2.1.4) is mandatory during the setup of the Control Node. It is recommended that a new user account (for example, named digs) be created for this purpose, instead of using an existing user account or the root account Installing the software As with the Client Node installation and Storage the Element installation, one may choose to install the software from either a pre-compiled binary package (if a suitable package is available) or from source code: 17

22 The installation process for a pre-compiled binary distribution is the same as for a client installation (see Section 3.1 for details). The installation process for source code is very similar to the client installation, except one should use the build target make-server: > ant make-server tee build.log Note that, having completed the unpack/build of the DiGS software, one also needs to set up the various configuration files and configure the shell environment for the DiGS package account, as for the client installation (Section 3.1) Creating the Control Node Configuration Files The Control Node hosts a number of configuration files that capture a description of the data grid in its entirety. These configuration files are used by a local daemon process, called the Control Thread, and by remote clients, running on the other nodes, to establish required information about the setup and status of the data grid. To ensure that the configuration files are accessible as required, one should make the files world-readable, though writable only by the DiGS Package Account. All Control Node configuration files should be created in the DiGS installation directory (for example, /home/digs/digs-3.1). The files are, as follows: qcdgrid.conf contains general configuration settings. mainnodelist.conf contains a description of the Storage Elements on the data grid. deadnodes.conf contains a list of those Storage Elements that are currently and unexpectedly not responding to Control Thread requests. disablednodes.conf contains a list of those Storage Elements that have been temporarily disabled by the data grid administrator: for example, during scheduled maintenance sessions. retiringnodes.conf contains a list of those nodes that are scheduled for removal from the data grid. group-mapfile contains a list of group memberships for each user certificate subject. This configuration file is discussed in Section The makeup of each of these configuration files (except for the group-mapfile, which is discussed in Section 3.4.5) is discussed, in turn, below. The qcdgrid.conf file The qcdgrid.conf file contains various settings, as follows: rc_port the port numbers on which the File Catalogue service is listening. qcdgrid_port the port numbers on which the Control Thread is listening. min_copies the default number of replicas for each user data file. disk_space_low the threshold free space (in bytes) below which a Storage Element should not be written to. disk_space_panic the threshold free space (in bytes) below which a Storage Element must not fall. administration the identity (that is, X.509 certificate distinguished name) of the data grid administrator. 18

23 admin_ the address of the data grid administrator. An example configuration is given below: rc_port=39281 qcdgrid_port=51000 min_copies=2 disk_space_low= disk_space_panic= administrator=/o=grid/o=ukhep/ou=epcc.ed.ac.uk/cn=james Perry The mainnodelist.conf file The mainnodelist.conf file contains a list of the Storage Elements on the data grid. For each node, it must/may include the following information: node the fully qualified domain name of the Storage Element. This element is required. site a human-readable name for the site at which the Storage Element is located. This can be an arbitrary text string. It is used by the Control Thread to check that two Storage Elements are located at different sites, when replicating data files. This element is required. disk the amount of disk space (in kilobytes), on the Storage Element, which is available for data grid usage. Initially, this attribute can be given an arbitrary numeric value, as it is automatically updated by the Control Thread as part of its monitoring remit. This attribute is required, though the value is maintained by the Control Thread. path the path containing the data directories on the Storage Element. This element is required. type The type of Storage Element. At the time of writing, globus and srm are supported. This element is required. inbox The full path to the inbox on the storage element. If this element is not supplied it will be assumed that the storage element does not contain an inbox. extrarsl this attribute is used to specify additional Globus Resource Specification Language clauses that are added to the RSL string, whenever a job is submitted to the Storage Element. This attribute can, for example, be useful for specifying additional environment variables that must be set. This element is optional. extrajsscontact this attribute specifies an extra string to be added to the job submission service contact string. This can be used to specify a non-standard Globus job manager, for example. This element is optional. gpfs a flag that should be set to 1 to enable GPFS support on the node 5, and 0 for otherwise. GPFS-based systems require the free space on the disk to be calculated in a different manner to normal Unix file servers. This element is optional. (Versions 1.x and 2.0 only) jobtimeout this attribute specifies the time-out value in seconds for (Globus GRAM-based) job submission operations to this node. This element is optional. ftptimeout this attribute specifies the time-out value in seconds for short file operations (for example, creating new directories, deleting files, and so on) on this node. This element is optional. copytimeout this attributed specifies the time-out value in seconds for long file operations (usually, data file transfers) on this node. This element is optional. 5 The DiGS development team recommend that you contact them, before enabling this flag, for a more detailed explanation of the support for GPFS. 19

24 endpoint this attribute is required for SRM storage elements but not for Globus SEs. It specifies the endpoint URL for contacting the SRM web service. In addition, as of DiGS version 3.0, entries are required which specify the total disk space allocated to DiGS on each data disk of each node. These properties are named data, data1, data2, etc. the same as the directories to which they refer, and their values are in kilobytes. An example mainnodelist.conf is given below: node=trumpton.ph.ed.ac.uk site=edinburgh disk= path=/disk/trumpton0/ukqcd.local/qcdgrid type=globus inbox=/home/digs/new data= data1= node=dylan.amtp.liv.ac.uk site=liverpool disk= path=/users/qcdgrid/dylan type=globus data= data1= node=pygrid1.swan.ac.uk site=swansea disk= path=/home/qcdgrid gpfs=1 type=globus data= data1= data2= data3= node=srm.epcc.ed.ac.uk site=edinburgh disk= path=/dpm/epcc.ed.ac.uk/home/digs inbox=/dpm/epcc.ed.ac.uk/home/digs/inbox type=srm endpoint=srm://srm.epcc.ed.ac.uk:8446/srm/managerv2 data= Other node status files The deadnodes.conf, disablednodes.conf and retiringnodes.conf files are managed by the software and should initially be created as empty files. For example, by running the following command sequence as the DiGS Package Account user: > cd ${DIGS_HOME} > touch deadnodes.conf > chmod 644 deadnodes.conf > touch disablednodes.conf 20

Globus Toolkit Manoj Soni SENG, CDAC. 20 th & 21 th Nov 2008 GGOA Workshop 08 Bangalore

Globus Toolkit Manoj Soni SENG, CDAC. 20 th & 21 th Nov 2008 GGOA Workshop 08 Bangalore Globus Toolkit 4.0.7 Manoj Soni SENG, CDAC 1 What is Globus Toolkit? The Globus Toolkit is an open source software toolkit used for building Grid systems and applications. It is being developed by the

More information

glite Grid Services Overview

glite Grid Services Overview The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,

More information

EMC Documentum Composer

EMC Documentum Composer EMC Documentum Composer Version 6.5 SP2 User Guide P/N 300-009-462 A01 EMC Corporation Corporate Headquarters: Hopkinton, MA 01748-9103 1-508-435-1000 www.emc.com Copyright 2008 2009 EMC Corporation. All

More information

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Why the Grid? Science is becoming increasingly digital and needs to deal with increasing amounts of

More information

Developing and Deploying vsphere Solutions, vservices, and ESX Agents. 17 APR 2018 vsphere Web Services SDK 6.7 vcenter Server 6.7 VMware ESXi 6.

Developing and Deploying vsphere Solutions, vservices, and ESX Agents. 17 APR 2018 vsphere Web Services SDK 6.7 vcenter Server 6.7 VMware ESXi 6. Developing and Deploying vsphere Solutions, vservices, and ESX Agents 17 APR 2018 vsphere Web Services SDK 6.7 vcenter Server 6.7 VMware ESXi 6.7 You can find the most up-to-date technical documentation

More information

vrealize Operations Manager Customization and Administration Guide vrealize Operations Manager 6.4

vrealize Operations Manager Customization and Administration Guide vrealize Operations Manager 6.4 vrealize Operations Manager Customization and Administration Guide vrealize Operations Manager 6.4 vrealize Operations Manager Customization and Administration Guide You can find the most up-to-date technical

More information

A Simple Mass Storage System for the SRB Data Grid

A Simple Mass Storage System for the SRB Data Grid A Simple Mass Storage System for the SRB Data Grid Michael Wan, Arcot Rajasekar, Reagan Moore, Phil Andrews San Diego Supercomputer Center SDSC/UCSD/NPACI Outline Motivations for implementing a Mass Storage

More information

Understanding StoRM: from introduction to internals

Understanding StoRM: from introduction to internals Understanding StoRM: from introduction to internals 13 November 2007 Outline Storage Resource Manager The StoRM service StoRM components and internals Deployment configuration Authorization and ACLs Conclusions.

More information

OGCE User Guide for OGCE Release 1

OGCE User Guide for OGCE Release 1 OGCE User Guide for OGCE Release 1 1 Publisher s Note Release 2 begins the migration to open standards portlets. The following has been published by the Open Grids Computing Environment: OGCE Release 2

More information

Progress in building the International Lattice Data Grid

Progress in building the International Lattice Data Grid Progress in building the International Lattice Data Grid Jefferson Lab, Newport News, VA 23606, USA E-mail: bjoo@jlab.org C.M. Maynard EPCC, School of Physics, Univeristy of Edinburgh, Edinburgh EH9 3JZ,

More information

Architecture Proposal

Architecture Proposal Nordic Testbed for Wide Area Computing and Data Handling NORDUGRID-TECH-1 19/02/2002 Architecture Proposal M.Ellert, A.Konstantinov, B.Kónya, O.Smirnova, A.Wäänänen Introduction The document describes

More information

Oracle Application Express: Administration 1-2

Oracle Application Express: Administration 1-2 Oracle Application Express: Administration 1-2 The suggested course agenda is displayed in the slide. Each lesson, except the Course Overview, will be followed by practice time. Oracle Application Express:

More information

FUSION REGISTRY COMMUNITY EDITION SETUP GUIDE VERSION 9. Setup Guide. This guide explains how to install and configure the Fusion Registry.

FUSION REGISTRY COMMUNITY EDITION SETUP GUIDE VERSION 9. Setup Guide. This guide explains how to install and configure the Fusion Registry. FUSION REGISTRY COMMUNITY EDITION VERSION 9 Setup Guide This guide explains how to install and configure the Fusion Registry. FUSION REGISTRY COMMUNITY EDITION SETUP GUIDE Fusion Registry: 9.2.x Document

More information

Scientific data management

Scientific data management Scientific data management Storage and data management components Application database Certificate Certificate Authorised users directory Certificate Certificate Researcher Certificate Policies Information

More information

Using the Horizon vrealize Orchestrator Plug-In

Using the Horizon vrealize Orchestrator Plug-In Using the Horizon vrealize Orchestrator Plug-In VMware Horizon 6 version 6.2.3, VMware Horizon 7 versions 7.0.3 and later Modified on 4 JAN 2018 VMware Horizon 7 7.4 You can find the most up-to-date technical

More information

Developing and Deploying vsphere Solutions, vservices, and ESX Agents

Developing and Deploying vsphere Solutions, vservices, and ESX Agents Developing and Deploying vsphere Solutions, vservices, and ESX Agents Modified on 27 JUL 2017 vsphere Web Services SDK 6.5 vcenter Server 6.5 VMware ESXi 6.5 Developing and Deploying vsphere Solutions,

More information

Installation and Administration

Installation and Administration Introduction to GT3 Background The Grid Problem The Globus Approach OGSA & OGSI Globus Toolkit GT3 Architecture and Functionality: The Latest Refinement of the Globus Toolkit Core Base Services User-Defined

More information

Setting Up Resources in VMware Identity Manager (On Premises) Modified on 30 AUG 2017 VMware AirWatch 9.1.1

Setting Up Resources in VMware Identity Manager (On Premises) Modified on 30 AUG 2017 VMware AirWatch 9.1.1 Setting Up Resources in VMware Identity Manager (On Premises) Modified on 30 AUG 2017 VMware AirWatch 9.1.1 Setting Up Resources in VMware Identity Manager (On Premises) You can find the most up-to-date

More information

EMC Documentum Composer

EMC Documentum Composer EMC Documentum Composer Version 6.0 SP1.5 User Guide P/N 300 005 253 A02 EMC Corporation Corporate Headquarters: Hopkinton, MA 01748 9103 1 508 435 1000 www.emc.com Copyright 2008 EMC Corporation. All

More information

NexentaStor VVOL

NexentaStor VVOL NexentaStor 5.1.1 VVOL Admin Guide Date: January, 2018 Software Version: NexentaStor 5.1.1 VVOL Part Number: 3000-VVOL-5.1.1-000065-A Table of Contents Preface... 3 Intended Audience 3 References 3 Document

More information

Migrating vrealize Automation 6.2 to 7.2

Migrating vrealize Automation 6.2 to 7.2 Migrating vrealize Automation 6.2 to 7.2 vrealize Automation 7.2 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition.

More information

BlackPearl Customer Created Clients Using Free & Open Source Tools

BlackPearl Customer Created Clients Using Free & Open Source Tools BlackPearl Customer Created Clients Using Free & Open Source Tools December 2017 Contents A B S T R A C T... 3 I N T R O D U C T I O N... 3 B U L D I N G A C U S T O M E R C R E A T E D C L I E N T...

More information

Developing and Deploying vsphere Solutions, vservices, and ESX Agents

Developing and Deploying vsphere Solutions, vservices, and ESX Agents Developing and Deploying vsphere Solutions, vservices, and ESX Agents vsphere 6.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced

More information

UNIT IV PROGRAMMING MODEL. Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus

UNIT IV PROGRAMMING MODEL. Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus UNIT IV PROGRAMMING MODEL Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus Globus: One of the most influential Grid middleware projects is the Globus

More information

The Grid Monitor. Usage and installation manual. Oxana Smirnova

The Grid Monitor. Usage and installation manual. Oxana Smirnova NORDUGRID NORDUGRID-MANUAL-5 2/5/2017 The Grid Monitor Usage and installation manual Oxana Smirnova Abstract The LDAP-based ARC Grid Monitor is a Web client tool for the ARC Information System, allowing

More information

Working with Basic Linux. Daniel Balagué

Working with Basic Linux. Daniel Balagué Working with Basic Linux Daniel Balagué How Linux Works? Everything in Linux is either a file or a process. A process is an executing program identified with a PID number. It runs in short or long duration

More information

IBM z Systems Development and Test Environment Tools User's Guide IBM

IBM z Systems Development and Test Environment Tools User's Guide IBM IBM z Systems Development and Test Environment Tools User's Guide IBM ii IBM z Systems Development and Test Environment Tools User's Guide Contents Chapter 1. Overview......... 1 Introduction..............

More information

Setting Up Resources in VMware Identity Manager. VMware Identity Manager 2.8

Setting Up Resources in VMware Identity Manager. VMware Identity Manager 2.8 Setting Up Resources in VMware Identity Manager VMware Identity Manager 2.8 You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/ If you have comments

More information

Developing and Deploying vsphere Solutions, vservices, and ESX Agents

Developing and Deploying vsphere Solutions, vservices, and ESX Agents Developing and Deploying vsphere Solutions, vservices, and ESX Agents vsphere 5.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced

More information

MANAGEMENT AND ANALYSIS. ( Version ) Stephan Pabinger,

MANAGEMENT AND ANALYSIS. ( Version ) Stephan Pabinger, QPCR: APPLICATION FOR REAL-TIME PCR DATA MANAGEMENT AND ANALYSIS AND ADMINISTRATION MANUAL ( Version 0.9.11 ) Author: Stephan Pabinger, stephan.pabinger@tugraz.at Graz University of Technology Institute

More information

vcloud Director Administrator's Guide

vcloud Director Administrator's Guide vcloud Director 5.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

Setting Up Resources in VMware Identity Manager

Setting Up Resources in VMware Identity Manager Setting Up Resources in VMware Identity Manager VMware Identity Manager 2.7 This document supports the version of each product listed and supports all subsequent versions until the document is replaced

More information

Oracle Fusion Middleware Installing and Configuring Oracle Business Intelligence. 12c ( )

Oracle Fusion Middleware Installing and Configuring Oracle Business Intelligence. 12c ( ) Oracle Fusion Middleware Installing and Configuring Oracle Business Intelligence 12c (12.2.1.3.0) E96171-01 May 2018 Oracle Fusion Middleware Installing and Configuring Oracle Business Intelligence, 12c

More information

dcache Introduction Course

dcache Introduction Course GRIDKA SCHOOL 2013 KARLSRUHER INSTITUT FÜR TECHNOLOGIE KARLSRUHE August 29, 2013 dcache Introduction Course Overview Chapters I, II and Ⅴ christoph.anton.mitterer@lmu.de I. Introduction To dcache Slide

More information

Forescout. eyeextend for IBM BigFix. Configuration Guide. Version 1.2

Forescout. eyeextend for IBM BigFix. Configuration Guide. Version 1.2 Forescout Version 1.2 Contact Information Forescout Technologies, Inc. 190 West Tasman Drive San Jose, CA 95134 USA https://www.forescout.com/support/ Toll-Free (US): 1.866.377.8771 Tel (Intl): 1.408.213.3191

More information

dcache Introduction Course

dcache Introduction Course GRIDKA SCHOOL 2013 KARLSRUHER INSTITUT FÜR TECHNOLOGIE KARLSRUHE August 29, 2013 dcache Introduction Course Overview Chapters I, II and Ⅴ Christoph Anton Mitterer christoph.anton.mitterer@lmu.de ⅤIII.

More information

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia Grid services Dusan Vudragovic dusan@phy.bg.ac.yu Scientific Computing Laboratory Institute of Physics Belgrade, Serbia Sep. 19, 2008 www.eu-egee.org Set of basic Grid services Job submission/management

More information

EMC Documentum Composer

EMC Documentum Composer EMC Documentum Composer Version 6 SP1 User Guide P/N 300 005 253 A01 EMC Corporation Corporate Headquarters: Hopkinton, MA 01748 9103 1 508 435 1000 www.emc.com Copyright 2008 EMC Corporation. All rights

More information

IBM Spectrum Protect Version Introduction to Data Protection Solutions IBM

IBM Spectrum Protect Version Introduction to Data Protection Solutions IBM IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM IBM Spectrum Protect Version 8.1.2 Introduction to Data Protection Solutions IBM Note: Before you use this information

More information

ForeScout Extended Module for IBM BigFix

ForeScout Extended Module for IBM BigFix ForeScout Extended Module for IBM BigFix Version 1.0.0 Table of Contents About this Integration... 4 Use Cases... 4 Additional BigFix Documentation... 4 About this Module... 4 Concepts, Components, Considerations...

More information

CS 1653: Applied Cryptography and Network Security Fall Term Project, Phase 2

CS 1653: Applied Cryptography and Network Security Fall Term Project, Phase 2 CS 1653: Applied Cryptography and Network Security Fall 2017 Term Project, Phase 2 Assigned: Tuesday, September 12 Due: Tuesday, October 3, 11:59 PM 1 Background Over the course of this semester, we will

More information

Linux Administration

Linux Administration Linux Administration This course will cover all aspects of Linux Certification. At the end of the course delegates will have the skills required to administer a Linux System. It is designed for professionals

More information

Using the VMware vcenter Orchestrator Client. vrealize Orchestrator 5.5.1

Using the VMware vcenter Orchestrator Client. vrealize Orchestrator 5.5.1 Using the VMware vcenter Orchestrator Client vrealize Orchestrator 5.5.1 You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/ If you have comments

More information

Grid Architectural Models

Grid Architectural Models Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers

More information

Guidelines on non-browser access

Guidelines on non-browser access Published Date: 13-06-2017 Revision: 1.0 Work Package: Document Code: Document URL: JRA1 AARC-JRA1.4F https://aarc-project.eu/wp-content/uploads/2017/03/aarc-jra1.4f.pdf 1 Table of Contents 1 Introduction

More information

Boundary control : Access Controls: An access control mechanism processes users request for resources in three steps: Identification:

Boundary control : Access Controls: An access control mechanism processes users request for resources in three steps: Identification: Application control : Boundary control : Access Controls: These controls restrict use of computer system resources to authorized users, limit the actions authorized users can taker with these resources,

More information

Tutorial 1: Introduction to Globus Toolkit. John Watt, National e-science Centre

Tutorial 1: Introduction to Globus Toolkit. John Watt, National e-science Centre Tutorial 1: Introduction to Globus Toolkit John Watt, National e-science Centre National e-science Centre Kelvin Hub Opened May 2003 Kelvin Building Staff Technical Director Prof. Richard Sinnott 6 RAs

More information

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.

g-eclipse A Framework for Accessing Grid Infrastructures Nicholas Loulloudes Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac. g-eclipse A Framework for Accessing Grid Infrastructures Trainer, University of Cyprus (loulloudes.n_at_cs.ucy.ac.cy) EGEE Training the Trainers May 6 th, 2009 Outline Grid Reality The Problem g-eclipse

More information

ARCHER Data Services Service Layer

ARCHER Data Services Service Layer ARCHER 1.0 ARCHER Data Services Service Layer System Administrator s Guide ICAT & MCAText Installation Configuration Maintenance ARCHER Data Services Service Layer... 1 About ARCHER Data Services Service

More information

AppDev StudioTM 3.2 SAS. Migration Guide

AppDev StudioTM 3.2 SAS. Migration Guide SAS Migration Guide AppDev StudioTM 3.2 The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2006. SAS AppDev TM Studio 3.2: Migration Guide. Cary, NC: SAS Institute Inc.

More information

Knowledge Discovery Services and Tools on Grids

Knowledge Discovery Services and Tools on Grids Knowledge Discovery Services and Tools on Grids DOMENICO TALIA DEIS University of Calabria ITALY talia@deis.unical.it Symposium ISMIS 2003, Maebashi City, Japan, Oct. 29, 2003 OUTLINE Introduction Grid

More information

Using the Horizon vcenter Orchestrator Plug-In. VMware Horizon 6 6.0

Using the Horizon vcenter Orchestrator Plug-In. VMware Horizon 6 6.0 Using the Horizon vcenter Orchestrator Plug-In VMware Horizon 6 6.0 You can find the most up-to-date technical documentation on the VMware Web site at: https://docs.vmware.com/ The VMware Web site also

More information

ForeScout Extended Module for IBM BigFix

ForeScout Extended Module for IBM BigFix Version 1.1 Table of Contents About BigFix Integration... 4 Use Cases... 4 Additional BigFix Documentation... 4 About this Module... 4 About Support for Dual Stack Environments... 5 Concepts, Components,

More information

Introduction to GT3. Overview. Installation Pre-requisites GT3.2. Overview of Installing GT3

Introduction to GT3. Overview. Installation Pre-requisites GT3.2. Overview of Installing GT3 Introduction to GT3 Background The Grid Problem The Globus Approach OGSA & OGSI Globus Toolkit GT3 Architecture and Functionality: The Latest Refinement of the Globus Toolkit Core Base Services User-Defined

More information

Multi-Tenancy in vrealize Orchestrator. vrealize Orchestrator 7.4

Multi-Tenancy in vrealize Orchestrator. vrealize Orchestrator 7.4 Multi-Tenancy in vrealize Orchestrator vrealize Orchestrator 7.4 Multi-Tenancy in vrealize Orchestrator You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/

More information

Distribute Call Studio applications to Unified CVP VXML Servers.

Distribute Call Studio applications to Unified CVP VXML Servers. is one of the Cisco Unified Customer Voice Portal (CVP) components and a web-based interface using which you can configure other Unified CVP components and devices in the Unified CVP solution. Use to perform

More information

IBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM

IBM Tivoli Storage Manager Version Introduction to Data Protection Solutions IBM IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM IBM Tivoli Storage Manager Version 7.1.6 Introduction to Data Protection Solutions IBM Note: Before you use this

More information

Design of Distributed Data Mining Applications on the KNOWLEDGE GRID

Design of Distributed Data Mining Applications on the KNOWLEDGE GRID Design of Distributed Data Mining Applications on the KNOWLEDGE GRID Mario Cannataro ICAR-CNR cannataro@acm.org Domenico Talia DEIS University of Calabria talia@deis.unical.it Paolo Trunfio DEIS University

More information

VMware Identity Manager Connector Installation and Configuration (Legacy Mode)

VMware Identity Manager Connector Installation and Configuration (Legacy Mode) VMware Identity Manager Connector Installation and Configuration (Legacy Mode) VMware Identity Manager This document supports the version of each product listed and supports all subsequent versions until

More information

NCP Secure Enterprise Management for Linux Release Notes

NCP Secure Enterprise Management for Linux Release Notes Major Release: 5.00 r39572 Date: May 2018 Prerequisites The following distributions and databases with the associated Connector/C drivers are supported with this release: Linux distribution Database Driver

More information

Using the MyProxy Online Credential Repository

Using the MyProxy Online Credential Repository Using the MyProxy Online Credential Repository Jim Basney National Center for Supercomputing Applications University of Illinois jbasney@ncsa.uiuc.edu What is MyProxy? Independent Globus Toolkit add-on

More information

Front Office and VMware. Automating server provisioning from your Service Catalog

Front Office and VMware. Automating server provisioning from your Service Catalog Front Office and VMware Automating server provisioning from your Service Catalog Contents 1.0 Introduction... 3 2.0 How it works... 3 2.1 Virtual Machine Templates... 3 2.2 Front Office Configuration...

More information

Contents Overview... 5 Downloading Primavera Gateway... 5 Primavera Gateway On-Premises Installation Prerequisites... 6

Contents Overview... 5 Downloading Primavera Gateway... 5 Primavera Gateway On-Premises Installation Prerequisites... 6 Gateway Installation and Configuration Guide for On-Premises Version 17 September 2017 Contents Overview... 5 Downloading Primavera Gateway... 5 Primavera Gateway On-Premises Installation Prerequisites...

More information

VMware AirWatch Content Gateway for Linux. VMware Workspace ONE UEM 1811 Unified Access Gateway

VMware AirWatch Content Gateway for Linux. VMware Workspace ONE UEM 1811 Unified Access Gateway VMware AirWatch Content Gateway for Linux VMware Workspace ONE UEM 1811 Unified Access Gateway You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/

More information

Setting Up Resources in VMware Identity Manager (SaaS) Modified 15 SEP 2017 VMware Identity Manager

Setting Up Resources in VMware Identity Manager (SaaS) Modified 15 SEP 2017 VMware Identity Manager Setting Up Resources in VMware Identity Manager (SaaS) Modified 15 SEP 2017 VMware Identity Manager Setting Up Resources in VMware Identity Manager (SaaS) You can find the most up-to-date technical documentation

More information

Intellicus Cluster and Load Balancing- Linux. Version: 18.1

Intellicus Cluster and Load Balancing- Linux. Version: 18.1 Intellicus Cluster and Load Balancing- Linux Version: 18.1 1 Copyright 2018 Intellicus Technologies This document and its content is copyrighted material of Intellicus Technologies. The content may not

More information

GUT. GUT Installation Guide

GUT. GUT Installation Guide Date : 17 Mar 2011 1/6 GUT Contents 1 Introduction...2 2 Installing GUT...2 2.1 Optional Extensions...2 2.2 Installation using the Binary package...2 2.2.1 Linux or Mac OS X...2 2.2.2 Windows...4 2.3 Installing

More information

Connector for Microsoft SharePoint 2013, 2016 and Online Setup and Reference Guide

Connector for Microsoft SharePoint 2013, 2016 and Online Setup and Reference Guide Connector for Microsoft SharePoint 2013, 2016 and Online Setup and Reference Guide Published: 2018-Oct-09 Contents 1 Microsoft SharePoint 2013, 2016 and Online Connector 4 1.1 Products 4 1.2 Supported

More information

Installing and Configuring VMware Identity Manager Connector (Windows) OCT 2018 VMware Identity Manager VMware Identity Manager 3.

Installing and Configuring VMware Identity Manager Connector (Windows) OCT 2018 VMware Identity Manager VMware Identity Manager 3. Installing and Configuring VMware Identity Manager Connector 2018.8.1.0 (Windows) OCT 2018 VMware Identity Manager VMware Identity Manager 3.3 You can find the most up-to-date technical documentation on

More information

Question No: 1 In which file should customization classes be specified in the cust-config section (under mds-config)?

Question No: 1 In which file should customization classes be specified in the cust-config section (under mds-config)? Volume: 80 Questions Question No: 1 In which file should customization classes be specified in the cust-config section (under mds-config)? A. web.xml B. weblogic.xml C. adf-config.xml D. adfm.xml Question

More information

Siebel Application Deployment Manager Guide. Version 8.0, Rev. A April 2007

Siebel Application Deployment Manager Guide. Version 8.0, Rev. A April 2007 Siebel Application Deployment Manager Guide Version 8.0, Rev. A April 2007 Copyright 2005, 2006, 2007 Oracle. All rights reserved. The Programs (which include both the software and documentation) contain

More information

BlackBerry Enterprise Server for IBM Lotus Domino Version: 5.0. Administration Guide

BlackBerry Enterprise Server for IBM Lotus Domino Version: 5.0. Administration Guide BlackBerry Enterprise Server for IBM Lotus Domino Version: 5.0 Administration Guide SWDT487521-636611-0528041049-001 Contents 1 Overview: BlackBerry Enterprise Server... 21 Getting started in your BlackBerry

More information

Bedework Calendar Deployment Manual

Bedework Calendar Deployment Manual Bedework Calendar Deployment Manual Bedework version 3.1 Last modified: July 30, 2006 Bedework Deployment Manual The Bedework Deployment Manual contains instructions for customizing and installing a production

More information

EUROPEAN MIDDLEWARE INITIATIVE

EUROPEAN MIDDLEWARE INITIATIVE EUROPEAN MIDDLEWARE INITIATIVE VOMS CORE AND WMS SECURITY ASSESSMENT EMI DOCUMENT Document identifier: EMI-DOC-SA2- VOMS_WMS_Security_Assessment_v1.0.doc Activity: Lead Partner: Document status: Document

More information

vcloud Director Administrator's Guide vcloud Director 9.0

vcloud Director Administrator's Guide vcloud Director 9.0 vcloud Director 9.0 You can find the most up-to-date technical documentation on the VMware Web site at: https://docs.vmware.com/ The VMware Web site also provides the latest product updates. If you have

More information

BIG-IP Access Policy Manager : Portal Access. Version 12.1

BIG-IP Access Policy Manager : Portal Access. Version 12.1 BIG-IP Access Policy Manager : Portal Access Version 12.1 Table of Contents Table of Contents Overview of Portal Access...7 Overview: What is portal access?...7 About portal access configuration elements...7

More information

GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide

GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide GT 4.2.0: Community Scheduler Framework (CSF) System Administrator's Guide Introduction This guide contains advanced configuration

More information

Zadara Enterprise Storage in

Zadara Enterprise Storage in Zadara Enterprise Storage in Google Cloud Platform (GCP) Deployment Guide March 2017 Revision A 2011 2017 ZADARA Storage, Inc. All rights reserved. Zadara Storage / GCP - Deployment Guide Page 1 Contents

More information

QuickStart: Deploying DataSynapse GridServer 5.0 on HP Clusters

QuickStart: Deploying DataSynapse GridServer 5.0 on HP Clusters QuickStart: Deploying DataSynapse GridServer 5.0 on HP Clusters 1 Overview... 2 2 DataSynapse: Quick Overview... 2 2.1 GridServer 5.0 Grid Topology... 5 2.2 The GridServer Administration Tool... 5 3 Requirements...

More information

Advanced Service Design. vrealize Automation 6.2

Advanced Service Design. vrealize Automation 6.2 vrealize Automation 6.2 You can find the most up-to-date technical documentation on the VMware website at: https://docs.vmware.com/ If you have comments about this documentation, submit your feedback to

More information

Relativity Data Server

Relativity Data Server Relativity Data Server Micro Focus The Lawn 22-30 Old Bath Road Newbury, Berkshire RG14 1QN UK http://www.microfocus.com Copyright Micro Focus 2009-2015. All rights reserved. MICRO FOCUS, the Micro Focus

More information

Polarion 18.2 Enterprise Setup

Polarion 18.2 Enterprise Setup SIEMENS Polarion 18.2 Enterprise Setup POL005 18.2 Contents Overview........................................................... 1-1 Terminology..........................................................

More information

Deploying Rubrik Datos IO to Protect MongoDB Database on GCP

Deploying Rubrik Datos IO to Protect MongoDB Database on GCP DEPLOYMENT GUIDE Deploying Rubrik Datos IO to Protect MongoDB Database on GCP TABLE OF CONTENTS INTRODUCTION... 1 OBJECTIVES... 1 COSTS... 2 BEFORE YOU BEGIN... 2 PROVISIONING YOUR INFRASTRUCTURE FOR THE

More information

AutoForm plus R6.0.3 Release Notes

AutoForm plus R6.0.3 Release Notes 0 Release Notes AutoForm plus R6.0.3 Release Notes AutoForm plus R6.0.3 Release Notes...1 1 General Information...2 2 Installation Instructions...3 Front-End and Back-End Windows...3 Prerequisites...3

More information

Bitnami Re:dash for Huawei Enterprise Cloud

Bitnami Re:dash for Huawei Enterprise Cloud Bitnami Re:dash for Huawei Enterprise Cloud Description Re:dash is an open source data visualization and collaboration tool. It was designed to allow fast and easy access to billions of records in all

More information

Using the VMware vrealize Orchestrator Client

Using the VMware vrealize Orchestrator Client Using the VMware vrealize Orchestrator Client vrealize Orchestrator 7.0 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by

More information

Prototype DIRAC portal for EISCAT data Short instruction

Prototype DIRAC portal for EISCAT data Short instruction Prototype DIRAC portal for EISCAT data Short instruction Carl-Fredrik Enell January 19, 2017 1 Introduction 1.1 DIRAC EGI, first European Grid Initiative, later European Grid Infrastructure, and now simply

More information

vcloud Director Administrator's Guide

vcloud Director Administrator's Guide vcloud Director 5.1.1 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

DIGIPASS Authentication for Check Point VPN-1

DIGIPASS Authentication for Check Point VPN-1 DIGIPASS Authentication for Check Point VPN-1 With Vasco VACMAN Middleware 3.0 2007 Integration VASCO Data Security. Guideline All rights reserved. Page 1 of 51 Disclaimer Disclaimer of Warranties and

More information

A Federated Grid Environment with Replication Services

A Federated Grid Environment with Replication Services A Federated Grid Environment with Replication Services Vivek Khurana, Max Berger & Michael Sobolewski SORCER Research Group, Texas Tech University Grids can be classified as computational grids, access

More information

VII. Corente Services SSL Client

VII. Corente Services SSL Client VII. Corente Services SSL Client Corente Release 9.1 Manual 9.1.1 Copyright 2014, Oracle and/or its affiliates. All rights reserved. Table of Contents Preface... 5 I. Introduction... 6 Chapter 1. Requirements...

More information

WORKSTATION SETUP GUIDE FOR ACCESSING THE MIBGAS PLATFORM. Date: 15/12/2016. Version: 4.1 MERCADO IBÉRICO DEL GAS

WORKSTATION SETUP GUIDE FOR ACCESSING THE MIBGAS PLATFORM. Date: 15/12/2016. Version: 4.1 MERCADO IBÉRICO DEL GAS WORKSTATION SETUP GUIDE FOR ACCESSING THE MIBGAS PLATFORM Date: 15/12/2016 Version: 4.1 Alfonso XI, 6. 28014 Madrid (España) www.mibgas.es T (+34) 91 268 26 01 CONTENTS 1 INTRODUCTION 2 2 PRIOR REQUISITES

More information

CA IdentityMinder. Glossary

CA IdentityMinder. Glossary CA IdentityMinder Glossary 12.6.3 This Documentation, which includes embedded help systems and electronically distributed materials, (hereinafter referred to as the Documentation ) is for your informational

More information

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI

Setup Desktop Grids and Bridges. Tutorial. Robert Lovas, MTA SZTAKI Setup Desktop Grids and Bridges Tutorial Robert Lovas, MTA SZTAKI Outline of the SZDG installation process 1. Installing the base operating system 2. Basic configuration of the operating system 3. Installing

More information

DataMan. version 6.5.4

DataMan. version 6.5.4 DataMan version 6.5.4 Contents DataMan User Guide 1 Introduction 1 DataMan 1 Technical Specifications 1 Hardware Requirements 1 Software Requirements 2 Ports 2 DataMan Installation 2 Component Installation

More information

TIBCO Spotfire Statistics Services Release Notes

TIBCO Spotfire Statistics Services Release Notes TIBCO Spotfire Statistics Services Release Notes Software Release 6.5 April 2014 Two-Second Advantage 2 Important SOME TIBCO SOFTWARE EMBEDS OR BUNDLES OTHER TIBCO SOFTWARE. USE OF SUCH EMBEDDED OR BUNDLED

More information

SAML-Based SSO Configuration

SAML-Based SSO Configuration Prerequisites, page 1 SAML SSO Configuration Task Flow, page 5 Reconfigure OpenAM SSO to SAML SSO Following an Upgrade, page 9 SAML SSO Deployment Interactions and Restrictions, page 9 Prerequisites NTP

More information

GUT. GUT Installation Guide

GUT. GUT Installation Guide Date : 02 Feb 2009 1/5 GUT Table of Contents 1 Introduction...2 2 Installing GUT...2 2.1 Optional Extensions...2 2.2 Installing from source...2 2.3 Installing the Linux binary package...4 2.4 Installing

More information

Lab 1 1 Due Wed., 2 Sept. 2015

Lab 1 1 Due Wed., 2 Sept. 2015 Lab 1 1 Due Wed., 2 Sept. 2015 CMPSC 112 Introduction to Computer Science II (Fall 2015) Prof. John Wenskovitch http://cs.allegheny.edu/~jwenskovitch/teaching/cmpsc112 Lab 1 - Version Control with Git

More information

AMGA metadata catalogue system

AMGA metadata catalogue system AMGA metadata catalogue system Hurng-Chun Lee ACGrid School, Hanoi, Vietnam www.eu-egee.org EGEE and glite are registered trademarks Outline AMGA overview AMGA Background and Motivation for AMGA Interface,

More information