GriPhyN-LIGO Prototype Draft Please send comments to Leila Meshkat

Size: px
Start display at page:

Download "GriPhyN-LIGO Prototype Draft Please send comments to Leila Meshkat"

Transcription

1 Technical Report GriPhyN GriPhyN-LIGO Prototype Draft Please send comments to Leila Meshkat Kent Blackburn, Phil Ehrens, Albert Lazzarini, Roy Williams Caltech Ewa Deelman, Carl Kesselman, Gaurang Mehta, Leila Meshkat, Laura Pearlman USC/ISI Bruce Allen, Scott Koranda UWM 1

2 1 Prototype Overview Goal Resources used in Prototype Prototype Logic Replica Catalog Replica Catalog Structure Supported Virtual Data Requests User Interface LIGO-LW Syntax Data and metadata flow Virtual Data Request (VDR) Syntax and Semantics Time interval Channels Response Request Interpretation Security Explanation of Myproxy Planner Planner Overview User request Replica Catalog Query Planning Scenarios Scenario 1: Data is Available at User-requested location Scenario 2: Data is Available but not at the User-requested location Scenario 3: User-requested data is unavailable Scenario 3.1: Full frame is located at UWM Scenario 3.2: Full frame is located at ISI: Scenario 3.3: The full frame is available at both ISI and UWM Scenario 4: Full frame does not exist Executor Conclusions and Future Directions References

3 1 Prototype Overview 1.1 Goal The goal of this prototype is two-fold: first, to develop initial prototypes for the Virtual Data Toolkit in support of the Virtual Data concepts, and second to provide a scientifically meaningful, grid-enabled environment for the LIGO collaboration. This prototype was demonstrated at the SC 2001 in November. MyProxy server xml Cgi interface HTTP frontend Frame P. Charlton Desired Result : Transformation Catalog Power spectrum of seismic channels Planner Monitoring Replica Catalog GridCVS Replica Selection G-DAG (DAGMan) Prototype exclusive Executor CondorG/ DAGMan Logs In design Globus component In integration GridFTP GRAM/LDAS GridFTP GridFTP GRAM GridFTP GRAM/LDAS LDAS at UWM UWM SC floor Compute Resource LDAS at Caltech Figure 1: Prototype Overview Figure 1 depicts the overall logic and design of the prototype. The user makes a request for data via a GUI, this request is translated into an XML syntax and is used as input for the Request Manager (RM). The RM surveys the available data and compute resources and makes a plan to satisfy the request. The plan is specified in the DAGMan format and presented to the Condor-G system [1] for execution. Condor-G executes the plan by 3

4 accessing various storage and compute resources. The resulting data is moved to the desired, user-specified location and registered in the replica catalog. 1.2 Resources used in Prototype The LIGO prototype resources which are shown at the bottom of figure 1 and in figure 2 include: Storage Resources: workstation at Caltech, 21 terabytes of commodity disk on the UWM Linux cluster, cluster storage at ISI. Compute Resources: LDAS at Caltech, LDAS at UWM, Planner and job submission server using Condor-G at ISI Replica Catalog: An LDAP (Light Weight Directory Access Protocol) on cluster front end node (dc-user@isi.edu) at ISI: ldap://dc-user.isi.edu:9129 Web Server: apache on dc-user@isi.edu at ISI UW M ISI Replica Catalog Caltech LDAS LDAS Storage Resource Storage Resource Storage Resource Figure 2: Existing Resources for the prototype 2 Prototype Logic The user makes a request for virtual data via a web portal. The user is able to specify the virtual channel name, the start time, where the resulting data should be placed and the format of the resulting data. The format of the data can be XML (specifically XSIL[2]), or Frame, which is a standard data format for the gravitational wave community. The request is translated into XML and is passed to the Request Manager (RM). 4

5 The Request Manager interprets the requests and constructs a DAG that fully represents the computations and data movements necessary to satisfy the request. Some of the subtasks of the RM are: Authenticate user. Make an optimal plan for satisfying the user s request using the available resources. (The planner is described in section 6) Map request into a sequence of grid operations that need to be performed. Submit the plan to Condor-G. Return the status of the request and the location of the data to the user. The RM sends the DAG to Condor-G. The DAG can contain LDAS commands, gridftp[11] commands, arbitrary analysis pipeline commands, updates to the Replica Catalog (RC) and other local commands such as FrCopy (not implemented yet) Condor-G executes the specified data movements, submits jobs to LDAS via a GRAM (Globus Resource Allocation Manager) [5] interface, and updates the replica catalog. Details about the DAG can be found in section 7. LDAS or another analysis system perform computation. Data is moved to the user-specified location and the user is then notified of the request s completion and the updates performed to the catalog. 3 Replica Catalog 3.1 Replica Catalog Structure Ou=isi.edu, o=netscaperoot Rc=Ligo_Replica_Catalog lc= lc= lc= lc= Caltech UWM Caltech UWM Caltech UWM Caltech UWM ISI ISI ISI ISI :Raw and materialized data :Materialized data only Figure 3: Replica Catalog Structure 5

6 The Replica Catalog maintains mappings from logical file names to physical file locations. The logical file names are grouped into logical collections. When trying to find the physical location of a file, the collection name and logical file name need to be provided. Following is the structure for the catalog: 1. Collections are based on the instrument name and time interval, the instruments are: LO and HO and H1. 2. The time interval for a collection is 200 seconds. For instance a collection might contain files [ ]. This is shown in figure Each collection will contain frame files of 1 second in duration, both full channel and single channel frames. The data is located at: workstation at Caltech, storage cluster at ISI, storage cluster at UWM. We have four collections, each containing a 200 second time intervals of data. The distribution of the raw and materialized data for each of these collections is shown in figure 2. The acronym lc is used to denote the logical collections. 4 Supported Virtual Data Requests For the purpose of the prototype, we decided to focus on a raw data set that represents the Hanford system in lock. We identified the locked interferometer (IFO) data as being present in run E2 and being composed of 1800 to 2000 seconds of full frame (raw) data. Several raw channels have been identified as "interesting": H2:LSC-AS_Q H2:LSC-AS_DC H2:PSL-FSS_FAST_F H0:PEM-LVEA_SEISX H0:PEM-LVEA_SEISY H0:PEM-LVEA_SEISZ H0:ALL H2:ALL The names are determined based on the instrument used for collecting the data (H0, H2), and the channel names (LSC-AS_Q, LSC-AS_DC, ) ALL is a valid channel name used to specify the entire frame file. For the purposes of this prototype, we have used the data collected in the time interval [ ] and classified it into four collections, each consisting of 200 seconds of data. The user uses a web based portal to make a request. The web page where the user makes the request is shown in figure 3. The input parameters include the channel name, the time in GPS, the output data location and the return data type. 6

7 Figure 4: User request web page 5 User Interface 5.1 LIGO-LW Syntax We used the LIGO-LW as the syntax for formatting the request that the user makes to the system and the response that the user receives from the system. The XML is built from the LIGO-LW dialect, which is essentially the same as the XSIL dialect [2][4]. LIGO-LW has a small number of elements, and each of them can have a "Name" and a "Type" attribute. The ones we use are mostly these: 1. <Param> has the semantics of a keyword-value pair. To express "x=3" we would say <Param Name="x">3</Param>. 2. <LIGO-LW> makes a collection. By nesting these elements, we can make a hierarchy. There is a special meaning to the "Type" attribute of this element, as this is used to extend the language. We could build a date extension to LIGO-LW as follows: 7

8 <LIGO-LW Type="Date"> <Param Name="Day">11</Param> <Param Name="Month">8</Param> <Param Name="Year">2001</Param> </LIGO-LW> When the parsing software gets the Type of a LIGO-LW element, it expects to build an object of that type from the enclosed elements Data and metadata flow Initially, a JobID number is allocated to the user s request. As the request is being processed, the user receives feedback as to the location of the requested data files and the various transfer and transforms that are being conducted on it. The various scenarios that can occur are discussed in section 7. When processing is complete, the status request gives a final status response (FSR) indicating whether or not the user s request has been satisfied and the output location of the desired data files. The JobID allocated to a request becomes invalid after the request has been processed completely. 5.2 Virtual Data Request (VDR) Syntax and Semantics The request is in XML 1.0 format, with the standard LIGO-LW DTD (Document Type Definition). Thus the first two lines are: <?xml version="1.0"?> <!DOCTYPE LIGO-LW SYSTEM "LIGO- LW.dtd"> A VDR begins with a LIGO-LW element of type VirtualDataRequest, like this: <LIGO_LW Name="banana" Type="VirtualDataRequest"> There may be other LIGO-LW elements of different types in the document, which will be ignored by the VDR parser. There may be multiple VDR's in the XML document, which are parsed in an arbitrary order. The system connects the job-id s with the name attributes of the VDR to provide a clear mapping between the ID s and the VDR s. 8

9 Semantically, there are four attributes to the VDR: channel name, timestamp, output location (including output filename) and output format. Each of these attributes is unique. The following sections list the element types that are relevant to the VDR Time interval This is part of the specification of what data is to be retrieved from the data system. It is expressed by two of the following three elements being present. Note that <Time> is a base type in LIGO-LW <Time Name="StartTime" Type="GPS"> </Time> The default type of the Time element is GPS (seconds since January ), but other formats are possible, as detailed in the LIGO-LW specifications cited above Channels The list of channels to be requested from the data archive comprises both channel names and the interferometers at which the channels were recorderd. This is specified by a LIGO-LW extension of type "ChannelSpecification". Currently we will assume that each interferometer records the same set of channels, each with the same name, frequency, etc. The data to be obtained is a Cartesian product of a list of interferometers with a list of channels. The list of interferometers is specified by a comma-separated list of names, and the channels are fully specified. If multiple channels are requested, each channel name has to be listed For example: <LIGO_LW Type="ChannelSpecification"> <Param Name="Detector">LHO,LLO</Param> <Param Name="RegEx">H2%3ALSC- AS_Q</Param> </LIGO_LW> Response The semantic content of this section of the VDR is currently: What format the response data should be, The location in the grid where the data should be put, Hints about caching, expiration, etc.(to be included in the future) This information is specified in a LIGO-LW extension of type "ResponseSpecification": <LIGO-LW Type="ResponseSpecification"> 9

10 This may contain a Param of type "ResponseFormat", and currently it can have the values "LIGO-LW" and "Frame", for example <Param Name="ResponseFormat">Frame</Param> If the requested response format is "LIGO-LW", then data may be returned with the TimeSeries extension of LIGO-LW, as specified in the papers cited above. This response can be seen using the Xlook viewer [1]. If the requested format is "Frame", then the FSR will still be a LIGO-LW formatted document, which includes a citation to the location of the requested Frame file on the grid. The response location, if present, may be specified as a Param, for example: <Param Name="ResponseLocation"> gridftp://dataserver.phys.uwm.edu/albert/myfile.fra me </Param> <Param Name= ResponseDomain > uwm.edu</param> This location is specified in URL format, but the resources must be writeable by the user, by using gridftp. 6 Request Interpretation Initially, the Request Manager provides a job ID to the user and logs the full XML request in its local log file. Following is an example XML file of a user s request: <!DOCTYPE LIGO-LW SYSTEM "LIGO-LW.dtd"> <LIGO_LW Name="banana" Type="VirtualDataRequest"> <Time Name="StartTime" Type="GPS"> </Time> <Time Name="EndTime" Type="GPS"> </Time> <LIGO_LW Type="ChannelSpecification"> <Param Name="Detector">LHO,LLO</Param> <Param Name="RegEx">H2%3ALSC-AS_Q</Param> </LIGO_LW> <Param Name="ResponseDomain"> isi.edu </Param> <Param Name="ResponseFormat"> xml 10

11 The user request is then mapped into a logical file name. This is done by concatenating the channel name and the timestamp. For example the logical file name corresponding to the request for the channel H2: LSC-AS-Q and the timestamp would be H2:LSC-AS-Q F. Next, the Request Manager will ask the user (via an ssl connection) for the user s name and password. This information will be used to access the MyProxy server where the user s certificate resides. The steps used to setup the MyProxy server and use it are described in the next section. 7 Security The security in the prototype is implemented using Globus-GSI [3] (Grid Security Infrastructure), which is a means of authenticating using X509_certificates and a MyProxy [4] Server. The MyProxy [4] server takes a standard Globus [3] proxy certificate and stores it in its cache with added credentials to it. The server can then at runtime upon request and by providing the correct credentials create a delegated proxy certificate, which can give limited access to the Grid resources. 7.1 Explanation of Myproxy The Myproxy package provides a secure method for portal users to access resources using a limited proxy using the GSI. As part of the package, a myproxy-server is set up on a trusted host for a site or application specific portal in order to maintain delegated credentials for users that are valid for a chosen duration. After delegating a credential to the myproxy-server, users can retrieve a delegated credential from the myproxy-server via the web server that can be used for job submission, file transfer, etc. In the event that the web server is broken into, only the time limited delegated credentials are compromised so the hacker would potentially have access to other resources for a short amount of time. The GSI also includes provisions for Certificate Revocation Lists (CRL) as a means to deny users access. The following diagram helps to illustrate how the Myproxy package works: Figure 5: Myproxy package 11

12 The user runs myproxy-init and provides their GSI passphrase to create a proxy. The user enters an additional password (that will later be used in the browser) of their own choosing to delegate a proxy to the myproxy-server. Path A uses the GSI to perform mutual authentication between the client and the myproxy-server. Using any web browser capable of HTTPS, the user enters the password in (1) to get a proxy from the myproxy-server. The myproxy-server does further delegation making the proxy that's used on the web server valid for less than the original lifetime on the myproxy-server. Path B uses the GSI to perform mutual authentication between the web server and the myproxy-server using either the web server's credentials (typically a Verisign or Thawte certificate) or a Globus certificate 8 Planner The Planner determines where on the Grid the relevant data is stored, where the best place is to do the computing and how to route the output to the user requested location. The planner may not necessarily start with raw data: the requested product may be already partially or completely computed. The data movement and computing jobs are controlled by the Condor software, a fault tolerant execution environment for collections of tasks with specified dependencies. Connection to the LDAS systems and data fetching is controlled by the Grid security. 8.1 Planner Overview The function of the planner is to find an optimal plan for satisfying the user s request using the available resources. The plan would depend on the location of the requested data files and the operations that need to be conducted on them. For the purpose of this prototype, we assume that the possible locations are ISI, UWM and Caltech. The user requested location can be any of the above locations and the possible operations include copying files, extracting channels from full frames and updating the replica catalog. 8.2 User request The user of the LIGO-GriPhyN prototype makes a request for virtual data via a web portal. The virtual channel name, the start time and the location where the resulting data should be placed is also specified by the user. Following is an example user request that the prototype uses for demonstration purposes. Channel Name: H2:LSC-AS_Q Timestamp: Output Location: (Y) ISI: dc-n1.isi.edu/output/ligo Data is copied in the collection ligo Output File name: H-H2:LSC-AS_Q F (internal to catalog) 12

13 8.3 Replica Catalog Query Upon the receipt of the user request, the planner needs to query the replica catalog to find the location of the data. In order to identify the collection that needs to be searched, we can use the following algorithm: If we want the data for time instant TIC Let X=TIC Mod 200; Y=TIC X The data is in the collection with starting point Y. Therefore the logical collection in question has the data for the times [Y-Y+199] 8.4 Planning An overview of the plan used for satisfying the users request is shown in figure 5. The information contained in the user request includes the collection name, the channel name, the full frame file name and the time interval. The replica catalog is queried for the file identified with this information. If the file is available at the user specified location, the user will be pointed to the exact physical location of the file. If the file is available but at a different location, a copy of it will be transferred to the user-specified location and the user will be given a pointer to it. If, on the other hand, the file is unavailable, the replica catalog needs to be queried for the full frame file, and the user-requested channel needs to be extracted from the full frame file and a sent to the user-requested location. The replica catalog is updated every time a file is copied to a new location. We shall elaborate on the plan for satisfying the user s request under each of these conditions in section

14 Collection name Channel name Filename.F Time interval Query Replica Catalog File not found File found at location X Query Replica Catalog for Filename.f File found at location X Error: File Does not exist X= Y* X!= Y* X= ISI X= UWM X=ISI & UWM Scenario 1 Scenario 2 Y=ISI Y=UWM * Y=User requested location Scenario 3.1 Scenario 3.2 Scenario 3.1 Scenario 3.2 Figure 6: Plan Overview 14

15 8.5 Scenarios Scenario 1: Data is Available at User-requested location In this case we proceed as follow: Plan for Scenario 1 Query Replica Catalog X=Y (output location is same as user requested - location) Point user to channel ( at the location) Example request is available at ISI. In this case, the user is given following pointers: Scenario 2: Data is Available but not at the User-requested location In this case, the data needs to be copied to the user-specified location and the replica catalog needs to be updated with the new location. After querying the replica catalog for the data and learning about the location of the existing data (Y), the plan is specified in a DAG and submitted to DAGMan for further execution. The plan for scenario 2 would be as follow: 15

16 Plan for scenario 2 Query Replica Catalog Data is available at location Y!=X Make a plan and submit to DAGMAN: Copy from X to Y Update Replica Catalog With location Y Point user to channel In this case, the requested data is available at the storage facility at UWM and therefore needs to be copied to the desired directory at ISI. The replica catalog then needs to be updated with the information about the new storage location path at ISI. The concrete form of the DAG would be as follow Scenario 2 DAGMan(Concrete form) globus_url_copy H-H2:LSC-AS_Q F From dataserver.phys.uwm.edu/ To dc-n1.isi.edu/output/ligo globus-replica -catalog -host dc-n1.isi.edu/output/ ligo logicalfile H-H2:LSC-AS_Q F -location ISI 16

17 The globus-url-copy command indicates that the data in the source file should be copied to the destination file. The globus-url-copy command in DAGman format would have the following information in it: #Command: executable = path/globus-url-copy arguments = #Source: gsiftp:// dataserver.phys.uwm.edu/grid_outgoing/h- H2:LSC- AS_Q F #Destination: gsiftp://dc-n1.isi.edu/output/h-h2:lsc-as_q f The globus-replica-catalog command indicates that the host computer has obtained a copy of the logical file name specified at the given location. It would have the following format: #Command: executable = path/globus-replica-catalog arguments = -host dc-n1.isi.edu/output/ligo logicalfile H-H2:LSC-AS_Q F -location ISI Scenario 3: User-requested data is unavailable Scenario 3.1: Full frame is located at UWM In this case, the initial query to the Replica Catalog does not result in a match. This indicates that the requested channel has not been extracted from the original full frame file. Therefore, there should be a new query to identify the location of the full frame file. Once the location of the full frame file is identified, it should be sent to the nearest LDAS system for extraction of the user specified file. We assume that geographic proximity is the criteria for selecting the LDAS to which the full frame must be sent for extraction purposes. The plan for scenario 3.1 is as follow: 17

18 Plan for scenario 3.1 Query Replica Catalog Data is unavailable Look into Replica Catalog to locate full fram at X. If X== UWM then Send frame file to the LDAS system at UWM. Extract the requested channel Send the channel to the storage system Update replica catalog Point user to channel Following is the abstract DAGMan that is designed for the execution of the plan: Send file to LDAS At UWM Abstract DAGMan Extract requested channel Send Channel to storage At Y The concrete form of this DAGMan is as follow: Update Replica Catalog With Y 18

19 globus_url_copy H: F From dataserver.phys.uwm.edu/ To dataserver.phys.uwm.edu/ LDAS_extract H-H2:LSC-AS_Q F From dataserver.phys.uwm.edu/ To dataserver.phys.uwm.edu/ Concrete DAGman globus_url_copy H-H2:LSC-AS_Q F From dataserver.phys.uwm.edu/ To dc-n1.isi.edu/output/ligo globus -replica -catalog -host dc-n1.isi.edu/output/ ligo logicalfile H -H2:LSC-AS_Q F -location ISI The globus-url-copy command indicates that the full frame file needs to be copied from the storage at UWM to the LDAS system there. #Command: executable = path/globus-url-copy arguments = #Source: gsiftp:// dataserver.phys.uwm.edu/griphyn_test/grid_outgoing/h: F #Destination: gsiftp://dataserver.phys.uwm.edu/griphyn_test/ldas/ldas-outgoing/jobs/grid_incoming/h: F The LDAS-extract command indicates that the user-requested channel must be extracted from the full frame file. #Command: globusrun -o r #From: dataserver.phys.uwm.edu/jobmanager-ldas 19

20 executable=/opt/gldas/bin/extract arguments= --start_time: end_time: channel H2:LSC-AS_Q --input_frame: H F -output_frame:h-h2:lsc-as_q f -frameformat= frame or xml The next globus-url-copy command indicates that the extracted channel must be sent to the user requested location #Command: executable = path/globus-url-copy arguments = #Source: Gsiftp://dataserver.phys.uwm.edu/griphyn_test/LDAS/ldas-outgoing/H-H2:LSC-AS_Q F #Destination: gsiftp://dc-n1.isi.edu/output/ H-H2:LSC-AS_Q F The replica catalog then needs to be updated with the information about the extracted file and it s location.: Command: executable = path/globus-replica-catalog arguments = -host dc-n1.isi.edu/output/ ligo logicalfile H-H2:LSC-AS_Q F -location ISI Scenario 3.2: Full frame is located at ISI: In this case, because of the proximity of Caltech to ISI, the LDAS at Caltech must be used for extraction purposes. The plan for satisfying the user request would be as follow: 20

21 Plan for Scenario 3.2 If X == ISI then Send X to the LDAS system at Caltech. Extract the requested channel Send channel to the storage system at u requested location Y Update replica catalog at user requested location Y Point user to channel. The abstract form of the DAGMan would look like this: Copy frame file to LDAS At Caltech Extract requested channel Abstract DAGMan Copy Channel to storage At Y Update Replica Catalog The concrete DAGMan would be as follow: 21

22 globus_url_copy H:Q F From dataserver.phys.caltech.edu/ To dataserver.phys.caltech.edu/ LDAS_extract H-H2:LSC-AS_Q F From dataserver.phys.caltech.edu/ To dataserver.phys.caltech.edu/ Concrete DAGman globus_url_copy H-H2:LSC-AS_Q F From dataserver.phys.caltech.edu/ To dc-n1.isi.edu/output/ligo globus -replica -catalog -host dc-n1.isi.edu/output/ ligo logicalfile H -H2:LSC-AS_Q F -location ISI The details of the commands specified in the concrete DAGMan are similar to those of scenario 3.1. The only difference is that the extraction process takes place at the LDAS system at Caltech instead of UWM Scenario 3.3: The full frame is available at both ISI and UWM In this case we consider the user-requested output location. If the user wants the data at ISI, then we send the full frame file to Caltech for extraction and the extracted data from Caltech to ISI (as in scenario 3.1). If, however, the user wants the data at UWM, then the data is sent to UWM for extraction (as in scenario 3.2). The plan for satisfying the user s request in this case is as follow: 22

23 Scenario 3.3 X==UWM and X==ISI If Y=ISI Go to Scenario 3.1 If Y=UWM Go to Scenario Scenario 4: Full frame does not exist In this case the user should be issued a warning message. The plan would be as follow: Scenario 4 Frame file does not exist Send following message: DATA DOES NOT EXIST Note that the current intelligence as well as the geographical cost model in the planner is hardwired. 9 Executor Once the submit files and the DAG are generated they are submitted to a Condor-G server which starts a DAGMan [5] which analyses the DAG and then submits jobs in the manner in which they are defined in the DAG. These jobs may run locally or on a condor pool (e.g. the third party data transfers) or via a GRAM interface to LDAS ( the transform jobs, which materialize the data). As the jobs in the plan finish they write the exit status of the job to a log file, which is parsed continuously by the planner. If any of the jobs in the plan fail then the execution of the job halts and the error is reported to the user via the web page. Condor-G has a restart facility available by which a job can be resubmitted and it can continue from where it stopped but this feature is currently not being used. If all the jobs in a plan succeed then the output file is generated and the replica catalog is updated to reflect the availability of the data at the new location. 23

24 The user then has the option to view the data in the browser or use the saved data. Following is an example of the output the user sees on the computer screen while the request is being processed: ************* Welcome leila *************** *************The JOB ID = 15557************ *******output type is = LIGO-LWD******** The desired channel name is H2:LSC-AS_Q with timestamp globus-replica-catalog -host "ldap://dc-user.isi.edu:9129/lc=ligo_ _ , rc=ligo_replica_catalog, ou=isi.edu, o=netscaperoot" -password iplanet.passwd -collection -findlocations./ligodemo/replica/search_list attributes uc >./LIGODEMO/replica/loc_list Searching the Replica Catalog for an existing file...completed The Requested file doesnt exist. Will materialize it. globus-replica-catalog -host "ldap://dc-user.isi.edu:9129/lc=ligo_ _ , rc=ligo_replica_catalog, ou=isi.edu, o=netscaperoot" -password iplanet.passwd -collection -findlocations./ligodemo/replica/search_list attributes uc >./LIGODEMO/replica/loc_list Searching the Replica Catalog for an existing file...completed Raw data is available at ISI Scenario 3.1 The Collection name is ligo_ _ The LDAS jobmanager is ldas-grid.ligo.caltech.edu:2119/jobmanager-ldas Writing sub file transfer_a2b_ sub...successful ************************WILL TRANSFER H F****************** from gsiftp://birdie.isi.edu/~/griphyn/ligodata/frames/h f to gsiftp://ldasgrid.ligo.caltech.edu/~/../../opt/gsiftp_root/gsiftp_root/griphyn_test/ldas/ldas_outgoing/jobs/grid_in coming/h f 24

25 My infile is H F, my outfile is H-H2:LSC-AS_Q xml15557 Writing Sub File transform sub...successful ****************Will apply Transformation ******************* Writing sub file transfer_a2b_ sub...successful **********************WILL GET RESULT H-H2:LSC-AS_Q xml *********************** from gsiftp://ldasgrid.ligo.caltech.edu/~/../../opt/gsiftp_root/gsiftp_root/griphyn_test/ldas/ldas_outgoing/jobs/grid_ou tgoing/h-h2:lsc-as_q xml15557 to gsiftp://birdie.isi.edu/~/griphyn/ligodata/frames/h-h2:lsc-as_q xml Writing Dag File ldas dag...successful Writing CHECK TRANSFER File CheckTransfer dag...Successful Writing Transfer File CheckExtract dag...Successful Checking your DAG input file and all submit files it references. This might take a while... Done File for submitting this DAG to Condor :./LIGODEMO/ldas dag.condor.sub Log of DAGMan debugging messages :./LIGODEMO/ldas dag.dagman.out Log of Condor library debug messages :./LIGODEMO/ldas dag.lib.out Log of the life of condor_dagman itself :./LIGODEMO/ldas dag.dagman.log Condor Log file for all jobs of this DAG :./LIGODEMO/logs/15557.log Submitting job(s). Logging submit event(s). 1 job(s) submitted to cluster Submitting Dag to Condor...Successful ************JOB SUBMITTED************* globus-replica-catalog -host "ldap://dc-user.isi.edu:9129/lc=ligo_ _ , rc=ligo_replica_catalog, ou=isi.edu, o=netscaperoot" -manager "cn=directory Manager, ou = isi.edu, o = NetscapeRoot" -password iplanet.passwd -collection -add-filenames./ligodemo/replica/update_list add-anyway 2>error1 Updating Collection ligo_ _ for file H-H2:LSC-AS_Q xml...SUCCESS globus-replica-catalog -host "ldap://dc-user.isi.edu:9129/lc=ligo_ _ , rc=ligo_replica_catalog, ou=isi.edu, o=netscaperoot" -manager "cn=directory Manager, ou = isi.edu, o = NetscapeRoot" -password iplanet.passwd -location ISI -add-filenames./ligodemo/replica/update_list add-anyway 2>error2 Updating Location for ISI file H-H2:LSC-AS_Q xml...SUCCESS *********The data is available at ************ 25

26 gsiftp://birdie.isi.edu/~/griphyn/ligodata/frames/h-h2:lsc-as_q xml To view the data through the browser click on the link VIEW OUTPUT 10 Conclusions and Future Directions The main objective of the prototype described in this paper is to demonstrate the feasibility of integrating the various LIGO data and computing resources via an underlying Globus infrastructure. Moreover, the security is provided based on the Grid Security Infrastructure (GSI). This prototype is based on the Virtual Data concept: the idea that users need not know where their data is stored, combined with the idea that users need not know whether data has been computed yet. There are services to discover what it will cost to obtain data and the quality of derived data. The Virtual Data Grid (VDG) will serve requests efficiently, subject to global and policy constraints, and with strong security. We envision increasing the sophistication of the planning mechanism, the amount of data available, and the number of channels in the future. In addition, more transformations will be supported and eventually the prototype will be extended into a tool. 11 References [1] Condor-G: A Computation Management Agent for Multi-Institutional Grids. J. Frey, T. Tannenbaum, M. Livny, I. Foster, S. Tuecke, Proceedings of the Tenth International Symposium on High Performance Distributed Computing (HPDC-10), IEEE Press, August 2001 [2] XSIL: Java/XML for Scientists by Roy Williams, California Institute of Technology; [3] [4] [5] [6] [7] [8] 26

27 [9] K. Blackburn, A. Lazzarini, P. Ehrens, P. Shawhan, R. Williams, E. Deelman, C. Kesselman, G. Mehta, L. Meshkat, L. Pearlman, B. Allen, S. Koranda. GriPhyN and LIGO: Realizing Virtual Data Concepts- SC2001 Denver Nov 2001 [10] [11] Ewa Deelman, Carl Kesselman, Gaurang Mehta. Transformation Catalog Design for GriPhyN Oct

LIGO Virtual Data. Realizing. UWM: Bruce Allen, Scott Koranda. Caltech: Kent Blackburn, Phil Ehrens, Albert. Lazzarini, Roy Williams

LIGO Virtual Data. Realizing. UWM: Bruce Allen, Scott Koranda. Caltech: Kent Blackburn, Phil Ehrens, Albert. Lazzarini, Roy Williams Realizing LIGO Virtual Data Caltech: Kent Blackburn, Phil Ehrens, Albert Lazzarini, Roy Williams ISI: Ewa Deelman, Carl Kesselman, Gaurang Mehta, Leila Meshkat, Laura Pearlman UWM: Bruce Allen, Scott Koranda

More information

UNICORE Globus: Interoperability of Grid Infrastructures

UNICORE Globus: Interoperability of Grid Infrastructures UNICORE : Interoperability of Grid Infrastructures Michael Rambadt Philipp Wieder Central Institute for Applied Mathematics (ZAM) Research Centre Juelich D 52425 Juelich, Germany Phone: +49 2461 612057

More information

Part IV. Workflow Mapping and Execution in Pegasus. (Thanks to Ewa Deelman)

Part IV. Workflow Mapping and Execution in Pegasus. (Thanks to Ewa Deelman) AAAI-08 Tutorial on Computational Workflows for Large-Scale Artificial Intelligence Research Part IV Workflow Mapping and Execution in Pegasus (Thanks to Ewa Deelman) 1 Pegasus-Workflow Management System

More information

Introduction to Grid Computing

Introduction to Grid Computing Milestone 2 Include the names of the papers You only have a page be selective about what you include Be specific; summarize the authors contributions, not just what the paper is about. You might be able

More information

Chapter 4:- Introduction to Grid and its Evolution. Prepared By:- NITIN PANDYA Assistant Professor SVBIT.

Chapter 4:- Introduction to Grid and its Evolution. Prepared By:- NITIN PANDYA Assistant Professor SVBIT. Chapter 4:- Introduction to Grid and its Evolution Prepared By:- Assistant Professor SVBIT. Overview Background: What is the Grid? Related technologies Grid applications Communities Grid Tools Case Studies

More information

Globus Toolkit Firewall Requirements. Abstract

Globus Toolkit Firewall Requirements. Abstract Globus Toolkit Firewall Requirements v0.3 8/30/2002 Von Welch Software Architect, Globus Project welch@mcs.anl.gov Abstract This document provides requirements and guidance to firewall administrators at

More information

Transparent Grid Computing: a Knowledge-Based Approach

Transparent Grid Computing: a Knowledge-Based Approach Transparent Grid Computing: a Knowledge-Based Approach Jim Blythe, Ewa Deelman, Yolanda Gil, Carl Kesselman USC Information Sciences Institute 4676 Admiralty Way, Suite 1001 Marina Del Rey, CA 90292 {blythe,

More information

The Problem of Grid Scheduling

The Problem of Grid Scheduling Grid Scheduling The Problem of Grid Scheduling Decentralised ownership No one controls the grid Heterogeneous composition Difficult to guarantee execution environments Dynamic availability of resources

More information

Mapping Abstract Complex Workflows onto Grid Environments

Mapping Abstract Complex Workflows onto Grid Environments Mapping Abstract Complex Workflows onto Grid Environments Ewa Deelman, James Blythe, Yolanda Gil, Carl Kesselman, Gaurang Mehta, Karan Vahi Information Sciences Institute University of Southern California

More information

Web-based access to the grid using. the Grid Resource Broker Portal

Web-based access to the grid using. the Grid Resource Broker Portal Web-based access to the grid using the Grid Resource Broker Portal Giovanni Aloisio, Massimo Cafaro ISUFI High Performance Computing Center Department of Innovation Engineering University of Lecce, Italy

More information

Credentials Management for Authentication in a Grid-Based E-Learning Platform

Credentials Management for Authentication in a Grid-Based E-Learning Platform Credentials Management for Authentication in a Grid-Based E-Learning Platform Felicia Ionescu, Vlad Nae, Alexandru Gherega University Politehnica of Bucharest {fionescu, vnae, agherega}@tech.pub.ro Abstract

More information

A Distributed Media Service System Based on Globus Data-Management Technologies1

A Distributed Media Service System Based on Globus Data-Management Technologies1 A Distributed Media Service System Based on Globus Data-Management Technologies1 Xiang Yu, Shoubao Yang, and Yu Hong Dept. of Computer Science, University of Science and Technology of China, Hefei 230026,

More information

UNIT IV PROGRAMMING MODEL. Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus

UNIT IV PROGRAMMING MODEL. Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus UNIT IV PROGRAMMING MODEL Open source grid middleware packages - Globus Toolkit (GT4) Architecture, Configuration - Usage of Globus Globus: One of the most influential Grid middleware projects is the Globus

More information

DIRAC distributed secure framework

DIRAC distributed secure framework Journal of Physics: Conference Series DIRAC distributed secure framework To cite this article: A Casajus et al 2010 J. Phys.: Conf. Ser. 219 042033 View the article online for updates and enhancements.

More information

Grid Computing Fall 2005 Lecture 5: Grid Architecture and Globus. Gabrielle Allen

Grid Computing Fall 2005 Lecture 5: Grid Architecture and Globus. Gabrielle Allen Grid Computing 7700 Fall 2005 Lecture 5: Grid Architecture and Globus Gabrielle Allen allen@bit.csc.lsu.edu http://www.cct.lsu.edu/~gallen Concrete Example I have a source file Main.F on machine A, an

More information

Planning and Metadata on the Computational Grid

Planning and Metadata on the Computational Grid Planning and Metadata on the Computational Grid Jim Blythe, Ewa Deelman, Yolanda Gil USC Information Sciences Institute 4676 Admiralty Way, Suite 1001 Marina Del Rey, CA 90292 {blythe, deelman, gil}@isi.edu

More information

Architecture Proposal

Architecture Proposal Nordic Testbed for Wide Area Computing and Data Handling NORDUGRID-TECH-1 19/02/2002 Architecture Proposal M.Ellert, A.Konstantinov, B.Kónya, O.Smirnova, A.Wäänänen Introduction The document describes

More information

Knowledge Discovery Services and Tools on Grids

Knowledge Discovery Services and Tools on Grids Knowledge Discovery Services and Tools on Grids DOMENICO TALIA DEIS University of Calabria ITALY talia@deis.unical.it Symposium ISMIS 2003, Maebashi City, Japan, Oct. 29, 2003 OUTLINE Introduction Grid

More information

Day 1 : August (Thursday) An overview of Globus Toolkit 2.4

Day 1 : August (Thursday) An overview of Globus Toolkit 2.4 An Overview of Grid Computing Workshop Day 1 : August 05 2004 (Thursday) An overview of Globus Toolkit 2.4 By CDAC Experts Contact :vcvrao@cdacindia.com; betatest@cdacindia.com URL : http://www.cs.umn.edu/~vcvrao

More information

Data Management 1. Grid data management. Different sources of data. Sensors Analytic equipment Measurement tools and devices

Data Management 1. Grid data management. Different sources of data. Sensors Analytic equipment Measurement tools and devices Data Management 1 Grid data management Different sources of data Sensors Analytic equipment Measurement tools and devices Need to discover patterns in data to create information Need mechanisms to deal

More information

Grid Compute Resources and Job Management

Grid Compute Resources and Job Management Grid Compute Resources and Job Management How do we access the grid? Command line with tools that you'll use Specialised applications Ex: Write a program to process images that sends data to run on the

More information

Globus Toolkit 4 Execution Management. Alexandra Jimborean International School of Informatics Hagenberg, 2009

Globus Toolkit 4 Execution Management. Alexandra Jimborean International School of Informatics Hagenberg, 2009 Globus Toolkit 4 Execution Management Alexandra Jimborean International School of Informatics Hagenberg, 2009 2 Agenda of the day Introduction to Globus Toolkit and GRAM Zoom In WS GRAM Usage Guide Architecture

More information

COMPUTE CANADA GLOBUS PORTAL

COMPUTE CANADA GLOBUS PORTAL COMPUTE CANADA GLOBUS PORTAL Fast, user-friendly data transfer and sharing Jason Hlady University of Saskatchewan WestGrid / Compute Canada February 4, 2015 Why Globus? I need to easily, quickly, and reliably

More information

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy

30 Nov Dec Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Advanced School in High Performance and GRID Computing Concepts and Applications, ICTP, Trieste, Italy Why the Grid? Science is becoming increasingly digital and needs to deal with increasing amounts of

More information

Design of Distributed Data Mining Applications on the KNOWLEDGE GRID

Design of Distributed Data Mining Applications on the KNOWLEDGE GRID Design of Distributed Data Mining Applications on the KNOWLEDGE GRID Mario Cannataro ICAR-CNR cannataro@acm.org Domenico Talia DEIS University of Calabria talia@deis.unical.it Paolo Trunfio DEIS University

More information

A Comparison of Two Methods for Building Astronomical Image Mosaics on a Grid

A Comparison of Two Methods for Building Astronomical Image Mosaics on a Grid A Comparison of Two Methods for Building Astronomical Image Mosaics on a Grid Daniel S. Katz, Joseph C. Jacob Jet Propulsion Laboratory California Institute of Technology Pasadena, CA 91109 Daniel.S.Katz@jpl.nasa.gov

More information

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid

THE GLOBUS PROJECT. White Paper. GridFTP. Universal Data Transfer for the Grid THE GLOBUS PROJECT White Paper GridFTP Universal Data Transfer for the Grid WHITE PAPER GridFTP Universal Data Transfer for the Grid September 5, 2000 Copyright 2000, The University of Chicago and The

More information

Inca as Monitoring. Kavin Kumar Palanisamy Indiana University Bloomington

Inca as Monitoring. Kavin Kumar Palanisamy Indiana University Bloomington Inca as Monitoring Kavin Kumar Palanisamy Indiana University Bloomington Abstract Grids are built with multiple complex and interdependent systems to provide better resources. It is necessary that the

More information

Meeting the Challenges of Managing Large-Scale Scientific Workflows in Distributed Environments

Meeting the Challenges of Managing Large-Scale Scientific Workflows in Distributed Environments Meeting the Challenges of Managing Large-Scale Scientific s in Distributed Environments Ewa Deelman Yolanda Gil USC Information Sciences Institute Scientific s Current workflow approaches are exploring

More information

A SEMANTIC MATCHMAKER SERVICE ON THE GRID

A SEMANTIC MATCHMAKER SERVICE ON THE GRID DERI DIGITAL ENTERPRISE RESEARCH INSTITUTE A SEMANTIC MATCHMAKER SERVICE ON THE GRID Andreas Harth Yu He Hongsuda Tangmunarunkit Stefan Decker Carl Kesselman DERI TECHNICAL REPORT 2004-05-18 MAY 2004 DERI

More information

Globus GTK and Grid Services

Globus GTK and Grid Services Globus GTK and Grid Services Michael Rokitka SUNY@Buffalo CSE510B 9/2007 OGSA The Open Grid Services Architecture What are some key requirements of Grid computing? Interoperability: Critical due to nature

More information

Layered Architecture

Layered Architecture The Globus Toolkit : Introdution Dr Simon See Sun APSTC 09 June 2003 Jie Song, Grid Computing Specialist, Sun APSTC 2 Globus Toolkit TM An open source software toolkit addressing key technical problems

More information

Grid Architectural Models

Grid Architectural Models Grid Architectural Models Computational Grids - A computational Grid aggregates the processing power from a distributed collection of systems - This type of Grid is primarily composed of low powered computers

More information

An OGSI CredentialManager Service Jim Basney a, Shiva Shankar Chetan a, Feng Qin a, Sumin Song a, Xiao Tu a, and Marty Humphrey b

An OGSI CredentialManager Service Jim Basney a, Shiva Shankar Chetan a, Feng Qin a, Sumin Song a, Xiao Tu a, and Marty Humphrey b UK Workshop on Grid Security Experiences, Oxford 8th and 9th July 2004 An OGSI CredentialManager Service Jim Basney a, Shiva Shankar Chetan a, Feng Qin a, Sumin Song a, Xiao Tu a, and Marty Humphrey b

More information

Using the MyProxy Online Credential Repository

Using the MyProxy Online Credential Repository Using the MyProxy Online Credential Repository Jim Basney National Center for Supercomputing Applications University of Illinois jbasney@ncsa.uiuc.edu What is MyProxy? Independent Globus Toolkit add-on

More information

Index Introduction Setting up an account Searching and accessing Download Advanced features

Index Introduction Setting up an account Searching and accessing Download Advanced features ESGF Earth System Grid Federation Tutorial Index Introduction Setting up an account Searching and accessing Download Advanced features Index Introduction IT Challenges of Climate Change Research ESGF Introduction

More information

STORK: Making Data Placement a First Class Citizen in the Grid

STORK: Making Data Placement a First Class Citizen in the Grid STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar University of Wisconsin-Madison May 25 th, 2004 CERN Need to move data around.. TB PB TB PB While doing this.. Locate the data

More information

Multiple Broker Support by Grid Portals* Extended Abstract

Multiple Broker Support by Grid Portals* Extended Abstract 1. Introduction Multiple Broker Support by Grid Portals* Extended Abstract Attila Kertesz 1,3, Zoltan Farkas 1,4, Peter Kacsuk 1,4, Tamas Kiss 2,4 1 MTA SZTAKI Computer and Automation Research Institute

More information

A Metadata Catalog Service for Data Intensive Applications

A Metadata Catalog Service for Data Intensive Applications Metadata Catalog Service Draft August 5, 2002 A Metadata Catalog Service for Data Intensive Applications Ann Chervenak, Ewa Deelman, Carl Kesselman, Laura Pearlman, Gurmeet Singh Version 1.0 1 Introduction

More information

LDAP Synchronization

LDAP Synchronization LDAP Synchronization Version 1.6 Corresponding Software Version Celonis 4.3 This document is copyright of the Celonis SE. Distribution or reproduction are only permitted by written approval of the Celonis

More information

glite Grid Services Overview

glite Grid Services Overview The EPIKH Project (Exchange Programme to advance e-infrastructure Know-How) glite Grid Services Overview Antonio Calanducci INFN Catania Joint GISELA/EPIKH School for Grid Site Administrators Valparaiso,

More information

Grid Technologies & Applications: Architecture & Achievements

Grid Technologies & Applications: Architecture & Achievements Grid Technologies & Applications: Architecture & Achievements Ian Foster Mathematics & Computer Science Division, Argonne National Laboratory, Argonne, IL 60439, USA Department of Computer Science, The

More information

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms

Grid Computing. MCSN - N. Tonellotto - Distributed Enabling Platforms Grid Computing 1 Resource sharing Elements of Grid Computing - Computers, data, storage, sensors, networks, - Sharing always conditional: issues of trust, policy, negotiation, payment, Coordinated problem

More information

Delivering Data Management for Engineers on the Grid 1

Delivering Data Management for Engineers on the Grid 1 Delivering Data Management for Engineers on the Grid 1 Jasmin Wason, Marc Molinari, Zhuoan Jiao, and Simon J. Cox School of Engineering Sciences, University of Southampton, UK {j.l.wason, m.molinari, z.jiao,

More information

Managing Grid Credentials

Managing Grid Credentials Managing Grid Credentials Jim Basney http://www.ncsa.uiuc.edu/~jbasney/ Senior Research Scientist Grid and Security Technologies National Center for Supercomputing Applications

More information

JOB SUBMISSION ON GRID

JOB SUBMISSION ON GRID arxiv:physics/0701101v2 [physics.comp-ph] 12 Jan 2007 JOB SUBMISSION ON GRID An Users Introduction Rudra Banerjee ADVANCED COMPUTING LAB. Dept. of Physics, University of Pune March 13, 2018 Contents preface

More information

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia

Grid services. Enabling Grids for E-sciencE. Dusan Vudragovic Scientific Computing Laboratory Institute of Physics Belgrade, Serbia Grid services Dusan Vudragovic dusan@phy.bg.ac.yu Scientific Computing Laboratory Institute of Physics Belgrade, Serbia Sep. 19, 2008 www.eu-egee.org Set of basic Grid services Job submission/management

More information

Data Placement for Scientific Applications in Distributed Environments

Data Placement for Scientific Applications in Distributed Environments Data Placement for Scientific Applications in Distributed Environments Ann Chervenak, Ewa Deelman, Miron Livny 2, Mei-Hui Su, Rob Schuler, Shishir Bharathi, Gaurang Mehta, Karan Vahi USC Information Sciences

More information

SDS: A Scalable Data Services System in Data Grid

SDS: A Scalable Data Services System in Data Grid SDS: A Scalable Data s System in Data Grid Xiaoning Peng School of Information Science & Engineering, Central South University Changsha 410083, China Department of Computer Science and Technology, Huaihua

More information

Managing Large-Scale Scientific Workflows in Distributed Environments: Experiences and Challenges

Managing Large-Scale Scientific Workflows in Distributed Environments: Experiences and Challenges Managing Large-Scale Scientific s in Distributed Environments: Experiences and Challenges Ewa Deelman, Yolanda Gil USC Information Sciences Institute, Marina Del Rey, CA 90292, deelman@isi.edu, gil@isi.edu

More information

Grid Compute Resources and Grid Job Management

Grid Compute Resources and Grid Job Management Grid Compute Resources and Job Management March 24-25, 2007 Grid Job Management 1 Job and compute resource management! This module is about running jobs on remote compute resources March 24-25, 2007 Grid

More information

J. Basney, NCSA Category: Experimental October 10, MyProxy Protocol

J. Basney, NCSA Category: Experimental October 10, MyProxy Protocol GWD-E J. Basney, NCSA Category: Experimental October 10, 2005 MyProxy Protocol Status of This Memo This memo provides information to the Grid community. Distribution is unlimited. Copyright Notice Copyright

More information

By Ian Foster. Zhifeng Yun

By Ian Foster. Zhifeng Yun By Ian Foster Zhifeng Yun Outline Introduction Globus Architecture Globus Software Details Dev.Globus Community Summary Future Readings Introduction Globus Toolkit v4 is the work of many Globus Alliance

More information

Michigan Grid Research and Infrastructure Development (MGRID)

Michigan Grid Research and Infrastructure Development (MGRID) Michigan Grid Research and Infrastructure Development (MGRID) Abhijit Bose MGRID and Dept. of Electrical Engineering and Computer Science The University of Michigan Ann Arbor, MI 48109 abose@umich.edu

More information

What makes workflows work in an opportunistic environment?

What makes workflows work in an opportunistic environment? What makes workflows work in an opportunistic environment? Ewa Deelman 1 Tevfik Kosar 2 Carl Kesselman 1 Miron Livny 2 1 USC Information Science Institute, Marina Del Rey, CA deelman@isi.edu, carl@isi.edu

More information

Downloading and Managing Firmware in Cisco UCS Central

Downloading and Managing Firmware in Cisco UCS Central Downloading and Managing Firmware in Cisco UCS Central This chapter includes the following sections: Downloading Firmware from Cisco.com, page 1 Firmware Library of Images, page 2 Configuring Firmware

More information

DIRAC Distributed Secure Framework

DIRAC Distributed Secure Framework DIRAC Distributed Secure Framework A Casajus Universitat de Barcelona E-mail: adria@ecm.ub.es R Graciani Universitat de Barcelona E-mail: graciani@ecm.ub.es on behalf of the LHCb DIRAC Team Abstract. DIRAC,

More information

A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS

A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS A RESOURCE MANAGEMENT FRAMEWORK FOR INTERACTIVE GRIDS Raj Kumar, Vanish Talwar, Sujoy Basu Hewlett-Packard Labs 1501 Page Mill Road, MS 1181 Palo Alto, CA 94304 USA { raj.kumar,vanish.talwar,sujoy.basu}@hp.com

More information

High Performance Computing Course Notes Grid Computing I

High Performance Computing Course Notes Grid Computing I High Performance Computing Course Notes 2008-2009 2009 Grid Computing I Resource Demands Even as computer power, data storage, and communication continue to improve exponentially, resource capacities are

More information

A Replica Location Grid Service Implementation

A Replica Location Grid Service Implementation A Replica Location Grid Service Implementation Mary Manohar, Ann Chervenak, Ben Clifford, Carl Kesselman Information Sciences Institute, University of Southern California Marina Del Rey, CA 90292 {mmanohar,

More information

Pegasus Workflow Management System. Gideon Juve. USC Informa3on Sciences Ins3tute

Pegasus Workflow Management System. Gideon Juve. USC Informa3on Sciences Ins3tute Pegasus Workflow Management System Gideon Juve USC Informa3on Sciences Ins3tute Scientific Workflows Orchestrate complex, multi-stage scientific computations Often expressed as directed acyclic graphs

More information

Guidelines on non-browser access

Guidelines on non-browser access Published Date: 13-06-2017 Revision: 1.0 Work Package: Document Code: Document URL: JRA1 AARC-JRA1.4F https://aarc-project.eu/wp-content/uploads/2017/03/aarc-jra1.4f.pdf 1 Table of Contents 1 Introduction

More information

A General Data Grid: Framework and Implementation

A General Data Grid: Framework and Implementation A General Data Grid: Framework and Implementation Wu Zhang, Jian Mei, and Jiang Xie Department of Computer Science and Technology, Shanghai University, Shanghai, 200072, China zhang@mail.shu.edu.cn, meijian_2003@yahoo.com.cn

More information

Grid Programming: Concepts and Challenges. Michael Rokitka CSE510B 10/2007

Grid Programming: Concepts and Challenges. Michael Rokitka CSE510B 10/2007 Grid Programming: Concepts and Challenges Michael Rokitka SUNY@Buffalo CSE510B 10/2007 Issues Due to Heterogeneous Hardware level Environment Different architectures, chipsets, execution speeds Software

More information

A Resource Discovery Algorithm in Mobile Grid Computing based on IP-paging Scheme

A Resource Discovery Algorithm in Mobile Grid Computing based on IP-paging Scheme A Resource Discovery Algorithm in Mobile Grid Computing based on IP-paging Scheme Yue Zhang, Yunxia Pei To cite this version: Yue Zhang, Yunxia Pei. A Resource Discovery Algorithm in Mobile Grid Computing

More information

A Cloud-based Dynamic Workflow for Mass Spectrometry Data Analysis

A Cloud-based Dynamic Workflow for Mass Spectrometry Data Analysis A Cloud-based Dynamic Workflow for Mass Spectrometry Data Analysis Ashish Nagavaram, Gagan Agrawal, Michael A. Freitas, Kelly H. Telu The Ohio State University Gaurang Mehta, Rajiv. G. Mayani, Ewa Deelman

More information

Grid Data Management

Grid Data Management Grid Data Management Data Management Distributed community of users need to access and analyze large amounts of data Fusion community s International ITER project Requirement arises in both simulation

More information

Grid Middleware and Globus Toolkit Architecture

Grid Middleware and Globus Toolkit Architecture Grid Middleware and Globus Toolkit Architecture Lisa Childers Argonne National Laboratory University of Chicago 2 Overview Grid Middleware The problem: supporting Virtual Organizations equirements Capabilities

More information

Applying Chimera Virtual Data Concepts to Cluster Finding in the Sloan Sky Survey

Applying Chimera Virtual Data Concepts to Cluster Finding in the Sloan Sky Survey Applying Chimera Virtual Data Concepts to Cluster Finding in the Sloan Sky Survey James Annis 1 Yong Zhao 2 Jens Voeckler 2 Michael Wilde 3 Steve Kent 1 Ian Foster 2,3 1 Experimental Astrophysics, Fermilab,

More information

Scheduling Data-Intensive Workflows onto Storage-Constrained Distributed Resources

Scheduling Data-Intensive Workflows onto Storage-Constrained Distributed Resources Scheduling Data-Intensive Workflows onto Storage-Constrained Distributed Resources Arun Ramakrishnan 1, Gurmeet Singh 2, Henan Zhao 3, Ewa Deelman 2, Rizos Sakellariou 3, Karan Vahi 2 Kent Blackburn 4,

More information

A Resource Discovery Algorithm in Mobile Grid Computing Based on IP-Paging Scheme

A Resource Discovery Algorithm in Mobile Grid Computing Based on IP-Paging Scheme A Resource Discovery Algorithm in Mobile Grid Computing Based on IP-Paging Scheme Yue Zhang 1 and Yunxia Pei 2 1 Department of Math and Computer Science Center of Network, Henan Police College, Zhengzhou,

More information

Advanced School in High Performance and GRID Computing November Introduction to Grid computing.

Advanced School in High Performance and GRID Computing November Introduction to Grid computing. 1967-14 Advanced School in High Performance and GRID Computing 3-14 November 2008 Introduction to Grid computing. TAFFONI Giuliano Osservatorio Astronomico di Trieste/INAF Via G.B. Tiepolo 11 34131 Trieste

More information

Microsoft Recertification for MCSE: Server Infrastructure. Download Full Version :

Microsoft Recertification for MCSE: Server Infrastructure. Download Full Version : Microsoft Recertification for MCSE: Server Infrastructure Download Full Version : https://killexams.com/pass4sure/exam-detail/ Answer: C QUESTION: 99 Your company has an office in New York. Many users

More information

Pegasus. Pegasus Workflow Management System. Mats Rynge

Pegasus. Pegasus Workflow Management System. Mats Rynge Pegasus Pegasus Workflow Management System Mats Rynge rynge@isi.edu https://pegasus.isi.edu Automate Why workflows? Recover Automates complex, multi-stage processing pipelines Enables parallel, distributed

More information

Federated Services for Scientists Thursday, December 9, p.m. EST

Federated Services for Scientists Thursday, December 9, p.m. EST IAM Online Federated Services for Scientists Thursday, December 9, 2010 1 p.m. EST Rachana Ananthakrishnan Argonne National Laboratory & University of Chicago Jim Basney National Center for Supercomputing

More information

Novell Access Manager

Novell Access Manager Setup Guide AUTHORIZED DOCUMENTATION Novell Access Manager 3.1 SP3 February 02, 2011 www.novell.com Novell Access Manager 3.1 SP3 Setup Guide Legal Notices Novell, Inc., makes no representations or warranties

More information

vcloud Director Administrator's Guide

vcloud Director Administrator's Guide vcloud Director 5.1.1 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions of

More information

Configuring SSL. SSL Overview CHAPTER

Configuring SSL. SSL Overview CHAPTER CHAPTER 8 Date: 4/23/09 This topic describes the steps required to configure your ACE (both the ACE module and the ACE appliance) as a virtual Secure Sockets Layer (SSL) server for SSL initiation or termination.

More information

A Compact Computing Environment For A Windows PC Cluster Towards Seamless Molecular Dynamics Simulations

A Compact Computing Environment For A Windows PC Cluster Towards Seamless Molecular Dynamics Simulations A Compact Computing Environment For A Windows PC Cluster Towards Seamless Molecular Dynamics Simulations Yuichi Tsujita Abstract A Windows PC cluster is focused for its high availabilities and fruitful

More information

Based on: Grid Intro and Fundamentals Review Talk by Gabrielle Allen Talk by Laura Bright / Bill Howe

Based on: Grid Intro and Fundamentals Review Talk by Gabrielle Allen Talk by Laura Bright / Bill Howe Introduction to Grid Computing 1 Based on: Grid Intro and Fundamentals Review Talk by Gabrielle Allen Talk by Laura Bright / Bill Howe 2 Overview Background: What is the Grid? Related technologies Grid

More information

COGNOS (R) 8 COGNOS CONNECTION USER GUIDE USER GUIDE THE NEXT LEVEL OF PERFORMANCE TM. Cognos Connection User Guide

COGNOS (R) 8 COGNOS CONNECTION USER GUIDE USER GUIDE THE NEXT LEVEL OF PERFORMANCE TM. Cognos Connection User Guide COGNOS (R) 8 COGNOS CONNECTION USER GUIDE Cognos Connection User Guide USER GUIDE THE NEXT LEVEL OF PERFORMANCE TM Product Information This document applies to Cognos (R) 8 Version 8.1.2 MR2 and may also

More information

Clouds: An Opportunity for Scientific Applications?

Clouds: An Opportunity for Scientific Applications? Clouds: An Opportunity for Scientific Applications? Ewa Deelman USC Information Sciences Institute Acknowledgements Yang-Suk Ki (former PostDoc, USC) Gurmeet Singh (former Ph.D. student, USC) Gideon Juve

More information

Creating Personal Adaptive Clusters for Managing Scientific Jobs in a Distributed Computing Environment

Creating Personal Adaptive Clusters for Managing Scientific Jobs in a Distributed Computing Environment 1 Creating Personal Adaptive Clusters for Managing Scientific Jobs in a Distributed Computing Environment Edward Walker, Jeffrey P. Gardner, Vladimir Litvin, and Evan L. Turner Abstract We describe a system

More information

Introduction to GT3. Introduction to GT3. What is a Grid? A Story of Evolution. The Globus Project

Introduction to GT3. Introduction to GT3. What is a Grid? A Story of Evolution. The Globus Project Introduction to GT3 The Globus Project Argonne National Laboratory USC Information Sciences Institute Copyright (C) 2003 University of Chicago and The University of Southern California. All Rights Reserved.

More information

The GridWay. approach for job Submission and Management on Grids. Outline. Motivation. The GridWay Framework. Resource Selection

The GridWay. approach for job Submission and Management on Grids. Outline. Motivation. The GridWay Framework. Resource Selection The GridWay approach for job Submission and Management on Grids Eduardo Huedo Rubén S. Montero Ignacio M. Llorente Laboratorio de Computación Avanzada Centro de Astrobiología (INTA - CSIC) Associated to

More information

A Login Shell interface for INFN-GRID

A Login Shell interface for INFN-GRID A Login Shell interface for INFN-GRID S.Pardi2,3, E. Calloni1,2, R. De Rosa1,2, F. Garufi1,2, L. Milano1,2, G. Russo1,2 1Università degli Studi di Napoli Federico II, Dipartimento di Scienze Fisiche, Complesso

More information

Scalable, Reliable Marshalling and Organization of Distributed Large Scale Data Onto Enterprise Storage Environments *

Scalable, Reliable Marshalling and Organization of Distributed Large Scale Data Onto Enterprise Storage Environments * Scalable, Reliable Marshalling and Organization of Distributed Large Scale Data Onto Enterprise Storage Environments * Joesph JaJa joseph@ Mike Smorul toaster@ Fritz McCall fmccall@ Yang Wang wpwy@ Institute

More information

Managing large-scale workflows with Pegasus

Managing large-scale workflows with Pegasus Funded by the National Science Foundation under the OCI SDCI program, grant #0722019 Managing large-scale workflows with Pegasus Karan Vahi ( vahi@isi.edu) Collaborative Computing Group USC Information

More information

Grid Computing Middleware. Definitions & functions Middleware components Globus glite

Grid Computing Middleware. Definitions & functions Middleware components Globus glite Seminar Review 1 Topics Grid Computing Middleware Grid Resource Management Grid Computing Security Applications of SOA and Web Services Semantic Grid Grid & E-Science Grid Economics Cloud Computing 2 Grid

More information

Load testing with WAPT: Quick Start Guide

Load testing with WAPT: Quick Start Guide Load testing with WAPT: Quick Start Guide This document describes step by step how to create a simple typical test for a web application, execute it and interpret the results. A brief insight is provided

More information

Planning the SCEC Pathways: Pegasus at work on the Grid

Planning the SCEC Pathways: Pegasus at work on the Grid Planning the SCEC Pathways: Pegasus at work on the Grid Philip Maechling, Vipin Gupta, Thomas H. Jordan Southern California Earthquake Center Ewa Deelman, Yolanda Gil, Sridhar Gullapalli, Carl Kesselman,

More information

Configuring SSL. SSL Overview CHAPTER

Configuring SSL. SSL Overview CHAPTER 7 CHAPTER This topic describes the steps required to configure your ACE appliance as a virtual Secure Sockets Layer (SSL) server for SSL initiation or termination. The topics included in this section are:

More information

Condor and Workflows: An Introduction. Condor Week 2011

Condor and Workflows: An Introduction. Condor Week 2011 Condor and Workflows: An Introduction Condor Week 2011 Kent Wenger Condor Project Computer Sciences Department University of Wisconsin-Madison Outline > Introduction/motivation > Basic DAG concepts > Running

More information

Cloud Computing. Up until now

Cloud Computing. Up until now Cloud Computing Lecture 4 and 5 Grid: 2012-2013 Introduction. Up until now Definition of Cloud Computing. Grid Computing: Schedulers: Condor SGE 1 Summary Core Grid: Toolkit Condor-G Grid: Conceptual Architecture

More information

Grid portal solutions: a comparison of GridPortlets and OGCE

Grid portal solutions: a comparison of GridPortlets and OGCE CONCURRENCY AND COMPUTATION: PRACTICE AND EXPERIENCE Published online 7 June 2007 in Wiley InterScience (www.interscience.wiley.com)..1112 Grid portal solutions: a comparison of GridPortlets and OGCE Chongjie

More information

Troubleshooting Grid authentication from the client side

Troubleshooting Grid authentication from the client side Troubleshooting Grid authentication from the client side By Adriaan van der Zee RP1 presentation 2009-02-04 Contents The Grid @NIKHEF The project Grid components and interactions X.509 certificates, proxies

More information

A Simple Mass Storage System for the SRB Data Grid

A Simple Mass Storage System for the SRB Data Grid A Simple Mass Storage System for the SRB Data Grid Michael Wan, Arcot Rajasekar, Reagan Moore, Phil Andrews San Diego Supercomputer Center SDSC/UCSD/NPACI Outline Motivations for implementing a Mass Storage

More information

Grid Computing Security: A Survey

Grid Computing Security: A Survey Grid Computing Security: A Survey Basappa B. Kodada, Shiva Kumar K. M Dept. of CSE Canara Engineering College, Mangalore basappabk@gmail.com, shivakumarforu@rediffmail.com Abstract - This paper provides

More information

GLOBUS TOOLKIT SECURITY

GLOBUS TOOLKIT SECURITY GLOBUS TOOLKIT SECURITY Plamen Alexandrov, ISI Masters Student Softwarepark Hagenberg, January 24, 2009 TABLE OF CONTENTS Introduction (3-5) Grid Security Infrastructure (6-15) Transport & Message-level

More information

Data Management for Distributed Scientific Collaborations Using a Rule Engine

Data Management for Distributed Scientific Collaborations Using a Rule Engine Data Management for Distributed Scientific Collaborations Using a Rule Engine Sara Alspaugh Department of Computer Science University of Virginia alspaugh@virginia.edu Ann Chervenak Information Sciences

More information