A Layer Based Architecture for Provenance in Big Data

Similar documents
SAF: A Provenance-Tracking Framework for Interoperable Semantic Applications

7 th International Digital Curation Conference December 2011

Towards Provenance Aware Comment Tracking for Web Applications

Based on Big Data: Hype or Hallelujah? by Elena Baralis

Provenance-aware Faceted Search in Drupal

Managing Rapidly-Evolving Scientific Workflows

Nowcasting. D B M G Data Base and Data Mining Group of Politecnico di Torino. Big Data: Hype or Hallelujah? Big data hype?

POMELo: A PML Online Editor

Data Provenance in Distributed Propagator Networks

Application of Named Graphs Towards Custom Provenance Views

Provenance-Aware Faceted Search in Drupal

Embedded Technosolutions

The International Journal of Digital Curation Volume 7, Issue

OPQL: A First OPM-Level Query Language for Scientific Workflow Provenance

Comparing Provenance Data Models for Scientific Workflows: an Analysis of PROV-Wf and ProvOne

Computer-based Tracking Protocols: Improving Communication between Databases

ProvDS: Uncertain Provenance Management over Incomplete Linked Data Streams

Introduction to Data Science

Provenance Map Orbiter: Interactive Exploration of Large Provenance Graphs

USC Viterbi School of Engineering

Implementation of a reporting workflow to maintain data lineage for major water resource modelling projects

A Cloud Framework for Big Data Analytics Workflows on Azure

Cloud Computing Techniques for Big Data and Hadoop Implementation

Wearable Technology Orientation Using Big Data Analytics for Improving Quality of Human Life

CE4031 and CZ4031 Database System Principles

Introduction to Text Mining. Hongning Wang

RESTORE: REUSING RESULTS OF MAPREDUCE JOBS. Presented by: Ahmed Elbagoury

PharmaSUG. companies. This paper. will cover how. processes, a fairly linear. before moving. be carried out. Lifecycle. established.

W3C Provenance Incubator Group: An Overview. Thanks to Contributing Group Members

Implementation of Privacy Mechanism using Curve Fitting Method for Data Publishing in Health Care Domain

Kepler/pPOD: Scientific Workflow and Provenance Support for Assembling the Tree of Life

Managing Exploratory Workflows

Fujitsu World Tour 2018

In the recent past, the World Wide Web has been witnessing an. explosive growth. All the leading web search engines, namely, Google,

PPKM: Preserving Privacy in Knowledge Management

Integrating Text and Graphics to Present Provenance Information

Collaborative provenance for workflow-driven science and engineering Altintas, I.

BIG DATA & HADOOP: A Survey

Continuous auditing certification

Tackling the Provenance Challenge One Layer at a Time

DataFinder A Scientific Data Management Solution ABSTRACT

Provenance Collection Support in the Kepler Scientific Workflow System

The 2018 (14th) International Conference on Data Science (ICDATA)

Visualization Architecture for User Interaction with Dynamic Data Spaces in Multiple Pipelines

Big Data - Some Words BIG DATA 8/31/2017. Introduction

Managing the Evolution of Dataflows with VisTrails Extended Abstract

Provenance for MapReduce-based Data-Intensive Workflows

Abstract Provenance Graphs: Anticipating and Exploiting Schema-Level Data Provenance

Data Governance for the Connected Enterprise

Financial Dataspaces: Challenges, Approaches and Trends

Bigtable. A Distributed Storage System for Structured Data. Presenter: Yunming Zhang Conglong Li. Saturday, September 21, 13

Performance Evaluation of a MongoDB and Hadoop Platform for Scientific Data Analysis

Survey Paper on Traditional Hadoop and Pipelined Map Reduce

Automatic Identification of Application I/O Signatures from Noisy Server-Side Traces. Yang Liu Raghul Gunasekaran Xiaosong Ma Sudharshan S.

Review - Relational Model Concepts

Lecture 25 Overview. Last Lecture Query optimisation/query execution strategies

Cybersecurity eit. Software. Certification. Industrial Security Embedded System

CE4031 and CZ4031 Database System Principles

Big Data for Compliance and Industry Trends. Nadeem Asghar Field CTO Financial Services - Hortonworks

The DataBridge: A Social Network for Long Tail Science Data!

/ Cloud Computing. Recitation 7 October 10, 2017

Updates through Views

Analysis of Approaches for Supporting the Open Provenance Model: A Case Study of the Trident Workflow Workbench

A Review to the Approach for Transformation of Data from MySQL to NoSQL

Integrating Oracle Databases with NoSQL Databases for Linux on IBM LinuxONE and z System Servers

MAPR DATA GOVERNANCE WITHOUT COMPROMISE

An Introduction to Big Data Formats

Fast and Effective System for Name Entity Recognition on Big Data

WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS

CIRGDISCO at RepLab2012 Filtering Task: A Two-Pass Approach for Company Name Disambiguation in Tweets

Performance Evaluation of NoSQL Databases

Science 2.0 VU Big Science, e-science and E- Infrastructures + Bibliometric Network Analysis

D DAVID PUBLISHING. Big Data; Definition and Challenges. 1. Introduction. Shirin Abbasi

When, Where & Why to Use NoSQL?

Open Data Standards for Administrative Data Processing

Taccumulation of the social network data has raised

GUEST EDITORS INTRODUCTION: SPECIAL ISSUE ON BIG DATA INTELLIGENCE AND COMPUTING

Copyright is owned by the Author of the thesis. Permission is given for a copy to be downloaded by an individual for the purpose of research and

Filtering Unwanted Messages from (OSN) User Wall s Using MLT

New research on Key Technologies of unstructured data cloud storage

Secure Conjunctive Keyword Ranked Search over Encrypted Cloud Data

ACCI Recommendations on Long Term Cyberinfrastructure Issues: Building Future Development

A Review on Privacy Preserving Data Mining Approaches

/ Cloud Computing. Recitation 8 October 18, 2016

Derivation Rule Dependency and Data Provenance Semantics

Election Analysis and Prediction Using Big Data Analytics

A Survey Paper on NoSQL Databases: Key-Value Data Stores and Document Stores

Developing a Research Data Policy

CS6200 Information Retrieval. Jesse Anderton College of Computer and Information Science Northeastern University

PLATFORM AND SOFTWARE AS A SERVICE THE MAPREDUCE PROGRAMMING MODEL AND IMPLEMENTATIONS

Efficient Algorithm for Frequent Itemset Generation in Big Data

Considerations of Analysis of Healthcare Claims Data

Survey on Incremental MapReduce for Data Mining

6 TOOLS FOR A COMPLETE MARKETING WORKFLOW

System Approach for Single Keyword Search for Encrypted data files Guarantees in Public Infrastructure Clouds

RETRACTED ARTICLE. Web-Based Data Mining in System Design and Implementation. Open Access. Jianhu Gong 1* and Jianzhi Gong 2

NEUROSEED WHITEPAPER. Version 1.1. May 1, 2018 Contents of the white paper are subject to changes and improvements

WebLab PROV : Computing fine-grained provenance links for XML artifacts

Principles of Dataspaces

Evolution of Database Systems

Transcription:

2014 IEEE International Conference on Big Data A Layer Based Architecture for Provenance in Big Data Rajeev Agrawal Ashiq Imran Cameron Seay Jessie Walker Computer Systems Technology North Carolina A&T State University Greensboro, USA ragrawal@ncat.edu Computer Science North Carolina A&T State University Greensboro, USA aimran@aggies.ncat.edu Computer Systems Technology North Carolina A&T State University Greensboro, USA cwseay@ncat.edu Mathematics and Computer Science University of Arkansas at Pine Bluff Pine Bluff, USA walkerjj@uapb.edu provenance to validate physical objects in arts, literary works, manuscript etc. Recently, the domain for provenance has gained significance attention in digital world and escience. The provenance of data is crucial for validating, debugging, auditing, evaluating the quality of data and determining reliability of data. In today s period of Internet, complex ecosystems of data are even more ubiquitous. It is not a surprise that data provenance has become a major research topic for conferences and workshops. Provenance has been typically considered by the database, workflows and distributed system communities. Capturing provenance can be a burdensome and labor intensive task. Provenance capture is often incorporated into a workflow system [1]. Provenance information has a broad range of critical application areas. For example, scientific data processing needs to keep track of data ownership and action performed in workflow to ensure the trust assigned to the output data. In business environments, provenance of documents is even more important, e.g., for regulatory compliance and litigations. Let us illustrate following scenario to get a better idea about provenance. Bob is a doctor of a hospital. Paul is a patient who visits to Bob for checkup. After diagnosis of Paul s health, Bob passes the prescription of Paul to another doctor Peter. After reviewing the prescription and Peter makes some changes, he sends it to another doctor James for doing some test. The provenance of the document should contain the chain (Bob, Peter, James, along with some information about the process names that prescribed in the document. So, provenance information is for keeping tracking the changes. Important information can be verified using previously collected provenance records. But provenance in big data is relatively unexplored topic and has not got much attention so far. Big data is expensive to manage and hard to extract value from. It is important to manage big data because more data may lead to more precise analysis and more precise analysis may lead to more confidence in decision making. Applying provenance in a different layer can provide various level of useful information [2]. In our paper, we propose a layer based architecture to use the provenance data. The main goal of the proposed architecture is to improve robustness and flexibility for the collection and management of provenance data. A lot of factors have been Abstract Big data is a new technology wave that makes the world awash in data. Various organizations accumulate data that are difficult to exploit. Government databases, social media, healthcare databases etc. are the examples of the big data. Big data covers absorbing and analyzing huge amount of data that may have originated or processed outside of the organization. Data provenance can be defined as origin and process of data. It carries significant information of a system. It can be useful for debugging, auditing, measuring performance and trust in data. Data provenance in big data is relatively unexplored topic. It is necessary to appropriately track the creation and collection process of the data to provide context and reproducibility. In this paper, we propose an intuitive layer based architecture of data provenance and visualization. In addition, we show a complete workflow of tracking provenance information of big data. Keywords-Provenance; Big data; Query; Visualization I. INTRODUCTION Big data is a prevalent term but not wholly new area in information technology. Many of the foremost research based agencies such as NASA, NSF, NIH along with various intelligence agencies have been involved with Big data. Social networking websites such as Facebook, Twitter, Linked-In, Google+, Youtube etc. are getting popular day by day. These sites deal with huge amount of data every day. Big data is getting importance not for its quantity of data rather the data can be used to extract a lot of information from it. Whenever we find some data, do we think what the source of data is? It is quite possible that data is copied from somewhere else. It is also possible that information is incorrect. The data we usually see on the web, such as, rating of a movie, smart phone, or news story, do we think about it, how much legitimate is the information? We have also learned to be careful before relaying on the information extracted from the Internet. For scientists, they need to have confidence on accuracy and timeliness on the data that they are working on. Data provenance refers the description of the origin, creation and propagation process of data. Data provenance is the lineage and derivation of the data. It stores ownership and process history about data objects. Provenance has been studied extensively in the past and people usually use 978-1-4799-5666-1/14/$31.00 2014 IEEE 1

considered for designing the architecture. Each layer has different role to play. We introduce new component visualization toolkit to visualize provenance data. Both administrator and user can access visualization toolkit to get a graphical overview in provenance information. We impose authorization for security and privacy on provenance data. Each user can only be able to view the changes that user made. Administrator can have access to all the users provenance data. Precisely, administrator has the flexibility to view provenance information of different user. It provides protection on provenance data. We use MongoDB [3] as a storage which will be discussed in later parts. The main contributions of this paper are as follows: A technique for capturing provenance in big data is provided. We present a layer based architecture of provenance collection and visualization. We introduce an access control mechanism in provenance data. The rest of the paper is organized as follows. Section II provides a discussion of related works on provenance. Section III describes detailed overview of our proposed layer based architecture, and the various terms associated with it. Section IV describes the overview of components of the proposed architecture. We summarize our architecture and discuss some future improvements of it and the various security and privacy issues related to trustworthy provenance in section V. Finally, the paper is concluded in section VI. II. RELATED WORKS Big data concept refers to a database which is continuously expanding that makes difficult to control and manage. The difficulty can be related to data capture, storage, search, sharing, analytics and visualization etc. Provenance in big data has been identified by a recent community whitepaper on the challenges and opportunities of big data [4]. Provenance has been studied extensively in the past, in particular in archival theory, and for the purpose of asserting authenticity of literary work, arts, manuscripts etc. Provenance has traditionally been collected at the workflow level often making it hard to capture relevant information about resource characteristics and is difficult for users to easily incorporate in existing workflows. Multiple provenance tools have been established for workflow e.g. VisTrails [5] and Kepler [6]. In recent years, we have got systems RAMP [7] that uses provenance collection from MapReduce workflows [8]. The use of provenance in determining the quality of scientific data and data provenance has also been shown [9]. Some applications have used provenance data in debugging (e.g., to track an erroneous data item back to the source from which it was derived), trust (e.g., by merging trust scores for the data in a data item's provenance), probabilistic data (the probability of a query result can be computed from the probabilities of the data items in its provenance [10, 11]. Provenance may be practiced across various domains where different standards are used, driving needs for interoperability which is currently satisfied by the Open Provenance Model (OPM) [12]. PROB [13], a toolkit for tracking provenance data has been recently proposed for Big data. This paper provides a diagram on scientific workflow on big data. Initially, from the subjects or phenomena, raw data is collected. After processing of the raw data, preprocessed data is made. This preprocessed data is ready to load into the cluster file system. After doing analysis, derived data is produced. This derived data can be visualized or described in publications. Finally, publications are distributed and shared. Provenance information can be used for security and privacy issues. Some challenges and solutions has been discussed about secure provenance in [14]. A framework for a family of Provenance-based Access Control (PBAC) models with extensions has been introduced that can handle traditional access control issues [15]. Big data provenance can be used to detect and evaluate performance blockages, to figure out performance metrics, and to test a system's ability to exploit commonalities in data and processing [16]. Provenance data can be useful to play the role of benchmarking. In addition, they show an example of monitoring and profiling using provenance information. Log files contain a lot of useful information that generate data objects. A model for detecting and accumulating data from log files has been used in [17]. A universal data provenance framework is proposed for data provenance information for real-world applications without any code modifications [2]. The ORCHESTRA [18], collaborative data sharing system describes provenance model and architecture. Our work tries to capture and represents from multiple layers. From the big data storage, it captures the provenance data. From the collection of provenance data, user and administrator can perform query in an intuitive manner. In addition, user and administrator can also visualize captured provenance information. III. LAYER BASED ARCHITECTURE Before starting to describe our layer based architecture, it is important to know what the advantages of layer based architecture are. The main advantage is that this architecture is possible to change the contents of any one of layers without having to make corresponding changes in any of the others. This enables parallel development of the different layers of the application. Complex application rules are easy to implement in application server. Any future changes can be easily incorporated. Figure 1 shows our proposed architecture for provenance collection, querying and visualization for scientific applications. We propose a layered architecture for exploring and collecting provenance. Our proposed architecture is divided into 3 layers. 2

1) Storage Layer: Vast number of information stores in Big data storage. This layer is for capturing provenance data and save it to provenance database. Storage layer collection component is responsible for capturing provenance data. Provenance information is extracted from big data storage. Provenance data itself is not big data. 2) Application Layer: Application layer is composed of two components: Query and Visualization toolkit. Query component is for accessing query from the provenance data. Visualization toolkit component supports graphical view of provenance data which may be useful for getting more knowledge about provenance information. 3) Access Layer: Access layer is for security and privacy purposes. This layer will provide privacy among users. If a user wants to check provenance record that user needs to login to verify her authorization. A user can only get access of her own provenance information whereas administrator can check every user s provenance information for security purposes. The amount of provenance data may differ according to the application and also the implementation done by the user. Provenance is captured in all forms and sizes, our architecture does not store provenance in a well-structured manner, focuses instead on the ability to write the captured provenance in a fast and efficient manner. It provides a way to collect all raw provenance data from the systems. Provenance data can be queried directly or further processed and transformed into different formats and more structured provenance. By separating the access layer, application layer, storage layer, optimizations on provenance collection is possible. A. Use case We consider a use case for provenance collection and storage system. The use case is for user provenance data which is stored into database. Whenever, user makes any changes on data, we keep track of the changes. For example, on each user session, user commits data into database. We need to capture user data submission. In our context, we capture user id, output of submission, execution time, user notations. Additionally, user can visualize his provenance information. B. Design Goals Less overhead: We need to process huge volume of data, so high performance is critical. It is necessary that provenance collection has minor impact on application s performance [19]. User annotations support: It is important to capture user notes or metadata. This is applicable for database and workflows as well as for big data. Thus, we need an interface that allows users to add their notes about the experiment and output. Visualization support: Provenance data can be used for security and privacy reasons. If we are able to visualize the provenance data, we will clearly retrieve the ownership and access history of data. Scalability: Scalability comes into play for big data. The volume of big data is growing exponentially. Provenance data is also rapidly increasing and therefore making it necessary to scale up provenance collection. Various data models support: So far data models for provenance system have been in structured databases. It is important to not only support for structured as well as semi structured provenance data models from users and systems because big data may not follow a particular structure. Secure Provenance: There is huge volume of data and information. It is important to maintain privacy and security of provenance information. Flexibility: A typical approach to deal with velocity of the data is to introduce data cleaning and integration in pay per use fashion. This may reduce the cost and consume less time [16]. C. Procedure Figure 1 shows the design of layered architecture. Our motive is to capture, query and visualize provenance information. After collecting provenance from big data storage, we keep our provenance information in provenance data storage. Initially, user needs to login to verify her identity whether she is an administrator or just a user. A user is an individual user who is not allowed viewing provenance information of other users. Administrator can view provenance information of all users. This is the only difference between an administrator and a user. Moreover, user submits query to provenance database. User gets the result of the query. However, administrator will get different result than a user. User can submit request to visualization toolkit and get chart of changes the user made in a timely manner. Administrator can also view provenance information using visualization toolkit. 3

Figure 1. Overview of Layer based Architecture Figure 2. A workflow diagram of tracking provenance in big data 4

IV. OVERVIEW OF COMPONENTS A. Provenenance Collection Figure 2 shows the workflow diagram of tracking provenance. In the beginning, we have big data subject or phenomenon such as the data of social networking site. We collect some portion of this paradigm of data. We get the raw data which may be structured or semi-structured. Then we process data using big data analytics and load the preprocessed data into databases. Databases can be of various types such as graph database and relational database. Relational database can be used for handling structured data. Graph database is useful for handling semi-structured or unstructured database since it contains graph like characteristics. Each node has a link with adjacent node. This allows graph database systems to utilize graph theory to very rapidly examine the connections and interconnectedness of nodes and how Netflix can recommend videos for you. After carefully analyzing and recognizing patterns or structure, we come to a decision that this data may be meaningful. Then we finally publish the data and distribute data to High Performance Computing (HPC) system. To clarify the shapes in the following figure 2, the parallelogram shape denotes data in different stage and rectangular shape denotes the process. Provenance data is captured whenever any changes occur from the beginning to the end. B. Storage The captured provenance data is sundry and wideranging. So, we need a data storage that must have capability of supporting semi-structured provenance document. In our case, we use an open source NoSQL data store that is MongoDB. It is document oriented and scalable. It was developed in C++. It is appropriate for our needs since we need to deal with semi-structured data. It also supports sharding. Sharding allows storing data records across multiple machines. MongoDB stores all data in documents, which are JSON-style data structures composed of field-andvalue pairs [3]. In our context, captured provenance data will be stored into MongoDB data store. MongoDB stores documents on disk in the BSON serialization format. BSON is a binary representation of JSON documents, though it contains more data types than JSON. Figure 3 shows the storage format of MongoDB. For file management, we use the GridFS specification for storing the actual file contents since it provides metadata and large file support. GridFS uses two collections for storing data. Large data objects are divided into chunk that are stored in chunk collection and metadata is stored in files collection. Figure 4 describes the prototype of a document files collection. Figure 3. Data storing format of MongoDB. Figure 4. GridFS files collection prototype in MongoDB. C. Query In our implementation, we consider building query interface to access the resulting provenance data. The query language is designed to abstract the underlying language used at the storage layer. This storage layer allows the user to query in an instinctive manner without knowing in depth knowledge. In our implementation, we have both command line and user interface for the users and administrators. The command line interface is meant to be used by users who are comfortable with the command line. On the other hand, user interface is meant supporting fairly open-ended provenance queries on the collection. D. Visualization Our approach includes visualization toolkit to provide an interactive and graphical view of provenance data. Visualization toolkit is different for administrator and user. For administrator, we consider two types of visualization: 1) User based visualization: This technique is only applicable for administrator. Administrator can have a view and select desired user to check provenance information. Figure 5 shows the user based visualization. 5

2) Time based visualization: It is applicable for both administrator and user. Figure 6 shows the time based visualizatin for administrator in a typical day. Figure 7 shows the time based visualization for user in monthly manner. In our design, administrator has the flexibilty of choosing time in hourly, daily, weekly, monthly basis. User can only view his provenance record in weekly or monthly basis. Figure 5. Provenance information (user based). Figure 6. Provenance information in a typical day (time based). Figure 7. Provenance information in a year by a user (time based). E. Administrator Administrator is a super user who has the authorization for checking global provenance information. Administrator has the capability to individually select user and view that user s provenance record. F. User User is an authorized person who has the capability to query and view his provenance information. Provenance data is a sensitive information. Without authorization, user can t check his provenance information. V. DISCUSSION In this paper, we describe layer based storage management architecture and workflow of tracking provenance. This supports both structured and semistructured provenance collection. Provenance captured in storage layer is a mixture of data and process provenance. It can be used for a number of use cases. Our approach is useful in some cases and falls short in others. We consider structured and semi-structured data. Our proposed architecture covers all of the design goals as discussed in section III. It is simple so it may not take much overhead. We provide multiple data model support to handle structured and semi structured data. We introduce visualization to have a clear view on provenance information. Since we use MongoDB database which support scalability. We include user annotation support which gives every user flexibility to add some custom notation in provenance information. Each user can write suitable notation according to his need. Provenance initiation is user controlled but entire process is automated. Our aim is to capture provenance close to raw format. We have introduced access control to preserve security and privacy of provenance data of each user. Administrator can visualize provenance information of all the users. Users can only visualize provenance of their own. Though for protecting provenance information only access control is not enough, we may need some additional policies. But, proper authorization can give protection of provenance information upto some extent. In future, we will impose other regulations and policies to make provenance data more secure. We may use digital signature and encryption method to increase security. The big data explosion has seen a number of NoSQL data stores for scalability. In our paper, we use MongoDB which support both structured and semi structured provenance document. The way we manage provenance data, it will be easy for the user to know where data objects are stored. Users can verify the NoSQL data store using regular expressions to search for previous job runs for a variety of information. VI. CONCLUSION In this paper, we discuss some design goals of provenance in big data. We present the design of a layer based architecture for provenance in big data. We divide the 6

architecture into 3 layers. Each layer has a significant role to play. We introduce access control mechanism to serve security aspect. Our design is applicable for simple queries. We have just proposed this design as capturing provenance record. We are in the process of implementing our architecture. In future work, we will add additional components to handle complex queries. We may use provenance information as a benchmark of performance measurement. We will measure the performance of our proposed architecture in High Performance Computing (HPC) systems. ACKNOWLEDGMENT This research work was funded by the Office of National Consortium of Data Science (NCDS). REFERENCES [1] Jiang, W., Hu, C., Pasupathy, S., Kanevsky, A., Li, Z., and Zhou, Y.: Understanding Customer Problem Troubleshooting from Storage System Logs, in Editor (Ed.)^(Eds.): Book Understanding Customer Problem Troubleshooting from Storage System Logs (edn.), pp. [2] Gessiou, E., Pappas, V., Athanasopoulos, E., Keromytis, A.D., and Ioannidis, S.: Towards a universal data provenance framework using dynamic instrumentation : Information Security and Privacy Research (Springer, 2012), pp. 103-114 [3] MongoDB, http://www.mongodb.org, Last accessed: July, 2014 [4] Bertino, E., Bernstein, P., Agrawal, D., Davidson, S., Dayal, U., Franklin, M., Gehrke, J., Haas, L., Halevy, A., and Han, J.: Challenges and Opportunities with Big Data, 2011 [5] Callahan, S.P., Freire, J., Santos, E., Scheidegger, C.E., Silva, C.T., and Vo, H.T.: VisTrails: visualization meets data management, in Editor (Ed.)^(Eds.): Book VisTrails: visualization meets data management (ACM, 2006, edn.), pp. 745-747 [6] Ludäscher, B., Altintas, I., Berkley, C., Higgins, D., Jaeger, E., Jones, M., Lee, E.A., Tao, J., and Zhao, Y.: Scientific workflow management and the Kepler system, Concurrency and Computation: Practice and Experience, 2006, 18, (10), pp. 1039-1065 [7] Park, H., Ikeda, R., and Widom, J.: Ramp: A system for capturing and tracing provenance in mapreduce workflows, 2011 [8] Dean, J., and Ghemawat, S.: MapReduce: simplified data processing on large clusters, Communications of the ACM, 2008, 51, (1), pp. 107-113 [9] Buneman, P., and Davidson, S.B.: Data provenance--the foundation of data quality, in Editor (Ed.)^(Eds.): Book Data provenance--the foundation of data quality (2013, edn.), pp. [10] Aggarwal, C.C.: Trio A System for Data Uncertainty and Lineage : Managing and Mining Uncertain Data (Springer, 2009), pp. 1-35 [11] Karvounarakis, G., and Green, T.J.: Semiring-annotated data: queries and provenance?, ACM SIGMOD Record, 2012, 41, (3), pp. 5-14 [12] Moreau, L., Clifford, B., Freire, J., Futrelle, J., Gil, Y., Groth, P., Kwasnikowska, N., Miles, S., Missier, P., Myers, J., Plale, B., Simmhan, Y., Stephan, E., and Van den Bussche, J.: The Open Provenance Model core specification (v1.1), Future Generation Computer Systems, 2011, 27, (6), pp. 743-756 [13] Korolev, V., Joshi, A., Korolev, V., Grasso, M.A., Joshi, A., Grasso, M.A., Dalvi, D., Das, S., Korolev, V., and Yesha, Y.: PROB: A tool for Tracking Provenance and Reproducibility of Big Data Experiments, Reproduce'14. HPCA 2014, 2014, 11, pp. 264-286 [14] Hasan, R., Sion, R., and Winslett, M.: Introducing secure provenance: problems and challenges, in Editor (Ed.)^(Eds.): Book Introducing secure provenance: problems and challenges (ACM, 2007, edn.), pp. 13-18 [15] Park, J., Nguyen, D., and Sandhu, R.: A provenance-based access control model, in Editor (Ed.)^(Eds.): Book A provenance-based access control model (IEEE, 2012, edn.), pp. 137-144 [16] Glavic, B.: Big Data Provenance: Challenges and Implications for Benchmarking : Specifying Big Data Benchmarks (Springer, 2014), pp. 72-80 [17] Ghoshal, D., and Plale, B.: Provenance from log files: a BigData problem. Proc. Proceedings of the Joint EDBT/ICDT 2013 Workshops, Genoa, Italy2013 pp. Pages [18] Green, T.J., Karvounarakis, G., Ives, Z.G., and Tannen, V.: Provenance in ORCHESTRA, 2010 [19] Cheah, Y.-W., Canon, R., Plale, B., and Ramakrishnan, L.: Milieu: Lightweight and Configurable Big Data Provenance for Science, in Editor (Ed.)^(Eds.): Book Milieu: Lightweight and Configurable Big Data Provenance for Science (IEEE, 2013, edn.), pp. 46-53 7