Lenses 2.1 Enterprise Features PRODUCT DATA SHEET

Similar documents
Deploying SQL Stream Processing in Kubernetes with Ease

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara

Security and Performance advances with Oracle Big Data SQL

Overview. Prerequisites. Course Outline. Course Outline :: Apache Spark Development::

Hortonworks DataFlow Sam Lachterman Solutions Engineer

Event Streams using Apache Kafka

Oracle Big Data SQL. Release 3.2. Rich SQL Processing on All Data

MAPR DATA GOVERNANCE WITHOUT COMPROMISE

Kafka Connect the Dots

Bring Context To Your Machine Data With Hadoop, RDBMS & Splunk

microsoft

The SMACK Stack: Spark*, Mesos*, Akka, Cassandra*, Kafka* Elizabeth K. Dublin Apache Kafka Meetup, 30 August 2017.

Is NiFi compatible with Cloudera, Map R, Hortonworks, EMR, and vanilla distributions?

Docker Universal Control Plane Deploy and Manage On-Premises, Your Dockerized Distributed Applications

Qualys Cloud Platform

Syncsort DMX-h. Simplifying Big Data Integration. Goals of the Modern Data Architecture SOLUTION SHEET

Search Engines and Time Series Databases

Developing Microsoft Azure Solutions (70-532) Syllabus

StreamSets Control Hub Installation Guide

Transformation-free Data Pipelines by combining the Power of Apache Kafka and the Flexibility of the ESB's

Down the event-driven road: Experiences of integrating streaming into analytic data platforms

HDP Security Overview

HDP Security Overview

Big Data Hadoop Developer Course Content. Big Data Hadoop Developer - The Complete Course Course Duration: 45 Hours

Oracle Big Data Connectors

A10 HARMONY CONTROLLER

Oracle NoSQL Database Enterprise Edition, Version 18.1

Datameer for Data Preparation:

Data Acquisition. The reference Big Data stack

TEN LAYERS OF CONTAINER SECURITY

Information empowerment for your evolving data ecosystem

FAQs. Business (CIP 2.2) AWS Market Place Troubleshooting and FAQ Guide

Increase Value from Big Data with Real-Time Data Integration and Streaming Analytics

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Enabling Universal Authorization Models using Sentry

Un'introduzione a Kafka Streams e KSQL and why they matter! ITOUG Tech Day Roma 1 Febbraio 2018

Architecting Microsoft Azure Solutions (proposed exam 535)

Mesosphere and Percona Server for MongoDB. Jeff Sandstrom, Product Manager (Percona) Ravi Yadav, Tech. Partnerships Lead (Mesosphere)

Developing Microsoft Azure Solutions (70-532) Syllabus

Fluentd + MongoDB + Spark = Awesome Sauce

Provide Real-Time Data To Financial Applications

Hortonworks Data Platform

CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM

Data Acquisition. The reference Big Data stack

MODERN BIG DATA DESIGN PATTERNS CASE DRIVEN DESINGS

A day in the life of a log message Kyle Liberti, Josef

The Hadoop Ecosystem. EECS 4415 Big Data Systems. Tilemachos Pechlivanoglou

Mesosphere and Percona Server for MongoDB. Peter Schwaller, Senior Director Server Eng. (Percona) Taco Scargo, Senior Solution Engineer (Mesosphere)

BIG DATA COURSE CONTENT

We are ready to serve Latest Testing Trends, Are you ready to learn?? New Batches Info

Developing Microsoft Azure Solutions (70-532) Syllabus

IBM Data Replication for Big Data

Modernizing Business Intelligence and Analytics

In-memory data pipeline and warehouse at scale using Spark, Spark SQL, Tachyon and Parquet

Hadoop & Big Data Analytics Complete Practical & Real-time Training

Innovatus Technologies

Introduction to Hadoop. High Availability Scaling Advantages and Challenges. Introduction to Big Data

CERTIFICATE IN SOFTWARE DEVELOPMENT LIFE CYCLE IN BIG DATA AND BUSINESS INTELLIGENCE (SDLC-BD & BI)

The Now Platform Reference Guide

IoT Sensor Analytics with Apache Kafka, KSQL and TensorFlow

AT&T Flow Designer. Current Environment

Red Hat JBoss Middleware Integration Products Roadmap. Ken Johnson Director, Product Management, Red Hat

Oracle Data Integrator 12c: Integration and Administration

Building Event Driven Architectures using OpenEdge CDC Richard Banville, Fellow, OpenEdge Development Dan Mitchell, Principal Sales Engineer

How to Keep UP Through Digital Transformation with Next-Generation App Development

Overview SENTINET 3.1

Oracle NoSQL Database Enterprise Edition, Version 18.1

Energy Management with AWS

Hadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved

API Gateway Version October Concepts Guide

Schema Registry Overview

@unterstein #bedcon. Operating microservices with Apache Mesos and DC/OS

Hortonworks DataFlow

WHITEPAPER. MemSQL Enterprise Feature List

Big Data Architect.

Oracle Data Integrator 12c: Integration and Administration

Using Apache Phoenix to store and access data

Cisco Tetration Analytics

70-532: Developing Microsoft Azure Solutions

Axway API Gateway. Version 7.4.1

Blended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a)

STATE OF MODERN APPLICATIONS IN THE CLOUD

Introducing Kafka Connect. Large-scale streaming data import/export for

Big Data. Big Data Analyst. Big Data Engineer. Big Data Architect

Data pipelines with PostgreSQL & Kafka

How to choose the right approach to analytics and reporting

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

Cloudline Autonomous Driving Solutions. Accelerating insights through a new generation of Data and Analytics October, 2018

Spatial Analytics Built for Big Data Platforms

WHITE PAPER. Reference Guide for Deploying and Configuring Apache Kafka

Let the data flow! Data Streaming & Messaging with Apache Kafka Frank Pientka. Materna GmbH

Overview. : Cloudera Data Analyst Training. Course Outline :: Cloudera Data Analyst Training::

Franck Mercier. Technical Solution Professional Data + AI Azure Databricks

Big Data Analytics using Apache Hadoop and Spark with Scala

Oracle Enterprise Manager 12c IBM DB2 Database Plug-in

Sentinet for BizTalk Server SENTINET

Flash Storage Complementing a Data Lake for Real-Time Insight

Apache Ignite - Using a Memory Grid for Heterogeneous Computation Frameworks A Use Case Guided Explanation. Chris Herrera Hashmap

VMware Cloud on AWS Technical Deck VMware, Inc.

by Cisco Intercloud Fabric and the Cisco

Transcription:

Lenses 2.1 Enterprise Features PRODUCT DATA SHEET 1

OVERVIEW DataOps is the art of progressing from data to value in seconds. For us, its all about making data operations as easy and fast as using the email. The main pillars are: Data accessibility Build real time flows Ensure compliance and security Get more from you data with Lenses and DataOps, leverage your data and existing data skills to iterate quickly and move to production faster. Focus on: Leveraging your whole organisation Data pipelines over infrastructure Configuration over code Data analytics over engineering Operations made easy Lenses, a DataOps platform, provides cutting edge visibility into data streams, integration, processing, operations and governance for the modern data-driven enterprise. Accelerate time to value, open up data streams to a wide audience, enable rapid construction and deployment of data pipelines at scale, for enterprises, with governance and security. DataOps is currently transforming data management. As a result, many end up with customised DIY solutions with inhouse engineering teams spending most of their time building infrastructure and tooling. Data should be in the hands of its users, as simple as their email, enabling innovation and minimising time to value. Building a data-driven culture allows all data personas to participate across the entire business. Lenses makes this ACCESS DATA Self-service SQL for universal data access, data preparation via different clients & channels For All Users Data Consumers Data Engineers Data Officers DATA FLOWS Free up time from building infrastructure with self service SQL, Kubernetes integration and monitoring flows out of the box. DATA GOVERNANCE & SECURITY We take care of the enterprise features required to operate your platform such as fine grade access control, multi-tenancy features, authorisation and authentication integrations, data obfuscation. APACHE KAFKA DISTRIBUTIONS APACHE KAFKA Apache Kafka 0.10.2.0+ CONFLUENT Confluent Platform 3.2.1+ with Apache Kafka 0.10.2.0+ CLOUDERA Cloudera Manager 5.9.x - Cloudera Distribution of Kafka 2.2.0 - Based on 0.10.2.0+ Cloudera Manager 5.13.x - Cloudera Distribution of Kafka 3.0.0 - Based on 0.11.0 HORTONWORKS HortonWorks HDP 2.6.1+ Apache Kafka 0.10.1.1+ MANAGE, SECURE, AND GOVERN YOUR DATA SECURITY LDAP, Kerberos, TLS Certificates, Robe based access CLIENTS JDBC, ReduxJS, Java, Python DATA LINEAGE Auditing, Processor & Application Topology, Schema Management ETL READY 25+ Landoop Kafka Connectors to simplify your streaming ETL data pipelines MONITOR 150+ Key Indicator metrics of your infrastructure, continuously monitored, 60 Days low level infrastructure metrics ALERT Setup and Receive alerts 2 All rights reserved. Lenses is trademark and registered trademarks of Landoop Ltd. in the United Kingdom and other countries. All other brand names, product names, or trademarks belong to their respective owners.

Features & Description Manage Data Lenses SQL Engine Transform data using SQL, regular expressions, nested data operators, etc View more Supported Formats Complex Types Data Preparation Insert Data Primitive Types, JSON, XML, CSV, Avro, Protobuf, Custom Serializers Lenses SQL support for complex types, arrays, unions, structs, maps Apply transformations, calculations, and aggregations Insert data into topics with payload suggestions.i.e. Testing event-driven microservices Delete Data Replay offsets Delete data from compacted topics Edit consumer commit offsets to replay messages Access Data Web UI Endpoints CLI JDBC Driver Native Clients Continuous Queries Query Management Access Metadata Avro Schemas Subscribe & Query Live or Retained Data Streams and view data on the browser. Intelisense SQL Editor Quick Filter Navigate to Partitions & Offsets Time travelling Access Metadata (ie. key, partitions, timestamp, offsets, headers) Different Data Views: Tree, Grid, Raw Download query result data sets Rest / WebSocket endpoints to integrate with data via Lenses. You can fire queries and subscribe to data for your custom data integrations with security applied. Lenses Command Line Interface to automate running queries and fit your CI/CD stream JDBC driver plugin to access data in motion for your custom apps or BI tools Python client for Jupyter notebooks and more Redux middleware for building real-time dashboards on live queries Scalable continuous queries over Kubernetes with stream processing capabilities. See more on Lenses SQL Engine. View and Manage live queries from users via the admin CLI. Metastore engine: Lenses identifies the structure of your data for the key and the value and represents them in a schema, for any data format, which can also be updated. Key/Value types: Lenses autodetect the data format of the key and value data format. Users can also override types for each topic. AVRO messages, accommodate for schema evolution, and enforce a data contract on your messages, being a fast and useful data serialization. Lenses integrates with widely used Schema Registries to manage schemas with RBAC. 3

Data Analysis & Preparation Know your data Access Data Structure & Organise Data Cleanse Data Transform Data Move Data Extract Insights Analytics Distributed Joins Self-service Infrastructure Anonymization Discover, Explore and Query your Data to understand how events and signals from the are captured Easy access to data irrespective of the underlying data format or serialization protocol Organize and structure your data in efficient data formats Filter, map and prepare the data for your business domain Apply the necessary transformations and aggregations and enrich data on the fly Move data between applications and source and target data systems Perform fast SQL operations for instant response to BI dashboards, reports, and machine learning applications. Lenses SQL integrates with native clients like JDBC driver, ReduxJS middleware for custom frontend apps, Python for integration with Jupyter or other notebooks. Continuous scalable queries for real-time filtering, enrichment and data transformation. Perform counts and anomaly detection use cases with a simple query. Output data to multiple systems and generate live dashboards to meet your SLAs Express massively parallel distributed joins, without having to worry about partitioning schemantics Move your SQL logic into production in seconds, and enjoy monitoring alerting and control over your data pipelines Anonymize and Obfuscate data while it transitions across your staging area Data Transformation & Stream Processing SQL Processors SQL Processors Scalability Native scale via Connect Workers Native scale via Kubernetes SQL Processors Capabilities SQL processors are continuously unbounded SQL queries, specifically for streaming data. Each query is providing its internal topology which is visualised in Lenses UI as interactive nodes, where we have live input and output data. SQL processors provide 3 execution modes: In process: which is good for development purposes Kubernetes: where each runner is managed by a Kubernetes pod Connect: where each runner is orchestrated by Kafka connect workers Distribute your stream-processing jobs to Kafka-Connect clusters. Lenses SQL stream processors provide a native integration with Connect workers to scale up or down. The number of processor runners can simply be specified via Lenses UI. Run and manage your data stream processing application within Kubernetes. Lenses SQL stream processors provide a native integration with Kubernetes to scale up or down. The number of processor runners can simply be specified via Lenses UI. Lenses manages the pod lifecycle and exposes the logs for debugging purposes. Self-service data preparation Topic to Topic transformations Filter and Enrich data Join the stream (with repartition support) Aggregations 4

Processing with Custom Apps Monitoring Alerting Lenses provides support for custom streaming applications. You can integrate the custom application nodes in the global topology view and leverage internal flow views and monitoring metrics for each app. Native support for Spark, Kafka Streams and Akka Streams For each processing entity (SQL or Custom) Lenses collects performance metrics for which threshold alerts can be applied. For example data in/out, rates, data size. Threshold alerts can be specified for the input data for each SQL processor Data Pipelines / Streaming ETL ready Global Topology Lenses Source Connectors Lenses Sink Connectors SQL for Ingestion Custom Connectors Change Data Capture SQL Stream Processing Custom Stream Processing Manage Lifecycle Import / Exports Multi-tenancy Alerting Lenses identifies the related entities (topics, connectors, processors) and constructs the pipeline for each flow. Topology has interactive nodes and Bring data onto Kafka from multiple sources: Twitter, BlockChain, MQTT, JMS, RDBMS, FTP, File location, Yahoo, Bloomberg, Blockchain, Cassandra, CoaP, ReThink and more. Popular distributed NoSQL and caching systems: Cassandra, DocumentDB, Elastic Search, SolR, HDFS, RDBMS, HazelCast, InfluxDB, MongoDB, HBase, Kudu, Redis, Pulsar, VoltDB, Rethink and others Lenses connectors (see https://github.com/landoop/stream-reactor) support SQL configuration to give extra capabilities and integrate with target systems Lenses is compatible with all Kafka connectors as it implements the connect API. Lenses makes connectors available for spin up instances as well as available in the global topology. Extract Change Data Capture (CDC) events from MainFrame or RDBMS via bin-logs. SQL processors can be part of your ETL Flow (see Data Transformation) Hook custom applications to your flows. Native support for Spark, Kafka Streams and Akka Streams and compatible with other types. (see Data Transformation) Admin actions for Connectors & Processors, to edit, scale, start/stop. Lenses CLI supports import/export of different elements as configuration, to create repeatable flows and promote across environments. Lenses provides multitenancy support via blacklisting/whitelisting which can protect the relevant flows via topic-centric configuration. Restricts the access to involved connectors, processors and consumers. Apply data consumption threshold alerts for each entity involved: consumers, sink connectors, SQL processors Security Role Based Access Basic Authentication LDAP Kerberos TLS Certificates Lenses supports role-based authorisation to protect your cluster and your data, especially in a multi-tenant environment. Also provides a no-data role for operation and support teams Token-based authentication scenarios Integrate via memberof or pluggable LDAP, or AD (Active Directory) and provide the people in your organization role-based access to Kafka streaming data JaaS, Utilise Kerberos Authentication, Authorization, Single sign-on Authenticate and Authorise Clients and Application via signed TLS certificates 5

ACLs Quotas Multi-tenancy Manage Access Control Lists (Kafka ACLs) via Lenses interface / CLI Manage Kafka quotas for clients via Lenses Interface / CLI Blacklist and Whitelist capabilities for user groups. It applies topic-centric security which can restrict access to various topics and its involved pipelines Governance & Lineage Audit Logs Audit Data access Data Schemas Topologies Query Topologies Immutable audit logs to track all CRUD actions in a who did what and when. Lenses also audits actions happen outside Lenses, by periodically monitoring your infrastructure. Compliance is hard when it comes to data access so Lenses is logging all queries and data access for its users. Lenses keeps its own metastore for data types and schemas and also integrates with widely used Schema Registries to poll the schemas, including schema management with RBAC. Visualizes history of Avro schemas and ensures evolution. See Browse Data. Application / Topology level view of data-pipelines, to highlight relations of data with private or GDPR regulated data. Visualise the internal data processing flow structure for continuous queries and custom applications that processing streaming data. Kafka Monitoring Lenses Services overview Data Metrics Consumer Lag Metrics API Prometheus Exporters Grafana Dashboards Monitor Client Apps Alert Manager Infrastructure Alerts Threshold Alerts Monitor the health status of the services involved in your infrastructure exposing basic metrics via JMX (Brokers, Schema Registry, Connects, Zookeeper) Topic Data metrics for infrastructure, to identify the most used topics, how topic traffic looks like and how data have been distributed across partitions. Ensure sync replicas. Calculated and exposed consumer lag for each consumer group and its instances. Lenses exposes calculated metrics via an endpoint for further integration Templated agents for exporting critical key-metrics of the infrastructure, and reporting to Prometheus. Collection of Grafana templates, curated over many years with suggestions and additions from our clients in production environments. Monitor Cluster overview, Kafka Metrics, Infrastructure, Consumer/Producer rates, Client Apps. Our metrics exporter works with your custom JVM applications and we give visual metrics for each application Integrates with Alert Manager for alert routing management Built-in and configurable infrastructure and domain alerts Create custom alerts for the different consumer groups. 6

More Resources Lenses SQL Engine for Browsing & Processing Data Lenses Resources About Us Landoop creator of Lenses, is a London based company. Our aim is to help companies open up their business data to all relevant users seamlessly by giving them the power of data operations. DataOps is currently transforming data management. Building a data-driven culture mandates that all data personas work with data, enabling participation from the entire business. Organisations are trying hard to expose their data via platform teams. As a result, many end up with customised DIY solutions with in-house engineering teams spending most of their time building infrastructure and tooling. Data should be in the hands of its users, as simple as their email, to enable innovation and minimise time to value. Give Lenses a spin at www.landoop.com Contact us at info@landoop.com All rights reserved. Lenses is trademark and registered trademarks of Landoop Ltd. in the United Kingdom and other countries. All other brand names, product names, or trademarks belong to their respective owners. 7