Data Integration Platform for Streaming Analytics​

To address their Big Data processing requirements, a client wanted to create a custom-built Data integration platform to provide near real-time access for processing and analytics of mainframe and streaming data.

However, the traditional Data Warehousing approach of ETL and staging areas limited their ability to deliver results to the business. An innovative solution was sought, which is where Alumnus was brought in. ​

We developed an end-to-end data transfer pipeline for static data from varied sources (Mainframes, applications using Cassandra / Teradata / Redis / JDBC / HDFS etc.)  to non-mainframe target technologies.

Extract Data (Mainframe, JDBC, Hadoop, RabbitMQ, SNMP, NetFlow), Transform, Load & Analyze (MongoDB, Cassandra, HDFS, Kafka)

In addition, a pipeline was also custom-built for real-time transfer and visualization of streaming data from varied technologies (RabbitMQ / SNMP / sFlow / Netflow etc.) to a Kafka-based target system.  ​

This streaming analytics capability was delivered on an AWS platform using Java / Nodejs technology while avoiding the cost and delay of traditional ETL and staging processes.