XenonStack

A Stack Innovator

Post Top Ad

Showing posts with label iot. Show all posts
Showing posts with label iot. Show all posts

Thursday, 26 December 2019

12/26/2019 05:34:00 pm

Predictive Healthcare Analytics Platform


Predictive Healthcare Analytics Platform

Predictive Healthcare Analytics Platform and solutions is a part of advanced analytics that is used to predict future events. Predictive analytics practices many ways from data mining, statistics, modeling, machine learning, and artificial intelligence to examine current data to make forecasts about the future.
  • Good healthcare boosts the economy of the nation. Precision medicine along with Big Data is leveraging in building better patient profiles as well as predictive models to diagnose and treat diseases.
  • TeleMedicine and AI in healthcare is indeed a miracle remotely performing treatment of patients using Pattern Recognition, optimizing duty allocation, monitoring live data.
  • Real-Time Big Data for Infection Control to predict and prevent infections through networks creating safer environments.
  • Patient Data Analytics for a patient dealing and preventing readmissions and better pharmaceutical supply chain management and delivery.

Challenges for Building Predictive Analytics Platform

  • Interface for the patient to search nearby doctors by particular Healthcare categories.
  • Enable patient visibility to see doctor’s availability online and communicate via text chat, audio or video call.
  • Visible allotment number to the patient in the waiting queue.
  • Communicate with the doctor as well as test or medicine suggestions to the patient.
  • Interface for the patient to contact nearby labs to collect a sample and upload test reports on the server followed by the push notification when the report is ready.
  • Share report with a doctor followed by a prescription to the patient.
  • Search for nearby medical stores and place an order for the prescription got from the doctor.

Solution Offerings for Real-Time Monitoring

Develop a Healthcare platform to fully automate using the latest technologies and distributed Agile development methods.

Real-Time Monitoring of User’s Events

Apache Kafka & Spark Streaming to achieve high concurrency, set up low latency messaging platform Apache Kafka to receive Real-Time user requests from REST APIs (acting as Kafka producer).
Apache Spark Streaming (processing and Computing engine) Spark-Cassandra connector, stored 1 million events per second in Cassandra. Built Analytics Data Pipeline using Kafka and Spark Streaming to capture user’s clicks, cookies, and other data to know users better.
Microservices using Spring Cloud, NetFlix OSS, Consul, Docker, and Kubernetes
Develop REST API’s using Microservices architecture with Spring Cloud and Spring Boot Framework using Java language. Moreover, use Async support of the Spring framework to create Async controllers that make REST API easily scalable.
Spring to deploy REST and use Kubernetes for secure containers and their management. For API gateway, use NetFlix Eureka Server which acts as a proxy for REST API and the lot of Microservices, Consul as DNS enables auto-discovery of Microservices.
12/26/2019 05:29:00 pm

Solutions for Building IoT based Smart Meters




Introduction to Smart Meters

In the Energy Upgrade solution, IoT is playing a significant role. The use of smart meters is increasing, which enables the intelligent and efficient use of energy at homes and businesses. Many grid power supply companies, small and large industries, private residential sector are also implementing the smart solution for energy efficiency and sustainability.

Business Challenge for Building the Analytics Platform

We need to build a complete analytical solution that can be used for the energy-saving recommendation based on the usages for the large buildings and industries. Also, the challenges were to filter the results found on floors, buildings, heat, water, electric. Along with the dashboard, alerting for usage also should be used based on usage.

Solution Approach for Building IoT based Smart Meters

Complete Smart meter based analytical dashboards which include -
  • Recommendation for energy saving
  • Predictive results for Energy Bills
  • Real-time alerting on some defined alerting rules
  • Analytical results on the base of historical data

Wednesday, 18 December 2019

12/18/2019 05:35:00 pm

Smart Manufacturing and IoT Solutions

Introduction to Smart Manufacturing

Smart manufacturing is a technology-driven approach that uses Internet-connected machinery to observe the production process. The purpose of SM is to recognize the possibilities for automating processes and use data analytics to enhance manufacturing execution.
  • Machines have threshold values for the different parameters like Temperature, Oil pressure, and Amperage which should not be crossed.
  • Implementation of the Internet of Things (IoT) enables proactive maintenance breaks on the machines to enable manufacturing for smart manufacturing. The collection of data in Real-Time is an excellent solution.
  • Smart Manufacturing industries reduce the risks

Challenge for Building the IoT Platform

Scalable Solution for smart manufacturing to manage all the machines at a single central point and handle sizeable Real-Time streaming data from sensors that can provide the alerts and trigger to turn off the motor in very minimum time.

Solution Offered Real-Time Data and IoT Platform

Installation of the sensor to collect each value (Temperature, Oil pressure, Amperage) in Real-Time using Google Cloud Platform.
Google Cloud IoT Core to ingest data from all sensors attached to different machines. The data collected from the sensors contain the machine identification number to differentiate the received data.
Google Cloud IoT Core sends the collected data to Google Cloud Pub/Sub. This data stream routes at multiple locations. Store the raw data into BigQuery and also to Google Cloud Function.
Detect whether the collected data values are higher than the threshold values. Implement Google Cloud Function and Google Pub/Sub as the trigger. Google Cloud function is an event-driven serverless compute platform to deploy Function As a Service which is auto-scalable, highly available and fault-tolerant.
Cloud function triggers the configuration changes to the Google Cloud IoT Core if data values are higher than the threshold value, for another sensor which controls the motor of the machine.
Google Cloud IoT Core sends the trigger to the device to turn off the machine and have a maintenance break.

Technology Stack

  • Google Pub/Sub
  • Google IoT Core
  • Google Cloud Function
  • Google Cloud DataFlow

Source: 
XenonStack/Use-Cases

Wednesday, 26 April 2017

4/26/2017 10:43:00 am

Enabling Real Time Analytics For IoT



What is Fast Data?


A few years ago, we remember the time when it was just impossible to analyze petabytes of data. Then emergence of Hadoop made it possible to run analytical queries on our huge amount of historical data.

As we know Big Data is a buzz from last few years, but Modern Data Pipelines are constantly receiving data at a high ingestion rate. So this constant flow of data at high velocity is termed as Fast Data.

So Fast data is not about just volume of data like Data Warehouses in which data is measured in GigaBytes, TeraBytes or PetaBytes.

Instead, we measure volume but with respect to its incoming rate like MB per second, GB per hour, TB per day. So Volume and Velocity both are considered while talking about Fast Data.

What is Streaming and Real-Time Data


Nowadays, there are a lot of Data Processing platforms available to process data from our ingestion platforms. Some support streaming of data and other supports true streaming of data which is also called Real-Time data.

Streaming means when we are able to process the data at the instant as it arrives and then processing and analyzing it at ingestion time. But in streaming, we can consider some amount of delay in streaming data from ingestion layer.

But Real-time data needs to have tight deadlines in the terms of time. So we normally consider that if our platform is able to capture any event within 1 ms, then we call it as real-time data or true streaming.

But When we talk about taking business decisions, detecting frauds and analyzing real-time logs and predicting errors in real-time, all these scenarios comes to streaming. So Data received instantly as it arrives is termed as Real-time data.
 

Stream & Real Time Processing Frameworks


So in the market, there are a lot of open sources technologies available like Apache Kafka in which we can ingest data at millions of messages per sec. Also Analyzing Constant Streams of data is also made possible by Apache Spark Streaming, Apache Flink, Apache Storm.


Spark Streaming



















Apache Spark Streaming is the tool in which we specify the time-based window to stream data from our message queue. So it does not process every message individually. 

We can call it as the processing of real streams in micro batches.
Whereas Apache Storm and Flink have the ability to stream data in real-time.

Why Real-Time Streaming


As we know that Hadoop, S3 and other distributed file systems are supporting data processing in huge volumes and also we are able to query them using their different frameworks like Hive which uses MapReduce as their execution engine.

Why we Need Real-Time  Streaming?


A lot of organizations are trying to collect as much data as they can regarding their products, services or even their organizational activities like tracking employees activities through various methods used like log tracking, taking screenshots at regular intervals.

So Data Engineering allows us to convert this data into structural formats and Data Analysts then turn this data into useful results which can help the organization to improve their customer experiences and also boost their employee's productivity.

But when we talk about log analytics, fraud detection or real-time analytics, this is not the way we want our data to be processed.The actual value data is in processing or acting upon it at the instant it receives.

Imagine we have a data warehouse like hive having petabytes of data in it. But it allows us to just analyze our historical data and predict future.

So processing of huge volumes of data is not enough. We need to process them in real-time so that any organization can take business decisions immediately whenever any important event occurs. This is required in Intelligence and surveillance systems, fraud detection etc.

Earlier handling of these constant streams of data at high ingestion rate is managed by firstly storing the data and then running analytics on it.

But organizations are looking for the platforms where they can look into business insights in real-time and act upon them in real-time.

Alerting platforms are also built on the top of these real-time streams. But Effectiveness of these platform lies in the fact that how truly we are processing the data in real-time.

Use Of Reactive Programming & Functional Programming


Now when we are thinking of building our alerting platforms, anomaly detection engines etc on the top of our real-time data, it is very important to consider the style of programming you are following.

Nowadays, Reactive Programming and Functional Programming are at their boom.

So, we can consider Reactive Programming as subscriber and publisher pattern. Often, we see the column on almost every website where we can subscribe to their newsletter and whenever the newsletter is posted by the publisher, whosoever have got subscription will get the newsletter via email or some other way.

So the difference between Reactive and Traditional Programming is that the data is available to the subscriber as soon as it receives. And it is made possible by using Reactive Programming model.

In Reactive Programming, whenever any events occur, there are certain components (classes) that had registered to that event. So instead of invoking target components by event generator, all targets automatically get triggered whenever any event occurs.

Now when we are processing data at high rate, concurrency is the point of concern. So the performance of our analytics job highly depends upon memory allocation/deallocation. So in Functional Programming, we don’t need to initialize loops/iterators on our own.

We will be using Functional Programming styles to iterate over the data in which CPU itself takes care of allocation and deallocation of data and also makes the best use of memory which results in better concurrency or parallelism.

Streaming Architecture Matters


While Streaming and Analyzing the real-time data, there are chances that some messages can be missed or in short, the problem is how we can handle data errors.

So, there are two types of architectures which are used while building real-time pipelines.
  • Lambda Architecture:

    This architecture was introduced by Nathan Marz in which we have three layers to provide real-time streaming and compensate any data error occurs if any. The three layers are Batch Layer, Speed layer, and Serving Layer.
    lambda architecture







  





Continue Reading the full Article At - XenonStack.com/Blog

Tuesday, 14 March 2017

3/14/2017 12:48:00 pm

Why We Need Modern Big Data Integration Platform




Data is everywhere and we are generating data from different Sources like Social Media, Sensors, API’s, Databases.

Healthcare, Insurance, Finance, Banking, Energy, Telecom, Manufacturing, Retail, IoT, M2M are the leading domains/areas for Data Generation. The Government is using BigData to improve their efficiency and distribution of the services to the people.

The Biggest Challenge for the Enterprises is to create the Business Value from the data coming from the existing system and from new sources. Enterprises are looking for a Modern Data Integration platform for Aggregation, Migration, Broadcast, Correlation, Data Management, and Security.

Traditional ETL is having a paradigm shift for Business Agility and need of Modern Data Integration Platform is arising. Enterprises need Modern Data Integration for agility and for an end to end operations and decision-making which involves Data Integration from different sources, Processing Batch Streaming Real Time with BigData Management, BigData Governance, and Security.


BigData Type Includes:
  • What type of data it is
  • Format of content of data required
  • Whether data is transactional data, historical data or master data
  • The Speed or Frequency at which data made to be available
  • How to process the data i.e. whether in real time or in batch mode


5 V’s to Define BigData



5vs of big data










 

Additional 5V’s to Define BigData


additional 5vs of big data



 

Data Ingestion and Data Transformation


Data Ingestion comprises of integrating Structured/unstructured data from where it is originated into a system, where it can be stored and analyzed for making business decisions. Data Ingestion may be continuous or asynchronous, real-time or batched or both.

Defining the BigData Characteristics: Using Different BigData types, helps us to define the BigData Characteristics i.e how the BigData is Collected, Processed, Analyzed and how we deploy that data On-Premises or Public or Hybrid Cloud.

  • Data type: Type of data
    • Transactional
    • Historical
    • Master Data and others

  • Data Content Format: Format of data
    • Structured (RDBMS)
    • Unstructured (audio, video, and images)
    • Semi-Structured

  • Data Sizes: Data size like Small, Medium, Large and Extra Large which means we can receive data having sizes in Bytes, KBs, MBs or even in GBs.

  • Data Throughput and Latency: How much data is expected and at what frequency does it arrive. Data throughput and latency depend on data sources:
    • On demand, as with Social Media Data
    • Continuous feed, Real-Time (Weather Data, Transactional Data)
    • Time series (Time-Based Data)

  • Processing Methodology: The type of technique to be applied for processing data (e.g. Predictive Analytics, Ad-Hoc Query and Reporting).

  • Data Sources: Data generated Sources
    • The Web and Social Media
    • Machine-Generated
    • Human-Generated etc

  • Data Consumers: A list of all possible consumers of the processed data:
    • Business processes
    • Business users
    • Enterprise applications
    • Individual people in various business roles
    • Part of the process flows
    • Other data repositories or enterprise applications

modern big data integration platform

 

Major Industries Impacted with BigData



industries impacted with big data

 

What is Data Integration?


Data Integration is the process of Data Ingestion - integrating data from different sources i.e. RDBMS, Social Media, Sensors, M2M etc, then using Data Mapping, Schema Definition, Data transformation to build a Data platform for analytics and further Reporting. You need to deliver the right data in the right format at the right timeframe.

BigData integration provides a unified view of data for Business Agility and Decision Making and it involves:

  • Discovering the Data
  • Profiling the Data
  • Understanding the Data
  • Improving the Data
  • Transforming the Data

A Data Integration project usually involves the following steps:

  • Ingest Data from different sources where data resides in multiple formats.
  • Transform Data means converting data into a single format so that one can easily be able to manage his problem with that unified data records. Data Pipeline is the main component used for Integration or Transformation.
  • MetaData Management: Centralized Data Collection.
  • Store Transform Data so that analyst can exactly get when the business needs it, whether it is in batch or real time.

modern big data integration platform

 

Why Data Integration is required


  • Make Data Records Centralized: As data is stored in different formats like in Tabular, Graphical, Hierarchical, Structured, Unstructured form. For making the business decision, a user has to go through all these formats before reaching a conclusion. That’s why a single image is the combination of different format helpful in better decision making.
  • Format Selecting Freedom: Every user has different way or style to solve a problem. User are flexible to use data in whatever system and in whatever format they feel better.
  • Reduce Data Complexity: When data resides in different formats, so by increasing data size, complexity also increases that degrade decision making capability and one will consume much more time in understanding how one should proceed with data.
  • Prioritize the Data: When one have a single image of all the data records, then prioritizing the data what's very much useful and what's not required for business can easily find out.
  • Better Understanding of Information: A single image of data helps non-technical user also to understand how effectively one can utilize data records. While solving any problem one can win the game only if a non-technical person is able to understand what he is saying.
  • Keeping Information Up to Date: As data keeps on increasing on daily basis. So many new things come that become necessary to add on with existing data, so Data Integration makes easy to keep the information up to date.

Continue Reading The Full Article At - XenonStack.com/Blog