Page 122 - Building Big Data Applications
P. 122

Chapter 6   Visualization, storyboarding and applications  119


                   Each sequence will create its own set of audit trails and the logs from these audits are
                 very essential for the teams to have access to and analyze. The audit logs can be accessed
                 and analyzed using Apache Flume or Apache Elastic or Splunk depending on what the
                 enterprise wants to accomplish during the analysis. This is where we will use machine
                 learning and neural networks the most as the logs can be varying in size, content and
                 actions. The processing of unsupervised data is more useful to learn from without
                 actually interfering the actual view of the execution. This flow is an essential step in
                 creation of the actual visualization for the team.
                   Each audit analysis will produce the exact user interactions, system interactions and
                 resulting steps that was the behavior of the system and the user. Think of this in research
                 with the data from CERN particle accelerators and the discovery of the “god particle”. The
                 actual scattering of the data and its outcomes were not executable repeatedly, however if
                 the logs are analyzed and outcomes understood, the discovery can be accelerated with the
                 right corrections applied. This type of activity is what research teams have always wanted
                 and have never been successful in applying their team skills together.


                 Let us look at some of the use cases of big data
                 applications

                 Clickstream analysisdwhen you engage in an Internet search for a product, you can see
                 that along with results for the product you searched, you are provided details on sales,
                 promotions, and coupons for the product in your geographical area and the nearest 10
                 miles, and also promotions being offered by all internet retailers. Analyzing the data
                 needed to create the search results tailored to meet your individual search results, we can
                 see that the companies, who have targeted you for a promotional offer or a discount
                 coupon. How did these retailers offer me a coupon? The answer to this is the big data
                 analytic application in the retailer which provides details on web users’ behaviors gathered
                 with the process of clickstream analysis. The big data application is architected to provide
                 instant analytics on any user or users’ behaviors based on specific geographies or multiple
                 geographies compared in a mash-up. To make these analytics work seamlessly and provide
                 multiple user connects with offers, the data integration in this architecture is complex.
                 What changes here is the actual mechanism of building this application itself.
                   For search to be executed, we need to have an interface that will use a search box and
                 a submit button. This is great as all search engines provide you an interface. What
                 happens next is the intricate process of why an engine like Google is a success while Bing
                 or Yahoo is a distant second or third. The application development process here has to
                 boil down to all levels of details. For search to happen, the following big data application
                 strategy needs to be written as foundations:

                   Step 0dFoundational web interface screen
                     Search text box
                     Submit button
                     Minimal logo
   117   118   119   120   121   122   123   124   125   126   127