Archives For Big Data

Understanding the DNS of Data Science


Data Science is the competitive advantage of the future for organizations interested in turning their data into a product through analytics. Industries from health, to national security, to finance, to energy can be improved by creating better data analytics through Data Science.

The Field Guide to Data Science

Download the full book


What if:

… you could identify which employees are likely to turnover in the next 6 months?

… you could understand which transformers are likely to fail during our next major storm?

… your call center agents could delight customer with the best next-step recommendations?

Big Data is the catch phrase of the day.  Everyone has heard of it, most have a vague idea what it means, some have a clear grasp of what it can really do, and few can execute it effectively.  I’ve been doing a lot of reading on Big Data and there is certainly no lack of resources on the topic.  As I read through many reports, white papers, press releases, magazine articles and company presentations (all available via a quick Google search) it occurred to me that visually communicating the value of Big Data is challenging because of the need to convey different concepts simultaneously.  The most popular category by far are plot charts on an X-Y axis.  These charts plot analytical complexity against some sort of business value measurement in a positive correlation that looks entertainingly similar to human evolution charts we’ve all seen, with man becoming more upright and intelligent with time.

Less popular, but also useful, are a bulls-eyes, Venn diagrams and an stacked area triangles. Regardless of graphic representation, they all follow the progression from What Happen(descriptive analytics), Why Did It Happen (correlation analytics), What Will Happen Next(predictive analytics), and What Should I Do About It (prescriptive analytics).  Which one do you prefer


Why We Like It: This chart is unique in that it goes all the way back to the beginning when data is first created and gathered in raw form.  So much of the resources needed to develop prescriptive analytics takes place in the very early stages of the process, and it’s nice that this graphic gives it a mention.  The overwhelming majority of data available for analysis does make it to the final predictive/prescriptive model.  If each circle represented the amount of actual data at that stage, the raw data circle (and cleaned data circle) would dwarf all the others, so thank you SAP for giving data its due.


Read More:




Data Science Venn Diagram v2.0

You can see a static image of the final viz below and check out the full story and interactive version in The Evolution of Languages on Twitter.

Languages of tweets

Languages of tweets




Big Data Architect, Requirements and Responsibilities to get the Job


  • Lead strategic consulting engagements for Perficient that quickly deliver high-value solutions for our clients.  This involves full lifecycle delivery of advanced business intelligence and data warehousing systems.
  • Represent Perficient as the primary on-site technical contact at client sites.
  • Be a subject matter expert in data warehouse appliances to include Big Data and Analytical solutions.
  • Provide technical and managerial leadership in a team that designs and develops path breaking large-scale cluster data processing systems
  • Assisting Perficient’s  customers develop strategies that maximize the value of their data
  • Establish Perficient’s thought leadership in the OSS and Appliances big data space by contributing white-papers, blog entries, and perspectives.

To succeed in this role you should:

  • Have broad set of technology skills to be able to design and build robust solutions for analyze big data
  • Ability to learn quickly as the industry grows
  • Understand business drivers and impacts
  • Be able to grasp the problem at hand and recognize appropriate approach, tools and technologies to solve it
  • Ability to be objective in evaluating technical solutions
  • Have a process driven mindset.
  • Leverage social media to establish thought-leadership


  • BS, MS in Computer Science, Math, or relevant Business Administration specialty
  • 7+ Years dealing with leading edge business intelligence and data warehousing systems.
  • Balance of industry and consulting experience.
  • In depth understanding of different architected approaches to managing and querying  large data sets
  • Experience leading data warehouse and/or big data platform selection projects
  • Hands on with one or more high-scale or distributed RDBMS and Big Data platforms (SAP HANA, IBM Netezza/Big Insights Infosphere, HP Vertica, Oracle Exadata and Big Data Appliance)
  • Hands-on experience with ETL (Extract-Transform-Load) tools (e.g. Informatica, Talend, Pentaho)
  • Hands-on experience with BI tools and reporting software (e.g. SAP BI, Microstrategy, Pentaho)
  • Knowledge of cloud computing infrastructure (e.g. Amazon Web Services EC2, Elastic MapReduce) and considerations for scalable, distributed systems
  • Knowledge of leading edge statistical software packages (e.g. SAS, SPSS, KXEN)
  • Knowledge of Hadoop and NoSQL platforms (e.g. key-value stores, graph databases, RDF triple stores)
  • Above average interpersonal, communication and presentation skills – must be able to explain technical concepts and analysis implications clearly to a wide and non-technical audience
  • Self-starter, with a keen interest in technology and highly motivated towards success
  • Ability bring clarity to ambiguous situations
  • Proven ability to work in fast-paced, agile environments

Big Data Benefits:What companies get from Analytics

- Better Social Influencer Marketing
- More accurate business insights
- Segmentation of customer base
- Identifying sales and market opportunites
- Automated decisions for real time processes
- Detection of fraud
- Wuantification of risks
- Better planning and forecasting
- Identifiying cost drivers

What companies get from Analytics

What companies get from Analytics

According to Tableau Soft ware the Top 10Trends in Business Intelligence for 2014 are :

Top 10 trends in business intelligence for 2014

Top 10 trends in business intelligence for 2014

In this interview David Reed from The Data Governance Forum speaks about the impact of big data on data governance, the key concerns of customers and their rights. Reed also shares his views on why it is important to establish a data governance culture and practice that is sustainable.

This interview forms part of PEX Network’s exclusive Big Data & Analytics Handbook. To download the full ebook, visit

Prof César A. Hidalgo from MIT Media Lab talks to Digital 2014 in advance of the 2014 ICT Conference held in Wales on 9+10 June 2014.

Big Data on AWS

March 25, 2014 — Leave a comment

Big Data on AWS


Big Data on AWS introduces you to cloud-based big data solutions and Amazon Elastic MapReduce (EMR), the AWS big data platform. In this course, we show you how to use Amazon EMR to process data using the broad ecosystem of Hadoop tools like Pig and Hive. We also teach you how to create big data environments, work with Amazon DynamoDB and Amazon Redshift, understand the benefits of Amazon Kinesis, and leverage best practices to design big data environments for security and cost-effectiveness.

Course Objectives

This course is designed to teach you how to:

  • Understand Apache Hadoop in the context of Amazon EMR
  • Understand the architecture of an Amazon EMR cluster
  • Launch an Amazon EMR cluster using an appropriate Amazon Machine Image and Amazon EC2 instance types
  • Choose appropriate AWS data storage options for use with Amazon EMR
  • Know your options for ingesting, transferring, and compressing data for use with Amazon EMR
  • Use common programming frameworks available for Amazon EMR including Hive, Pig, and Streaming
  • Work with Amazon Redshift to implement a big data solution
  • Leverage big data visualization software
  • Choose appropriate security options for Amazon EMR and your data
  • Perform in-memory data analysis with Spark and Shark on Amazon EMR
  • Choose appropriate options to manage your Amazon EMR environment cost-effectively
  • Understand the benefits of using Amazon Kinesis for big data

Intended Audience

This course is intended for:

  • Partners and customers responsible for implementing big data environments, including:
    • Data Scientists
    • Data Analysts
    • Enterprise, Big Data Solution Architects


We recommend that attendees of this course have:

  • Basic familiarity with big data technologies, including Apache Hadoop and HDFS
    • Knowledge of big data technologies such as Pig, Hive, and MapReduce is helpful but not required
  • Working knowledge of core AWS services and public cloud implementation
    • Students should complete the AWS Essentials course or have equivalent experience
  • Basic understanding of data warehousing, relational database systems, and database design

Delivery Method

This course will be delivered through a mix of:

  • Instructor-Led Training (ILT)
  • Hands-on Labs

Hands-On Activity

This course allows you to test new skills and apply knowledge to your working environment through a variety of practical exercises


  • 3 Days

Course Outline

This course will cover the following concepts on each day:


Day 1

  • Overview of Big Data and Apache Hadoop
  • Benefits of Amazon EMR
  • Amazon EMR Architecture
  • Using Amazon EMR
  • Launching and Using an Amazon EMR Cluster
  • High-Level Apache Hadoop Programming Frameworks
  • Using Hive for Advertising Analytics

Day 2

  • Other Apache Hadoop Programming Frameworks
  • Using Streaming for Life Sciences Analytics
  • Overview: Spark and Shark for In-Memory Analytics
  • Using Spark and Shark for In-Memory Analytics
  • Managing Amazon EMR Costs
  • Overview of Amazon EMR Security
  • Exploring Amazon EMR Security
  • Data Ingestion, Transfer, and Compression

Day 3

  • Using Amazon Kinesis for Real-Time Big Data Processing
  • AWS Data Storage Options
  • Using DynamoDB with Amazon EMR
  • Overview: Amazon Redshift and Big Data
  • Using Amazon Redshift for Big Data
  • Visualizing and Orchestrating Big Data
  • Using Tableau Desktop or Jaspersoft BI to Visualize Big Data

Public Data Sets available at to pratice with your BigData tools

Running a Big Data Infrastructure: Five Areas That Need Your Attention
Download Free book (Don’t need to signup)

1. Elasticity
Scalability generally refers to the ability of a technology to support increasing demand. Typically, in order to scale infrastructure, more
machines are added. Elasticity refers to the ability of a technology to flex up or down depending on demand at any given time. Working on an
elastic substrate allows a company to utilize one machine for lots of operations, not just Big Data, without affecting the budget.

2. Reliability
Business users count on the availability of data at regular intervals to efficiently run the company. Reliability begins with how effectively data
enters the infrastructure and ends with predictable data delivery. Companies, such as Facebook, utilize internal service level agreements
(SLAs) to ensure data availability across teams.

3. Self-Service Tools

Business analysts require data to perform their job functions, yet without a deep technical skill
set, quickly and easily accessing the required data can be practically impossible. In the past,
companies resolved this issue by hiring teams to act as liaisons between the data engineers and
business users. Today, self-service tools provide ready access to data through a user-friendly
interface. Productivity is increased across the board making it possible for more people in the
organization to make data-driven decisions and making it possible for data teams to support
larger and larger numbers of business users.

4. Monitoring

Active monitoring of the infrastructure minimizes issues and maximizes availability. How much
monitoring? The more, the better. With the introduction of self-service tools, more users relying on the data will be impacted by
inefficiencies or other concerns. Monitoring provides insight into how the system is operating and quickly identifies minor glitches that can be
corrected before they develop into
major problems.

5. Open Source

Open source technology evolves rapidly, and new features, such as faster queries, could result
in increased productivity, improve cluster availability, or even competitive advantage in the
marketplace. Data engineers must stay on top of the latest versions to ensure that the
infrastructure runs at peak performance. Working with a knowledgeable vendor can help data
engineers mitigate the risk of missing even one update.


Big Data Tools for Data Developers and Data Analyst


1. Splice Machine


A real-time SQL-on-Hadoop database, Splice Machine takes Big Data beyond analytics with the ability to derive real-time, actionable insights for rapid decision-making. Not only can Splice Machine process real-time updates, but it offers the ability to utilize standard SQL and is capable of scaling out on commodity hardware. Splice Machine can be used in circumstances where MySQL or Oracle can’t scale.

Key Features: 

  • SQL-99 compliant, with standard ANSI SQL
  • Easily scales from gigabytes to petabytes using cost-effective, commodity hardware
  • Real-time updates with transactional integrity
  • Distributed computing architecture
  • Multiple Version Concurrency Control (MVCC)

Cost: Contact for a quote

2. Palantir

Palantir was founded in 2004 by a group of former PayPal employees and Stanford computer scientists. The company has doubled in size every year to date, but strives to maintain its startup culture. Offering a suite Big Data solutions for integrating, visualizing and analyzing information, Palantir’s product line emphasizes scalability, security, ease of use, and collaboration. Palantir’s solutions are most commonly used in intelligence, defense, financial and law enforcement applications, but it’s quickly growing in other verticals.

Key Features: 

  • Solutions for integrating, visualizing and analyzing data
  • Serves a multitude of industries with custom solutions
  • Exploit and analyze data
  • Extract data from multiple sources
  • Privacy and data protection policies
  • Simplify workflows by integrating data into a single dashboard

Cost: Contact for a quote

3. Attivio

As enterprises are coping with a broader variety of information sources, eliminating information silos is critical to gaining comprehensive insights and identifying key relationships among data. Attivio’s Active Intelligence Engine combines Big Data and Big Content to analyze everything, including human-generated text through advanced text analytics. Combined with universal indexing and automatic ad-hoc JOIN, Attivio is a powerful solution for making valuable connections between all your data.

Key Features: 

  • Combines Big Data and Big Content
  • Eliminates information silos
  • Adds context and signals from human-generated information sources
  • Supports BI/data visualization tools
  • In-engine analytics

Cost: Contact for a quote

4. Google Charts

Google Charts is a free tool with a wide range of capabilities for visualizing data from a website. From simple charts to complex hierarchical tree maps, Google Charts offers a gallery with a multitude of pre-configured chart types to choose from. Google Charts is easily implemented by embedding simple JavaScript code on a website, yet it offers complex functionality with the ability to connect to dashboards, sort, modify, and filter data, connect to a database or integrate with and pull data from a website. Take it a step further by implementing the Chart Tools Datasource protocol and allow other entities to source data from your website.

Key Features: 

  • Charts exposed as JavaScript classes
  • Customize to match the look and feel of a website
  • Charts populated using DataTable class
  • Sort, modify, filter data
  • Populate data from a variety of sources

Cost: FREE

5. Mortar


Mortar is a “general purpose platform for high-scale data science” designed to help data scientists spend more time analyzing their data and deriving actionable insights, instead of dedicating valuable time to building infrastructure and re-configuring systems. With Mortar, you can build a custom-built recommendation engine in days, not months.

Key Features: 

  • Open-source tools for building a recommendation engine
  • Built on Hadoop and Apache Pig
  • Create, test and run jobs from in-browser IDE
  • Snapshots monitor changes and progress
  • Instant feedback on code for fewer bugs


  • Public – FREE
  • Solo – $99/month
  • Team – $499/month

6. SAP


SAP’s HANA platform can be combined with Apache Hadoop for the ability to integrate and analyze massive loads of data in real time. The platform makes it possible to derive actionable insights by making valuable connections between all types of information, from a multitude of sources. Combine SAP HANA with applications that leverage Big Data insights to quickly create additional revenue streams and improve operations.

Key Features: 

  • Infinite storage
  • Flexible data management for all types of data
  • Discover insights with analytics solutions
  • Runs processes 1,000 to 100,000 times faster in-memory
  • SAP IQ analytics holds the Guinness World Record for data loading

Cost: Contact for a quote

7.  Cambridge Semantics


Collecting, integrating, and analyzing Big Data doesn’t have to be a major effort. Cambridge Semantics makes it all possible with the Anzo Software Suite, an open platform for building Unified Information Access (UIA) solutions. That means replacing the information silos that leave data isolated and useless with a powerful, seamless data integration machine that streamlines data collection and enables sophisticated analysis for rapid decision making. And you can implement all this within hours or days — not the typical weeks or months required for an initiative at this level.

Key Features: 

  • Combine data from a multitude of sources
  • Customized, interactive web dashboards for analysis
  • Share spreadsheets in sync automatically
  • Useful for CRM, billing, project management and more

Cost: Contact for a quote

8. Fusion Charts


Just because development is in your blood doesn’t mean you should embrace complexity when simpler solutions are readily available. Fusion Charts enables you to create sophisticated, cross-device compatible JavaScript charts with animation, rich interactivity, and impressive design with ease. Don’t spend your valuable time on child’s play; with Fusion Charts, you can spend more time on complex development tasks and deliver even better results.

Key Features: 

  • Developer resources center
  • Interactive zooming and scrolling
  • Real-time charts and gauges
  • Multi-lingual charts
  • Visually editable charts and gauges
  • Linked charts and a variety of effects

Cost: Contact for a quote

9. MarkLogic


MarkLogic is built to support the world’s biggest data loads, bringing all types of relevant content back to users who can turn it into action. With real-time updates and alerts, connections between information make new opportunities immediately obvious. MarkLogic is ideal for enterprises that count on revenue through paid content search. With geographic data combined with content, location relevance is built in, and geographic boundaries make advanced data filtering possible.

Key Features: 

  • Range of Big Data solutions
  • Speeds development
  • Flexible APIs
  • NoSQL
  • Real-time analysis and updates
  • Bring all types of content back to end users

Cost: Contact for a quote

10. Syncsort


Syncsort offers a range of products and solutions to help you tap into Big Data. Hadoop, Linux, Unix and Windows, and Mainframe solutions, Syncsort’s product lineup offers a solution to meet practically any configuration needs. A GUI-based solution, Syncsort enables developers to create solutions for collecting, processing, and distributing more data in less time.

Key Features: 

  • Solutions for Hadoop, Mainframe, Windows, Linux, Unix
  • Lowers the barriers to Hadoop adoption
  • Eliminates the need for custom code for Hadoop implementation
  • High-performance sorting
  • Improve efficiency

Cost: Contact for a quote

11. DataStax


DataStax helps companies like Netflix, Healthcare Anytime, eBay, and even Adobe harness the power of Big Data with less effort and at a lower cost than traditional solutions. Tapped as the first alternative to Oracle, DataStax provides the constant uptime and lightning speed required for modern customer-facing applications. When you need the capacity to handle massive data loads at maximum speed for real-time analysis, DataStax packs a major punch with a robust visual query tool for developers.

Key Features: 

  • Visual query tool for developers
  • Create and run Cassandra Query Language (CQL) queries and commands
  • Visually navigate and interact with data clusters
  • Works with DataStax Community and Enterprise editions

Cost: Contact for a quote

12. Guavus


Need to create a rich, engaging and meaningful customer experience? Guavus drives better decision making with powerful analytics capabilities combined with advanced data science and the ability to distill data in real time to derive actionable insights at the precise moment of opportunity. Through continuous correlation and ongoing analysis, vast amounts of static and dynamic data are handled with ease and revealing opportunities to generate more revenue, reduce overhead costs, and monetize new streams.

Key Features: 

  • Analyze-First Analytics Architecture
  • Analyze high-volume data streams in near real time
  • Handles multiple data sources with ease
  • Continual data analysis from moment of capture

Cost: Contact for a quote

13. mongoDB


An open-source document database, mongoDB is ideal for developers who want precise control over the final results and processes for handling Big Data. With full index support, you have flexibility to index any attribute and scale horizontally without compromising functionality. Rich, document-based queries and GridFS for storing files of any size without the risk of compromising your stack, mongoDB is a scaleable, flexible, and powerful solution for Big Data.

Key Features: 

  • Open-source platform
  • Document-oriented storage
  • Flexible aggregation and data processing
  • Full index support
  • GridFS

Cost: FREE

14. Infochimps Cloud


A cloud service solution for Big Data, Infochimps Cloud makes it possible to deploy Big Data applications rapidly and without the typical time commitment. For applications requiring real-time analysis, multi-source streaming data, a NoSQL database, or a Hadoop cluster, Infochimps Cloud offers a solution that facilitates rapid implementation. Real-time analytics, ad hoc analytics, and batch analytics comprise Infochimps Cloud’s three essential cloud services.

Key Features: 

  • Integrate with any data source – CRM solutions, etc.
  • Log analysis
  • Mobile data analytics
  • Fraud detection and risk analysis
  • Ad targeting
  • Customer insights via social media sources, website clickstreams and more

Cost: Contact for a quote

15. Pentaho


Pentaho brings IT and business users together by joining data integration and business analytics for integrating, visualizing, analyzing and blending Big Data in ways never before possible for better business results. When you need the ability to put robust information at your users’ fingertips in real time and at a reasonable cost, Pentaho’s open, embeddable and extensible analytics platform makes it easy to visualize, explore, and predict — turning data into value.

Key Features: 

  • High-volume data processing
  • Adaptive Big Data layer
  • Data mining and predictive analysis
  • Instaview – data to insights in 3 steps

Cost: Contact for a quote

16. Karmasphere


Designed for teams of analysts who need to explore and analyze Big Data on Hadoop, Karmasphere is a key solution for self-service analytics on Hadoop. With companies seeking more effective and efficient ways to make use of Big Data, users can turn data into real business value by discovering insights and influencing outcomes.

Key Features: 

  • Organized dashboard
  • SQL data explorer
  • 250-plus pre-packaged Hadoop algorithms
  • SAS, SPSS and R Analytic Models
  • Dynamic data lenses for self-service analytics

Cost: Contact for a quote

17. Placed


Placed facilitates data collection from offline sources, enabling enterprises to derive actionable insights through a combined analysis of both offline and online behavior and data metrics. Placed targeting and placed attribution facilitates better results from mobile advertising by mapping the relationship between people and places by capitalizing on Big Data capabilities.

Key Features: 

  • Measure visitation trends over time
  • Measures 100 million locations a day, across more than 100,000 opted-in US smartphones
  • Inference Pipeline references a place database with nearly 300 million features for the US alone
  • Largest repository of offline insights into the paths and behaviors of consumers
  • Audience segmentation by demographics and other data points
  • Affinity modeling for understanding relationships between data
  • Monitor and understand how consumer behavior changes over time

Cost: Contact for a quote

18. Upsight


Upsight, formerly Kontagent, provides actionable analytics for developers to understand what’s happening with your apps and derive actionable insights from data to impact acquisition, engagement, retention and revenue. The platform also enables the creation of targeted in-app and out-of-app metrics in line with KPIs.

Key Features: 

  • Free, enterprise-grade analytics
  • Unlimited data storage
  • Data mining with Hadoop
  • Measure anything from social apps to games and mobile dating apps
  • Funnel analysis
  • Cohort explorer
  • Predictive LTV


  • FREE – Analytics and unlimited data storage, 250k push
  • Core – $500/month - Custom Events up to 100k MAU, 500k push
  • Pro – $2,000/month - Custom Events up to 250K MAU, 1M push
  • Enterprise – Starting at $3,000/month - Unlimited Data Storage & Custom Events + Data Mine + Predictive LTV + A/B, unlimited push

19. Talend


Talend Open Studio is “a powerful and versatile set of open source products for developing, testing, deploying and administrating data management and application integration projects.” Providing a unified environment for managing the full lifecycle, even across enterprise boundaries, Talend enables developers to reclaim their productivity with a fully integrated platform for joining data integration, data quality, MDM, application integration and big data.

Key Features: 

  • Data integration at a cluster scale
  • No need to write or maintain code
  • Works with leading Hadoop distributions
  • Pull source data from anywhere, including NoSQL

Cost: FREE

20. Jaspersoft


Connect and visualize data for Hadoop Analytics, MongoDB Analytics, Cassandra Analytics, and other platforms in one central repository. Using Big Data, developers can configure reports, analytics, dashboards, and more, without having to migrate data to multiple databases.

Key Features: 

  • Real-time analytics
  • Integrate all your data
  • Blend data through innovative data virtualization metadata layer or raditional data warehouse using ETL
  • Present integrated visualizations and dashboards within your apps
  • Create intuitive design tools for non-designers to create visualizations


  • Free
  • Jaspersoft for AWS – Less than $1/hour

21. Keen IO


With powerful APIs for gathering all the data you need and deriving the actionable insights you need to drive your business forward, Keen IO is a powerful, flexible, and scaleable solution that’s literally Big Data, easy-to-implement and at your fingertips.

Key Features: 

  • Send as much data as you want, from any source
  • Set up event data on any action, such as upgrades, impressions or purchases
  • Arbitrary JSON format
  • Custom properties


  • Developer – FREE – 50,000 events/month
  • Startup – $20/month – 100,000 events/month
  • Growth – $125/month – 1M events/month
  • Premium – $300/month – 4M events/month
  • Professional – $600/month – 8M events/month
  • Business – $1,000/month – 15M events/month
  • Enterprise – $2,000/month – 50M events/month
  • Custom – Negotiable – 50M – 100B events/month

22. Skytree


High-performance machine learning on Big Data for advanced analytics, Skytree offers the ideal platform for fully exploiting the opportunities presented by Big Data. With a multitude of industry-focused solutions as well as solutions encompassing everything from predictive analytics to algorithmic pricing, Skytree is a comprehensive Machine Learning platform emphasizing the growing importance of Predictive Analytics in Big Data.

Key Features:

  • Business Analytics range from value analytics to fraud detection and what-if analytics
  • Marketing Analytics offer solutions ranging from ad optimization to lead scoring and recommender systems
  • Only general purpose scalable Machine Learning system on the market
  • Highest accuracy on the market; unprecedented speed and scale
  • Power Packs modules are plugged into the Skytree Server Foundation

Cost: Contact for a quote

23. Tableau Software


Tableau was launched by a computer scientist, an Academy Award-winning professor, and a business leader with a passion for data. This perfect trio created a powerful suite of solutions designed to put more data at users’ fingertips — and help them understand it in more meaningful ways. With an advanced query language for powerful visualizations, the ability to natively query databases, cubes, warehouses, and more, a lightning-fast in-memory analytics database designed to eliminate silos and more, Tableau addresses every corner of Big Data demands.

Key Features: 

  • In-memory analytics database eliminates memory silos
  • Leverages the complete memory hierarchy from disk to L1 cache
  • Tableau Public – free tool bringing data to life on the web
  • Touch, swipe and tab functionality for mobile
  • Easily layer in additional data sources
  • Access any data with a few clicks


  • Tableau Public – FREE
  • Tableau Desktop, Tableau Server, and Tableau Online – Contact for a quote

24. Splunk


Splunk harnesses all the machine data created by websites, applications, servers, networks, sensors, mobile devices, and other sources to monitor actions, activities and events, analyzing those data sources to derive actionable insights. Splunk is a self-contained software package downloadable and functional on any device.

Key Features: 

  • Derive insights from Big Data with speed and simplicity
  • Works on most major Hadoop distributions, including including first-generation MapReduce and YARN
  • Splunk Hadoop Connect enables bi-directional integration
  • Real-time collection, indexing, and analyzing


  • Splunk Storm – FREE cloud service for developers
  • Splunk Enterprise – Perpetual License – Starts at $4,500 for 1 GB/day, plus annual support fees
  • Splunk Enterprise – Term License – Start at $1,800 per year, which include annual support fees
  • Hunk – One-year term license of Hunk starts at $2,500 per Hadoop TaskTracker or Compute Node with a minimum of ten TaskTrackers or Compute Nodes
  • Splunk Cloud – Annual subscription pricing, data volumes of 5GB/day to 1TB/day

25. Platfora


Platfora hides the complex nature of Hadoop, making it simpler for enterprises to discover and understand facts in their business across events, actions, behaviors and time. Built by Silicon Valley veterans who have built market-leading companies around big ideas, the Platfora team understands the power of Big Data and aims to change customers’ lives with Platfora as they’ve done with companies in the past.

Key Features: 

  • Vizboards for self-service, interactive data visualization
  • Analytics Engine, In-Memory Accelerator, and Hadoop Processor
  • Entity-centric data catalog
  • Build interest-driven pipelines of facts
  • Analyze data iteratively with segmentation
  • Collaboration features
  • On-premise or cloud deployment

Cost: Contact for a quote

26. Continuuity


Continuuity enables developers to build Big Data applications quickly, easily, and seamlessly, deploying instantly on-premise or to the cloud. It’s all made possible through simple APIs that can be used with virtually any platforms.

Key Features: 

  • User-implemented real-time stream processors (Flows)
  • Process a batch of data objects with the same transaction
  • More than one instance possible with each Flowlet
  • Programmatic control with REST interfaces
  • Three partitioning strategies to choose from
  • DataSets for higher-level abstractions

Cost:  Contact for a quote

27. BitDeli


BitDeli is an analytics tool for GitHub, enabling developers to gather data on who is viewing their repositories, where and when. With a one-click install, you can easily add analytics to your repositories and start gathering valuable data, including aggregate statistics across all forks for a given repository.

Key Features: 

  • One-click install
  • Automatically generated pull requests
  • Trending badge indicator shows repository popularity
  • Global rankings for comparison
  • Fork aggregation for a broad picture of project health

Cost: Based on GitHub Enterprise pricing

28. Flurry


Flurry is an end-to-end solution for analyzing consumer behavior, advertising to the right audience, at the right time, and discovering new ways to monetize audiences. Flurry makes use of 3.5 billion app session reports per day totaling more than 3 terabytes to provide valuable insights for app developers, such as a deep understanding of the user base, engagement benchmarks, and other key metrics.

Key Features: 

  • Demographic estimations
  • App engagement benchmarks
  • App category and consumer interests
  • World’s largest app-audience data set
  • Reach more than 250 million customers per month
  • Data-powered targeting


  • Flurry Analytics – FREE
  • Flurry AppCircle, FlurryPersonas, Flurry AppSpot – Contact for a quote

29. Spring Data /Pivotal


Spring Data
Spring Data is a set of projects designed to make it easier to use new data access technologies and provide improved support for relational database technologies. An umbrella project with many sub-projects designed for specific databases, Spring Data is an excellent source of Big Data tools to streamline the use of Big Data in the modern enterprise environment.

Key Features: 

  • Support for Hadoop, mongoDB, Data Rest and more
  • Also provides consulting services
  • Customized, all-in-one Eclipse-based distribution
  • Tool suites for ready-to-use solutions

Cost: FREE

30. Hortonworks


Hortonworks offers a complete distribution for fully harnessing the power of Hadoop. With Hortonworks, you can process and analyze everything from Sentiment to Sensors with a 100% open-source, enterprise-grade distribution of Hadoop for every platform.

Key Features: 

  • Interact with all data, in multiple ways, simultaneously
  • Stable, tested, complete package of all services required for the platform
  • Integrates with other tools
  • HDP is built and supported by original architects, builders and operators of Hadoop

Cost: FREE

31. StatsMix


A complete dashboard for all your business needs, StatsMix dashboards are customized to your specific requirements, with the data you need from sources like Salesforce, MySQL, Google Analytics, and other tools and services.

Key Features: 

  • Chart and track anything
  • Measure KPIs
  • Track any metric with API
  • Automatic social monitoring
  • Share metrics and dashboards via email, embed them, or create guest accounts
  • Aggregate metrics to eliminate silos
  • Custom dashboards


  • Basic – $24/month – 100k API requests
  • Standard – $49/month – 300k API requests
  • Pro – $99/month – 1M API requests
  • Premium – $199/month – 3M API requests
  • Enterprise – $499/month – 8M API requests

32. Pervasive


Pervasive offers a number of Big Data tools, including several solutions for Hadoop and a free RushLoader for Hadoop. From DataFlow Analytics to ParAccel Dataflow ETL/DQ for designing end-to-end ETL and quality data workflows, Pervasive is a Big Data power suite.

Key Features: 

  • Partnership with Actian for powering Big Data 2.0
  • Predictive Analytics for Big Data
  • Simple interface for loading massive amounts of data at rapid speeds
  • Fastest data-crunching engine in the world


  • RushLoader for Hadoop – FREE
  • ParAccel Dataflow Loader for Hadoop – FREE for 12 months
  • All other products – Contact for a quote

33. InfiniDB


Real-time enablement of Big Data insights is yours with InfiniDB. A 100% open-source platform, you can harness the power of Big Data without the typical cost.

Key Features: 

  • Three open-source versions available
  • Completely MySQL-accessible
  • Familiar, MySQL interface for large-scale, ad hoc BI
  • Dimensional and predictive analytics
  • Integrates with the Hadoop ™ Distributed File System (HDFS)
  • Real-time, ad hoc analytics within an Apache Hadoop cluster

Cost: FREE

34. GridGrain


GridGrain reimagines in-memory computing for a competitive edge in the modern business environment. Nikita Ivanov and Dmitriy Setrakyan share a passion for high-performance computing, a shared vision on which they based the first release of GridGrain in 2007. The list of features, functionality and capabilities of these solutions is astounding.

Key Features:

  • In-Memory Data Grid
  • Supports SQL, K/V, MongoDB, MPP, MapReduce
  • Hyper Clustering
  • Zero Deployment
  • Advanced Security
  • Fault Tolerance
  • Load Balancing
  • Customizable Event Workflow
  • Programmatic Querying
  • Minimal or no integration
  • No ETL required
  • Eliminate MapReduce overhead
  • Works with any Hadoop distribution

Cost: Contact for a quote

 35. DeepDive


A new type of system to help developers analyze data on a deeper level, DeepDive is an open-source project with a simple four-step process for writing applications on the platform. With calibrated probabilities for every assertion it makes, DeepDive is designed to navigate around the problematic nature of human error in development.

Key Features: 

  • Handles large amounts of data from multiple sources
  • Write simple rules and offer feedback on prediction accuracy
  • “Distantly” learns, rather than requiring a tedious machine-learning process for training predictions
  • Scaleable, high-performance inference and learning engine

Cost: FREE

36. Lavastorm Analytics


Business turn to analytics to develop faster, better, and cheaper methods and processes. Interestingly, that’s precisely what Lavastorm offers in its platform: faster, better, and cheaper analytics for achieving business goals. Enterprises are demanding more capabilities and faster speeds, and Lavastorm eliminates the need for a disjointed approach with visualization tools, spreadsheets, BI applications, databases and other information silos with a seamless solution delivering end-to-end analytics.

Key Features: 

  • Reduce analytic development by 90% or more
  • Large volumes of data in short amounts of time
  • Reuse and share analytics knowledge across teams
  • Detect hard-to-find issues with 40% less false positives
  • Visibility control for management and executives

Cost: Contact for a quote

37. SpagoWorld


From business intelligence to middleware, SpagoWorld offers a range of solutions for enterprises — all on an open-source platform. SpagoWorld’s Big Data BI solution enables the collection of massive quantities of data, in rapid timeframes, for use across SpagoWorld’s other platforms for further analysis and business intelligence derivatives.

Key Features: 

  • Extract data from various platforms, from database and analytics platforms to NoSQL databases or enriched distributions
  • Supports real-time analysis of streaming data
  • Charts, reports, thematic maps, cockpits
  • Translate information to self-service BI
  • Reporting, multi-dimensional analysis
  • Ad hoc reporting
  • Location intelligence
  • Real-time dashboards and console

Cost: FREE

38. RapidMiner


A code-free zone, RapidMiner provides advanced analytics with no programming required for configuration. With all-new application wizards for churn reduction, sentiment analysis, predictive maintenance and direct marketing in RapidMiner 6.0, this tool is one of the fastest advanced analytics solutions available.

Key Features: 

  • Hundreds of methods for data integration
  • Runs on every major platform
  • No programming required
  • Drag-and-drop interface

Cost: Contact for a quote

39. Orange

Orange is an open-source data visualization and analysis tool for both novices and experts. Data mining is conducted either through visual programming or Python scripting, with components for machine learning and ad-ons for bioinformatics and text mining.

Key Features: 

  • Remembers choices and makes suggestions
  • Intelligently chooses communication channels between widgets
  • Packed with visualization options from bar charts to dendograms
  • Integration and data analytics
  • Combine widgets to design the framework of your choice
  • Toolbox with more than 100 widgets

Cost: FREE

40. OpenDataSoft


OpenDataSoft is a comprehensive discovery tool with maps, charts, and graphs to explore public data sets. A cloud-based platform, OpenDataSoft is designed for seamless and unlimited data publishing, sharing, and resuse.

Key Features: 

  • Reuse data through APIs and apps models
  • Collect data from any source
  • Read and understand all formats
  • Make databases findable and reusable
  • Standard access formats
  • Interactive & shareable visualization
  • API factory
  • Web extensions and open source

Cost (pricing based on Euros): 

  • FREE – Civic initiatives and academic projects
  • 200/month – 100k records, 20K UI/API queries/day
  • 700/month – 10M records, 100K UI/API queries/day
  • Contact for a quote – Unlimited records, UI/API queries/day


41. Angoss


A comprehensive marketing analytics solution, Angoss offers real-time Big Data insights for a variety of verticals and business sectors. From credit scoring to opportunity and lead scoring, fraud deterrence and claims management, Angoss is capable of capturing and analyzing data for a multitude of applications.

Key Features:

  • Automated workflows to develop scorecards
  • Select the most predictive variables
  • Advanced predictive modeling
  • Angoss Decision and Strategy Trees
  • Data preparation and profiling
  • Model validation and deployment

Cost: Contact for a quote

42. Mu Sigma



Mu Sigma is one of the world’s largest Decision Sciences and analytics firms, helping companies to institutionalize data-driven decision making by harnessing Big Data. With a set of proprietary platforms to enable rapid decision-making and comprehensive data collection and integration that eliminates information silos, Mu Sigma is a powerful tool for machine learning, operational research, artificial intelligence, and more.

Key Features: 

  • Hosts Mu Sigma problem DNAs
  • Real-time analytics and event stream processing
  • Load models into an enterprise ecosystem for consumption
  • Embedded advanced analytics engine
  • Influence analysis and topic modeling
  • Sentiment evaluation
  • Easily scaled on commodity hardware

Cost: Contact for a quote

43. ERwin



A collaborative data-monitoring environment, ERwin offers an intuitive, graphical interface with a centralized view of key definitions, enabling the leveraging of data as strategic business asset. The product is comprised of a number of editions designed for different stakeholders within an organization, providing a targeted level of information availability and display and configurations for better understanding and usability.

Key Features: 

  • Achieve business agility through model-driven collaboration
  • Collaborate via web or desktop
  • Active model templates and naming standards
  • Display themes, custom data types, macro language and API
  • Custom reporting
  • Metadata integration tools


  • CA ERwin Data Modeler Standard Edition r9.5 – Product plus 1 Year Enterprise Maintenance – $4,794
  • CA ERwin Data Modeler Standard Edition r9.5 – Product plus 3 Years Enterprise Maintenance – $6,392
  • CA ERwin Data Modeler Workgroup Edition r9.5 – Product plus 1 Year Enterprise Maintenance – $6,708
  • CA ERwin Data Modeler Workgroup Edition r9.5 – Product plus 3 Years Enterprise Maintenance – $8,944
  • CA ERwin r9.5 Data Modeler for Microsoft SQL Azure – Product plus 1 Year Enterprise Maintenance – $1,679.94
  • CA ERwin r9.5 Data Modeler for Microsoft SQL Azure – Product plus 3 Years Enterprise Maintenance – $2,239.92
  • CA ERwin r9.5 Web Portal Standard Edition 1-5 Users – Product plus 1 Year Enterprise Maintenance – $8,399.70
  • CA ERwin 9.5 Web Portal Standard Edition 1-5 Users – Product plus 3 Years Enterprise Maintenance – $11,199.60

44. HPCC Systems


A proven and battle-tested platform for manipulating, querying, transforming, and data warehousing Big Data, HPCC Systems solves Big Data problems facing modern enterprises in any vertical.

Key Features: 

  • Processing clusters use off-the-shelf hardware
  • Clusters typically homogeneous, but not required
  • Distributed, Thor, Roxie file systems
  • Linux operating system
  • Build multi-key, multi-field (aka compound) indexes on DFS files
  • Data warehouse capabilities for structured queries and data analysis applications
  • Supports thousands of users with sub-second response time, depending on application

Cost: FREE

45. pmOne


Offering Big Data and Business Intelligence solutions, pmOne’s cMORE enables users to quickly build, flexibly grow and efficiently administer solutions. It leverages and extends SQL Server functionality, as well as that of Excel, SharePoint, and other components in the Microsoft BI stack.

Key Features: 

  • Simplified standard and ad hoc reporting
  • Credible alternative to SAP-based data warehouse
  • Consistent reporting company-wide
  • Personalize reports; distribute books
  • Easy access to SAP data and other systems
  • Based on Microsoft BI

Today, Quentin Clark, corporate vice president of the Data Platform Group, announced that SQL Server 2014 has released to manufacturing and Windows Azure HDInsight supporting Hadoop 2.2 is now generally available. SQL Server 2014 delivers hybrid cloud enhancements that make it easy to extend your on-premises database environment to the cloud. HDInsight delivers incremental performance and data compression improvements while leveraging the power of YARN.

Windows Azure HDInsight is Microsoft’s 100 percent Apache Hadoop-based distribution forWindows Azure. It allows customers to dynamically provision Hadoop clusters to process data and leverage the elastic scale of Windows Azure. Customers can leverage their existing Windows Azure Blob storage or store data in the native Hadoop HDFS file system and dynamically provision Hadoop clusters to process data and leverage the elastic scale of Windows Azure.


Predictive analytics and other categories of advanced analytics are becoming a major factor in the analytics market. We evaluate the leading providers of advanced analytics platforms that are used to build solutions from scratch.

Advanced Analytics Platforms

Advanced Analytics Platforms