Deploying Hadoop on EC2 with Whirr
Apache Whirr is a set of tools to deploy cloud services. It can be used on Amazon Elastic Cloud(EC2), Rackspace Cloud and many other cloud providers. Requirement You need to have an account on Amazon...
View ArticlePlaying with Hadoop Pig
Hadoop Pig is a tool to manipulate data from various sources (CSV file, MySQL, MongoDB, …) using a procedural language (Pig Latin). It can run standalone or distributed with Hadoop. Unlike Hive, it can...
View ArticlePlaying with the Mahout recommendation engine on a Hadoop cluster
Apache Mahout is an open source library which implements several scalable machine learning algorithms. They can be used among other things to categorize data, group items by cluster, and to implement a...
View ArticlePlaying with Apache Hive, MongoDB and the MTA
Apache Hive is a popular datawarehouse system for Hadoop that allows to run SQL queries on top of Hadoop by translating queries into Map/Reduce jobs. Due to the high latency incurred by Hadoop to...
View ArticlePlaying with Apache Hive and SOLR
As described in a previous post, Apache SOLR can perform very well to provide low latency analytics. Data logs can be pre-aggregated using Hive and then synced to SOLR. To this end, we developed a...
View ArticleInstalling and comparing MySQL/MariaDB, MongoDB, Vertica, Hive and Impala...
A common thing a data analyst does in his day to day job is to run aggregations of data by generally summing and averaging columns using different filters. When tables start to grow to hundreds of...
View ArticleUsing the Mahout Naive Bayes Classifier to automatically classify Twitter...
In this post, we are going to categorize the tweets by distributing the classification on the hadoop cluster. It can make the classification faster if there is a huge number of tweets to classify. To...
View Article