Monthly Archives: April 2013


BigSQL Up and Running

BigSQL Up and Running All Big Data projects are a collection of related projects, each suited to a specific function to implement that Big Data solution. BigSQL can be used as a Big Data Data Warehousing solution. In addition to the archetypical Big Data components (Hadoop Distributed File System and MapReduce), there are additional projects [...]

By | April 30th, 2013|Categories: Big Data, BigSQL, Hadoop, High Availability, OpenSource, PostgreSQL|Tags: , , , , , , , , , |Comments Off on BigSQL Up and Running

What is Hive

I recently stumbled across a good clean post from IBM on Hive. It's a couple paragraphs and a nice little 3 minute video. Enjoy!

By | April 26th, 2013|Categories: Uncategorized|Comments Off on What is Hive

Debugging PostgreSQL w Eclipse on OSX

Tonight I worked on getting PostgreSQL running in the Juno Eclipse debugger on Max OSX. It was relatively easy to do by following this Working with Eclipse post on the PostgreSQL wiki site. The cool thing is that the above Wiki post works nicely even though: It is a Linux example rather than OSX It [...]

By | April 26th, 2013|Categories: Uncategorized|Tags: |Comments Off on Debugging PostgreSQL w Eclipse on OSX


BigSQL. From data to information, from information to insight. A state-of-the-art Big Data Warehouse solution that is fast, secure and continuously available. BigSQL will scale from your desktop to the cloud. Run real time OLAP directly from the worlds most secure RDBMS. Get started with BigSQL right now. You can immediately put BigSQL to work [...]

Announcing Hadoop FDW 1.0.7

We are proud to announce the availability of version 1.0.7 of our Haddoop Foreign Data Wrapper for PostgreSQL. Its 100% free and 100% open source. Check it out at:

By | April 20th, 2013|Categories: Big Data, BigSQL, Hadoop, PostgreSQL|Comments Off on Announcing Hadoop FDW 1.0.7

PostgresHA 9.2.4 Released

We are proud to announce that our 100% free and open source Postgres High Availability Bundle (PostgresHA = PostgreSQL + pgHA) is now available as a Developers Sandbox for OSX and Linux 64. This proven solution leverages PostgreSQL's native streaming replication technology and pgBouncer. It efficiently supports multiple read-only slaves, seamless DDL replication, and various [...]

By | April 20th, 2013|Categories: High Availability, OpenSource, PostgreSQL|Comments Off on PostgresHA 9.2.4 Released

PostgreSQL Clustering with Postgres HA

PostgreSQL Clustering made easy with PostgresHA. Authored jointly with Scott Mead, Sr. Architect OpenSCG and Architect of PostgresHA. PostgresHA is Highly Available (HA) PostgreSQL. High Availability High availability is not new. Financial trading systems and the telephone network were federally mandated to provide highly availability. People didn’t do HA because it was a fun project, [...]

By | April 17th, 2013|Categories: High Availability, OpenSCG, OpenSource, PostgreSQL|Tags: , , , , , , , , |Comments Off on PostgreSQL Clustering with Postgres HA

Big Data Implementations

Big Data Implementations. Big Data is conceptually a collection of data and the means to store (a file system) and access it (a runtime). This means that all of the components are swappable including the very core components that were the archetypal definition of Big Data – the file system (HDFS) and the execution environment [...]

By | April 16th, 2013|Categories: Big Data, BigSQL, Fun thoughts, Hadoop|Tags: , , , |Comments Off on Big Data Implementations

The Hadoop Ecosystem

The Hadoop Ecosystem. Big Data is more than just the Apache Hadoop kernel; it is a collection of related projects. Some of them are mandatory (you cannot do Big Data without them) and others are optional (like data loaders and flow processing languages). The mandatory Hadoop projects are HDFS, MapReduce and Common.  Common is a [...]

By | April 16th, 2013|Categories: Big Data, Hadoop, High Availability, OpenSource|Tags: , , , , , |Comments Off on The Hadoop Ecosystem


Hadoop - HDFS and MapReduce. Hadoop came out in 2008 and the primary components of it (HDFS and MapReduce) conceptually look like this. Hadoop, HDFS and MapReduce MapReduce splits work across many machines. Map takes the job, breaks it into many small reproducible pieces and sends them to different machines and runs them [...]

By | April 16th, 2013|Categories: Big Data, Fun thoughts, Hadoop|Tags: , , , , |Comments Off on Hadoop