Open source software, such as the Apache HadoopⓇ standard within the Big Data realm, has become the default and dominate choice when companies are choosing to deploy software. Not too long ago, most executives didn’t necessarily see how much of their operations run on open source. Now, they do. There are three key reasons why, and we outline these in this post along with an upcoming webinar on an Open Source Playbook for 2016.
I recently attended PGconf in Vienna, Austria, where we announced the open-sourcing of Pivotal Greenplum, which has become the first open source massively parallel data warehouse. Now known as Greenplum Database in it's open source form, anyone can clone the github repo and build the product, but there is another segment of the community that just wants to try out the functionality of the product without going through that process. For that group, we now have the Pivotal Greenplum Sandbox Virtual Machine that allows a free trial of the open source Greenplum Database, the commercially available Pivotal Greenplum Command Center management tool, Apache MADlib (incubating), PostGIS, PL/R, PL/Perl, and PL/Java into an easy-to-use virtual machine which runs in either VirtualBox or VMware Fusion.
Welcome to the November 2015 edition of the Build Newsletter. For this edition, we start out with the big Dell acquisition news, talk about our donation of 5 million lines of code to open source, discuss unicorns and disruption, and provide a huge list of Cloud Native development updates.
It is an exciting time for customers using data-related software—with more data being processed an analysed than ever. To do this effectively you need some heavy duty technology. And one of the best for a long time has been the Pivotal Greenplum Database. Now, this database is available as open source—giving customers even more flexibility, and enabling a flourishing community of developers and contributors. In this episode, host Simon Elisha explores more about what this means and how you can be a part of it.
Today, Pivotal unveiled the first massively parallel processing (MPP) data warehouse to open source. The release of Greenplum Database is significant for two important reasons. First, while it is the only open source option for data warehousing, it also leads the industry with query optimization technology that is multiple times faster than any other commercial offering today. Second, it completes Pivotal's 10 month transition to engage more openly with all the communities we affect—and helping companies of all types on their digital transformation journey. With nearly 10 million lines of code released this year, this release signals to the industry that a sea change has happened, and the days of closed source, vendor-lock in, legacy models are coming to an end.
On Saturday, October 17th, Open Data Platform Initiative founding partners Hortonworks and Pivotal co-sponsored a HackFest for Apache Ambari. Participants were directed to use Apache Ambari’s extensible framework to write an Ambari Service or create customized Views to help understand, view, and manage cluster resources in a multi-tenant Apache Hadoop environment. Here’s some of the projects and ideas that resulted from the hackathon.
After several big events and announcements at Spring One, ApacheCon Europe, Strata + Hadoop NYC, VLDB, and more, we have tons of industry news and updates and a list of upcoming events. As always, we give architects, engineers, and digital leaders a perspective on digital transformation from Pivotal’s Cloud Native viewpoint. Read on for some key open source news, hear the highlights in the Spring ecosystem, get a sense of the latest in Cloud Native platforms, and peer into some of the most interesting IoT stuff going on—like 3D printed chocolate based on open source patterns! Lastly, software is one thing, but meatware is another—digital transformation means we have to update the meatware.
Today, Pivotal announced it has open sourced HAWQ and MADlib, contributing them to the Apache Software Foundation (ASF) where they are now officially listed as incubating. In this post, Pivotal’s data engineering leader, Gavin Sherry explains why HAWQ and MADlib are needed to create a Hadoop Native SQL infrastructure, and why the only way forward to do that is through open governance and and curation managed by the ASF.
Last Saturday, Pivotal hosted a "small" PostgreSQL Conference in Beijing. At least that is what we initially asked for: "a little Meetup, one or two speakers". Over 100 people showed up and discussed some pressing issues regarding the database market in China, including Postgres relationship with products like Pivotal Greenplum Database, MySQL and Apache Hadoop, as well as many of the open source community efforts across the broader community.
In this post, Pivotal’s Chief Scientist, Jignesh Patel, recaps the key talks and papers contributed by Pivotal at the 41st annual VLDB conference. Covering common table expressions, improvements from hardware/software collaboration, a new perspective on benchmarks, R and hardware utilization issues, and mapping relational learning to relational algebra, Jignesh explains and provides reference links to all of the topics. He also provides links to the presentations being covered again at several meetups.