Splice Machine Moving Its Database to Open Source
If you want to test the database at scale at incredible loads, you can do so with the community edition. You can even go live on the community edition, but there are a number of operational features that will be held back for the enterprise edition, like backup features, other security features and authentication features. These are mostly features that operational people in the organization—not the developers—care about. These features are essentially related to governance and reliability, and Splice Machine will charge a per-node price for them. Meanwhile, the company is actively seeking contributors and thought leaders in database architecture and distributed systems to help guide and support the transition, develop best practices and help shape next-generation features that best suit the open-source community, Zweben said. "The evolution of Splice Machine from being the first transactional RDBMS on Hadoop, to incorporating Apache Spark as an analytical engine, has been amazing to watch as a member of their Advisory Board," said Mike Franklin, former chair of the School of Computer Science at the University of California, Berkeley, and incoming chair of Computer Science at the University of Chicago. "Our AmpLab at Berkeley has initiated many open-source projects, including Apache Spark, Apache Mesos and Alluxio—formerly Tachyon. I applaud Splice Machine in taking the leap and joining the open-source community." "We are making the transition to open source to build a larger community around Splice Machine," Zweben said in a statement. "Our whole team is eagerly anticipating the contributions that going open source can enable. We also look forward to being more active within the open-source communities beyond our participation around HBase and Spark."Splice Machine announced its plans to go open source at the Spark Summit 2016 conference, where the company is in beta on its Spark version. Zweben said the Spark version offers an important message to users in that Splice Machine is a dual-engine database that allows users to do both operational workloads and analytical workloads. "At the Spark Summit, almost everyone there is thinking Spark is just for analytics," Zweben said. "But now you'll be able to run an application on Spark, meaning you'll be able to change data on the fly in real-time using standard data management capabilities with transactional integrity and immediately be able to analyze that using Spark." Until now, you could never build an analytical application on Spark, he added. "All you could do was analyze the data you had in your repository," he said. "So this is an important contribution to the Spark community."
Apache Spark is an open-source cluster computing framework used for processing and analyzing big data. Apache HBase, an open-source distributed database modeled after Google's BigTable, is one of the key building blocks of Splice Machine and is what enables real-time updates on top of the Hadoop Distributed File System (HDFS).