Cloud’s New Language Set to Bloom

By Nicole Hemsoth

April 26, 2010

MIT’s Technology Review provided some in-depth coverage on the top ten emerging technologies for the coming year. On a list that includes solar fuel, green concrete, and light-trapping photovoltaics, the publication tacked on the concept of “cloud programming” — a revolutionary movement that is a attributed in large part to the work of Joseph Hellerstein and his team at the University of California, Berkeley.

Hellerstein and other researchers at Berkeley have created a language called “Bloom,” which has received a great deal of attention following its implementation in Hadoop. Hellerstein was able to make key alterations to Hadoop to render it less likely to hit critical points of failure and simplify scheduling processes among other unique “quick fixes.” The special ingredient in this language that makes it so functional and useful is simplicity — it casts aside the need to program around dependencies and unknowns, allowing for innovation and creativity versus what amounts to educated guesswork for programmers. While it’s not clear what immediate life-altering changes will be evident in HPC and enterprise computing upon the release date, the Bloom language is making cloud programming far more intuitive and applicable for a community far larger than a cluster of specialized infrastructure engineers. By removing the complex barriers to cloud programming, there is no end in sight for new developments for cloud-specific platforms. The language will be released in full later in 2010, but it is already gathering great interest from cloud infrastructure providers, among others.

As the MIT Technology Review describes, since prior to Hellerstein’s work, languages processed data in static batches, but they were unable to process data that was changing in real-time. Hellerstein’s “fix” for this problem is “to build into the language the notion that data can be dynamic, changing as it’s being processed. This sense of time enables a program to make provisions for data that might be arriving later — or never.”

Writing cloud applications is nothing short of a programming nightmare, even for the most experienced programmers. This is due to the fact that managing and tracking data across time is filled with such a host of unknowns that it can be nearly impossible to do accurately. Furthermore, since the miniscule errors that threaten programming tasks are difficult to detect (let alone repeat and debug), the process of writing cloud applications is thwarted at worst and messy at best.

Hellerstein’s idea is geared at finding a way to tidy the programming process by altering the database programming languages so they can be easily unified and the focus can be on creativity and innovation versus the time-consuming task of mere data tracking.

The following is a transcript from HPC in the Cloud’s interview with Joseph Hellerstein:

HPCc: Can you give some background on the Boom and Bloom project — how the idea developed, how long ago you started working on the language?

Joseph Hellerstein: This project began about eight years ago when we were spending time looking at some of the simple infrastructure involved in Internet search (which included Web crawlers and  Google’s PageRank algorithm for ranking pages to give better answers) and we noticed that in both cases, the algorithms they were using looked a lot like walking along a graph — the Web is like a graph of Web pages that are connected to one another and the crawler will follow the links and “walk” around on a network or graph of nodes. The page rank algorithm for ranking walks the same sort of graph but in its own particular way. Both of these algorithms were things that in classic logic language anyway, you could easily write in a few lines. So back in 2003 we were wondering why a programmer should have to work so hard to build these when they could be written so simply. With this in mind we set out to build an infrastructure so that a basic three-line program could achieve those goals and perform those tasks.

We did some initial things with that concept. For instance, we built a deep Web crawler, a crawler for the Gnutella P2P file-sharing network — and when we done we realized that what we were doing was actually a lot like routing on a network; that what a web crawler does is it moves its attention from node to node on the internet something like a packet as it is sent from node to node on the internet from a source to a destination.

After thinking about this, we wondered if it would be possible for us to implement TCP/IP in the same way we implemented these crawlers and use very simple logic programs. The more we thought about it, the more excited we got because the answer to that was yes, as we started thinking, we got excited because a lot of the basic protocol routing on the internet and wireless networks could be set at a much higher level in these logic languages. From then on, we set ourselves to task of building a system and language to let people build network protocols.

So all of this takes us to close to the beginning of the new century when we were doing a lot of work with what we called declarative networking, which make it easy for people to build network protocols. And when we hit this point, we had this really cool language for people who built plumbing for the internet but for these folks who already like how they’ve been doing things, there just wasn’t a lot of demand. So we were at a point in the research where we had this really great idea but the application domain it was developed for either isn’t that interesting or those who could make use of it aren’t interested.

At about this time that we had this developed and weren’t seeing a lot of uptake on the idea, cloud computing started getting a lot attention and at some point, collectively really, a light bulb went off as we realized that cloud is the perfect environment — that we needed an easier way to program things that would go across lots of different computers. And everybody needed this, not just a select group of infrastructure engineers. Cloud computing is where we saw a critical need for a programming paradigm shift.

HPCc: So this new language was not driven by a specific need or to answer to the problem of programming difficulty across large numbers of computers — it evolved on its own?

Hellerstein: No, it took on a life of its own, but actually, to back up a bit, it is a nice story about research and long-term payoff.

The roots of this go back to the 70s and 80s in database theory — back in the 70s there were some early papers saying that we should be able to program with simple mathematical logic, with just writing down rules of inference — of course, this was considered to be something of a curiosity. From this, there was a language called Prolog, which had some brief excitement and the database community picked it up in a theoretical context and developed a cleaner language called DataLog. There was a lot of controversy surrounding all of this because the theorists were wring about the esoteric principles behind this and the applied people in the field, some of whom were very influential, were telling them it was useless and had no utility — they said this in print. The applied folks and theorists — there was some bad blood there.

In early 200 when we were playing with crawlers and routing we thought that stuff might be useful in practice. Coincidentally, at the same time, a bunch of other groups in different research areas (security, complier analysis, natural language processing) started picking up DataLog and started to revive this esoteric language. In short, we’re part of the payoff of work that went on when I was in high school and college.

HPCc: What major problems does this language solve for programmers and what will it mean for large-scale computing in the cloud?

Hellerstein: First of all, the hard part about programming multiple computers at once arguably boils down to the interweaving of actions in time across the different computers since they’re all working at the same time. A second part of the problem has to do with the fact that people’s versions of data are often different across different machines.

We can put this problem in the form of a simple analogy: You can imagine writing down a recipe you want someone to cook. If you know it’s going to be a single cook you write down the order of operations and it works fine. But you take that same recipe to a commercial kitchen at a hotel and break down the pieces of the recipe and assign them to sets of different chefs, the coordination is a little different. You need to make sure each step is complete at the right time to be handed off to the next chef and that the coordination from one step to the next works. When you do things this way — have multiple “chefs” it just becomes a more complicated problem.

So to bring it back to the concept at hand, when people think about programming a thousand computers they write a single program that when run it a thousand times. When you do this, you just hope the interweavings of the actions it takes work out right — reasoning it out is very tricky because you can’t control the speed every computer runs at, or at which the messages travel to communicate and coordinate and can’t control the possibility that some of them will crash in the middle. Programming around these eventualities is really hard — even great programmers find this hard. Debugging these is also hard because the problem might be based on very subtle errors that cannot be reproduced.

The other half of the problem has to do with data and the dependencies between different versions of data. If it’s looking at a copy of some piece of data and another machine is looking at a different copy of the piece they might come to different conclusions and take different actions. It’s worse than just copied data — if it’s dependency-related, it’s not that there are different copies — there are results of different copies, which leads to a mangled final product.

To put it simply, our what our work is trying to do is start with the data itself and get people to talk about what should happen to the data step-by-step through a program without ever having them specify at all how many machines are involved. So, when you ask a query of a database you describe what data you want — not how to get it. There’s a database engine underneath that gives the “how” and that database engine has built-in operations that can deal with the tasks of query processing in a parallel fashion.

Query processing tasks are easy to parallelize — they rip apart the data into pieces and operate on the pieces independently. If we can get people to specify what the program should produce versus how it should do it, the system underneath will take care of the details. All the system underneath needs to do is partition data the way that a database query processor does.

HPCc: Can you describe how the Bloom language worked in a test project like the Hadoop modifications you made?

Hellerstein: Our work on Hadoop is not unlike the first things we did with crawlers; we looked at the Hadoop system and said, “it’s not that hard — so why is it such a big piece of software?”

Conceptually what we were doing was really simple; you think it shouldn’t be hard but it’s very big software and is complex and hard to change. We wanted to show our language was real so that everyone could see it in practice. So what we did, we completely rebuilt the Hadoop file system from scratch in one of our earlier languages to show we could do it and that it was in every way compatible with the Hadoop file system. We built the api-compatible version of Hadoop then we went ahead and said “what’s the yuckiest, hardest to modify pieces of Hadoop” and extracted that — which was a scheduling piece of Map Reduce, and we went ahead and put in a scheduler implemented in our language. It worked — we showed it could be smaller, simpler, and easier to make Hadoop better so we moved forward. Our next project was to examine the single point of failure in Hadoop — the point at which everything is lost due to this one master node. We talked to IBM and found that they’d lost terabytes upon terabytes of data. So we implemented PAXOS, which is a protocol to get the computers in a distributed environment to agree with each other and this one-month change altered Hadoop for the better.

Another aspect of Hadoop we looked at involved the master node in times of need; So the master node needs more hosepower if you run out of power or memory and you have to buy another bigger one, which is a very “un-Hadoopy” thing to do. The master is the gatekeeper to scaling — we saw we could partition the work of the master across many nodes. Upon realizing this, it took us 24 hours and we were able to parallelize this across machines. This kind of dramatic increase in ability is what we wanted to show.

HPCc: Where is the most interest in this coming from right now? Who are going to be the early adopters working with this in mission-critical or HPC environments?

Hellerstein: The people we’ve been talking to the most are those who are building cloud and data infrastructure, in part because those are the applications we’ve been focusing on. So the cloud research group at Microsoft, the people at SAP — enterprise computing is a natural place to think about this taking shape first within but it’s not likely to change very fast in practice is my guess. Enterprise as an industry is conservative because of the cost of entry is high.  The whole market moves slow — so even though those are the natural obvious applications and adopters we think about and it could be that something very lightweight is the first thing to pick this up. The people we have looking closely are all data people — people with big data problems but whoever thought something like Twitter would become the next big thing? There’s no way to tell.

HPCc: How are we going to see this manifest down the line? How is this is going to create a “boom” in cloud computing, even if it is starting at the top slowly and trickling down?

Hellerstein: When someone presents a platform with new capabilities, you probably don’t know what people will do with it yet — it’s wide open. For instance, when mobile phones with compute power started coming out there was some talk about email on the phones and location services, but people really didn’t know how mobile phones would change application developing and what kinds of applications we’d want to run on our phones. That’s a story that’s still developing, actually. In some senses, application developers are not aware of what the next “killer app” will be but people do have the tools to create it, thus it becomes about innovation. That’s where we are with cloud right now; what will happen when people have the power of a thousand computers at their disposal for development?

We’re not starting to try to develop the killer app — we’re showing how easy we can make it to build what we know already works.  The people who really need all thousand computers at once are who we protoyped this for: infrastructure services — the cloud infrastructure providers like Google and Amazon, who need to have large-scale data export and analytics systems. We showed it’s much easier to build those things in our environment than it was originally. That leaves the door open to innovation.

To read more about Joseph Hellerstein, go to his U.C. Berkeley page that has numerous links to papers and reports on this and other subjects relevant to the HPC and cloud community.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

European HPC Summit Week and PRACEdays 2018: Slaying Dragons and SHAPEing Futures One SME at a Time

June 20, 2018

The University of Ljubljana in Slovenia hosted the third annual EHPCSW18 and fifth annual PRACEdays18 events which opened May 29, 2018. The conference was chaired by PRACE Council Vice-Chair Sergi Girona (Barcelona Super Read more…

By Elizabeth Leake (STEM-Trek for HPCwire)

An Overview of ‘OpenACC for Programmers’ from the Book’s Editors

June 20, 2018

In an era of multicore processors coupled with manycore accelerators in all kinds of devices from smartphones all the way to supercomputers, it is important to train current and future computational scientists of all dom Read more…

By Sunita Chandrasekaran and Guido Juckeland

Cray Introduces All Flash Lustre Storage Solution Targeting HPC

June 19, 2018

Citing the rise of IOPS-intensive workflows and more affordable flash technology, Cray today introduced the L300F, a scalable all-flash storage solution whose primary use case is to support high IOPS rates to/from a scra Read more…

By John Russell

HPE Extreme Performance Solutions

HPC and AI Convergence is Accelerating New Levels of Intelligence

Data analytics is the most valuable tool in the digital marketplace – so much so that organizations are employing high performance computing (HPC) capabilities to rapidly collect, share, and analyze endless streams of data. Read more…

IBM Accelerated Insights

Preview the World’s Smartest Supercomputer at ISC 2018

Introducing an accelerated IT infrastructure for HPC & AI workloads Read more…

Lenovo to Debut ‘Neptune’ Cooling Technologies at ISC

June 19, 2018

Lenovo today announced a set of cooling technologies, dubbed Neptune, that include direct to node (DTN) warm water cooling, rear door heat exchanger (RDHX), and hybrid solutions that combine air and liquid cooling. Lenov Read more…

By John Russell

European HPC Summit Week and PRACEdays 2018: Slaying Dragons and SHAPEing Futures One SME at a Time

June 20, 2018

The University of Ljubljana in Slovenia hosted the third annual EHPCSW18 and fifth annual PRACEdays18 events which opened May 29, 2018. The conference was chair Read more…

By Elizabeth Leake (STEM-Trek for HPCwire)

Cray Introduces All Flash Lustre Storage Solution Targeting HPC

June 19, 2018

Citing the rise of IOPS-intensive workflows and more affordable flash technology, Cray today introduced the L300F, a scalable all-flash storage solution whose p Read more…

By John Russell

Sandia to Take Delivery of World’s Largest Arm System

June 18, 2018

While the enterprise remains circumspect on prospects for Arm servers in the datacenter, the leadership HPC community is taking a bolder, brighter view of the x86 server CPU alternative. Amongst current and planned Arm HPC installations – i.e., the innovative Mont-Blanc project, led by Bull/Atos, the 'Isambard’ Cray XC50 going into the University of Bristol, and commitments from both Japan and France among others -- HPE is announcing that it will be supply the United States National Nuclear Security Administration (NNSA) with a 2.3 petaflops peak Arm-based system, named Astra. Read more…

By Tiffany Trader

The Machine Learning Hype Cycle and HPC

June 14, 2018

Like many other HPC professionals I’m following the hype cycle around machine learning/deep learning with interest. I subscribe to the view that we’re probably approaching the ‘peak of inflated expectation’ but not quite yet starting the descent into the ‘trough of disillusionment. This still raises the probability that... Read more…

By Dairsie Latimer

Xiaoxiang Zhu Receives the 2018 PRACE Ada Lovelace Award for HPC

June 13, 2018

Xiaoxiang Zhu, who works for the German Aerospace Center (DLR) and Technical University of Munich (TUM), was awarded the 2018 PRACE Ada Lovelace Award for HPC for her outstanding contributions in the field of high performance computing (HPC) in Europe. Read more…

By Elizabeth Leake

U.S Considering Launch of National Quantum Initiative

June 11, 2018

Sometime this month the U.S. House Science Committee will introduce legislation to launch a 10-year National Quantum Initiative, according to a recent report by Read more…

By John Russell

ORNL Summit Supercomputer Is Officially Here

June 8, 2018

Oak Ridge National Laboratory (ORNL) together with IBM and Nvidia celebrated the official unveiling of the Department of Energy (DOE) Summit supercomputer toda Read more…

By Tiffany Trader

Exascale USA – Continuing to Move Forward

June 6, 2018

The end of May 2018, saw several important events that continue to advance the Department of Energy’s (DOE) Exascale Computing Initiative (ECI) for the United Read more…

By Alex R. Larzelere

MLPerf – Will New Machine Learning Benchmark Help Propel AI Forward?

May 2, 2018

Let the AI benchmarking wars begin. Today, a diverse group from academia and industry – Google, Baidu, Intel, AMD, Harvard, and Stanford among them – releas Read more…

By John Russell

How the Cloud Is Falling Short for HPC

March 15, 2018

The last couple of years have seen cloud computing gradually build some legitimacy within the HPC world, but still the HPC industry lies far behind enterprise I Read more…

By Chris Downing

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 19, 2018

Petaflop per second deep learning training performance on the NERSC (National Energy Research Scientific Computing Center) Cori supercomputer has given climate Read more…

By Rob Farber

Lenovo Unveils Warm Water Cooled ThinkSystem SD650 in Rampup to LRZ Install

February 22, 2018

This week Lenovo took the wraps off the ThinkSystem SD650 high-density server with third-generation direct water cooling technology developed in tandem with par Read more…

By Tiffany Trader

ORNL Summit Supercomputer Is Officially Here

June 8, 2018

Oak Ridge National Laboratory (ORNL) together with IBM and Nvidia celebrated the official unveiling of the Department of Energy (DOE) Summit supercomputer toda Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Hennessy & Patterson: A New Golden Age for Computer Architecture

April 17, 2018

On Monday June 4, 2018, 2017 A.M. Turing Award Winners John L. Hennessy and David A. Patterson will deliver the Turing Lecture at the 45th International Sympo Read more…

By Staff

Leading Solution Providers

SC17 Booth Video Tours Playlist

Altair @ SC17

Altair

AMD @ SC17

AMD

ASRock Rack @ SC17

ASRock Rack

CEJN @ SC17

CEJN

DDN Storage @ SC17

DDN Storage

Huawei @ SC17

Huawei

IBM @ SC17

IBM

IBM Power Systems @ SC17

IBM Power Systems

Intel @ SC17

Intel

Lenovo @ SC17

Lenovo

Mellanox Technologies @ SC17

Mellanox Technologies

Microsoft @ SC17

Microsoft

Penguin Computing @ SC17

Penguin Computing

Pure Storage @ SC17

Pure Storage

Supericro @ SC17

Supericro

Tyan @ SC17

Tyan

Univa @ SC17

Univa

Google Chases Quantum Supremacy with 72-Qubit Processor

March 7, 2018

Google pulled ahead of the pack this week in the race toward "quantum supremacy," with the introduction of a new 72-qubit quantum processor called Bristlecone. Read more…

By Tiffany Trader

Google I/O 2018: AI Everywhere; TPU 3.0 Delivers 100+ Petaflops but Requires Liquid Cooling

May 9, 2018

All things AI dominated discussion at yesterday’s opening of Google’s I/O 2018 developers meeting covering much of Google's near-term product roadmap. The e Read more…

By John Russell

Nvidia Ups Hardware Game with 16-GPU DGX-2 Server and 18-Port NVSwitch

March 27, 2018

Nvidia unveiled a raft of new products from its annual technology conference in San Jose today, and despite not offering up a new chip architecture, there were still a few surprises in store for HPC hardware aficionados. Read more…

By Tiffany Trader

Pattern Computer – Startup Claims Breakthrough in ‘Pattern Discovery’ Technology

May 23, 2018

If it weren’t for the heavy-hitter technology team behind start-up Pattern Computer, which emerged from stealth today in a live-streamed event from San Franci Read more…

By John Russell

HPE Wins $57 Million DoD Supercomputing Contract

February 20, 2018

Hewlett Packard Enterprise (HPE) today revealed details of its massive $57 million HPC contract with the U.S. Department of Defense (DoD). The deal calls for HP Read more…

By Tiffany Trader

Part One: Deep Dive into 2018 Trends in Life Sciences HPC

March 1, 2018

Life sciences is an interesting lens through which to see HPC. It is perhaps not an obvious choice, given life sciences’ relative newness as a heavy user of H Read more…

By John Russell

Intel Pledges First Commercial Nervana Product ‘Spring Crest’ in 2019

May 24, 2018

At its AI developer conference in San Francisco yesterday, Intel embraced a holistic approach to AI and showed off a broad AI portfolio that includes Xeon processors, Movidius technologies, FPGAs and Intel’s Nervana Neural Network Processors (NNPs), based on the technology it acquired in 2016. Read more…

By Tiffany Trader

Google Charts Two-Dimensional Quantum Course

April 26, 2018

Quantum error correction, essential for achieving universal fault-tolerant quantum computation, is one of the main challenges of the quantum computing field and it’s top of mind for Google’s John Martinis. At a presentation last week at the HPC User Forum in Tucson, Martinis, one of the world's foremost experts in quantum computing, emphasized... Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This