My wife found this great Upworthy video about rushing kills creativity. Showing my true software process geek, I realized that it would be misinterpreted that faster releases left less room for creativity. Nothing could be farther from the truth…here’s my first video blog!
“Double wide” is not a term I’ve commonly applied to servers, but that’s one of the cool things about this new class of servers that Dell, my employer, started shipping today.
My team has been itching for the chance to start cloud and big data reference architectures using this super dense and flexible chassis. You’ll see it included in our next Apache Hadoop release and we’ve already got customers who are making it the foundation of their deployments (Texas Adv Computing Center case study).
If you’re tracking the latest big data & cloud hardware then the Dell PowerEdge C8000 is worth some investigation.
Basically, the Dell C8000 is a chassis that holds a flexible configuration of compute or storage sleds. It’s not a blade frame because the sleds minimize shared infrastructure. In our experience, cloud customers like the dedicated i/o and independence of sleds (as per the Bootstrapping clouds white paper). Those attributes are especially well suited for Hadoop and OpenStack because they support a “flat edges” and scale out design. While i/o independence is valued, we also want shared power infrastructure and density for efficiency reasons. Using a chassis design seems to capture the best of both worlds.
The novelty for the Dell PowerEdge C8000 is that the chassis are scary flexible. You are not locked into a pre-loaded server mix.
There are a plethora of sled choices so that you can mix choices for power, compute density and spindle counts. That includes double-wide sleds positively brimming with drives and expanded GPU processers. Drive density is important for big data configurations that are disk i/o hungry; however, our experience is the customer deployments are highly varied based on the planned workload. There are also significant big data trends towards compute, network, and balanced hardware configurations. Using the C8000 as a foundation is powerful because it can cater to all of these use-case mixes.
That reminds me! Mike Pittaro (our team’s Hadoop lead architect) did an excellent Deploy Hadoop using Crowbar video.
Last week, my team at Dell led a world-wide OpenStack Essex Deploy event. Kamesh Pemmaraju, our OpenStack-powered solution product manager, did a great summary of the event results (200+ attendees!). What started as a hack-a-thon for deploy scripts morphed into a stunning 14+ hour event with rotating intro content and an ecosystem showcase (videos). Special kudos to Kamesh, Andi Abes, Judd Maltin, Randy Perryman & Mike Pittaro for leadership at our regional sites.
Clearly, OpenStack is attracting a lot of interest. We’ve been investing time in content to help people who are curious about OpenStack to get started.
While I’m happy to be fueling the OpenStack fervor with an easy on-ramp, our primary objective for the Deploy Day was to collaborate on OpenStack deployments.
On that measure, we have room for improvement. We had some great discussions about how to handle upgrades and market drivers for OpenStack; however, we did not spend the time improving Essex deployments that I was hoping to achieve. I know it’s possible – I’ve talked with developers in the Crowbar community who want this.
If you wanted more expert interaction, here are some of my thoughts for future events.
- Expert track did not get to deploy coding. I think that we need to simply focus more even tightly on to Crowbar deployments. That means having a Crowbar Hack with an OpenStack focus instead of vice versa.
- Efforts to serve OpenStack n00bs did not protect time for experts. If we offer expert sessions then we won’t try to have parallel intro sessions. We’ll simply have to direct novices to the homework pages and videos.
- Combining on-site and on-line is too confusing. As much as I enjoy meeting people face-to-face, I think we’d have a more skilled audience if we kept it online only.
- Connectivity! Dropped connections, sigh.
- Better planning for videos (not by the presenters) to make sure that we have good results on the expert track.
- This event was too long. It’s just not practical to serve Europe, US and Asia in a single event. I think that 2-3 hours is a much more practical maximum. 10-12am Eastern or 6-8pm Pacific would be much more manageable.
Do you have other comments and suggestions? Please let me know!
- Sponsors: Canonical & Dell
- Speakers: Joseph George – Dell, Ameet Paranjape - Canonical, Matt Ray – Opscode, Ziad Sawalha – Rackspace
- Attendance: 30 people
- Videos: Jeff Kramer > http://www.jeffkramer.com/2012/02/17/openstack-austin-meetup-videos/
Matt Ray talks about Chef
Matt Ray from Opscode presented some of the work with Chef and OpenStack. He talked about the three main chef repos floating around. He called out Anso’s original cookbook set that is the basis for the Crowbar cookbooks (his second set), and his final set is the emerging set of cookbooks in OpenStack proper. The third one is interesting and what he plans to continue working on to make into his public openstack cookbooks. These are an amalgamation of smokestack, RCB, Anso improvements, and his (Crowbar’s).
He then demoed his knife plugin (slideshare) to build openstack virtual servers using the Openstack API. This is nice and works against TryStack.org (previously “Free Cloud”) and RCB’s demo cloud. All of that is on his github repo with instructions how to build and use. Matt and I talked about trying to get that into our Crowbar distro.
There were some questions about flow and choice of OpenStack API versus Amazon EC2 API because there was already an EC2 knife set of plugins.
Ziad Sawalha talks about Keystone
He split his talk into two pieces, Incubation Process and Keystone Overview. He asked who was interested in what and focused his talk more towards overview than incubation.
Some key take-aways:
- Keystone comes from Rackspace’s strong, flexible, and scalable API. It started as a known quantity from his perspective.
- Community trusted nothing his team produced from an API perspective
- Community is python or nothing
- His team was ignored until they had a python prototype implementing the API
- At this point, comments on API came in.
- Churn in API caused problems with implementation and expectations around the close of Diablo.
- Because comments were late, changes occurred.
- Official implementation lagged and stalled into arriving.
- API has been stable since Diablo final, but code is changing. that is good and shows strength of API.
- Side note from Greg, Keystone represents to me the power of API over Code. You can have innovation around the implementation as long all the implementations have a fair ground work to plan under which is an API specification. The replacement of Keystone with the Keystone Light code base is an example of this. The only reason this is possible is that the API was sound and documented. (Rob’s post on this)
Ziad spent the rest of his time talking about the work flow of Keystone and the API points. He covered the API points.
- Client to Keystone, Keystone to Client for initial auth token
- Client to Middleware API for the services to have a front.
- Middleware to Keystone to verify and establish identity.
- Middleware to Service to pass identity
Not many details other then flow and flexibility. He stressed the API design separated protocol from actions and data at all the layers. This allows for future variations and innovations while maintaining the APIs.
Ziad talked about the state of Essex.
- RBAC (aka Role Based Access Control)
- Many backends
- Code replacement Keystone Light
- LDAP backend
- SQL backend
- AD backend
- Another backend
- Federation was planned but will most likely be pushed to G
- Federation is the ability for multiple independent Keystones to operate (bursting use case)
- Dependent upon two other federation components (networking and billing/metering)
- Ubuntu 11.04 support! Thanks to Justin Shepherd from Rackspace Cloud Builders!
- Alias names for nodes in the UI
- User managed node groups in the UI
- Ability to pre-populate the alias, description and group for a node (not integrated with DNS yet)
- Hadoop is working again – we addressed the missing Ganglia repo issue. Thanks to Victor Lowther.
Also, I’ve spun new open source ISOs with the new features. User beware!
I had a question about moving barclamps between solutions. Since Victor just changed the barclamp build to create a tar for each barclamp (with the debs/rpms), I thought it was the perfect time to explain the new feature.
You can find the barclamps on the Crowbar ISO under “/dell/barclamps” and you can install the TAR onto a Crowbar system using “./barclamp_install foo.tar.gz” where foo is the name of your barclamp.
Here’s a video of how to find and install barclamp tars:
Note: while you can install OpenStack into a Hadoop system, that combination is NOT tested. We only test OpenStack on Ubuntu 10.10 and Hadoop on RHEL 5.7. Community help in expanding support is always welcome!
I’m proud to announce that my team at Dell has open sourced our Apache Hadoop barclamps! This release follows our Dell | Cloudera Hadoop Solution open source commitment from Hadoop World earlier this month.
As part of this release, we’ve created nearly AN HOUR of video content showing the Hadoop Barclamps in action, installing Crowbar (on CentOS), building Crowbar ISOs in the cloud and specialized developer focused builds.
If you want to talk to the Crowbar team. We’re attending events in Boston 11/29, Seattle 11/30, and Austin 12/8.
Here are links to the videos:
- Hadoop Demo: http://youtu.be/eilF16KqRmg
- Install Crowbar from ISO: http://youtu.be/WAUKMlawrPw (prebuilt ISO)
- Basic ISO Build: http://youtu.be/FsOBaAiDgYs
- Advanced ISO Build: http://youtu.be/qvsfXPH5k5Q
Greg Althaus did a 90 minute Crowbar deep dive at this week’s Austin Cloud User Group. Brad Knowles recorded audio and posted it so I thought I’d share the link and my annotations. There are a lot more times to catch up with our team at Dell in Austin, Boston, and Seattle.
Video Annotations - (##:## time stamp)
- 00:00 Intros & Meeting Management
- 12:00 Joseph George Introduction / Sponsorship
- 16:00 Greg Starts – why Crowbar
- 19:00 DevOps slides
- 21:00 What does Crowbar do for DevOps
- make it easier to manage
- make it easier to repeat
- 24:00 What’s included – how we grow / where to start
- 27:20 Starting to show crowbar – what’s included as barclamps
- pluggable / configuration
- 28:10 What is a barclamp
- discussion about the barclamps in the base
- 34:30: We <3 Chef. Puppet vs Chef
- 36:00 Why barclamps are more than cookbooks
- 36:30 State machine & transitions
- Q&A Section
- 38:50 Reference Architectures
- 43:00 Barclamps work outside of Crowbar?
- 44:15 Hardware models supported
- 47:30 Storage Queston
- 49:00 HA progress
- 53:00 Ceph as a distributed cloud on all nodes
- 56:20 Deployer has a map of how to give out roles
- 58:00 Demo Fails
- 58:30 Crowbar Architecture
- 62:00 How Crowbar can be extended
- 63:00 Workflow & Proposals
- 65:40 Meta Barclamps
- 71:10 Chef Environments
- 73:40 Taking OpenStack releases and Environments
- 75:00 The case for remove recipes
- 77:33 Git Hub Tour
- 79:00 Network Barclamp deep dive
- 84:00 Adding switch config (roadmap topic)
- 86:30 Conduits
- 87:40 Barclamp Extensions / Services
- 89:00 Questions
- 89:20 Proposal operations
- 93:30 OpenStack Readiness & Crowbar Design Approach
- 93:10 Network Teaming
- 94:30 Which OS & Hypervisors
- 96:30 Continuous Integration & Tools
- 98:40 BDD (“cucumberesque”) & Testing
- 99:40 Build approach & barclamp construction
- 100:00 Wrap up by Joseph
This release raises the bar on open Hadoop deployments by making them faster, scalable, more integrated and repeatable.
These barclamps were developed in conjunction with our licensed Dell | Cloudera Solution. The licensed solution is for customers seeking large scale and professionally supported big data solutions. The purpose of the open barclamps (which pull the open source parts from the Cloudera distro) is to help you get started with Hadoop and reduce your learning curve. Our team invested significant testing effort in ensuring that these barclamps work smoothly because they are the foundational layer of our for-pay Hadoop solution.
Included in the Hadoop barclamp suite are Hadoop Map Reduce, Hive, Pig, ZooKeeper and Sqoop running on RHEL 5.7. These barclamps cover the core parts of the Hadoop suite. Like other Crowbar deployments (see OpenStack), the barclamps automatically discover the service configurations and interoperate. One of our team members (call him Scott Jensen) said it very simply “I can deploy a fully an integrated Hadoop cluster in a few hours. That friggin’ rocks!” I just can’t put it more eloquently than that!
I’ll post again when we flip the “open” bit and invite our community to dig in and help us continue to set the standards on open Hadoop deployments.
Barton posted these two videos of me talking about the release too:
Hadoop & Crowbar:
Dev’s Only Short:
This week at the OpenStack Design Summit and Conference in Boston, my team unveiled the Diablo+ Crowbar deployments. The OpenStack deployment that’s included with Crowbar reflects a collaborative effort between Dell, Opscode, and Rackspace and pulls packages from the Rackspace repository. It was important for us to use the Rackspace repos so that we could include integrated Keystone and Dashboard components that were omitted from the Diablo (current) release. Our decision to include these Essex (coming) components is based on customer feedback.
Since some of you cannot make it to the show and see the demo in person, we’ve captured it as a video for your enjoyment. The OpenStack deployment is available in our open source distribution. We are currently in QA for the overall solution so expect additional refinement as we progress towards our next OpenStack solution release.
REMINDER: Dell Hardware is NOT required to use Crowbar for OpenStack. The open source version has everything you need – the BIOS and RAID barclamps are optional (but handy).