Podcast: Diane Mueller talks OpenShift Origin

Diane Mueller, the gloriously self styled “PaaS Queen” on our OpenShift team sat down with me at Openstack Developer Summit in Portland almost two weeks ago now and talked everything OpenShift Origin. Boy was it a doozie.. This lady is seriously amazing when it comes to understanding the PaaS marketplace and she needs you to listen to her. She had me at go when we were recording, held me in the palm of her hand the entire session and I would follow her into battle in a heartbeat folks.

Use OpenShift ? Use Heroku ? Thinking about PaaS ? Then you don’t want to miss this podcast, Diane is a heavy hitter and she has really started to blaze a trail at Red Hat since joining us earlier this year. You know when you have one of those “Come to Jesus” moments, someone flicks a light on upstairs and stuff falls into place ? Diane has that ability to shake the wheat from the chaff and demonstrates it in todays show.

Come back soon for two major podcasts on Cloud legality with Kuan Ho, and a discussion with Brian Stevens CTO of Red Hat talking OpenStack and Open Source.

Download the podcast here in MP3 format only

Podcast: OpenStack Developer Summit 2013

rm

I’m in Portland, Oregon at the OpenStack Developer Summit 2013. Place is buzzing and we released Red Hat’s RDO – learn more by visiting the site, signing up for a user account and downloading everything you need – now.

So we recorded a load of podcasts – heres the first featuring two interviews with Dave Neary talking RDO and the drive behind Red Hat’s rolling OpenStack community release and how to get involved. Bill Bauman is up next talking RDO and OpenStack, talking Summit goodness, hear him at IBM’s tech conference next week in Amsterdam.

Download the podcast here in MP3 format only

OpenStack Developer Summit kicks off

Am sat at Openstack Developer Summit listening to Red Hat’s own Dave Neary talking about Personas. Really great presentation. Dave and I missed each other at FOSDEM so this week will get a recording done with him for an upcoming podcast.

If you are local to Portland came say hi, the great news this morning about our RDO community OpenStack release has the venue chattering and it’s adding to the importance or why OpenStack is the most important technology stack for the last decade.

20130415-101111.jpg
I am podcasting all week from Portland so if you want to be on my show on iTunes come grab me.

OpenStack Summit Red Hat discount code

I will be attending the OpenStack Summit in Portland in a few weeks time, it promises to be a great event with developers, users and adoptees from all over the world converging.

Red Hat are one of the principal sponsors of the event and as part of this I’d like to offer attendees a discount code on attendance registration. Usual registration is $600 but we’re offering a discount unique to Red Hat of $150 per attendee, which is a big saving.

I will be podcasting from the event so come say hi to me and the rest of the Red Hat team and talk to us about the cool stuff you can expect over the next weeks and months as we make some announcements around planned launches.

To get the discount visit the Eventbrite portal and during the registration process you will see a section prompting you to enter a promotional code above the green “Register” button.

Simply enter the code RedHat to get the discounted registration rate.

You can see the preliminary conference schedule here.

Look forward to seeing some of you in Portland. If you fancy recording a podcast look me up and let’s sit down and chat on mic.

Solving EU privacy issues with CloudForms

Over the last two and a half years it’s become clearer that despite best efforts there has been a bottleneck in the European Union’s ability to leverage their influence in development of new methodologies of increasing technology consumption or investment in EU cloud.

The clue to the problem lies very much in the lack of credible underlying support that surrounds the European Commissions cloud strategy that emerged in September 2012 that I’ve talked about here before. Their stated aims to increase the spread and adoption of Cloud Computing in EU states were slated to generate about €900bn of generated revenue and a speculative figure of an additional increase in headcount in IT related services by 3.8 million new hires. I’ve read the report in detail and it still makes no sense and just seems to be a finger in the wind (like many analyst reports we all read daily) as to them “taking the temperature of the industry as a whole.”

Maybe it was to buy more time until their slated 2014 time window when the assumption is that the common EU data protection regulations will be outlined. These will replace sovereign data protection acts such as that we take for granted in the UK and to understand the thinking of how that impacts on Cloud.

If we examine how that impacts, say on a company like Amazon, purely as an example, they currently have to implement working practices for AWS in the EU where applicable in contract terms for sovereign customers. These practices have to follow to the letter the data protection acts in France, Germany, Ireland, the UK etc. All those actual data protection acts can be see to be following a skeleton or outline of actual data protection directives issued by the EU but each with their own specific tailored requirements around statute in applicable sovereign territories. So currently it’s hard work for any provider of services to offer a blanket one size fits all across the EU, and the cost of sales and architecture is therefore increased as is cost of adoption for consumption of elastic services generically across multiple territories.

So the hope is we can look forward to 2014 expecting a unified approach to data protection and therefore investment and adoption of catalogue cloud services as an industry. There is no denying that if you have that territorial harmonisation of regulation across the EU it will make it easier for corporations and organisations to build compliance frameworks but also if we were to turn that on it’s head it will create a new raft of operational requirements.

Each member state will have to take on board their individual responsibilities for the legal statute required to make it work and that means additional challenges in Sweden, Germany, Spain and especially France. The workload alone on the part of data controllers facing new responsibilities are going to dramatically increase as well as the definition and creation of procedures and controls. The need to understand how to fit within a new skeleton regulation framework for the management of data privacy then needs to also fold in the needs to handle reporting. We now move to a theoretical world post 2014 where an organisation needs to file compulsory data breach notifications immediately at identification of a data loss or hack.

This all impacts on the lifecycle of cloud services and repudiation of data within contractual periods across multiple territories and potentially multiple providers in open hybrid cloud. This is one of the great facets of ManageIQ capabilities to tag and to “patrol” your complete Cloud fabric in order for you to be able to conform out the box today with responsibilities as a data controller or processor. CloudForms handles Cloud. It doesn’t matter whether thats defined as a public cloud sitting on a provider presence or a private cloud sat in your datacentre. If you’re serious about Cloud you need to have CloudForms in your corner.

An example of this, if for example you have a private cloud the new EU guidance adjudges you to be the processor responsible for data and in most EU states the controller as well and it becomes entirely your position of authority to control the access and protection to that data.

When you start moving those workloads and data upstream to a supplier such as a Red Hat Certified Cloud Provider partner the guidance is clear. The onus is on you the individual to examine at contract and actual practice level that your provider has both the security in place to protect you, but that isn’t enough. You need to be able to do more than just assume a contract keeps you safe without taking on the need for expensive audit procedures and a huge raft of risk registers and rolling pentests / conformance exercises with an often unwilling third party provider who assumes you were happy at the SLA level.

CloudForms combined with ManageIQ give you a single pane capability and the context tagging and reporting doesn’t actually care where your instance is running, be it on a raft of providers on ESX or KVM regardless of location, it just reports and keeps your cloud in line with your controls. It actually draws you in line with the EU regulations ahead of time.

So when the EU regulatory guidance actually becomes more than lipstick on a pig you can look smugly and realise that having implemented CloudForms and MiQ you were ahead of the game, and your business not impacted either by additional regulatory need and complex guidance having a negative impact on your growth.

Expect to read more about CloudForms in the coming weeks and months, for more information engage with your local Red Hat country representative.

Podcast – Mark McLoughlin talks OpenStack

markmSo last night I released a podcast that got a terrific number of downloads with Robyn Bergeron of Fedora and tonight I bring you one of the other podcasts I recorded recently at FOSDEM with Mark McLoughlin of Red Hat, who apart from being a seriously talented software engineer was also recently elected onto the Foundation Board of OpenStack. Terrific guy, passionate about Cloud, passionate about his team at Red Hat and very clued up.

Now we’d been trying to record this recently but diaries hadn’t synced and if you get a chance to record with Mark you take it. However, the Gods were against us everytime I hit record on my rig at FOSDEM. We were given what I can only describe as a bunker to record in, a room with no windows, with broken furniture, geriatric Red Cross volunteers positioned there to deal with dehydrated and broken geeks who had over done the Belgian beer intake the night before.

They didn’t respect we were recording and we had to cope with foil wrapped geeks, in fact the last podcast I managed to record was interupted by a very sick and ill Python developer being vocal into a bucket six feet from where I was recording. Not ideal. Next year I hope they can cater for us a LOT better. Seriously less than impressed.

The recording we did with Mark I knew from the minute I pressed record was going to be testing. I’ve tried to clean it up digitally to get rid of first aid traffic noise and echo and am still not happy with the quality by any degree. However Mark and I wanted to get it out to you and the consensus is that it’s good to go.

Come back soon for some far better sounding audio !

Download the podcast here in MP3 format only

Podcast: Robyn Bergeron talks Fedora

Today’s podcast is with Robyn Bergeron who is of course the Community Project Leader of the Fedora Project, the erstwhile evergreen Linux distribution sponsored by Red Hat.

Last June Robyn and I were in Boston together and I meant to get her in front of one of my microphones to record a podcast but it was the last day of Red Hat Summit and people were packing up and getting ready to disappear all points east and west and it never happened.

So it was a given that the first opportunity I had to record something with her turned into a forty five minute recording I’ve cut down to about 25 minutes or so for this podcast.  We talk Fedora of course, releases, release criteria and etiquette, conventions and community, we talk OpenStack, we talk Aeolus and JBoss and all things technical that make up Fedora’s capabilities as part of upstream RHEL.

Listen carefully and you may even hear John Mark Walker from Gluster.org muscle in on the recording. Do of course download and listen, or subscribe via iTunes, Stitcher Internet Radio, Podfeed or via the RSS using your client of choice.

Download the podcast here in MP3 format only

Fedora 18 is out and it’s Cloudy

So we’ve released the latest version of Fedora, release 18. It’s the hard work of Robyn Bergeron and her team of erstwhile project maintainers, community contributors, documentation editors and hundreds of people with Fedora carved in their hearts. Kudos to their efforts, this release has been a labour of love. I was on the phone to Robyn last week and we were talking about the herculean efforts of all those involved to get this release out – albeit later than planned because of the documented issues with the whole Microsoft bootloader crap.

This release also features latest release of oVirt 3.1 (listen to the podcast with Jon Benedict in the Podcast directory above), latest version of Eucalyptus appearing for the first time (ver 3.2), as well as the Folsom release of OpenStack and the very cool Red Hat sponsored HeatAPI that we’ve featured in the podcast with Steven Hardy recently.

Robyn and I will be recording a podcast in a few weeks at FOSDEM in Brussels talking Fedora 18 and I think I may even be doing some stuff more formally with the Fedora crew if everything aligns. Watch this space for more news if and when it happens.

A grumble first. The new installer has real issues. To say that you are replacing the existing Anaconda because its old doesn’t wash if you don’t look at the behaviour of your former installation engine scripting. It’s quite a big fail and hopefully this will be fixed and fixed fast. If you are installing over or upgrading  a previous version of Fedora and you have previously used LUKS/DMCrypt the partitioning tool doesn’t allow you to authenticate the underlying volume or mount it just telling you you have 900k or whatever free on your drive. Non sensical – all old versions of Anaconda supported mounting of encrypted partitions. So if you are installing and have LUKS/DMCrypt on a partition my advice is back it up to a drive and blast it away as otherwise you’re potentially going to be screaming at the installer screen. You could argue that to install a fresh squeaky clean F18 install its nice to start with a clean harddrive but in reality we’re all adults and thats just bonkers. You have to think out the box and think that a LOT of your existing users will be using disk encryption and where applicable you document and build on screen assistance to what is the worst partitioning logic I have ever seen in an installer.

Think back 12 years to Caldera’s LISA installer and the emergence of Anaconda circa 2000/1 replacing the libnewt traditional installer and what a breath of fresh air it was. This is a major major step back. Oh and the artwork really sucks. No idea what high school grad they got to use Inkscape in his/her lunchbreak. Poor.

So you’ve installed Fedora 18, what next ? I like to customise my Fedora and to do that I have always preferred to use Fedora Utils thats just moved to GitHub. Fedora Utils is the ongoing work of Satyajit Sahoo and it saves a lot of time and hassle to get you a box with restricted codecs and applications.

Once you’ve installed the codecs and tools installed from Fedora Utils drop to the console and immediately avoid GPG key failure error messages grab both these RPMs

http://download1.rpmfusion.org/free/fedora/rpmfusion-free-release-branched.noarch.rpm
http://download1.rpmfusion.org/nonfree/fedora/rpmfusion-nonfree-release-branched.noarch.rpm

In a console move to the directory that you’ve downloaded to and issue the following command for each RPM both FREE and NON FREE

sudo yum localinstall --nogpgcheck

Next step, again as root

# yum clean all
# yum check-update
# yum update

Once update then add the following tools (especially if you use HP all in one devices or Jetdirect printers).

yum install hplip-gui gthumb gimp pulsecaster audacity ardour

You’re good to go

Oh and remember to encrypt your boot partition, LUKS/DMCrypt is your friend – security costs nothing folks.

Great distro – major move – but the installer is just very poor as is the logic behind partitioning and disk mounting. Let’s the side down, but as with everything in Open Source we can get under the hood and fix it.

Cloud – The Hidden Costs of Saying No

In the mid 1970s one of Britain’s most eminent theoretical physicists, pioneer quantum physicist, pure mathematician, metaphysicist (now residing in New Jersey aged 89) wrote an article called “The Hidden Costs of Saying No”. That man was Freeman Dyson who to this day is one of the most forward thinking scientific engineers to have ever walked this earth. His work in nuclear physics and quantum computing all led to major advances in things that we take for granted today.

The paper he penned in 1974 first crept across my desk as a student in the early 1990s and it’s as probably completely relevant today as it was the day it left his typewriter on that spring day in the mid 1970s. At the time I remember reading it as part of work I was doing around environmental technologies and investment in technology. Didn’t think I’d ever be pulling it out again but it’s pertinence to where we are today in Cloud is enormous.

The argument made in the paper essentially argues eloquently that the price we pay for not doing something should be considered carefully during the decision process, I can pull some key phrases out the paper here – and thirty eight years on they still resonate and are applicable to technology processes today.

Freeman writes “it is not enough to count the hidden costs of saying yes to new enterprises. We must also learn to count the hidden costs of saying no. The costs of saying no may be high, although they are often uncertain and intangible. Our existing political processes introduce a strong bias into the consideration of new enterprises”

We need to know more accurately the costs of saying no, and we need procedures that allow a more realistic weighing of uncertainties when knowledge is lacking.”

In a nutshell Freeman described a lot of the things we take for granted working using Open Source methodologies and technologies. Daily development routines and complex decision making that is sped up by the adoption of combined global sharing of ever changing but singularly stable code trees. That code polished and brought to an enterprise market by Red Hat. Saying no to established costly proprietary ways of working that tie us into vendor lock in but also slow down the establishment and the stabilisation of future technology advancement.

In Cloud we have tangible decisions to take and those decisions impact on our core existing infrastructures as much as they do on our future road maps both for technology advancement and technology adoption. This also often the shapes of our businesses as we grow them either by harnessing more intelligent open ways of working. I am fortunate in working with some of the best and brightest from MIT and Dartmouth. I can count in my Rolodex (ok so artistic licence my email address book) some of the most capable technologists on the planet. We all, to a man (or woman) have grown careers and fostered approaches to our working days by sharing. We are daily creating pathways openly and in a community that stands up to be counted. In Cloud it’s no different.

The practical advancements in everything from storage to provisioning, from application deployment to lifecycle management rely on constant re-evaluation of process and often political dogma. We challenge regularly the recognised established old school ways of working in software environments to both satisfy the ever growing needs of our customerbase but also because as fast as we release stable supported code, the underlying network and physical hardware we rely on to pump those ones and zeroes changes. This then affords us new highways to motor down to deliver our cargoes certifying RHEL / RHEV / JBoss / Gluster in a supported release on these latest greatest technology environments. Honing, performance testing, securing and documenting to provide stable and polished deployable environments.

If we look at Platform as a Service, PaaS, a decade ago the decisions needed to get an application into a live environment to be consumed both internally and externally. This would have gone through a whole selection of change controls and also interpersonal relationships and decision making circuits within an organisation. All time consuming.

So we’re talking 2000-2, web based application development wasn’t in it’s infancy but the role of DevOps vs ITOps was a lot less equal than it was now. Silo’d mentality meant disproportionate decision making was often weighed to the advantage of the greater good rather than advancement. Open Cloud affords us now in 2012 the ability to utilise the best our developers can deliver and get that out in a safe and supportable manner and to do it using a range of tools, languages and libraries like never before – and to share this globally.

I truly believe the last proprietary technologies we will see in the datacentre are VMWare and to a lesser much smaller extent (because of current adoption levels) Microsoft’s HyperV cloud technology. While you could see that as a contentious statement the law of diminishing returns dictates that there is less available funding for IT projects globally, that our masters and our consumers are more savvy and expect more for their pound, euro, dollar, rupee, yen etc. To critically be able to do more with less headcount to be able to maintain what we have but to get to the next level with regards to being able to harness and deliver against business need.

The only way you can do that is openly.

The only way you can do that if you understand real world sane economics at a processor core level or at the application development and management level is openly. To therefore sink your investment into a proprietary core product and try and then stretch your IT architectures around something that makes you fit around it not you work to best advantage holds no credible place in the long term procurement strategy of the savvy CIO.

I was at VMWorld in 2012 in Barcelona. I probably will be blackballed for writing this and not get invited back but it reminded me of the same sort of protective over arching ethos of the Windows shows circa 2000-2001. When BackOffice was most at its hyped. BackOffice was great for those who needed a GUI to provision a file and print environment – to stand up a SQL database, to provision a mail system. It was the defacto go to environment of choice for those that counted their technical staff’s prowess by the number of trained staff who could click a mouse and read event viewer without falling into a coma. It was point and click enterprise computing at it’s most basic supplemented by “developers” who using Visual Basic and Visual Studio took runtimes and libraries of precompiled and often MSDN sourced libraries in order to get applications and databases to work. It wasn’t cutting edge. It wasn’t innovative and it’s a reason many of these organisations and system integrators got left behind both in growth and revenue by the more savvy tech startups who went Open and used code from Red Hat and the Open Source community.

For the companies that adopted an open strategy (the companies that have become the dot.com darlings and you rely on daily) they used Linux. They used Samba, they used Apache, they used Exim/Sendmail/Postfix for their mail as they spent money on people and research rather than on per seat licence or per mailbox licencing. They didn’t use Microsoft SQL or Oracle they just used MySQL or Postgres. The very rate they needed to develop using a paid for access model would have broken them but also the technology sucked (a lot) both from a performance perspective but also because you couldn’t get under the hood and tinker. They also contributed back – sharing information and sharing modifications for the common good. They challenged the hidden costs of saying no by embracing opportunity cost and common technical challenges rather than signing a EULA and waiting for the next MSDN CD box to arrive in the post.

The forward thinking companies who became the backbone of the internet relied on Linux. Not Solaris, not SCO, not Microsoft Windows NT or BackOffice. They deployed at speed and they were able to ride on the back of the speed of advancement of development environments such as PHP, Perl, Python, Java etc to get things done and to get things done stably and openly. These are the companies now with the banked revenue with the earnings figures and the technologies we consume (and Red Hat more often provide the support and provide the backbone to achieve stable platforms to base these technologies).

In Cloud we do the EXACT same thing, we build using what we have and we are brave enough to ask questions to understand what we have, and to understand what we need. We change the dynamic of IT by being brave enough to follow the example of Freeman Dyson.

In 1974 as part of this paper Dyson stated “Technology has always been, and always will be, unpredictable. Whenever things seem to be moving smoothly along a predictable path, some unexpected twist changes the rules of the game and makes the old predictions irrelevant.” A more visionary statement you are less likely to find in any management textbook or MBA guidance.

And for those who try to challenge and control or dictate privacy regulation or to impose territory or sovereign specific controls on Cloud. Beware, Freeman Dyson saw you coming nearly forty years ago and cleverly ties in William Blake’s writing he states with such punch that our elected officials in the EU and in other countries should take heed from:

The other lesson that we have to learn is that bureaucratic regulation has a killing effect on all creative endeavor. No matter how wisely framed and well intentioned, legal formalities tend to become inflexible. Procedures designed to fit one situation are applied indiscriminately to others. Regulations, whose purpose was to count the cost of saying yes to an unsound project, have the unintended effect of saying no to all projects that do not fit snugly into the bureaucratic system. Inventive spirits rebel against such rules and leave the leadership of technology to the uninventive. These are the hidden costs of saying no. To mitigate such costs, lawyers and legislators should carry in their hearts the other lesson that Blake has taught us: “One Law for the Lion and Ox is Oppression.”

Freeman Dyson didn’t invent Open technologies but he does talk a lot of sense. At Red Hat we like to think that the paradigm shift that is leading us to Cloud has a backbone that is empowered by open creative technological folk wearing red fedoras that want to understand how you can get to Cloud today, securely safely and empowering you to be able to answer the difficult question you may face when positioning open technologies vs proprietary stacks. How much does it cost to say no ? After reading this article you should be able to answer a lot of that question yourself. Working with Red Hat we’ll help you get the rest of those answers using the knowledge we’ve worked on over the last thirteen years in the enterprise marketplace.

I was also at GigaOM Structure in Amsterdam recently and Brian Stevens from Red Hat sat on stage doing a fireside chat and was asked the question whether Github was the new Linux ? The question made me giggle nervously as Linux is Linux – there was no polished answer but it relies on a company such as Red Hat and a thought leader such as Brian, Paul Cormier or Jim Whitehurst to continue to prove that our work and our mission statement shares the same punch as Freeman Dyson’s almost prophetical paper did way back in 1974.

The same level of expectation around our support of OpenStack being critical for open hybrid cloud aligned with our proven Red Hat stack and we are doing it openly and transparently as Platinum members of the OpenStack Foundation. You can get the latest Folsom technology preview by clicking here for RHEL users. Challenging to say no to proprietary working methodologies, aiding and maturing OpenStack and the whole Cloud paradigm.

There’s one future-proof cloud and it’s open

Happy Thanksgiving, thanks for reading. With grateful thanks to Freeman Dyson one of the greatest technologists in the modern world. The original Sheldon Cooper.

Podcast: Steven Hardy – HeatAPI

EXCLUSIVE OPENSTACK PODCAST WITH STEVEN HARDY !!!

Since the launch of the Technology Preview Red Hat OpenStack release a couple of months back a lot of people have acted surprised that Red Hat is actually the third largest contributor to OpenStack on a commit basis. We are 10000% committed to the success of OpenStack and some of our key developers are pushing key stable sources into the project.

One of these newer projects is HeatAPI which allows a lot of Amazon Web Services integration especially in the upcoming Fulsom release.

For a few months Steven Hardy and I have tried to find time in our schedules to record something and yesterday we managed to record a really good podcast that will appeal to all OpenStack developers.

Steven is also speaking at the London Developer Day (see my previous article) and we discuss this in the podcast.

Download the podcast here in MP3 and OGG formats