How failed: the technical aspects

(Also read parts 1, 2 and 3.)

A lot of how works is opaque. This makes it hard to say authoritatively where all the problems lie and even harder to say how they can be solved. Clearly my knowledge is imperfect and thus my critiques are not perfect either. I am left to critique what the press has reported and what has come out in public statements and hearings. I can make reasonable inferences but my judgment will be somewhat off the mark because of the project’s opacity.

It didn’t have to be this way. The site was constructed with the typical approach used in Washington, which is to contract out the mess to a bunch of highly paid beltway bandit firms. Give them lots of money and hope with their impressive credentials that something will emerge that is usable. It was done this way because that’s how it’s always done. Although lots of major software projects follow an open source approach, this thinking hasn’t permeated the government yet, at least not much of it. Open source projects mean the software (code) is available to anyone to read, critique and suggest improvements. It’s posted on the web. It can be downloaded, compiled and installed if someone has the right tools and equipment.

It’s not a given that open sourcing this project was the right way to go. Open source projects work best for projects that are generic and used broadly. For every successful open source project like the Apache Web Server there are many more abandoned open source projects that are proposed but attract little attention. Sites like are full of these.

In the case of, an open source approach likely would have worked, and resulted in a system that would have cost magnitudes less and would have been much more usable. It would still have needed an architectural committee and some governance structure and programmers as well, principally to write a first draft of the code. Given its visibility and importance to the nation it would have naturally attracted many of our most talented programmers and software engineers, almost all of who would have donated their time. Contractors would have still been needed, but many of them would have been engaged in selecting and integrating suggested code changes submitted by the public.

If this model had been used, there probably would have been a code repository on Programmers would have submitted changes through its version control system. In general, the open source model works because the more eyes that can critique and suggest changes to code, the better the end result is likely to be. It would have given a sense of national ownership to the project. Programmers like to brag about their genius, and some of our best and brightest doubtless would have tooted their own horns at their contributions to the code.

It has been suggested that it is not too late to open source the project. I signed a petition on asking the president to do just this. Unfortunately, the petition process takes time. Assuming it gets enough signers to get a response from the White House, it is likely to be moot by the time it is actively taken up.

I would have like to have seen the system’s architecture put out for public comment as well. As I noted in my last post, the architecture is the scaffolding on which the drywall (code) is hung. It too was largely opaque. We were asked to trust that overpaid beltway bandits had chosen the right solutions. It is possible that had these documents been posted early in the process then professionals like me would have added public comments, and maybe a better architecture would have resulted. It was constructed instead inside the comfort of a black box known as a contract.

After its deployment, we can look at what was rolled out and critique it. What programmers can critique is principally its user interface because we can inspect it in detail. The user interface is important, but its mistakes are also relatively easy to fix and have been fixed to some extent. For example, the user interface now allows you to browse insurance plans without first establishing an account. This is one of these mistakes that are obvious to most people. You don’t need to create an account on to shop for a book. This was a poorly informed political decision instead. Ideally someone with user interface credentials would have pushed back on this decision.

What we saw with the system’s rollout was a nice looking screen that was full of many things that had to be fetched through separate calls back to the web server. For example, every image on a screen has to be fetched as a separate request. Each Javascript library and cascading style sheet (CSS) file also has to be fetched separately. In general, all of these have to complete before the page can be usable. So to speed up the page load time the idea is to minimize how many fetches are needed, and to fetch only what is needed and nothing else. Each image does not have to be fetched separately. Rather a composite image can be sent as one file and through the magic of CSS each image can be a snippet of a larger image and yet appear as a separate image. Javascript libraries can be collapsed into one file, and compressed using a process called minifying.

When the rollout first happened there was a lot of uproarious laughter from programmers because of all the obvious mistakes. For example, images can be sent with information that tells the browser “you can keep this in your local storage instead of fetching it every time the user reloads the page”. If you don’t expect the contents of a file to change, don’t keep sending it over and over again! There are all sorts of ways to speed up the presentation of a web page. Google has figured out some tricks, for example, and has even published a PageSpeed module for the Apache web server. Considering these pages will be seen by tens or hundreds of millions of Americans, you would expect a contractor would have thought through these things, but they either didn’t or didn’t have the time to complete them. (These techniques are not difficult, so that should not be an excuse.) It suggests that at least for the user interface portion of the project, a bunch of junior programmers were used. Tsk tsk.

Until the code for is published, it’s really hard to assess the technical mistakes of the project, but clearly there were many. What professionals like myself can see though is pretty alarming, which may explain why the code has not been released. It probably will in time as all government software is in theory in the public domain. Most likely when all the sloppy code behind the system is revealed at last, programmers will be amazed that the system worked at all. So consider this post preliminary. Many of’s dirty secrets are still to be revealed.

How failed: the architectural aspects

(Read parts 12 and 4.)

If you were building a house that you didn’t quite know how it would turn out when you started, one strategy would be to build a house made of Lego. Okay, not literally as it would not be very livable. But you might borrow the idea of Lego. Each Lego part is interchangeable with each other. You press pieces into the shape you want. If you find out half way through the project that it’s not quite what you want, you might break off some of the Lego and restart that part, while keeping the part that you liked.

The architects of had some of this in their architecture: a “data hub” that would be a big and common message broker. You need something like this because to qualify someone for health insurance you have to verify a lot of facts against various external data sources. A common messaging system makes a lot of sense, but it apparently wasn’t built quite right. For one thing, it did not scale very well under peak demand. A messaging system is only as fast as its slowest component. If the pipe is not big enough you install a bigger pipe. Even the biggest pipe won’t be of much use if the response time to an external data source is slow. This is made worse because generally an engineer cannot control aspects of external systems. For example, the system probably needs to check a person’s adjusted gross income from their last tax return to determine their subsidy. However, the IRS system may only support ten queries per second. Throw a thousand queries per second at it and the IRS computer is going to say “too busy!” if it says anything at all and the transaction will fail. From the error messages seen on, a lot of stuff like this was going on.

There are solutions to problems like these and they lay in fixing the system’s architecture. The general solution is to replicate the data from these external sources inside the system where you can control them, and query replicas instead of querying the external sources directly. For each data source, you can also architect it so that new instances of it can be spawned on increased demand. Of course, this implies that you can acquire the information from the source. Since most of these are federal sources, it was possible, providing the Federal Chief Technology Officer used his leverage. Most likely, currency of these data is not a critical concern. Every new tax filing that came into the IRS would not have to be instantly replicated into a cloned instance. Updating the source once a day was probably plenty, and updating it once a month likely would have sufficed as well.

The network itself was almost certainly a private and encrypted network given that privacy data traverses it. A good network engineer will plan for traffic ten to a hundred times as large as the maximum anticipated in the requirements, and make sure that redundant circuits with failover detection and automatic switchover are engineered in too. In general, it’s good to keep this kind of architecture as simple as possible, but bells and whistles certainly were possible: for example, using message queues to transfer the data and strict routing rules to handle priority traffic.

When requirements arrive late, this can introduce big problems for software engineers. Based on what you do know though, it is possible to run simulations of system behavior early in the life cycle of the project. You can create a pretend data source for IRS data that, for example, always returns an “OK” while you test basic functionality of the system. I have no idea if something like this was done early on, but I doubt it. It should have been if it wasn’t. Once the interaction between these pretend external data sources was simulated, complexity could be added to the information returned by each source, perhaps error messages or messages like “No such tax record exists for this person” to see how the system itself would behave, with attention to the user experience through the web interface as well. The handshake with these external data sources has to be carefully defined. Using a common protocol is a smart way to go for this kind of messaging. Some sort of message broker on an application server probably has the business logic to order the sequence of calls. This too had to be made scalable so that multiple instances could be spawned based on demand.

This stuff is admittedly pretty hard to engineer, and is not the sort of systems engineering that is done every day, and probably not by a vendor like CGI Federal. But the firms and the talent are out there to do these things and would have been done with the proper kind of system engineer in charge. This kind of architecture also allows for business rule changes to be centralized, allowing for the introduction of different data sources late in the life cycle. Properly architected, this is one way to handle changing requirements, providing a business-rules server using business rules software is used.

None of this is likely to be obvious to a largely non-technical federal staff groomed for management and not systems engineering. So a technology advisory board filled with people who understand these advanced topics certainly was needed from project inception. Any project of sufficient size, scope, cost or of high political significance needs a body with teeth like this.

Today at a Congressional hearing officials at CGI Federal unsurprisingly declared that they were not at fault: their subsystems all met the specifications. It’s unclear if these subsystems were also engineered to be scalable on demand as well. The crux of the architectural problem though was clearly in message communications between these system components, as that is where it seems to break down.

A lesson to learn from this debacle is that as much effort needs to go into engineering a flexible system as goes into the engineering of each component. Testing the system early under simulated conditions, then as it matures under more complex conditions and higher loads would have detected these problems earlier. Presumably there would then have been time to address them before the system went live because it would have been a visible problem. System architecture and system testing is thus vital for complex message based systems like, and a top notch system engineering plan needed to be have been at its centerpiece, particularly since the work was split up between multiple vendors with each responsible for their subsystem.

Technical mistakes will be discussed in the last post on this topic.

How failed: the programmatic aspects

(Also read parts 1, 3 and 4.)

I am getting some feedback: isn’t really a failure. People are using the website to get health insurance, albeit not without considerable hassle at times. I’ll grant you that. I’ll also grant you that this was a heck of a technical challenge, the sort I would have gladly taken a pass on, even for ten times my salary. It’s a failure in that it failed to measure up to its expectations. President Obama said there would be “glitches”, but these were far more than glitches. If this were a class project, a very generous professor might give it a D. I’d give it a D-, and that’s only then after a few beers. Since I don’t drink to imbibe, I give it an F.

In the last post, I looked at the political mistakes that were made. Today I’ll look at the programmatic mistakes. I’m talking about how in general the program was managed.

Some of it is probably not the fault of the program or project manager. This is because they were following the law, or at least regulation. And to follow the law you have to follow the FAR, i.e. the Federal Acquisition Regulation. It’s the rules for buying stuff in the federal government, including contracted services. Violating the FAR can put you in prison, which is why any project of more than tiny size has a contracting officer assigned to it. In general, the government wants to get good value when it makes a purchase. Value usually but does not always translate into lowest price. With some exceptions, the government considers having contractors construct a national portal for acquiring health care to be the same as building a bridge. Put out the requirements for an open bid and select the cheapest source. Do this and taxpayers will rejoice.

This contract had a lot of uncertainty, which meant it had red flags. The uncertainty was manifested in many areas, but certainly demonstrated in requirements that were not locked down until this year. I’d not want to waste my time coding something that I might have to recode because the requirements changed. This uncertainty was reflected in how the contract was bid. It’s hard to bid it as a fixed price contract when you don’t know exactly what you are building. If you were building a house where every day the owner was directing changes to the design you wouldn’t expect builders to do it using a fixed price contract. Same thing here. It appears the contract was largely solicited as “time and materials”. This accounts in part for total costs, which at the moment are approaching half a billion dollars. This kind of work tends to be expensive by its nature. CGI Federal probably had the lowest cost per hour, which let it win the bid.

There is some flexibility in choosing a contractor based on their experience constructing things a lot like what you want built. CGI Federal is a big, honking contractor that gets a lot of its business in government contracts. Like most of these firms, it has had its share of failures.  A system of the size of is a special animal. I am not sure that any of the typical prime contractors in the government software space were qualified to build something like this, at least not if you wanted it done right.

There is some flexibility allowed in the statement of work (SOW), generally put together by the program manager with help from a lot of others. I don’t know precisely what rules applied to the contracting process here, but it is likely, probably by expending a lot of political capital, to create SOW that would have properly framed the contracting process so something actually usable could be constructed. A proper SOW should have included criteria for the contractor like:

  • Demonstrated experience successfully creating and managing very large, multi-vendor software projects on time that meet requirements that change late in the system life cycle
  • Demonstrated ability to construct interactive web-based software systems capable of scaling seamlessly on demand and interacting quickly with disparate data sources supplied by third parties

The right SOW would have excluded a lot of vendors, including probably CGI Federal but very possibly some of the big players in this game like Unisys, IBM and Northrop Grumman. Yes, many of these vendors have built pretty big systems, but they often come with records that are spotted at best, but whose mistakes are often overlooked. Until recently I used a Northrop Grumman system for my federal travel. They did build it, but not successfully. For more than a year the system was painfully slow and the user interface truly sucked.

Successfully building a system of this type, which was highly usable upon initial deployment, should qualify that contractor to bid on this work. Offhand I don’t know who would qualify. I do know whom I would have wanted to do the work: They know how to create large interactive and usable websites that scale on demand. Granted even Amazon Web Services is not perfect, with occasional outages of its cloud network, but we’re talking a hassle factor of maybe .1% compared to what users have experienced with They used to do this for other retailers but may have gotten out of that business. I would have appealed to their patriotic senses, if they had any, to get them to bid on this work. In any event, even if they had bid they did not get the contract. So there was a serious problem either with the SOW or the “one size fits all” federal contracting regulations the doubtlessly very serious contracting officer for this project followed.

The size of this project though really made building it in-house not an option. So a board consisting of the best in-house web talent and program management talent in the government should have overseen it. Others have noted that the team that constructed President Obama’s websites, used to win two elections, would have been great in this role. In any event, the project needed this kind of panel from the moment the statement of work (SOW) was put together through the life of the project, and that includes post deployment.

Probably what they would have told those in charge was things they did not want to hear, but should have heard. The project should be delivered incrementally, not all at once. It should not be deadline driven. Given the constantly changing requirements, risk management strategies should have been utilized throughout. When I talk about architectural and technical mistakes in future posts, I’ll get into some of these.

In short, this project was a very different animal: highly visible, highly risky, with requirements hard to lock down and with technical assumptions (like most states would build their own exchanges) far off the mark. You cannot build a system like this successfully and meet every rule in the FAR. It needed waivers from senior leaders in the administration to do it in a way that would actually work in the 21st century, rather than to follow contracting procedure modeled on the spendthrift acquisition of commodities like toilet paper. An exception might even have been needed to have been written into the ACA bill that became law.

Next: architectural mistakes. and the problems with interactive federal websites

Today’s Washington Post highlights problems with the new site, the website used by citizens to get insurance under the Affordable Care Act. The article also talks about the problems the federal government is having in general managing information technology (IT). As someone who just happens to manage such a site for the government, I figure I have something unique to contribute to this discussion.

Some of the problems the health care site are experiencing were predictable, but some were embarrassingly unnecessary. Off the top of my head I can see two clear problems: splitting the work between multiple contractors and the hard deadline of bringing the website up on October 1, 2013, no matter what.

It’s unclear why HHS chose to have the work done by two contractors. The presentation (web-side) was done by one contractor and the back end (server-side) was done by another. This likely had something to do with federal contracting regulations. It perhaps was seen as a risk mitigation strategy at the contracting level, or a way to keep the overall cost low. It’s never a great idea for two contractors to do their work mostly mindless of the other’s work. Each was doing subsystem development, and as subsystems it’s possible that each worked optimally. But from the public’s perspective it is just one system. What clearly got skipped was serious system testing. System testing is designed to test how the system behaves from a user’s perspective. A subset of system testing is load testing. Load testing sees how the system reacts when it is under a lot of stress. Clearly some of the requirements for initial use of the system wildly underestimated the traffic the site actually experienced. But it also looks like in an effort to meet an arbitrary deadline, load testing and correcting the problems from it could not happen in time.

It also looks like the use cases, i.e. user interaction stories that describe how the system would be used, were bad. It turned out that most initial users were just shopping around and trying to find basic information. It resulted in a lot of browsing but little in the way of actual buying. Most consumers, particularly when choosing something as complex as health insurance, will want to have some idea of the actual costs before they sign up. The cost of health care is obviously a lot more than just the cost of premiums. Copays can add thousands of dollars a year to the actual cost of insurance. This requires reading, study, asking questions of actual human beings in many cases, and then making an informed decision. It will take days or weeks for the typical consumer to figure out which policy will work best for them, which means a lot of traffic to the web site, even when it is working optimally.

The Post article also mentions something I noticed more in my last job than in my current one: that federal employees who manage web sites really don’t understand what they are managing. This is because most agencies don’t believe federal employees actually need experience developing and maintaining web sites. Instead, this work is seen as something that should be contracted out. I was fortunate enough to bring hands on skills to my last job, and it was one of the reasons I was hired. In general, the government sees the role of a federal employee to “manage” the system and for contractors to “develop and maintain” the system. This typically leads to the federal employee being deficient in the technical skills needed and thus he or she can easily make poor decisions. Since my last employer just happened to be HHS, I can state this is how they do things. Thus, it’s not surprising the site is experiencing issues.

Even if you do have a federal staff developing and maintaining the site, as I happen to have in my current job, it’s no guarantee that they will all have all the needed skills as well. Acquiring and maintaining those skills requires an investment in time and training, and adequate training money is frequently in short supply. Moreover, the technology changes incredibly quickly, leading to mistakes. These bit me from time to time.

We recently extended our site to add controls that give the user more powerful ways to view data. One of these is a jQuery table sorter library. It allows long displays of data in tables to be filtered and sorted without going back to the server to refresh the data. It’s a neat feature but it did not come free. The software was free but it added marginally to the time it took the page to fully load. It also takes time to put the data into structures where this functionality can work. The component gets slow with large tables or multiple tables on the same page. Ideally we would have tested this prior to deployment, but we didn’t. It did not occur to me, to my embarrassment. I like to think that I usually catch stuff like this. This is not a fatal problem in our case, but it is a little embarrassing, but only to the tune of a second or two extra for certain web pages to load. Still, those who have tried it love the feature. We’re going to go back and reengineer this work so that we only use it with appropriately sized tables. Still, the marginal extra page load time may be so annoying for some that they choose to leave the site.

Our site like is also highly trafficked. I expect that will get more traffic than our site, which is thirty to 40 million successful page requests per month. Still, scaling web sites is not easy. The latest theory is to put redundant servers “in the cloud” (commercial hosting sites) to use as needed on demand. Unfortunately, “the cloud” itself is an emerging technology. Its premier provider, Amazon Web Services, regularly has embarrassing issues managing its cloud. Using the cloud should be simple but it is not. There is a substantial learning curve and it all must work automatically and seamlessly. The federal government is pushing use of the cloud for obvious benefits including cost savings, but it is really not ready for prime time, mission-critical use. Despite the hassles, if high availability is an absolute requirement, it’s better to host the servers yourself.

The key nugget from the Post’s article is that the people managing these systems in many cases don’t have the technical expertise to do so. It’s sort of like expecting a guy in the front office of a dealership to disassemble and reassemble a car on the lot. The salesman doesn’t need this knowledge but to manage a large federal website you really need this experience to competently manage your websites. You need to come up from the technical trenches and then add managerial skills to your talents. In general, I think it’s a mistake for federal agencies to outsource web site development. Many of these problems were preventable, although not all of them were. Successful deployment of these kinds of sites depends to a large extent on having a federal staff knowing the right questions to ask. And to really keep up to date on a technology that changes so quickly, it’s better to have federal employees develop these sites for themselves. Contractors might still be needed, but more for advice and coaching.

Each interactive federal web site is its own unique system, as certainly is. The site shows the perils of placing too much trust in contractors and in having a federal managerial staff with insufficient technical skills. Will we ever learn? Probably not. Given shrinking budgets and the mantra that contracting out is always good, it seems we are doomed to repeat to these mistakes in the future.

Don’t say I didn’t warn you.

(Update: 10/28/13. This initial post spawned a series of posts on this topic where I looked at this in more depth. You may want to read them, parts 1, 2, 3 and 4.)

The government needs common sense contracting

If like me you work for the federal government, or even if you do not, there is a good chance you have contractors in your workplace. Love or loathe contractors they are a fact of work life for many of us. Arguably, our occupation of Iraq would not have succeeded without lining the pockets of contractors like Halliburton with billions of dollars. The Army gave up making their privates peel potatoes decades ago.

Contractors are often necessary. I would not want to replace my own roof, or make a roofer my employee just to get my roof replaced. The theory of contracting is it allows you to acquire either a specialized skill for a limited period or it allows others to perform routine services that are considered so ordinary that they can be easily replaced if they do not perform. You hire employees for those unique, domain related skills that you will need performed on a continuing basis.

Where I work there are many contractors that are truly disposable. After a couple years, they seem almost like part of the furniture. Then one day they disappear along with their contract. These include what I think of as people in blue (from their blue garments): guards, floor sweepers, restroom cleaners and the people who man the registers in our cafeteria. Then there are others that are technically contractors but sure feel like employees to me.

In my last job, contractors were so pervasive and the downsizing so extreme that federal employees like myself who just happened to be able to program a computer were not allowed to. Computer programming and design, even for large legacy systems that were poorly documented, were treated the same I treat my roof: hire a contractor. Instead, we project managers were tasked to make sure our systems were maintained and modified but having little understanding of how it actually worked. Consequently, keeping the contractors who could actually retrofit the system became key to our own job success. If some of these contractors left, our agency’s mission would have been severely impacted.

After a while, it became clear that the contractors had the vital domain knowledge and project managers never would because it was out of our scope of permitted duties. It was a very curious situation: the federal employees, people who normally hang around an agency for twenty years or more, would hop and skip other agencies where they felt more vested in their work. Meanwhile some of the key contractors had stayed there for twenty years or more and were effectively managing government systems. They were indispensable. The contracting agency changed, but they still sat at the same desks doing the same work, but drawing a salary from a different company. The smarter ones incorporated and sold the services of their “corporation” to the contracting agency for higher sums of money. Some of these people were making in effect GS-15 money for GS-12 work. It is nice work if you can get it because you effectively created your own small monopoly.

Throughout the federal government, contractors are doing work that they should not. The Washington Post today documented yet another example of contracting going awry. The Inspector General for the Department of Health and Human Services audited medical equipment claims that were charged to the taxpayer. It found an error rate of 29 percent. Who is doing this work? Contractors. Should they? Probably not.

Considering that Medicare costs hundreds of billions of tax dollars a year, an error rate of 29 percent is unacceptable. Some senior bureaucrat, probably to satisfy the current administration in power, which believes in maximum contracting out, decided that it did not want to make actual employees responsible for monetary judgments like what constitutes a valid government expense. This was a boneheaded decision that has since resulted in what appears to be a waste of billions of dollars annually. What incentive does a contractor have to excel when every few years their contract will be re-competed? Why should a contractor’s employee care too much when they are looking at the calendar and are pondering their next contract too? Why in particular should they care when they are making these decisions yet are not held directly accountable for their decisions? They will be paid regardless.

Since contracts are legal instruments, contractors excel and doing precisely what the contract says and typically have little incentive to go beyond it. Many in fact prefer to do less than what is required, on the hope that it will be too much hassle to hold them accountable. This results, coincidentally, in an improvement to their bottom line. That appears to be the case at HHS. This happens because contractors are not necessarily vested in their work, like an employee would be nor is there much fear of accountability. The result can and often does breed mediocrity. Mediocrity is driven by an obsession by the government to get the lowest cost. It operates on the assumption that the work is in essence rote, when it is often specialized, unique and enduring. Yet, year after year, as I look around it sure appears that some contractors are doing this kind of work. If they look like an employee and smell like an employee why are they not treated as an employee? Why not just hire them? You already know the answer: because it is politically incorrect.

My office is big enough where it recently opened its own health club. Plastered on the door is a prominent notice: the club is for employees not contractors. As you might expect, an employee get other perks too such as a generous retirement plan. Occasionally though a contractor gets a perk that an employee does not. In my agency when a contractor travels on official business, the travel time is billable, along with all their travel expenses. Employees are not entitled to overtime for travel on nights or weekends. (Over the last few years, we were allowed to claim travel time outside of our regular hours as compensatory time. In addition, I can keep what small frequent flyer miles I earn.)

Once upon a time, I learned that the relationship between an agency and a contractor was legally considered a “relationship of equals” rather than a supervisor-employee relationship. This makes sense if you are a homeowner and need your roof fixed, but in the workplace, it often makes little sense and is a distinction that is meaningful only to lawyers.

I think contracting rules should be rewritten so they meet the common sense test. Contracting can be kept for jobs that are low-skilled and truly interchangeable, such as pushing a broom. They can be kept for highly specialized jobs that are limited in time and scope, such as a technology assessment. They should not be used to perform judgmental work for which the government is legally responsible. They should not be used as service contracts for work that is domain specific, specialized and amorphous in nature.

Until that time if any when some common sense returns to government, can we at least allow the contractors to use our health club?

Dissing Excellence in Government

Why do we have governments? I’m serious. It shouldn’t be necessary to even ask this question. It should be obvious. But apparently some members of Congress haven’t grasped the basics. People form governments because there are certain things that can’t or shouldn’t be done by the private sector. It’s right there in the preamble to the United States Constitution. Our federal government exists to:

“… establish Justice, insure domestic Tranquility, provide for the common defense, promote the general Welfare, and secure the Blessings of Liberty to ourselves and our Posterity…”

So outsourcing our judiciary is out. We have decided that corporations should not determine if we go to war or how it will be managed. Also our government is empowered through law to promote the general welfare. It’s okay and constitutional for the government to engage in activities that make the country more prosperous and free, as long as it does it generally, i.e. for the public.

So you would think that the National Weather Service (NWS), which meticulously monitors our nation’s weather and provides sound scientific forecasts to the public, would be engaged in an inherently governmental mission. Well, at least in the eyes of some people, you would be wrong. In particular Pennsylvania Senator Rick Santorum (Republican, naturally) thinks the NWS needs to stop being so darn public with its information. Yes, although through your tax dollars we fund the NWS to the tune of about $617M a year, some like Sen. Santorum want you to pay again. He in particular has introduced S. 786, a bill “To clarify the duties and responsibilities of the National Oceanic and Atmospheric Administration and the National Weather Service, and for other purposes.”

Basically Santorum wants the NWS and its parent the National Oceanic and Atmospheric Administration (NOAA) to stop releasing those pesky routine forecasts and oh so convenient meteorological information. Instead he wants you to be forced to get the information from private sources like AccuWeather and The Weather Channel. The NWS should restrict itself to issuing “severe weather forecasts and warnings designed for the protection of life and property of the general public” and “hydrometeorological guidance and core forecast information.” Government civil servants would be prohibited from releasing any information “that might influence or affect the market value of any product, service, commodity, tradable, or business.” Whoa! That’s quite a way to stifle a civil servant!

In short, if someone else can make a dime off of it, NOAA and the NWS shouldn’t be doing it. But if some major tropical storm or tornado is headed our way it’s okay. And it’s still okay to put data out there for general use, but God forbid that it should be spun into actual useful information although it sounds like “guidance” is still okay. (“It will be hot in August in Texas.”) And don’t allow those government civil servants to actually use their professional education to turn information into knowledge, like make drought predictions. Save that for Santorum’s buddies at AccuWeather.

Why? Because Santorum claims this will stimulate the private sector innovation. He doesn’t want the government to provide this information in a timelier, cheaper or non-partisan manner. He wants you to cough up additional money to get this knowledge by subscribing to AccuWeather or watching all those annoying ads on

We so often hear that government is wasteful and bloated. But for less than $700M a year we have a National Weather Service that provides accurate and timely forecasts to all comers. (That’s less than $3 per year per person.) So what are the NWS and NOAA real crimes? What it amounts to is they are doing their job too well. All that valuable weather information is available in real time on their web sites. Oh Lord, the NWS has been too innovative. You can even get localized weather information available as an RSS news feed. And the NWS has done this despite flat or shrinking budgets.

So first civil servants get unfairly tarnished for being wasteful, bureaucratic, bloated and not thinking like the private sector. But when civil servants demonstrate extreme competence and entrepreneurial behavior, like apparently the many marvelous employees at the NWS, and do things faster, better and cheaper than the private sector, they are being bad. As a civil servant myself this really irks me. Man, we can’t win for losing! Apparently we aren’t living up to our stereotypes and that really irks some politicians.

Thankfully so far Santorum’s bill has no cosponsors. This means it is likely to die a quick death. But there are no guarantees in the weird Republican controlled times that we live in. You have to wonder what’s next: will the Secret Service be outsourced to Halliburton?

I am envious of my colleagues in NOAA and the NWS because they are doing exactly what I want to do at the U.S. Geological Survey. I manage NWISWeb, the system that puts out in real time water information collected by USGS’s National Water Information System (NWIS). Since I arrived about a year ago I’ve been building the case with management that we too need to make our peer reviewed water science data more broadly and easily available. NWISWeb is a great system but right now it is limited in that it serves our water data to be read by humans in a browser only. That hasn’t stopped lots of clever people in the private sector like the American Whitewater from figuring out ways to get our data and place it inside their products. And that’s fine with us. We make our data available equally to all comers. If the whitewater rafting people can figure out a way to show their members the local stream conditions on their web site more power to them.

I’d like to take the hassle out of getting the data though. I’d like us to offer our data using web services. This new technology would let computers grab and process our data on the fly without necessarily writing a lot of customized code, downloading files or scraping screens for content. I’d love for to display our stream flow and groundwater information in their maps. (Of course I’d like them to show the USGS logo too, so the public understands who is really gathering the data.) Particularly during periods of heavy flooding and hurricanes this information served in many places can save lives and reduce property damage.

We already have hydrologists working with NOAA and other organizations like They are working on models that can turn the number of feet a stream is over flood stage into a map that will show the surface area that would be underwater. But wait! It sounds like if a lot of senators think like Senator Santorum then the USGS would be in the data collection business exclusively. No point having your hard earned tax money used to infer any meaning from the data. So what if you live in a trailer next to a rapidly rising creek and are too busy watching Survivor to check stream conditions. If you don’t have your contract with AccuWeather to warn you about approaching floods that’s your problem. As a risk mitigation strategy, consider buying a life vest for every member of your family.

Time will tell whether Santorum’s bill lives or dies. And time will tell whether my idea will fly at the USGS. Like the NWS we work paycheck to paycheck. Since the Bush Administration wants to keep our budget flat money likely won’t be forthcoming for such an endeavor unless it is pulled from somewhere else. But if my executives are to take a clue from Senator Santorum, they might fear to be too innovative. Could be risky.

Instead of whining that the government is doing its job too well, AccuWeather should find ways to add value. Perhaps it could put more of its own sensors out in the field, collect different kinds of data and integrate it with the NWS data they are already getting cost free. AccuWeather is not being innovative. It is being anticompetitive.

I say let’s applaud NOAA and the NWS for their excellence and foresight. There actually is quite a lot of this innovation in the federal government if you look for it. Perhaps instead of giving agencies like these flat budgets they should be rewarded for their innovation with more money so they can do an even better job. They are after all clearly promoting the general welfare. Our founding fathers would be pleased.

Continue reading “Dissing Excellence in Government”