Risk Factor iconRisk Factor

Senate Condemns US Air Force ECSS Program Management’s Incompetence

IT Hiccups of the Week With no compelling IT system snafus, snags, or snarls last week to report on, we thought we’d return to an oldie but goodie project failure of the first order: the disastrous U.S. Air Force Expeditionary Combat Support System (ECSS) program.

The reason for our revisit is the public release a short time ago of the U.S. Senate staff report [pdf] into the fiasco.  Last December,  Senators Carl Levin and John McCain, respectively the chairman and ranking member of the Senate Armed Services Committee, requested the report. The request was made in the wake of the Air Force’s publication of the executive summary [pdf] of its own investigative report which apparently the Senators were not altogether happy with. You may recall that Levin and McCain christened the billion-dollar program failure—which the Air Force admitted failed to produce any significant military capability after almost eight years in development—as being “one of the most egregious examples of mismanagement in recent memory.” Given the number of massive DoD IT failures to choose from, that is saying something.

Not surprisingly, the Senate staff report identified basically the same contributing factors for the debacle as the internal Air Force report, albeit with different emphasis. Whereas the Air Force report listed four contributing factors for the ECSS program’s demise (poor program governance; inappropriate program management tactics, techniques, and procedures; difficulties in creating organizational change; and excessive personnel and organizational churn), the Senate staff report condensed them into three contributing factors:

  • Cultural resistance to change within the Air Force
  • Lack of leadership to implement needed changes; and
  • Inadequate mitigation of identified risks at the outset of the procurement.

The Senate report focused much of its attention on the last bullet concerning ECSS program risk mismanagement. In large part, the report blamed the calamity on the Air Force’s failure to adhere to business process reengineering guidelines “mandated by several legislative and internal DOD directives and [that] are designed to ensure a successful and seamless transition from old methods to new, more efficient ways of doing business.” From reading the report, one gets the image of an exasperated parent scolding a recalcitrant child: Congress seemed as miffed at the Air Force for ignoring its many IT-related best practices directives as for the failure itself.

Clearly adding to the sense of frustration is that the Air Force “identified cultural resistance to change and lack of leadership as potential [ECSS] problems in 2004” when the service carried out a mandated risk assessment as the program was being initially planned. Nevertheless, the risk mitigation approaches the service ended up developing were “woefully inadequate.” In fact, the report said that the Air Force identified cultural resistance as an ongoing risk issue throughout the program. However, the lack of action to address it permitted the “potential problem” to become an acute problem.

To its credit, the ECSS program did try to set out an approach in 2006 to try to contain the technical risks involved in developing an integrated logistics system to replace hundreds of legacy systems then in use across the Air Force. Two key risk reduction aspects of the plan were to “forego any modifications” to the Oracle software selected for ECSS and to “conduct significant testing and evaluation” of the system.  However, by the time the ECSS project was canceled in 2012, the report notes, Oracle’s software was not only being heavily customized, but it also wasn’t being properly tested.

Several things contributed to this 180 degree turn in project risk reduction, according to the report. One was partially a problem of the Air Force conducting what can only be called bait-and-switch procurement. As the report states:

"In its March 2005 solicitation, the Air Force requested an “integrated product solution.” The Air Force solicitation stated that it wanted to obtain “COTS [commercial off-the-shelf] software [that is] truly ‘off-the-shelf’: unmodified and available to anyone.” Oracle was awarded the software contract in October 2005, and provided the Air Force with three stand-alone integratable COTS software components that were “truly off the shelf.” Oracle also provided the Air Force with tools to put the three components together into a single software “suite,” which would “[require] a Systems Integrator (SI) to integrate the functions of the three [components].” Essentially, this meant the various new software pieces did not initially work together as a finished product and required additional integration to work as intended.

Furthermore,

"In December 2005, the Air Force issued its solicitation for a systems integrator (SI) … portrayed the three separate Oracle COTS software components, as a single, already-integrated COTS product which was to be provided to the winning bidder as government funded equipment (GFE). Confusion about the software suite plagued ECSS, contributing significantly to program delays. Not only was time and effort dedicated to integrating the three separate software components into a single integrated solution, but there were disagreements about who was responsible for that integration. While CSC [the system integrator] claimed in its bid to have expertise with Oracle products, the company has said that it assumed, that the products it would receive from the Air Force would already be integrated. Among the root causes of the integration-related delay was the Air Force’s failure to clearly understand and communicate program requirements.

Adding to the general confusion was the small issue of exactly how many legacy systems were going to be replaced. The report states:

"When the Air Force began planning for ECSS, it did not even know how many legacy systems the new system would replace. The Air Force has, on different occasions, used wildly different estimates on the number of existing legacy programs, ranging from “175 legacy systems” to “hundreds of legacy systems” to “over 900 legacy systems.”

Curiously, the Senate report doesn’t note that even if the Air Force was trying to get rid of “only” 175 legacy systems, that was still some 20 times more than the Air Force’s last failed ERP attempt a few years earlier. The staff report seems to assume that such a business process engineering undertaking was still feasible from the start (and during a period of conflict as well), which is a highly dubious assumption to be making.

Probably the most damning sentence in the whole report is the following:

"To date, the Air Force still cannot provide the exact number of legacy systems ECSS would have replaced."

Two years after ECSS was terminated, after two major investigations into why ECSS failed, and while the Air Force is actively engaged in planning for another try, this fact is still rather amazing.

I’ll let you read the report to dig through the other gory details involving the risk-related issues involving cultural resistance and lack of leadership, but suffice to say you have to wonder where top Air Force and Department of Defense leadership was during the eight years this project blunder unfolded. As I have noted elsewhere, the DoD CIO at the time claimed to be “closely” monitoring the program, and up to the day ECSS was terminated, the CIO viewed it as being only a moderately risky program.

There was the same lack of curiosity on the part of Congress as well, however. DoD ERP system developments have been well-documented by the US Government Accountability Office [pdf] for over two decades as being prone to self-immolation. But Congress has kept the money flowing to them anyway without bothering to perform much in the way of oversight. Predictably, the Senate report avoids looking into Congress's own role in permitting the ECSS failure to occur.

The Senate report goes on to list several other DoD ERP programs that are trying their best to imitate ECSS. In this time of tight government budgets, that list might actually move Congress to quit acting as a disinterested party to their future outcomes. In fact, Federal Computer Week ran an article last week that indicated the Senate Appropriations Defense Subcommittee was slicing $500 million dollars off of DoD’s IT budget, which is clearly a warning shot across DoD’s bow.

Another warning shot of note is that both Senators Levin and McCain have noted that: “No one within the Air Force and the Department of Defense has been held accountable for ECSS’s appalling mismanagement. No one has been fired. And, not a single government employee has been held responsible for wasting over $1 billion dollars in taxpayer funds.” The Senators have stated they plan to introduce legislation to hold program managers more accountable in the future.

I suspect—and dearly hope—that if another ECSS happens in defense (or in other governmental agencies or departments, for that matter), more than a few civil and military careers will be, like ECSS, terminated.

In Other News …

Birmingham England Traffic Wardens Unable to Issue Tickets

Chicago Car Sticker Enforcement Delayed After Computer Glitch

Ohio’s Lorain City Municipal Court Records are Computer "Nightmare"

Immigration System Crash Leads to Chaos at Santo Domingo’s Las Americas Airport

Texas TxTag Toll System Upgrade Causes Problems

Melbourne Members Equity Bank System Upgrade Issues Vexes Customers

Reservation System Issue Hits Las Vegas-based Allegiant Air Flights

Vancouver’s Skytrain Shutdown Angers Commuters

Computer Assigns Univ of Central Florida Freshman to Live in Bathrooms and Closets

Australia’s Woolworth Stores Suffers Store-wide Checkout Glitch

UK Retailer Marks & Spencer’s Revenue Results Smacked by Website Woes

IT Hiccups of the Week

We concentrate this week’s edition of IT snarls, snags, and snafus on the lessons being learned the hard way by Marks & Spencer—the U.K.'s largest clothing retailer and one of the top five retailers in the country—on what happens when your online strategy goes awry. What makes this more than a run-of-the-mill website goes bad story, at least in the U.K., is that as London's Daily Mail put it late last year, “Marks & Spencer, to coin a phrase, is not just any shop. It is the British shop, as much a part of our cultural heritage as the Women’s Institute, the BBC and the Queen.”

M&S launched with great fanfare a new £150 million website in February as a primary means to stem declining sales and profitability, as well as accelerate the achievement of the 128–year old company’s objective of being an international multichannel retailer. However, last week, CEO Marc Bolland announced shortly before the company’s annual meeting that on-going “settling in” problems with its website contributed to an 8.1 percent drop in online sales over the previous quarter. The decline in online sales, which was more than expected, helped M&S chalk up its 12th quarter in a row of declining sales in its housewares and clothing division.

Read More

Thousands of Bags Miss Flights at Heathrow Terminal 5 Again

IT Hiccups of the Week

Here's some glitch déjà vu from 2008, namely another baggage system miscue involving British Airways (BA) at Heathrow International Airport in London. As you may remember, in March 2008, BA and Heathrow operator British Airports Authority (now known as Heathrow Airport Holdings) opened the long-awaited BA Terminal 5 with great fanfare, with BAA loudly proclaiming the “world-class” baggage system was “tried, tested and ready to go.” No Denver International Airport baggage system-like problems for them! And BA's deservedly poor reputation as the top airline for losing luggage would finally be over.

Of course, such publicly-stated optimism over the reliability of automation is rarely left unpunished. Almost immediately, a massive meltdown of the baggage system on the first day of T5’s operation led to more than 28,000 passenger bags piled high across the terminal, hundreds more being lost, and some 15 percent of BA flights being cancelled over the course of nearly a week. It took three weeks before the majority of bags were reunited with passengers. The extreme embarrassment for both BA and Heathrow management because of the incident was acute, as was BA passenger rage, to say the least.

The nightmares of that week have slowly receded from BA passengers' memories. That is, until Friday, 27 June, when London papers like the Daily Mail reported that T5’s automated baggage system had suffered another major IT failure, with bags having to be handled manually again. As a result, thousands of BA passengers were sent (unknowingly) on their way without their luggage, including those passengers transiting through London via T5. The Mail quoted a BA spokesperson as saying, “On Thursday morning, the baggage system in Terminal 5 suffered an IT problem which affected how many bags could be accepted for each flight… We are very sorry for the difficulties this has caused and we have been working hard with the airport to make sure we reunite all of our customers with their luggage as quickly as possible.”

The BA spokesperson failed to point out that the phrase “how many bags could be accepted for each flight” actually meant no bags were accompanying their owners on an untold number of BA flights. BA also insisted to the press that they stop saying that passenger bags were lost; the bags merely “missed” their flights, BA pouted.  

A short two-paragraph Heathrow Airport Holdings press release did BA one better at trying to downplay the baggage system problem, stating that it affected only “some bags,” and that flights were in fact operating “normally.” You have to love press statements that are totally true but also totally disingenuous.

BA passengers on Thursday were naturally displeased at traveling without their bags, but at least they got to their destination, unlike those flying out of T5 last September, when another but very short-lived IT problem with the baggage system prevented hundreds of passengers from ever boarding their flights and had to be rebooked onto new ones, many the next day.

While BA passengers from June 27 were naturally miffed, what BA and Heathrow’s operator failed to make clear until early this week was that the “intermittent” IT problems with T5’s baggage system had actually begun on Thursday, 26 June and continued well into Sunday, 29 June. I am sure that many BA passengers flying out of T5 on June 28 and 29 would have changed airlines if they knew the full extent of the baggage problems. Conveniently, neither BA nor the airport operator came forward with the information about the multi-day operational problem until Tuesday, 1 July. Nor have they disclosed the total number of bags or passengers inconvenienced.

Both BA and Heathrow Airport Holdings are in damage control mode as BA passengers, many of them famous, have taken to social media to lambast them both. Many passengers, for example, have complained that when they finally did receive their bags, they had been ransacked with items stolen from them. Others complained that their journeys were over by the time their bags finally reached them.

BA put out another press release blaming international airline security rules for bags being opened as well as being delayed, and further promised to look into the ransacking claims. A BA spokesperson went on to apologize, stating that, “We are very sorry that this process is taking longer than anticipated, and we fully understand the frustration that this is causing.” Heathrow Airport Holdings new CEO John Holland-Kaye also apologized, saying the IT problem had taken too long to resolve and that airport needs “to do better.” Disclosing IT problems while they are occurring would be a good start.

The BA spokesperson went on to warn that it would still take “several days” before all the bags that “missed” their flights are reunited with their owners. BA also indicated that because of the number of bags involved, its bag tracking system was not working as it should, which could further add to the delays.

BA is reminding its customers flying out of T5 that, “You may wish to carry essential items in hand baggage where possible.” That is probably good advice. ComputerWorldUK reports that Heathrow Airport Holdings is remaining very tight-lipped over what caused the baggage system fault and why it took four days to fix it, which is rarely a good sign that everything is under control.

In Other News…

Florida’s DMV Computer System Back Online

Bombay Stock Exchange Recovers from Outage

New Zealand Exchange Suffers IT Glitch

DNS Error Hits British Telecom

Irish Drivers Avoid Parking Fines in County Clare Due to Computer Error

PayPal Error Blocks CERN and MIT anti-Spying ProtonMail Fundraising Efforts

Microsoft Anti-crime Operation Disrupts Legitimate Servers

UK Adult Content Filters Hit 20 Percent of Legal Popular Sites

Goldman Sachs Gets Court to Order Google to Block Misdirected Email

HHS IG Reports Say Federal and State Health Insurance Exchange Controls Very Weak

 

Outages Galore: Microsoft, Facebook, Oz Telecom Users are Unhappy Lot

IT Hiccups of the Week

We go on an IT Hiccups hiatus for a week and wouldn’t you know it, Facebook does a worldwide IT face plant for thirty minutes while mobile phone users of two of the three largest telecom providers in Australia, Optus and Vodafone, coincidentally suffer concurrent nationwide network outages for hours on the same day. Microsoft follows that with back-to-back Office 365-related outages, each lasting more than six plus hours. In addition, there were system operational troubles in Finland, India and New York to name but a few. So, we decided to focus this week’s edition of IT problems, snafus and snarls on the recent outbreak of reported service disruptions that happened around the world as well as those sincere sounding but ultimately vacuous apologies that always now accompany them. 

Our operational oofta review begins last Tuesday, when Microsoft’s Exchange Online was disrupted for some users starting from around 0630 to until almost 1630 or so East Coast time, leaving those affected without email, calendar and contact information capability. The disruption was somewhat embarrassing for Microsoft, which likes to tout that its cloud version of Office365 is effectively always available (or at least 99.9% of the time).

Read More

French Hospital's Computer “Bug” Trashes a Fortune in Perfectly Good Drugs

IT Hiccups of the Week

Last week saw another wave of healthcare-related IT malfunctions, problems, and issues being reported. This time, we turn our focus to a controversy currently capturing the attention of the French press: the startling admission by administrators at the university hospital in Rennes that perfectly good drugs and other medical supplies are being trashed as a result of technical issues with its relatively new automated pharmacy system.

What first drew my attention to this story was an English-language story appearing at the International Business Times that claimed the Centre Hospitalier Universitaire de Rennes—better known as CHU—has sent a staggering €10 million worth of medicine and other medical paraphernalia to its incinerator because of a “computer bug” in the pharmacy distribution “robot” that was installed at the hospital in 2010. According to the IBT story—which claimed as its source an investigative story published in The Monthly Rennes—the “bug led to duplication and storage problems, which caused pallet-loads of medication to be destroyed.”

Read More

900,000 Medi-Cal Applicants Stuck in California Healthcare Backlog Purgatory

IT Hiccups of the Week

Last week saw an emergency room full of healthcare-related IT problems, issues and challenges being reported.  We chose to concentrate this week’s edition of IT Hiccups on one which resembles a healthcare version of Hotel California where hundreds of thousands of California Medi-Cal health insurance applications have been checked in, but can’t seem to leave the confines of California government offices.

Read More

Fat Finger Flub Takes Down Cloud Computing Datacenter

IT Hiccups of the Week

A wide variety of IT-related blips, failures, and mistakes occurred last week. However, the most interesting IT Hiccups related story involved what was described as a “fat finger” error by an operator at the cloud computing service provider Joyent’s US-East-1 datacenter in Ashburn, Virginia. It disrupted operations for all of Joyent’s datacenter customers for at least twenty minutes last Tuesday. For a small number of unlucky Joyent customers, the outage lasted 2.5 hours. 

According to a post-mortem note by a clearly embarrassed Joyent, “Due to an operator error, all us-east-1 API [application programming interface] systems and customer instances were simultaneously rebooted at 2014-05-27T20:13Z (13:13PDT).” The reason for the reboot, Joyent explained, was that a system operator along with other Joyent team members were “performing upgrades of some new capacity in our fleet, and they were using the tooling that allows for remote updates of software. The command to reboot the select set of new systems that needed to be updated was mis-typed, and instead specified all servers in the datacenter. Unfortunately the tool in question does not have enough input validation to prevent this from happening without extra steps/confirmation, and went ahead and issued a reboot command to every server in us-east-1 availability zone without delay.”

Read More

1 Million Americans Likely Receiving Incorrect Federal Healthcare Subsidies

IT Hiccups of the Week

Last week saw an increase in the number and types of IT-related malfunctions, ooftas, and errors reported over the previous week. The most interesting IT Hiccups-related story of the lot was one from the Washington Post indicating that 3 million of the 8 million or so individuals enrolled for health insurance under the Affordable Care Act have a variety of “discrepancies” in their applications, including 1.1 million to 1.5 million applications with levels of income individuals are claiming that differ “significantly” with Internal Revenue Service (IRS) documents.

As a result, the Post estimates, the Obama Administration may be making improper subsidies payments for more than 1 million Americans to help pay for their health insurance plans, including hundreds of thousands likely receiving higher subsidy amounts than they are entitled. Furthermore, because of computer-related issues, it may be well into late summer before the situation can be rectified. As of today, the Administration has not lowered (or increased) anyone’s subsidy amount because of these discrepancies, the Post reports.  A person who is currently receiving too high a subsidy will be required to pay it back to the IRS by April 2015, however. Given that those receiving subsidies require low incomes in the first place, a person who is innocently getting too high a subsidy may be facing a massive, and potentially bankrupting, tax bill next year.

And for those found to be lying on their applications in an attempt to receive a higher subsidy than they deserve, not only are they going to get a nasty payment demand from the IRS,  but they also face a $250 000 civil fine. Even an “honest error” on an insurance enrollment application can result in a civil fine of $25 000, although I doubt that will ever be imposed, because it would be a surefire way to discourage lots of subsidy eligible individuals from signing up for health insurance.

Paying out health insurance subsidies (or even approving health insurance) without first fully verifying income (and other application information) was never supposed to happen. According to the original plan, the “back-end” office IT systems linking the different government agencies and departments such as the IRS, Department of Homeland Security, the Social Security Administration, the Veteran’s Administration,  and so on were to be in place by 1 October so that the income, immigration status, citizenship, age, other health insurances received, etc. could be verified almost immediately upon receipt of a person’s health insurance enrollment application.

However, because of all the troubles trying to get the ACA website itself working by 1 October and afterwards, the Administration decided early on to defer the back-end office systems development.  Henry Chao, deputy chief information officer at the Centers for Medicare and Medicaid Services testified in front of the US Congress last November that 30 to 40 percent of the back-end systems work still needed to be completed. While everyone knew the systems were behind schedule, to say that Chao's admission took Congress by surprise is an understatement.

Kathleen Sebelius, the Health and Human Services Secretary at the time, promised a startled Congress that the back-end systems would be in place by mid-January, so not to worry. That timetable, however, soon slipped to mid-March, and now, as the Post reports, to late summer. Some $121 million has been budgeted since the beginning of the year to try to get these back-end systems working, which comes on top of the unknown tens of millions of dollars previously spent on their development.

Interestingly, when the systems' development timetable slipped from mid-January to mid-March, the Administration admitted that if the back-end systems were not completed by the March date,  “the entire healthcare reform program [could] be jeopardized” because the “issuance of [incorrect] payments to health plans ... could seriously put them at financial risk; potentially leading to their default and disrupting continued services and coverage to consumers.”

Another reason why the lack of back-end systems could place ACA in jeopardy is that without them, the Federal government cannot accurately determine how many people have even paid their health insurance premiums. The lack of valid premium payment data along with improper subsidy payments could lead to wrongly predicting the “Risk Adjustment, Reinsurance, and Risk Corridor [pdf], potentially putting the entire health insurance industry at risk," Administration documents stated.

The Obama Administration naturally is trying to play down its own dire warnings about paying out subsidies without first verifying a person’s eligibility or knowing who has and hasn’t paid their premiums, which has angered many Republicans in Congress, some of whom who are calling for a suspension of unverified subsidy payments. The Administration is unlikely to do that, but if it can’t get those back-end systems up and running in the next few months, there will likely be an increasingly large political—not to mention huge financial—cost to be paid.

Speaking of cost, a story by Politico reports that $475 million has been spent so far on the failed Oregon, Massachusetts, Maryland, and Nevada health insurance exchanges alone. Both Maryland and Massachusetts, however, now want even more money to fix their broken exchanges, while the FBI is trying to determine whether there was fraud involving the Oregon exchange effort. Nevada announced last week that it has decided to throw in the towel on its own exchange, and will use the Federal one instead, at least for 2015. Rumor has it that Rhode Island, which has a working exchange, may soon decide to move to the Federal exchange as well. In Rhode Island’s case, the future costs of operating and maintaining its exchange is starting to look increasingly unaffordable.

The Politico story also says that according to the Kaiser Foundation’s calculation, the Federal government has given state governments some $4.698 billion since 2011 to support their exchange development efforts.  In addition, according to recently released figures, the Federal government has so far obligated some $834 million to create its exchange and back end support systems, and will need another $200 million in Fiscal Year 2015 to maintain it, which is about double what was thought to be required to develop the exchange in the first place.

Finally, last week the real story of the first day ACA enrollments came to light thanks to a Freedom of Information Act lawsuit by Judicial Watch. Whereas it was long thought that only 6 people were able to enroll that first day, in reality, only one person was ever able to enroll for health insurance coverage, even though 43 208 accounts were created, government records show.

In Other News ...

Oregon Secretary of State Website Error Delays Congressional Vote Reporting

eVoting Machines Malfunction in Pennsylvania’s Westmoreland County

Hotel Reservation System Error Overbooks Rooms for Tennessee Ironman Competition

Computer Malfunction Causes CSX Train to Block New York Road

IT Ticketing Problem Fixed for Glasgow 2014 Commonwealth Games

Massachusetts Still Trying to Fix New Unemployment System Issues

Tacoma Washington Schools Experience More Online Testing Woes

New York City Places “Fundamentally Wrong” 911 System Implementation on Hold

U.S. Immigration Court System Outage Enters Week Six

IT Hiccups of the Week

It’s been another slow news week in the world of IT-related errors, problems and failures. The most interesting IT Hiccup of the week involves a puzzling six-week outage affecting the computer system that supports U.S. Executive Office for Immigration Review (EOIR) administered courts across the country.

About a month ago, the New York Post ran a story describing a “tech meltdown” involving five servers located in Falls Church, Virginia, on 12 April. The servers are an integral part of the computer system supporting the 59 immigration courts administered by the EOIR. According to the Department of Justice, the EOIR “primarily decides whether foreign-born individuals, who are charged by the Department of Homeland Security (DHS) with violating immigration law, should be ordered removed from the United States or should be granted relief or protection from removal and be permitted to remain in this country.”

As a result of the outage, which also took out the mandatory electronic registry for accredited immigration attorneys and representatives, the 260-plus immigration court judges (and staff) have had to return to pencil-and-paper methods. As described at JDSUPRA.com, the problem has kept “court clerks from accessing court records, entering new ones in the system, and making audio recordings of hearings.”

Read More

British Columbia's Integrated Case Management System Falls Over

IT Hiccups of the Week

After the previous week’s deluge, last week saw a return to a more “normal” number of reported IT-related malfunctions, errors and problems. The most significant involves British Columbia’s controversial CAD $182 million Integrated Case Management (ICM) system. It has been plagued with so many operational issues that government social workers using the system complain that it has been essentially unusable for over a week, the Vancouver Sun reported.

The ICM (pdf) is a new  government IT system aimed at streamlining the management of computer files across the three British Columbia government ministries (Social Development and Social Innovation; Children and Family Development; and Technology, Innovation, and Citizens’ Services) that provide social services to “poor children, disabled people and troubled families racked by addiction, mental illness and violence.” The ICM system is being introduced in four phases. Phase 1 occurred in November 2010; Phase 4, which is the final phase, is scheduled for roll-out at the end of this year.  

I previously noted that with the introduction of Phase 2 in April 2012, social workers bitterly complained that the ICM system kept freezing, data routinely disappeared, and the system was extremely cumbersome to use. Government officials at the time tried to play down the problems, characterizing them as just the routine “challenges” accompanying any new IT system introduction. But independent assessment reports later confirmed the validity of the complaints. The assessment also indicated a rather dysfunctional project development that needed immediate correcting.

Phase 3 of the ICM system was introduced a year ago March, and the complaints about the system started to subside. That is, until a couple of weeks ago, when all the old performance problems seemed to reappear with a vengeance. Social workers were livid that they were once more having to deal “with blank screens, and the inability to access to their clients’ addresses and warnings that may have been part of the file” which forced them to revert to paper and pencil methods. Exacerbating their anger was the B.C. government’s decision not to disclose publicly for almost a week that there were new problems occurring with the ICM system.

That information blackout also raised the ire of Mary Ellen Turpel-Lafond, the province’s independent watchdog for children and youth, who said ICM's ongoing problems were placing vulnerable children and adults, especially those who might need protection, at risk. She also called for a new independent assessment of the ICM system's performance and development.

After the ICM problems became public early last week, British Columbia’s Technology Minister Andrew Wilkinson claimed they were nothing more than “intermittent slowdowns.” Further, despite all the warnings by Turpel-Lafond as well as the political opposition about vulnerable children, the problems encountered caused only minor impacts on the delivery of critical government services. B.C. Children’s Minister Stephanie Cadieux also strongly rejected the notion that children were being placed at risk, the Vancouver Sun reported.

On Wednesday, in a bit of delicious irony, the Times Colonist newspaper reported that Wilkinson confidently told reporters that all the issues with the system were now fixed, so, please would everyone just move along. However, within an hour of saying the ICM was up and smoothly running again, an opposition party official informed a very surprised Wilkenson (and apparently those same reporters) that the ICM had crashed and burned again.

Opposition party members had earlier in the week pointedly reminded Wilkenson that even while the ICM system was having its previously undisclosed troubles, he was boasting to the legislative assembly about the government’s “enviable record of deployment of top IT services.” The opposition members wanted to know whether Wilkinson was just out of touch with what was going on in his area of responsibility, or whether he had misled the assembly. Given the latest meltdown, opposition members were now inquisitive about his lack of up-to-date knowledge on ICM's status since they knew more about it than he did.

A highly chastened Wilkenson told the Times Colonist that he was “unhappy” about the latest ICM outage, and implied that he was misled about the system's true state.  He added that, “This is clearly a system that’s unstable. We’re going to get to the bottom of this, we’re going to report back to the house and to the people of British Columbia what’s gone on, and we’re going to seek our remedies.”

Wilkenson didn’t offer a timetable as to when the ICM would indeed be fixed, or when he would report on what was really going on with the ICM, however.

It now looks like the Auditor General of British Columbia, who is currently conducting an IT audit of the ICM system, will have one more issue to investigate. He may also want to ask Wilkenson why the government is claiming that the ICM system has been “available and functioning for the benefit of British Columbians nearly 100 percent of the time” since April 2012, when it demonstrably has not.

In Other News…

Bank of Montreal Apologizes over MasterCard Problems

U.S. State Department Reports New Green Card Lottery Glitch

Thousands of Northern Ireland Nurses Underpaid for Five Months

New Election Software Delays Voting Report in North Carolina

Maybank in Malaysia Apologizes for Multi-day Nationwide Operational Problems

Advertisement

Risk Factor

IEEE Spectrum's risk analysis blog, featuring daily news, updates and analysis on computing and IT projects, software and systems failures, successes and innovations, security threats, and more.

Contributor
Willie D. Jones
 
Load More