Written byChristopher O’Malley
Everyone agrees that modernization of federal IT infrastructure and systems is necessary, and the swift passage of the Modernizing Government Technology Act will help. But one area of disagreement is the viability of “legacy” technology and what exactly should be done about it.
The government has been reporting that 75 to 80 percent of the federal IT budget is spent on running legacy (or existing) systems. That may sound alarming to those who aren’t familiar with the inner-workings of a large IT organization. However, the percentage is in-line with the industry average. Gartner says the average distribution of IT spending between run, grow and transform activities — across all industries — is 70 percent, 19 percent and 11 percent respectively. Those numbers have been consistent over the past decade.
In addition, the mere fact that a technology system is legacy, or old, does not mean that it isn’t effective. When it comes to working code, the longer it’s been running, the better it becomes as bugs and inefficiencies are eliminated over time. CIOs from the Department of Health and Human Services, the IRS and the Defense Department are just a few that have said that their legacy systems, written in Assembler and COBOL, are well written and can be kept current through ongoing stewardship.
Unfortunately, some technology leaders look past these facts and advocate an overall “rip, rewrite and replace” strategy. In my nearly 30 years with major IT companies, I’ve seen how blanket replacement strategies and new, unproven technologies have fared. More often than you might expect, the result is a waste of precious time, money and energy on an outcome that doesn’t achieve the stated goals. Consider Pennsylvania’s lawsuit against IBM, where a major modernization project to deal with its legacy technology resulted in complete failure.
So, with $500 million in taxpayer dollars ready to be spent, it’s imperative that we take a look at legacy modernization approaches that have worked and those that have failed.
The cloud is not the answer to every technological need.
A blanket re-platforming of core legacy applications is an enormously expensive and risky undertaking. The key is understanding when the cloud is appropriate and when other platforms offer the optimal approach. Large private organizations with the same “can’t fail” reliability needs as the government have long depended on the mainframe for their systems of record. Is their mainframe code “legacy?” Well, it may have started decades ago, but the hardware and software have been continually updated. Ninety-two of the world’s top 100 banks still use mainframes and wouldn’t think of migrating off the platform.
Cloud-first has become a misguided proxy for what really matters, which is continuously improving the citizen experience in engaging federal services. This is the perfect time for federal CIOs to think deeply and critically about a cloud-first strategy, and ask themselves if they are really doing what’s in the best interest of U.S. citizens or wasting time and taxpayer dollars chasing the “next big thing.” This fundamental question should drive every federal CIO’s vision and mission.
The two platform approach
What is working well is a hybrid approach we call two platform IT, which smartly uses the best of both worlds in appropriate situations. Part one takes advantage of the mainframe’s unmatched security, reliability and performance-at-scale for transaction processing and other systems of record — for example, the U.S. federal tax system. Part two uses the cloud for other services or applications that, while essential, don’t require the near-perfect reliability the mainframe offers – for example, HR applications.
The mainframe modernization issue
It is true that the experienced mainframe workforce is primarily made up of Baby Boomers who are retiring. What’s not true is that this workforce needs to be replaced with employees of equal experience. A new generation of post-modern mainframe solutions bring agility and ease of use to the mainframe. This includes everything from updating the mainframe interface, thereby making it familiar to new IT workers, to speeding up the development processes supporting the mainframe, which has not traditionally been known for rapid development cycles. The result is a multi-platform system that can keep pace with the demands of our mobile, digital economy.
It’s also important to note that a well-managed and maintained mainframe is also the most secure computing infrastructure available, requiring much less protection from outside attack. If you’re thinking “What about the massive OPM breach? Wasn’t that a mainframe problem?” Actually, no. The distributed systems surrounding the mainframe allowed the breach, and if simple, available security measures had been taken, it could have been avoided. The post-modern mainframe requires 69 percent less effort to secure than other systems, according to IBM research.
So as the MGT’s new funding and flexible budgeting opportunities are clarified in the coming months, federal CIOs will have to interpret the law to best leverage their legacy systems and determine what changes will truly serve the citizens paying for them.
Is it old and in need of replacing? Or is it a proven technology that simply requires modernization? MGT dollars will be best spent with careful consideration to these questions.
Chris O’Malley is CEO of Compuware, where he is responsible for setting the company’s vision, mission and strategy. With 30 plus years of IT experience, Chris has led the company’s transformation into becoming the “mainframe software partner for the next 50 years.”