Many apps flunk security check before move to cloud

CEOs and the technologists who work for them like to say the applications they rely on are safe as houses

CEOs and the technologists who work for them like to say the applications they rely on -- especially the kind custom-written by specialists at banks and investment companies with fortunes behind them--are safe as houses.

And they are, if you're talking about houses in Louisiana when the Gulf starts lashing hurricanes and tarballs.

Almost 60 percent of all the applications brought to security testing and risk-analysis company Veracode during the past 18 months couldn't meet the minimum standards for acceptable security, even when the criteria were dialed down to accommodate applications that don't pose a great security risk, according to Samskriti King, vice president of product marketing at the company. Web-based apps carry their own special set of risks.

"There are far more people on Web projects because they're often easier to develop; many components are already available so you can stand up Web applications very easily," King says. "Developer education usually focuses on applications generated and used in one place, but Web applications could touch many places, so a vulnerability in one component could manifest in many places if it's reused."

Unfortunately, developers trained with software that's generated and used in one location with a single set of servers often don't understand the precautions needed for Web applications that take code, data, and elements of the interface from many servers, she says.

[ For more background on securing Web-based apps, see 5 Problems with SaaS Security. ]

The typical number of security flaws, especially in legacy or other homegrown software, must be taken into account by cloud-computer service providers, says Thomas Kilbin, CEO of cloud and hosted-server provider Virtacore Systems. After all, he says, customers who want on-demand compute capacity don't want to rewrite all their applications just to run in an environment designed to save money and add convenience.

"Our customers are taking apps they had running in their back office and moving them to private clouds for the most part," Kilbin says. "They are not developing any apps geared towards only working in a cloud IaaS/SaaS model. We secure these apps via a number of methods, traditional firewalls, app specific firewalls from Zeus, etc."

Keeping Web-based apps secure can be particularly tough for smaller IT teams.

"The cloud model is more threat-rich than the shared hosting model, mainly because in shared hosting the core OS and apps -- php, perl, mysql--are kept updated by the service provider," Kilbin says. "In the cloud, the customer has to keep the core OS updated, along with the application stacks, in addition to their code."

Most customers don't have the expertise or the time to do so, Kilbin says.

Some 2,922 applications were examined by Veracode in the past 18 months, with the results detailed in the company's recently released State of Software Security Report: The Intractable Problem of Insecure Software.

Some of the applications sent to Veracode for testing come from ISVs or corporate programmers in the last stages of development. Another big chunk comes from developers who have to present certifications or risk analyses before closing a deal with government agencies or heavily regulated industries.

Old App Flaws Revealed Before Web Moves

Increasingly, however, Veracode is testing software that clients have used for a long time or are very confident in, but are now migrating to a cloud or Web-based service environment. The requests often come from corporate IT executives who turn out to be wrong in believing that their secure, homegrown applications are either homegrown or secure, especially when they're moved into multi-site environments for the first time.

Both commercial and open-source applications failed Veracode's tests more often than homegrown -- at 65 per cent and 58 per cent respectively. Homegrown applications failed 54 per cent of the time, Veracode reports.

Software written by outsourcing firms missed the mark an astonishing 93 per cent of the time, Veracode says.

Even applications being used by banks and financial service companies failed 56 per cent of the time on initial submission, though the criteria are tougher for those applications, because problems in those apps would create more havoc than, say, in an internally developed server-monitoring application, King says.

Internal developers shouldn't be comparatively complacent, however, King says. Though internal apps are generally assumed to be made of 70 per cent homegrown code, reuse of code, objects and procedures is so common that between 30 per cent and 70 per cent of the code in homegrown applications actually came from commercial software.

Internal developers are also unaccountably unaware of the most common exploits likely to be used against Web-fronting applications, resulting in an 80 percent failure rate for Web applications, which are tested against the list of 10 most-common security threats published and publicized by the Open Web Application Security Project (OWASP), King says.

"At that point it just comes down to developer education," King says.

Cross-site scripting is the most common security flaw in all the types of software Veracode tests, but is most noticeable in Web- and cloud-based software, King says.

But the time it takes to fix problems and get an application to an acceptable level of security has dropped drastically from 30 to 80 days a year or two ago to only 16 days now, mainly because developers of all stripes are putting greater emphasis on security, software quality, and shortening their time to market, King says.

There aren't any shortcuts, but Veracode does have some suggestions for IT teams to counter the most consistent app security problems:

1. Design apps assuming they'll link cross-site; secure those links and the processes that launch them.

Cross-site scripting (XSS) accounts for 51 per cent of all vulnerabilities, according to Veracode. Apps written in .net have an abnormally high number of XSS issues because many .net controls don't automatically encrypt data before sending or storing it. Check and encrypt all points of output. Inadequate or absent encryption in non-.net applications also created problems, but are easy to fix once the source of in-the-clear data broadcasts are identified.

2. Focus your efforts on the greatest source of vulnerabilities.

You can assume software from any provider is likely to have vulnerabilities, but put extra Q/A and security analysis effort into code from outsourced programming services, ISVs and components from either of those that find their way into homegrown applications.

3. Verify security of the application itself in a cloud or SaaS environment.

Whether the customer or the service provider supplies the application, check it for flaws or vulnerabilities in a realistic cloud/SaaS/shared-resource environment, not just in a workgroup on a LAN. Security in the cloud platforms is still evolving, and the skills to write secure code for them is not widespread. Stick extra red flags on this part of your project plan.

4. Location is irrelevant. New criteria are impact, impact, impact.

A printer-management application with a flaw that allows hackers to draft a LaserJet into a bot army can cause headaches. An accounting, customer-data-management or cashflow-automation app with a backdoor can put you out of business. Use Level of Risk as a multiplier to determine how important a particular app is to evaluate, and how much time or money you should spend getting it fixed.

5. Don't ignore the basics.

The 10 most common attacks on Web applications are listed here by OWASP. The 25 most significant security errors that appear in applications are listed here. They're easy to read and come with extra help to fix or avoid errors already known by everyone who might want to hack your systems.

Show Comments