Secure Coding mailing list archives
Secured Coding
From: "Greenarrow 1" <Greenarrow1 () msn com>
Date: Thu, 12 Feb 2004 17:06:15 +0000
"The security products industry has created some great defenses for protecting technology that can be walled off from non-trusted outsiders. Firewalls, VPNs and strong authentication are mature technologies that work well to wall off vulnerable software where possible. But security product defenses fall short when protecting technology that needs to be exposed to non-trusted (or less trusted) outsiders. These are your potential customers, current customers, partners and suppliers. Web applications and e-mail are examples of this type of software and are a major source of security vulnerabilities. The class of software that can't rely on network defenses needs to take care of its own security. The source of the problem needs to be the source of the solution, the software itself. Currently the software industry is creating secure software in reactive mode. Every time you download a patch and update your computer to make it more secure, you are downloading a correction to a piece of software your computer runs. The timeline leading up to the correction usually goes like this: Vendor ships software with latent security flaw. Vulnerability researcher discovers the flaw through manual testing and reports it to the vendor. A maintenance engineer at the vendor reproduces the flaw and tracks down the place in the source code where the original programmer made a coding error. The engineer fixes the problem in the source code, builds a patch and runs a regression testing suite to make sure the fix didn't break anything else. The vendor issues a patch and notifies customers. Attackers develop exploits and compromise vulnerable computers. Customer downloads patch, potentially runs his/her own test suite and then deploys the patch on each vulnerable computer. If there was a way to identify the problem in the source code before the software shipped to customers, large expenses would be saved by both vendors and customers. A NIST study, "The Economic Impacts of Inadequate Infrastructure for Software Testing, 2002," put the cost of fixing a bug in the field at $30,000 vs. $5,000 during coding. That study only takes into account the vendor's cost. A much larger cost is borne by software users: the cost of cleaning up worms, viruses and other intrusions, and keeping systems patched. For minor vulnerabilities customer costs are in the millions. For major worm outbreaks the costs can range into the billions. Luckily we are not doomed to a costly reactive approach. There is a way to prevent most security flaws during the original production of the software. It's called secure coding. There are well known classes of coding flaws that any programmer can easily learn to identify and avoid. Most of it is just good programming practices such as correctly sizing buffers, checking function return codes, and using platform security and crypto API's properly. Most insecure code is simply sloppy code. Software customers can save time and money by demanding that their vendors fix flaws up front with secure coding and not subject them to costly and seemingly endless worm and virus remediation and patching regimens." Regards, Greenarrow1 InNetInvestigations-Forensics
Current thread:
- Secured Coding Greenarrow 1 (Feb 12)
- Re: Secured Coding Chris Wysopal (Feb 12)
- Administrivia: Registration require sites (was Re: Secured Coding) Dave Aronson (Feb 13)
- Re: Administrivia: Registration require sites (was Re: Secured Coding) Alex Lambert (Feb 14)
- Administrivia: Registration require sites (was Re: Secured Coding) Dave Aronson (Feb 13)
- <Possible follow-ups>
- RE: Secured Coding jjchryan (Feb 12)
- Re: Secured Coding Chris Wysopal (Feb 12)