AV-TEST Product Review and Certification Report – Q4/2010
During October, November and December 2010 we continuously evaluated 22 security products using their default settings. We always used the most current publicly-available version of all products for the testing. They were allowed to update themselves at any time and query their in-the-cloud services. We focused on realistic test scenarios and challenged the products against real-world threats. Products had to demonstrate their capabilities using all components and protection layers.

Internet Security
| Version | 2011 |
| Platform | Windows Vista (SP2, 32 bit) |
| Report | 104850 |
| Date | Q4/2010 |
Protection
Protection against malware infections
(such as viruses, worms or Trojan horses)
More information
Industry average![]() | October | November | December | |
| Protection against 0-day malware attacks from the internet, inclusive of web and e-mail threats (Real-World Testing) 77 samples used | 82% | 96.0% | 85.0% | 91.0% |
| Blocking of malware on or post execution (Dynamic Detection Testing) 20 samples used | 55% | 90.0% | ||
| Detection of a representative set of malware discovered in the last 2-3 months (AV-TEST reference set) 642,693 samples used | 96% | 100% | 100% | 100% |
| Detection of widespread malware (according to the WildList) 30,532 samples used | 100% | 100% | 100% | 100% |
| Protection Score | 5.5/ 6.0 | |||
Repair
Cleaning and repair of a malware-infected computer
Industry average![]() | October | November | December | |
| Removal of all active components of widespread malware (according to the WildList) from a computer 22 samples used | 82% | 95.0% | ||
| Removal of further malicious components and remediation of critical system modifications made by malware 22 samples used | 45% | 59.0% | ||
| Detection of deliberately hidden active malware (Rootkits and stealth malware) 10 samples used | 100% | 100% | ||
| Removal of deliberately hidden active malware (Rootkits and stealth malware) 10 samples used | 60% | 50.0% | ||
| Repair Score | 5.0/ 6.0 | |||
Usability
Impact of the security software on the usability of the whole computer
(lower values indicate better results)
More information
Industry average![]() | October | November | December | |
| Average slow-down of the computer by the security software in daily use Number of test cases: 13 | 221% | 150s | ||
| False detections of legitimate software as malware during a system scan (false positives) 533,690 samples used | 7 | 6 | 10 | 18 |
| False warnings of certain actions during the installation and use of legitimate software 20 samples used | 1 | 0 | ||
| False blockings of certain actions during the installation and use of legitimate software 20 samples used | 1 | 0 | ||
| Usability Score | 3.5/ 6.0 | |||
Evaluation based on a point system
All products can achieve a maximum of 6 points each in the three categories of protection, performance and usability. This means 18 points are the best possible test result.
At 10 points or higher, a product is awarded the AV-TEST seal of approval.
At 17.5 points or higher, AV-TEST also issues the "TOP PRODUCT" award. Additional details ›
outstanding | satisfactory | insufficient | certificates |






















































