AV-TEST Product Review and Certification Report
During January, February and March 2011 we continuously evaluated 22 security products using their default settings. We always used the most current publicly-available version of all products for the testing. They were allowed to update themselves at any time and query their in-the-cloud services. We focused on realistic test scenarios and challenged the products against real-world threats. Products had to demonstrate their capabilities using all components and protection layers.
Protection
Protection against malware infections
(such as viruses, worms or Trojan horses)
Protection against 0-day malware attacks from the internet, inclusive of web and e-mail threats (Real-World Testing)
107 samples used
84.0
78.0
87.0
91.0
Blocking of malware on or post execution (Dynamic Detection Testing)
29 samples used
62.0
21.0
Detection of a representative set of malware discovered in the last 2-3 months (AV-TEST reference set)
522,634 samples used
97.0
98.0
99.0
99.0
Detection of widespread malware (according to the WildList)
24,106 samples used
100
100
100
100
Protection Score
3.5/ 6.0
Repair
Cleaning and repair of a malware-infected computer
Removal of all active components of widespread malware (according to the WildList) from a computer
21 samples used
86.0
95.0
Removal of further malicious components and remediation of critical system modifications made by malware
21 samples used
48.0
62.0
Detection of deliberately hidden active malware (Rootkits and stealth malware)
14 samples used
79.0
64.0
Removal of deliberately hidden active malware (Rootkits and stealth malware)
14 samples used
57.0
36.0
Repair Score
3.5/ 6.0
Usability
Impact of the security software on the usability of the whole computer
(lower values indicate better results)
Average slow-down of the computer by the security software in daily use
Number of test cases: 13
171
79s
False detections of legitimate software as malware during a system scan (false positives)
841,767 samples used
False warnings of certain actions during the installation and use of legitimate software
20 samples used
False blockings of certain actions during the installation and use of legitimate software
20 samples used
Usability Score
5.5/ 6.0
Evaluation based on a point system
All products can achieve a maximum of 6 points each in the three categories of protection, performance and usability. This means 18 points are the best possible test result.
At 10 points or higher, a product is awarded the AV-TEST seal of approval.
At 17.5 points or higher, AV-TEST also issues the "TOP PRODUCT" award.


