2007-12-01
Abstract
Fewer than half the products submitted for this VB100 review on Windows 2000 made the grade - largely thanks to some pesky polymorphic file infectors and a rash of false positives. John Hawes has all the details.
Copyright © 2007 Virus Bulletin
Windows 2000 is getting a little long in the tooth, having been superseded within two years of its release by Windows XP – whose slightly shinier surfaces seemed so revolutionary back in 2001 – and this year by the even shinier Vista. Despite its age and rather drab looks, Win2k soldiers gamely on, serving its purpose perfectly adequately for plenty of users and still being the operating system of choice in many homes and businesses.
For the developers of security products this represents something of a challenge. New platform versions will inevitably present plenty of new hurdles, with tweaks needed to various parts of the products, not least the interfaces, to keep pace with the ever-improving look and feel of computer desktops. But while all this newness is being added there is also a duty for developers to keep in touch with the old.
While many (but by no means all) security vendors, including Microsoft itself, have retired support for the Win9x family, Windows 2000 (currently held in an ‘extended support’ period by Microsoft) remains too big a market to drop, and its close proximity to current market leader Windows XP has meant that, in most cases, little extra work is needed to ensure mutual compatibility. Of course, with most development and QA eyes firmly on the more common, more recent platforms, bugs and troubles on older versions are more likely. However, with yet another bumper crop of products to slog through in a somewhat short month, I hoped that the products would prove as stable, reliable and trouble free as the platform itself.
Windows 2000 has been sitting on Service Pack 4 for several years, and as usual with VB100 tests the platform was used in a fairly bare state with no further updates added unless required by a specific product.
The installation and setup of Windows 2000 was thus a fairly straightforward task, familiar from countless previous ventures down the same path, and complicated only by a lack of support for some components in the fairly new machines preferred for VB100 testing. Rather than face several weeks testing at low resolution, extra drivers were added to fully enable the modern graphics cards, as well as network interfaces, but otherwise the systems were left untouched. I expected some products to require updates, such as upgrading Internet Explorer or Windows Installer to more recent versions, but these changes were not made by default in order to ensure that products with such requirements could easily be identified.
The test sets were based on the most recent WildList available on 26 October, with the product submission deadline a few days later. This month, a spurt of hard work from the WildList Organization meant that the September WildList was available in plenty of time to be included, and it was upon this list that the main test set was based. With a large number of new additions by recent standards, replicating and validating samples for the set was a bigger job than usual, but helped by the preponderance of familiar old names: large numbers of W32/Rbot and W32/Sdbot, with plenty of W32/Agobot and W32/Rontokbro and other similar items. There were a few less common additions, including plenty of file infectors, mainly from the W32/Looked and W32/Fujacks families, but including a W32/Virut variant which promised to present significant challenges in detection.
Also of note was the fact that, for the first time in a while, there was not a single new W32/Mytob variant to be added – a sign, perhaps, that this family is finally showing its age.
With a lot of lab time taken up with additions to the core set, expansion of the other test sets was limited. A sprinkling of items were added to the collection of worms and bots (mostly yet more variants of the major families) and the existing polymorphic test sets were expanded.
The clean set was enlarged with the usual selection of items, mostly from popular and recently released software packages on common download sites.
To assist in the presentation of speed results a small new set of files was added. With products offering some wildly different sets of default settings, the archive test has long presented problems when showing speed measurements, with products that do not scan inside archives unfairly showing better speeds than their more thorough rivals. To guide readers in interpreting these results, a set of common archive types has been created at various depths of nesting, with the Eicar test file at the bottom of each. A plain, uncompressed copy of the test file was added to check that it was indeed included in the detection, and as an extra, another copy renamed to a random extension was added to test scanning of non-standard filenames. I created a rather arbitrary cut-off point, deciding that products should detect at least five levels deep in at least four of the eight archive types included in the set in order to be include in the 'all files' speed graphs, and below this level a product’s scan times would only be included on the ‘default settings’ display.
Czech Republic-based AEC has been doing pretty well with its Trustport product, achieving some impressive scores in numerous tests thanks to its multi-engine approach. The product submitted here, the anti-virus component, is not available as a standalone product but is part of the Trustport Workstation suite, along with a swathe of other security solutions, and is rolled into a range of server and gateway products.
Installation of the product hit an immediate, if not unexpected stumbling block in the form of the requirement for Internet Explorer 6 or newer. While this is not an extravagant demand, it does raise a small concern – it’s more than possible that a user, having restored a system to an old safe state (perhaps using a rescue CD provided by the system retailer), would be in the position of running a bare Windows 2000 installation, and would thus have to spend quite some time online in an entirely unprotected state to acquire the required updates. Given the scare stories that estimate the average infection time for an unprotected system connected to the web to be as little as ten minutes, this window of exposure could be unacceptable.
Once installed, Trustport presents a solid and reliable appearance with its graphics depicting well shielded footsoldiers – an image backed up by the multi-engine scanner at its heart. The product’s makeup has changed somewhat since its last appearance, with the BitDefender engine included in earlier versions replaced by those of Dr.Web and VirusBlokAda – an interesting selection, not least as it includes an engine which has yet to appear on the VB test bench. A lot of heuristic technology hinted at a high risk of false positives, but could be expected to ensure pretty thorough coverage of infected items.
Tests were carried out easily, with the speed tests particularly straightforward as the default action is to scan all files, including the contents of archives, both on demand and on access. The new set of archive types was detected in depth, although neither of the engines implemented on access seemed capable of penetrating .LZH files – the on-access mode uses only two of the available scanning engines, though more can be added by the more paranoid user as long as they have the available processing power. Of course, multiple engines are unlikely to achieve the best speeds or lowest overheads, and speed figures here showed a pretty hefty drain on resources.
The many engines spotted a fairly large number of potentially unwanted items in the clean sets, a large number of which were system tools from Sysinternals, and all of which were labelled in the log with the rather stark and worrying ‘Infected!’. However, their full definitions described them more accurately as tools or programs. As feared a few full false positives were also flagged, spoiling the product’s chances of winning another VB100 award. More surprisingly, a few samples of the new W32/Virut variant were missed on access, indicating that these were likely to prove a problem for at least a few more products as testing continued.
Please refer to the PDF for test data
Agnitum’s product is fairly recent and almost certainly developed since the arrival of Windows XP. It showed no signs of requiring any extra software – at least until halfway through the installation, when an error message revealed the absence of a required DLL. This did not seem to be a fatal problem, and the install continued to the requested reboot, on return from which the system froze in an unresponsive state. Reimaging and trying the installation again with the extra DLL in place led to a much more complex installation process, with a series of configuration pages to be worked through before reaching the reboot phase. Again, the system failed to return – even safe mode seemingly inaccessible – and the developers were called for assistance. Investigation indicated that the problem related to the rather modern systems being used for the test, and when the test image was ported to more humble hardware there were no such difficulties.
With no clear way of circumventing the problems on the main systems, tests proceeded minus the speed test, which would have been all but meaningless on the considerably slower hardware.
The product looked good and proved pleasant to work with, offering a wide range of modules which sadly went unexplored. With good detection across the test sets and no false positives generated in the clean sets, Agnitum earns a VB100 award.
Please refer to the PDF for test data
Alwil’s product is one of the more dependable regulars in VB’s tests, and while the interface is far from my favourite, its intricacies no longer cause too many difficulties. Some admirably solid results were achieved on scanning the new archive set, with neither archived nor renamed copies of the Eicar test file spotted in default modes, but everything detected with the archive and ‘all files’ settings turned on.
Speeds on demand were good, although on-access times were harder to come by – the product does not check files on simple opening, and on-access results for the infected sets were taken by copying the collection to the system across the network.
Results were pretty much as expected for avast!, with some older items missed but little from the more up-to-the-minute sets. Full coverage of the WildList was achieved, but hopes of a VB100 award were dashed by a single false positive in the clean set.
Please refer to the PDF for test data
AntiVir is another solid performer in VB’s comparative testing, with an excellent history both in detection and speed, and it did well again here. The product is pleasingly laid out and simple to use, with the installer especially rapid and problem-free, and the tests zipped along at a similarly impressive rate. The archive sets were covered fully by default on demand, and almost so on access, with the rather odd exception of a few files in the .ACE format – while most were spotted, including the deepest nested to 10 levels, levels 3, 5 and 8 were missed.
Infected items were covered pretty well, with only a small number of polymorphic samples of rather rare and obscure variants missed. With the WildList covered in full, including those pesky Virut samples, only false positives could stop Avira claiming another award, and unluckily, two files were indeed erroneously flagged as infected, denying Avira the chance to add to its collection of VB100 awards this month.
Please refer to the PDF for test data
The BitDefender product stated that a better version of the Windows Installer was needed to install it – but as a pleasant surprise it also informed me that it had a copy handy and would install it for me. The pleasurable moment soon passed though, when after a reboot and a second attempt at installing, it was found that IE6 would also be needed and on that count I would have to fend for myself. I had also been informed that Update Rollup 1 was required for the product to function – but a quick check without this generated no warnings from the product, and left the on-access functionality crippled, despite a comforting green tick insisting that all protection was active.
After several reboots therefore, I was finally able to get to work, and initial scans proceeded quite happily, with no false positives spotted on demand and most of the archive types detected easily, although .TGZ and self-extracting zips were only delved into to a depth of eight levels.
Scanning of the infected sets proved simple and highly successful, with a tiny number of misses and no false positives, thus earning BitDefender another VB100 award.
Please refer to the PDF for test data
Installing Bullguard confirmed a suspicion I had had all along – that the requirement for upgrades to Internet Explorer (already made by a few products and likely to crop up at least a few more times before I was done) is purely for cosmetic reasons. Bullguard has no such dependency, and installed smoothly on the bare system with no need for any extra work on my part.
The user experience was not adversely affected by the lack of modern display technology, and the tests proceeded nicely, recording similar times and detection rates to BitDefender, whose engine the product is based on. The archive results were likewise the same, with .TGZ and self-extractors limited to eight levels but everything else covered. With admirable detection rates – missing barely a handful of samples per set, none of which were in the WildList set – and no false positives, Bullguard earns its second VB100.
Please refer to the PDF for test data
A few hiccups occurred during the installation of CA’s home-user product, starting with the seemingly inevitable need to upgrade the browser (a minimum of version 5.5 this time). I also noted that some other items come along with the product, including the Yahoo! Toolbar, and that the browser homepage was set to Yahoo!, which I found rather surprising. I was positively upset by the fact that the boxes to accept these changes were checked by default – since the VB100 testing protocol requires default settings, this meant agreeing to yet more EULAs, which in traditional CA style must be scrolled all the way through before they can be accepted.
The design of the product itself is pretty slick, with clear and easy controls, and despite my misgivings about the optional extras I found myself quite liking it. Configuration was fairly minimal, but the defaults made sense, with archive scanning switched on for on-demand scanning (.ACE files not scanned) and off for on-access scanning except for a single level of the ubiquitous .ZIP (and its twin sister .JAR, essentially zip renamed).
Speeds were very good indeed, and detection generally good, but in the WildList set several items were missed including some W32/Rbot variants and the entire set of the W32/Viruts. CA thus misses out on a VB100 award here.
Please refer to the PDF for test data
CA’s more grown-up product, the corporate-targeted eTrust, did not complain about the browser in use during the installation, but I found myself needing to upgrade regardless when I later found that some of the popup screens in the options areas of the interface lacked their vital control buttons.
This interface has never been a favourite of mine, but its usual slowness under Windows XP was somewhat less intrusive under 2000. Accessing logs was as tricky as ever, with large ones occasionally overwhelming the display system and leaving me with blank browser windows and no option to export to a text file. As usual I simply removed the raw files to a Linux machine and stripped out the required data.
The logs indicated much better coverage of the WildList by eTrust than by its sister product, hinting that the home-user product submitted may have been using some slightly older definition data. Archive scanning was a little odd, with a maximum of nine levels checked on demand and none on access, despite the GUI inferring that they should be. Speeds were very good, and without any false positives eTrust succeeds where CA AV failed, and wins another VB100 award.
Please refer to the PDF for test data
Dr.Web proved much less problematic, with a simple installer requiring no extra fiddling and another very pleasing interface, laid out with impressive clarity and logic as well as being appealing to the eye. Running through the tests was quite enjoyable as a result, which was a good thing as they did take some time – Dr.Web is a very thorough product, delving deeply into files before passing them as clean. On demand, archives were not scanned by default. However, .CHM help files, of which a few are included in the clean set, are checked in all their many sub-parts, which explains the relatively low throughput, rendered even lower when full archive scanning is activated. Full archive scanning covered everything but .ACE to a depth of 10 levels.
Detection rates were excellent across the test sets until the WildList tripped the product up with several misses, including those pesky W32/Virut samples. A couple of false positives added to Dr.Web’s problems, and the product unfortunately misses out on a VB100 once more.
Please refer to the PDF for test data
Nod32 has undergone a bit of a revolution recently, with a spanking new interface introduced to coincide with the launch of version 3, and that of its big sister Smart Security (see VB, November 2007, p.19). However, ESET opted to give the ever-reliable version 2.7 one last hurrah this month.
Installing and using the product has never been too difficult, and as usual testing sped through in remarkable time, with the usual excellent results. Speeds were as fast as ever, although archives could not be scanned on access, and detection was at the expected near flawless level, with only a single rather obscure and highly polymorphic sample missed. With the WildList fully covered and no false positives, ESET adds yet another VB100 award to its groaning trophy cabinet.
Please refer to the PDF for test data
Fortinet’s desktop product remains little changed since I first encountered it, presenting a serious-looking interface with a wealth of security functions accessed via a string of tabs. During installation the product complained about a missing DLL file, but presumably this related to some other part of the product, as the anti-virus seemed as solid and robust as ever.
Usability was similarly problem-free, and scanning times were decent for the level of thoroughness offered by the default settings, detecting the majority of the nested archives without the need for adjustment. Detection was splendid almost across the board until those troublesome Virut samples reared their ugly heads, with two missed detections being enough to prevent Fortinet from winning another VB100 award.
Please refer to the PDF for test data
F-PROT is perhaps the simplest product on test this month, with a fairly basic interface providing access to straightforward anti-virus scanning and cleaning and no additional bells and whistles. This made testing pretty straightforward, and everything zoomed through in good time, with the more in-depth speed tests skipped on access thanks to a dearth of configuration.
Detection was as top-class as ever, with just about everything taken in the engine’s stride, but a single false positive showed up in the clean set, a file apparently highly similar to a known malicious item, meaning that Frisk joins the growing list of vendors narrowly failing to reach the VB100 standard this month.
Please refer to the PDF for test data
F-Secure’s current product is another highly familiar one which took little time to get set up and going. The in-depth scanning with multiple technologies meant speed times were not the best, even though archives could not apparently be scanned deeper than five levels. While running sizeable scans, the interface choked up a few times, lingering unresponsive at the very last stage of the scanning process, with only a reboot able to bring it back in touch with the user. Logging was also a little pesky, with sizeable chunks of information apparently missing from logs exported from the viewer interface.
However, detection was excellent, and there were no false detections, and F-Secure thus comfortably earns another VB100 award.
Please refer to the PDF for test data
GDATA’s product is another multi-engine beast, which for this submission at least seems to have dropped the familiar ‘AVK’ name. The interface seemed unchanged however – a clear and well-laid-out thing which is always a pleasure to operate.
Of course, the multiple engines meant that scanning speeds were slow, even on access, but depth of scanning and accuracy are clearly the product’s strengths, and with barely any misses and no false positives GDATA also wins a VB100 award for its collection.
Please refer to the PDF for test data
Wildly popular AVG, the free home-user version of which seems to be in almost every home these days, has always been a little fiddly for my liking, but whether it has been tweaked a little or I’ve just grown used to it, in this test I found the interface perfectly reasonable and even quite pleasant to work with.
Configuration was a little short for the on-access scanner, but elsewhere everything worked fine, with very good if not great detection in the infected sets, including flawless coverage of the WildList despite those difficult polymorphic samples. With no false positives either, Grisoft also wins another VB100 award.
Please refer to the PDF for test data
Ikarus has had some problems in its recent entries in VB100 comparative reviews, but earlier issues with its interface seem to have been resolved – on this occasion everything ran fine and stably with no difficulty. Even the updates to the Windows Installer and the .NET framework required by the product were provided thoughtfully as part of the submission and installed automatically as part of the setup process.
Configuration of scanning is somewhat limited by the interface, but the default setting of scanning up to three levels into archive files seems sensible, and speeds were fairly good across the sets.
Detection was a little improved on previous efforts, but a handful of samples of each of two Virut variants in the set proved undetectable, and a rash of false positives added to Ikarus’s woes. There were also a fair number of items labelled ‘not-a-virus: Monitor.Win32.Keylogger’, which for now I have generously recorded as ‘suspicious’ rather than full false positive detections, but which certainly seem a little suspect themselves. Despite these problems the product seems to be improving fast and looks a likely candidate to qualify for a VB100 award sometime soon.
Please refer to the PDF for test data
Iolo returns to the test bench for another stab after being denied a VB100 by a whisker a few months ago. The product is well designed and pleasant to use, and although it requires IE6 to operate, it politely offers to go online and fetch a copy.
As with many of the products aimed more squarely at the home user, configuration was somewhat limited, with on-access scanning barely adjustable and actions on discovering malware restricted to delete, disinfect or quarantine. With logging also absent, I allowed the product to delete the virus collections from the system, which left only a few samples in most sets but also many of the two Virut strains along with another file infector, W32/Expiro. Iolo will therefore have to try again for the VB100 award, which should be well within its grasp with just a little more work.
Please refer to the PDF for test data
Kaspersky is a much more seasoned product, version 7 of the product having dropped a lot of the cuddly cartoonishness of the previous offering and presenting a sterner but glossier face to the world. Usability has not been diminished however, and few problems were encountered other than some slowness exporting particularly large logs to file.
Detection rates were excellent as ever, with the new nested set detected very neatly. With no false positives spotted, all looked good until a single item was missed on access. This, an instance of W32/Autorun added recently to the list, could be detected by the product on demand, but was not scanned on access unless the ‘scan installation packages’ option was activated. Kaspersky thus narrowly misses out on a VB100 this time.
Please refer to the PDF for test data
Kingsoft achieved a VB100 award in its previous appearance in VB (see VB, August 2007, p.13). The product this time seemed little changed, with the interface nicely laid out and appearing pretty stable, but experiencing some difficulties in the log viewer when faced with unfamiliar locales – only US English is supported, and others cause a nasty crash.
Scanning speeds were rather average, and configuration absent on access, but false positives and even suspicious flags were encouragingly absent throughout the clean sets. The infected sets were less well covered, in particular the older items, and in the WildList set several nasties were missed, including most of the files infected with Virut and Expiro, as well as several W32/SDbot variants. Kingsoft thus falls short of the required standard this time, and will have to try again to achieve its second VB100 award.
Please refer to the PDF for test data
McAfee’s desktop product is another that seems to have remained relatively unchanged for some time, and its performance was similarly predictable.
Scanning times were decent, with archives ignored by default in both modes but thoroughly handled if requested; detection was impeccable, with nothing missed anywhere and no false positives. VirusScan wins a VB100 award effortlessly.
Please refer to the PDF for test data
Perhaps unsurprisingly, Forefront makes use of all available Microsoft technology and requires numerous updates to be in place before it will install. The rollup package, an improved version of the installer, and an update to the Agent API are all required. It also uses the event log to record its activities rather than providing its own system, which I found a little awkward, but the server-side management system doubtless provides a more usable form of information management.
Configuration was rather minimal, which again may be explained by the absence of the management side of things, but defaults were sensible and testing ran without difficulties. With nothing of significance missed and no false positives, Forefront qualifies for a VB100 award.
Please refer to the PDF for test data
Microworld Technologies presents a fairly comprehensive product, including the Kaspersky engine alongside a range of its own protection technologies. The product’s default settings lean towards the paranoid, with on-access defaults including all archive types. With a well designed interface providing for all my needs, testing thus took little of my own time, but quite a bit for the system, as clean sets were probed deeply.
Detection of the infected sets was excellent, eScan managing to avoid the problem which upset Kaspersky’s own product, and comfortably earning a VB100 award.
Please refer to the PDF for test data
Norman’s is another interface which has grown on me after struggling to understand its complexities in earlier tests. The only lingering annoyance is the lack of information on scan progress, with there being no progress bar or count of files scanned so far.
Speeds were reasonable, and detection levels decent, with most misses on old and obscure items. However, two files in the clean sets were flagged as nondescript malware by the heuristics, thanks to the use of a rather unusual packer, and again some of those tricky Virut samples were missed, leaving Norman just short of the mark for the VB100 award this month.
Please refer to the PDF for test data
PCTools is a relative newcomer to VB100 comparative testing, taking its first award just a few months ago (see VB, June 2007, p.10).
The plain anti-virus product, based on the VirusBuster engine, offers a reasonable level of configuration and a pleasant user experience for the most part. The logging presented rather a strange problem though – opening logs from the interface brought up a ‘file in use’ error, and they could thus only be accessed by copying the files and opening the copies. Some good detection rates were shown, but also some remarkably slow times in the speed tests, even with the default on-demand settings scanning archives to a depth of one level only. However, with nothing missed in the WildList and no false positives, PCTools AV wins itself a second VB100 award.
Please refer to the PDF for test data
Spyware Doctor is PCTools’ rather more venerable anti-spyware product, now available with anti-virus functionality rolled in, and while the interface closely resembles the plain AV product there were a number of differences. Logging seemed to be limited to a small file size, meaning that larger scans needed to be split up into chunks to acquire the necessary data, while on-access scanning seemed not to be sparked by simple file opening, which meant the product had to be excluded from the on-access speed test. On-demand times were considerably better than those of its sister product, despite defaults including all archive types (apart from the rather obscure .LZH) to a depth of at least 10 levels.
Detection rates differed slightly too, and in the clean set the anti-spyware side of things detected a single false positive, thus denying Spyware Doctor a VB100 despite full coverage of the WildList.
Please refer to the PDF for test data
Quick Heal (which is now the name of both the product and its vendor, having recently changed from CAT) is another well designed product. It zipped through speed tests in good time and could only be cajoled into scanning to a depth of five levels, into a limited selection of archive types. A few nasty crashes occurred during the scanning of infected sets, but they were handled better on a second attempt, and while detection was a little short on the older sets nothing more important was missed, and false positives were also absent. Quick Heal thus earns itself a VB100 award.
Please refer to the PDF for test data
UK-based Redstone produces a managed-service protection product, of which this is a simple client version using the .NET framework for its interface. Running the product is a straightforward business, with a simple menu accessed via the system tray. Configuration is a little more fiddly, requiring the tweaking of registry settings, but the submission came with a prepared set of useful entries, enabling testing to proceed without too many problems.
The product is based on the Kaspersky engine, and detection rates were thus at the top end of the scale, while speed times were more average. A few difficulties were encountered, including the absence of logging and some odd behaviour on demand, when the ‘always delete’ option seemed to be ignored for a few items, resulting in a string of popups requesting confirmation before deleting.
False positives were absent, but the W32/Autorun sample which tripped up Kaspersky was also missed here, in both modes, and Redstone will thus have to try again before gaining a VB100 award.
Please refer to the PDF for test data
Another newcomer to the VB100 test bench, China-based Rising has developed a considerable profile outside its home country in recent years, and it was with some excitement that I took my first look at its product. First impressions were excellent, with the product looking very clean and stylish, clearly laid out and easy to use.
Speed test results were fairly good, and stability seemed solid too, but during on-access scanning of the infected sets the product seemed to stop blocking after 10,000 samples or so. The test was retried at a slower pace. The problem did not recur, and results were thus obtained, showing the expected high numbers of misses in older sets but little in the newer areas. Two misses in the WildList, both single samples from sets of file-infectors, and a pair of false positives in the clean sets, were enough to spoil Rising’s chances of qualifying for the VB100 at first attempt, but it is another likely candidate to make the grade pretty soon.
Please refer to the PDF for test data
Sophos is among the most regular of VB100 entrants, with its product little changed in the half-dozen Windows tests I have performed in my time here, and as usual setting it up and running the tests were simple tasks. Speeds were very good, even with archive scanning turned up to the maximum available five levels, and after a false positive upset things last time (see http://www.virusbtn.com/virusbulletin/archive/2007/10/vb200710-comparative)the clean sets were cleared with only a few hacker tools alerted on as possible security risks.
Detection was at its usual high levels, with almost everything covered, but again in the WildList set those Virut samples proved too difficult, and Sophos is denied the VB100 for the second time in a row.
Please refer to the PDF for test data
This month Symantec presented a totally different looking product from those seen in previous tests, considerably more colourful and less severe. The cosmetic enhancements required IE6, and after the installer had aborted requesting this update it left something lingering behind, which meant the IE6 installer insisted on a reboot before it could run itself. However, after several reboots to set up, tests continued apace.
Speeds were reasonable, although configuration was somewhat less in-depth than in previous submissions and archives could only be scanned to a depth of three levels, with .ACE and .TGZ ignored. However, detection was excellent, with nothing missed, and without false positives either Symantec earns another VB100 award.
Please refer to the PDF for test data
OfficeScan also required IE6 in order to operate the web console which provides much of the product’s configuration, although options were available to delegate some control to the simpler local interface. Testing slipped rapidly along, flipping between the two control systems as required, and times were good and detection rates decent, although the renamed Eicar test file was not spotted with the default settings. Some older sample sets were a little short, but more seriously two Virut samples were missed, one each of the two variants causing most trouble here, and Trend is thus denied an award this time.
Please refer to the PDF for test data
Bringing up the alphabetical rear, VirusBuster presented its usual colourful and reasonably usable product, which provided adequate configuration options and its usual slightly fiddly system of setting up scanning jobs. These jobs showed good scanning speeds, and pretty thorough detection across the sets; with those troublesome Virut variants taken in its stride, and without any sign of a false positive VirusBuster takes home another VB100 award.
Please refer to the PDF for test data
Having expected numerous problems to have arisen from the aging platform, these proved to be limited to the chore of installing extras before products could install or operate properly.
In fact, far more difficulties were thrown up by another rather old issue, the polymorphic file-infecting virus. With modern malware trends having tended for some time towards the non-self-replicating, or at least towards static worms which simply drop identical copies of themselves around the place, old-style file infectors have been making something of a comeback lately. W32/Detnat, W32/Looked (aka Viking), W32/Fujacks, and of course the more tricky polymorphic type, W32/Polip and W32/Virut, all lurk on the WildList and some of them have made a considerable impression on global prevalence charts in recent months. This month’s Virut addition revealed deficiencies in detection for several products, the vendors of which have all been informed of the problem, which should have been resolved by most in advance of the publication of this review.
A swathe of products have also fallen to another problem which has shown a rising trend lately: false positives. A relatively small addition to the clean test sets threw up several individual examples (few of the files that were false alarmed on affected more than one product, or more specifically one engine), and in some cases several files were misidentified by a single product.
The result has been one of the poorest scores for some time in a VB comparative, with fewer than half the entrants making the grade, and another trend – the inclusion of third-party engines in products – magnifying the scale of the problem. Hopefully the shock of so much devastation caused by a few polymorphic viruses will ensure virus labs remain on their guard and encourage more thorough checking of detection for file-infecting items in future.
Test environment. Tests were run on identical machines with AMD Athlon64 3800+ dual core processors, 1GB RAM, 40GB and 200 GB dual hard disks, DVD/CD-ROM and 3.5-inch floppy drive, all running Microsoft Windows 2000 Professional SP4. Agnitum Outpost was tested on a 1.6 GHz Intel Pentium machine with 512 MB RAM and is thus excluded from speed measurements.
Any developers interested in submitting products for VB's comparative reviews should contact [email protected]. The current schedule for the publication of VB comparative reviews can be found at http://www.virusbtn.com/vb100/about/schedule.xml.