The trial version is fully functional but lasts for 30 days which should be enough time for you to test out its functions against your sites.
Yes, it supports the requirements for the number of urls greater than 50000 as defined in the protocol. When the number of urls are greater than 50000, it automatically creates an index page and then creates site maps site1.xml, site2.xml and so on
The speed of the crawler will mainly depend on your internet connection speed and on how many pdf's or word documents there are on the site. However, on a reasonably fast broadband connection, you could expect a rate of over a hundred pages per minute.
It is able to crawl and index normal web files of type, html, asp, aspx, php and so on. It is also able to index pages that are dynamically generated from databases.
The text from files of type pdf, Microsoft Word and rtf documents is extracted and used as the extract text. Also, titles of Word and Excel files are extracted and used as the title of the document.
Yes, it creates text log files for each crawled site detailing each web page and the links found on the page. It also creates an error log file detailing any errors such as 404 file not found.
The site map filenames are defined in the 'Crawl settings' tab.
If the filename entered is sitemap.html, then the output filename for a single file will be sitemap.html. If the filename entered is sitemap.php, then the output filename will be a single file sitemap.php. For multiple files, if the filename entered is sitemap.php, then the multiple a-z files will be identified as sitemap_a.php, sitemap_b.php and so on.