Blue Crab 4.8.9 Released

May 26, 2007 in Software (E)


Blue Crab 4.8.9 Released - May 25, 2007 - With Blue Crab you can download all the content including HTML, PDF, graphics, video, file archives, etc., or use selective filtering to restrict downloads to specific kinds of files. For example, you can choose to save only the JPEG images Blue Crab finds, or just the PDFs.

Blue Crab has a special feature called the "Image Grabber" which you can use to easily download just the graphics on a website. Moreover, you can view a slide show of the images as they are downloaded. You also have the option of "flattening" the images, i.e. putting them all into one folder, or preserving the folder structure on the server (just as when downloading a complete website for offline viewing.)

Starting with a single URL Blue Crab begins traversing the site by following the links on the HTML pages it finds. Blue Crab does not stray off the domain of the starting URL; in other words it won't download the whole web! Moreover, you can restrict Blue Crab to a subset of the given website by specifying strings that must match parts of the URL.

Blue Crab is gentle on the server. It processes only one URL at a time and may be configured to grab resources at preset time intervals.

Major Features:

* Unique URL extraction algorithms yield very complete downloads compared to other products of this type.
* You can download websites either for offline browsing, via re-linking of URLs, or to create virtual copies of them on your hard drive, i.e. backup a website as-is.
* Supports both HTTP, HTTPS protocols as well as server realms. Detailed Grabber window shows server HTTP header response fields.
* Optionally display web pages as they are crawled using either your own web browser or Blue Crabs own web window.
* You can restrict a crawl to just the initial domain, or allow the program to "stray" into offsite URLs.
* Blue Crab can remap "dynamic" URLs which contain path and search arguments so that resulting disk files are navigable offline.
* Image URLs into JPEG files (or any other graphic file type supported by Quicktime) to create complete pictures of web pages not possible simply by taking a single screenshot yourself.
* Blue Crab provides its URL imaging ability as a service menu item available from other applications.
* URLs can be imaged into files, or onto the clipboard.
* Built-in directory search displays found files as a hierarchical list, with double-clickable entries.
* A "batch window" enables you to download your own list of files. Simply drag and drop URLs into the batch window, or import them from an HTML file , or text file.
* Grabbing in "images-only" mode just downloads image files, optionally "flattening" the results (i.e. putting all images into one folder.)
* Reusable configuration settings control what is grabbed, saved or crawled:
* Filter by size, date, file type or even content
* Filter by filename, extension or path
* Bookmarks window for storing frequently crawled websites.
* Supports form submissions to enable you to begin a crawl from a login page.
* Supports cookies for more accurate downloads.
* Generate site maps that consist of a hierarchical display of URLs. Such maps can be optionally filtered by extension.
* Find email addresses.
* Customizable user-agent (spoofing) enables you to download websites which are platform specific.
* Optionally receive email notifications of Blue Crab's progress during a lengthy crawl.
* Dock badging provides visual feedback or progress when the program is in the background.

Limit Point Software has produced software exclusively for the Mac community for over ten years. The application areas include web crawling, bulk email, Quicktime editing, document search and much more. A novelty is our collection of "Utility Software" which consists of a bundle of independent small programs which lack feature bloat as each targets a specific task.


Print this page | PDF | TXT | Read other releases by this member.
web | crawler | download | website | collect | images | search