I bet about all of the internet users life, sometime it happens where we think if only we could download a full website so that we can check it even when we don’t have internet connection! There might be lots of reason for doing that but mostly it can be to check some sites content entirely all together without the hassle of reloading pages or may be some other reasons. Well whatever your necessity to doing so – It’s quite possible to do that! You just need to know some tools which will help you to do that. Guess what? It’s free also. Yay! that’s all we can say when we get something free yet useful
Okay fellows, in this article I am exactly going to reveal some secret tools to you More specifically, we’ll show you how to download an entire website using free software.
HTTrack is a free (GPL, libre/free software) and easy-to-use offline browser utility.
It allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your computer. HTTrack arranges the original site’s relative link-structure. Simply open a page of the “mirrored” website in your browser, and you can browse the site from link to link, as if you were viewing it online. HTTrack can also update an existing mirrored site, and resume interrupted downloads. HTTrack is fully configurable, and has an integrated help system.
Download it from here: http://www.httrack.com/page/2/en/index.html
How to use: Well, this manual is the best solution http://www.httrack.com/html/index.html
GNU Wget is a free utility for non-interactive download of files from the Web. It supports HTTP, HTTPS, and FTP protocols, as well as retrieval through HTTP proxies.
Wget is non-interactive, meaning that it can work in the background, while the user is not logged on. This allows you to start a retrieval and disconnect from the system, letting Wget finish the work. By contrast, most of the Web browsers require constant user’s presence, which can be a great hindrance when transferring a lot of data.
Wget can follow links in HTML and XHTML pages and create local versions of remote web sites, fully recreating the directory structure of the original site. This is sometimes referred to as “recursive downloading.” While doing that, Wget respects the Robot Exclusion Standard (robots.txt). Wget can be instructed to convert the links in downloaded HTML files to the local files for offline viewing.
Wget has been designed for robustness over slow or unstable network connections; if a download fails due to a network problem, it will keep retrying until the whole file has been retrieved. If the server supports regetting, it will instruct the server to continue the download from where it left off.
Most of the Linux distribution comes with Wget installed, so you don’t have to do anything to install Wget.
Using Wget to download entire website:
Create directory where you are planing to store the website content: mkdir /home/nikesh/linuxpoison
use following command to download the website:
wget -r -Nc -mk http://latestcrunchs.com/
-r Turn on recursive retrieving
-N Turn on time-stamping
-m Create a mirror
-k Convert the link
After completion, all content will get downloaded into your directory for offline viewing.
I hope you enjoyed reading this article and will give it a try and keep all the favorite websites in your pocket through your USB disk! Please let me know if you find this article useful. Please also let us know if you know any more tools that do exactly the same task or even better. I look forward to hearing from your about this article named “Downloading Entire Website Using Free Software - a how to guide to download website fully for offline browsing using software” through comment. Till then take care and have fun in your life