Create publicly available web page archives with

Martin Brinkmann
Apr 22, 2015

Web pages can change from one moment to the next. Entire websites may go down and take contents with them, content may be edited or removed, or sites may become unavailable because of technical issues.

If you need access to information, or want to save a copy of them to make sure you can access them at all time, then you have several options at your disposal to do so.

Probably the easiest is to save the web page to your local system. Hit Ctrl-s while on it, pick a descriptive name and local directory, and all of the contents are saved to the computer you are working on. Extensions like Mozilla Archive Format improve that further by saving all contents to a single file.

Another option is to take a screenshot of the page or part of it instead. This works as well, has the advantage that you save a single file but the disadvantage that you cannot copy text.

Tip: Firefox users hit Shift-F2, type screenshot and then enter to create a screenshot of the active web page. Chrome users can save web pages as PDF documents natively instead.

The third local option comes in the form of website archivers. Programs like Httrack crawl websites for you and save all contents to a local directory that you can then browse at anytime even without Internet connection.

Remote options can be useful as well. The most popular is without doubt offered by as it creates automatic snapshots of popular Internet pages that you can access then. Want to see one of the first versions of Ghacks? Here you go.

The downside is that you cannot control what is saved. is a free service that helps you out. To use it, paste a web address into the form on the services main page and hit submit url afterwards.

The service takes two snapshots of that page at that point in time and makes it available publicly.

The first takes a static snapshot of the site. You find images, text and other static contents included while dynamic contents and scripts are not.

The second snapshot takes a screenshot of the page instead.

An option to download the data is provided. Note that this downloads the textual copy of the site only and not the screenshot.

A Firefox add-on has been created for the service which may be useful to some of its users. It creates automatic snapshots of every web page that you bookmark in the web browser after installation of the add-on.

Word of warning: All snapshots are publicly available. While pages that require authentication cannot be saved by the service, it may still take snapshots of pages that you may not want to reveal to the public.

An option to password protect snapshots or protect them using accounts would certainly be useful in this regard.

The service can prove useful in other situations. For instance, if you cannot access a resource on the Internet, then you may still access it by using instead. While that provides access to text and image information only, it should be sufficient in most cases.

Closing Words is a useful but specialized service. It works well right out of the box but would benefit from protective features or an optional account system. All in all though, it can be quite handy at times to save web page information permanently in another location on the Internet.

Review Date
Reviewed Item
Author Rating

Tutorials & Tips

Previous Post: «
Next Post: «


  1. Anonymous said on June 21, 2022 at 2:10 pm

    look wikipedia article , as service has more alternative domains , including anonymous .onion domain

  2. rahiel said on July 28, 2015 at 2:50 pm

    That Firefox add-on is also available for Chrome: It recently got the ability to locally archive websites in MHTML format.

  3. Marc said on April 27, 2015 at 10:35 pm

    Can an entire domain sanpshot be downloaded out of the Wayback Machine? It seems you can’t. The WM contain sites that no longer exist I wish I could have them offline.
    Edit: and found the answer cool
    Edit2: for some reason I can’t view an offline but archived website

  4. juju said on April 25, 2015 at 9:28 am

    I wonder what happend to Coral Content Distribution Network ( It was also useful for short term archiving.

  5. interstellar said on April 24, 2015 at 4:57 am

    Thank you, Martin.
    I did not know about this useful resource.

    (Ghacks only site whitelisted here!).
    Support Ghacks! whitelist it…

  6. KK said on April 23, 2015 at 2:51 am

    You can also enable native MHTML whole page archiving in Chrome.


    Then scroll down to:
    Save Page as MHTML Mac, Windows, Linux
    Enables saving pages as MHTML: a single text file containing HTML and all sub-resources.

    Click “Enable” and restart the browser.

  7. loki said on April 23, 2015 at 1:59 am

    Another good offline alternative for the Firefox users is the Scrapbook add-on:

  8. fred said on April 23, 2015 at 12:11 am

    I’m curious to hear whether, after one has asked or other 3rd-party site to make an archived copy… is the content reflected in the archived copy “admissable as evidence” during legal proceedings? Would a judge instruct a court clerk to retrieve content from the cited archive URL to serve as an evidentiary copy? Is there yet a precedent for this (either way — in favor of, or denying, admissability)

  9. Falck said on April 22, 2015 at 11:37 pm

    >The downside is that you cannot control what is saved.
    That’s not entirely true:
    There’s even a bookmarklet and Firefox addon.

  10. Eli said on April 22, 2015 at 8:20 pm

    “Tip: Firefox users hit Shift-F2, type screenshot and then enter to create a screenshot of the active web page. Chrome users can save web pages as PDF documents natively instead.”

    This only takes the screenshot of your view so…

    Bonus Tip: If you instead type “screenshot –fullpage” without the quotes it will take a screenshot of the entire page, useful for long pages with a scroll bar.

  11. mike said on April 22, 2015 at 7:28 pm

    Guys, just make sure you don’t “abuse” the service to link to news sources which are deemed unworthy of pageviews / traffic. Please use or instead, where archival isn’t the main purpose :)

    1. Zeus said on April 23, 2015 at 2:43 am

      I think that practice is so petty. Either the source isn’t worth linking to in the first place, or you feel people should read the full text of an article, in which case the authors deserve their ad revenue.

      It’s like the difference between summarizing a bad movie’s plot to warn people against watching it, and pirating the film and distributing it to everyone you know just so you can all nod and agree it’s not worth the money.

Leave a Reply

Check the box to consent to your data being stored in line with the guidelines set out in our privacy policy

We love comments and welcome thoughtful and civilized discussion. Rudeness and personal attacks will not be tolerated. Please stay on-topic.
Please note that your comment may not appear immediately after you post it.