Loading old pages

Author Message
Rey 12/09/2014 03:24 pm
Hi,

Is it possible to load an old page of a project?

I have several projects that are scheduled to download new/modified files nightly, and was just wondering what happens when a website takes content off and I end up downloading a new page without the old links.

Please advise.

Thanks
Oleg Chernavin 12/10/2014 12:25 pm
If the links to old pages disappear, these old pages will be still on your hard disk.

Best regards,
Oleg Chernavin
MP Staff
Rey 12/11/2014 12:40 pm
Thanks for the reply.

I'm more concerned when pages get updated. How can I access the previous version of a page?

For instance, a new index.html now redirects to a new site, instead of listing links as before. I'd like to load the old index.html, if possible.

Thanks

Oleg Chernavin 12/11/2014 02:23 pm
You may use the Project Properties dialog - File Copies section to keep old versions when the update is downloaded.

Oleg.
Bane 03/21/2015 06:11 pm
"Project Properties dialog: Advanced - File Copies"

I'd like to know:

1) OEP somehow determined there were changes to a website page(s), and it downloaded only new versions (I think). Is there a way to determine what changed between those versions? If not that would be a good feature. I think a page that displays the current date/time would always make the page look like it had been changed.

2) Is there a way to view a prior website version in its entirety, with older pages displayed instead of the newer versions? Currently we have to view a single old page at a time, and links don't work.
Oleg Chernavin 03/21/2015 06:15 pm
1. Yes, we had an idea for such feature, but not many people asked for it. Also, it would be hard to properly visualize the changes list.

2. Also a rarely asked feature. You may simply use a macro to save a site to a different folder on every download:

DD=c:\download\{:year}{:0month}{:0day}\

Oleg.
Bane 03/21/2015 06:34 pm
Yes, we could save an entire site to multiple folders. But, that means ALL files would be downloaded, whether they had changed or not. That includes videos, PDFs, etc.

I forgot to mention. This is a WIX site. WIX makes extensive use of AJAX.

OEP recognizes the URLs but apparently doesn't capture the pages.

The only way I know to download all the pages in the site is to provide the SEO URL of *each* page.

The pages aren't formatted properly, and many of the linked-to URLs aren't resolved, but at least I can capture the main text of the website.

I used the AJAX SEO format for each page. The format is:
-- current version:
http://www.WebSiteName.com/?_escaped_fragment_=page-name-A/xy11
http://www.WebSiteName.com/?_escaped_fragment_=page-name-B/fg23
-- prior versions:
http://www.WebSiteName.com/?_escaped_fragment_=page-name-A/xy11_YYMMDDNN

The Google writeup for SEO pages is:
https://developers.google.com/webmasters/ajax-crawling/

It would be great for OEP to generate a set of static HTML that will format the pages properly. It would be nice if it functioned properly, but that's too much to ask for.

A guide to get HTML snapshots of each page is:
https://developers.google.com/webmasters/ajax-crawling/docs/html-snapshot
Oleg Chernavin 03/24/2015 06:31 pm
Is it possible to look at the real site to understand it better? You may send it to support@metaproducts.com.

Thank you!

Oleg.