We're in the process of shutting down The Conversations Network (including the IT Conversations podcast). The plan is to render a static-HTML version of our websites for permanent hosting at the Internet Archive.
What's the easiest way to generate static HTML from the roughly 5,000 dynamic pages currently generated dynamically from PHP?
I know we could tweak the code to cache the PHP output, write it to files, then walk the sitemaps to generate every page. But I wonder if there are any options we should consider. Any tools for doing this and scraping the HTML as-is? (Something other than Acrobat Pro?)
Unfortunately, we also have a fair number of Ajax calls, which are going to make this more difficult. I imagine we'll have to un-Ajax them first.