The way to solve something like this is to get down to the basics.
Your code seems to run OK up step 30.
the URL for the first request is:
if you try that url by itself (in browser), it brings you to a wayback machine error page:
If you can find the actual url, then you can go from there (use dead-eye's code)
NOTE: a session is a good idea, but not strictly needed to download zip files, I do it all the time.
Your code seems to run OK up step 30.
the URL for the first request is:
Output:http://web.archive.org/web/20031124231537/http://www.projectai.com:80/libraries/download.php?fileid={3810}
so try that by utself with requests:import requests url = 'http://web.archive.org/web/20031124231537/http://www.projectai.com:80/libraries/download.php?fileid={3810}' response = requests.get(url) print('status code: {}'.format(response.status_code)) if response.status_code == 200: print('saving page') with open('results.html', 'wb') as fp: fp.write(response.content)it returns a 404 error which is:
Quote:404 Not Found
The requested resource could not be found but may be available in the future. Subsequent requests by the client are permissible.
if you try that url by itself (in browser), it brings you to a wayback machine error page:
Output:Hrm.
The Wayback Machine has not archived that URL.
This page is not available on the web
because page does not exist
Try it!If you can find the actual url, then you can go from there (use dead-eye's code)
NOTE: a session is a good idea, but not strictly needed to download zip files, I do it all the time.