I want to be able to get the list of all URLs that a browser will do a GET request for when we try to open a page. For eg: if we try to open cnn.com, there are multiple URLs within the first HTTP response which the browser recursively requests for.
I'm not trying to render a page but I'm trying to obtain a list of all the urls that are requested when a page is rendered. Doing a simple scan of the http response content wouldn't be sufficient as there could potentially be images in the css which are downloaded. Is there anyway I can do this in python?
It's likely that you'll have to render the page (not necessarily display it though) to be sure you're getting a complete list of all resources. I've used
PyQT
andQtWebKit
in similar situations. Especially when you start counting resources included dynamically with javascript, trying to parse and load pages recursively withBeautifulSoup
just isn't going to work.Ghost.py is an excellent client to get you started with PyQT. Also, check out the QWebView docs and the QNetworkAccessManager docs.
Ghost.py returns a tuple of (page, resources) when opening a page:
resources
includes all of the resources loaded by the original URL as HttpResource objects. You can retrieve the URL for a loaded resource withresource.url
.I guess you will have to create a list of all known file extensions that you do NOT want, and then scan the content of the http response, checking with "if substring not in nono-list:"
The problem is all href's ending with TLDs, forwardslashes, url-delivered variables and so on, so i think it would be easier to check for stuff you know you dont want.