I want to download all accessible html files under www.site.com/en/. However, there are a lot of linked URLS with post parameters on the site (e.g. pages 1,2,3.. for each product category). I want wget NOT to download these links. I'm using
-R "*\?*"
But it's not perfect because it only removes the file after downloading it.
Is there some way for example to filter the links followed by wget with a regex?