How to get all image, pdf and other files links from a website?

I am working on an application where I have to

1) get all the links of website
2) and then get the list of all the files and file extensions in each
of the web page/link.

I am done with the first part of it :slight_smile:
now I have to get the all the files/file-extensions in each of the
page.

Can anybody guide me how to parse the links/webpage and get the file-
extensions in the page?

Is it me or has this particular homework question turned up a few times already?

Hint: This has been asked and answered before quite recently
(yesterday even) so try reading the mailing list.