Page 1 of 1
How to get a list of links in a web page ?
Posted: Mon Oct 27, 2003 7:54 pm
by fonkfonk
Hello,
I'm planing to create a prog that would allow to generate a list of links in a webpage.
I assume I would have to download the page and then parse the <a href=""></a> lines.
Is there a way to recursevely read all the links "on line" (I found nothing for this particular purpose)...
Thanks for any idea.
Pierre
Posted: Mon Oct 27, 2003 9:40 pm
by plouf
there are some com ways DCOM ithink but dunno more
personally i believe the sipmpler way will be to downloads the whole page as .html and repeatively seek for "href=" reference
Posted: Mon Oct 27, 2003 11:26 pm
by Max.
Maybe the source code of Freak's IETool can give you a hint:
viewtopic.php?t=2698
It's an extension to the MS Internet Explorer which allows you to compile selected text of a HTML pages with PureBasic.
Next I'd check for is FloHimself's Regular Expression Library; ideal to parse for links with an expression like
Don't have a link for Flo's lib handy, but if you search for Regular Expression, you should come up with a result quickly.
Posted: Tue Oct 28, 2003 1:49 pm
by freak
> Maybe the source code of Freak's IETool can give you a hint:
my tool only simulates CTRL+C to copy the text, you can't read all the
contents of the page like that.
Timo
Posted: Tue Oct 28, 2003 5:09 pm
by fonkfonk
Well, it seems that plouf and I agree on the same idea ...
And this is more easy for me ! :roll:
Thanks all for taking the time to find a solution !
Regards,
Pierre