The professional, friendly Java community. 21,500 members and growing!

The Java Programming Forums are a community of Java programmers from all around the World. Our members have a wide range of skills and they all have one thing in common: A passion to learn and code Java. We invite beginner Java programmers right through to Java professionals to post here and share your knowledge. Become a part of the community, help others, expand your knowledge of Java and enjoy talking with like minded people. Registration is quick and best of all free. We look forward to meeting you.

Crawling through a site: simulating a browser as close as possible

Hi guys!
I'm creating a piece of software that basically crawls through several pages getting ALL the links on the page.
You may say "yeah, that's easy... you can download the page's source code via an httpClient and then use HtmlParser, JerichoParser, etc.", but the problem is that those are decent parsers for HTML links but VERY SLOW parsers for links on JavaScript scripts and other technologies.
That's the reason of my title. I need to crawl through ALL of the links on a page, but on a decent speed... JUST as a browser does.

My questions are:
1. Recommendations on a parser to do this job? I already tried with JerichoParser, and it is pretty slow (the JS part)
2. Which variety of links and technologies will I encounter (besides HTML and JS)? I need a parser that handles all of them on an efficient (and easy!) way.