I would like to setup the crawler to crawl a website, let say blog, and fetch me only the links in the website and paste the links inside a text file. Can you guide me step by step for setup the crawler? I am using Eclipse.
Asked
Active
Viewed 2,729 times
1 Answers
0
Jsoup will do everything you need wrt html parsing. Jsoup is a java api for handling html source code. You can get
- Table, with which you can parse each and every row or column.
- List of all the links and source imports to that html(imports like css and js files).
- Data of particular tag.
and more.
For your purpose, here is the example code.
Hope this will help you.

Ahamed
- 39,245
- 13
- 40
- 68