Creating Code For Simplicity but Logic For Need
In this article we will learn how to extract all the links from a webpage using a web client. At the end of this article you will be able to create an application that can extract links from pages and filter those links on the basis of parameters you want. So without wasting much time let’s dive directly into the code.
Creating the Link Grabber
So we are creating a link grabber. For that we need some logic and it’s always a good idea to clarify the logic before creating something. So let’s define the logic.
The logic is:
Before proceeding let’s code the preceding logic.
The following is the code for the grabber.
public partial class Form1 : Form
private void button1_Click(object sender, EventArgs e)
WebBrowser wb = new WebBrowser();
wb.Url = new Uri(textBox1.Text);
wb.DocumentCompleted += wb_DocumentCompleted;
void wb_DocumentCompleted(object sender, WebBrowserDocumentCompletedEventArgs e)
HtmlDocument source = ((WebBrowser)sender).Document;
private void extractLink(HtmlDocument source)
HtmlElementCollection anchorList = source.GetElementsByTagName(“a“);
foreach (var item in anchorList)
That’s it; all done. Now you have successfully made a link grabber. You can further extend it by adding a filter to it. In my next part I will show how to add a filter and how to download files. Thanks for reading and don’t forget to comment and share.