First, we will scrape HTML links from Wikipedia.org. This is permitted by Wikipedia's GPL license, and this demonstration is fair use. Here we see code that downloads the English Wikipedia page.
Note It opens Wikipedia and downloads the content at the specified URL. Part 2 uses my special code to loop over each link and its text.
Example 2. Here I show a simple class that receives the HTML string and then extracts all the links and their text into structs. It is fairly fast, but I offer some optimization tips further down. It would be better to use a class.
MatchCollection This example first finds all hyperlink tags. We store all the complete A tags into a MatchCollection.
Step 2 The code loops over all hyperlink tag strings. In the algorithm, the next part examines all the text of the A tags.
HREF This attribute points to other web resources. This part is not failsafe, but almost always works.
Returns The method returns the List of LinkItem objects. This list can then be used in the foreach-loop from the first C# example.
C# program that scrapes HTML
static void Main()
// 1.// URL: http://en.wikipedia.org/wiki/Main_PageWebClient w = new WebClient();
string s = w.DownloadString("http://en.wikipedia.org/wiki/Main_Page");
// 2.foreach (LinkItem i in LinkFinder.Find(s))
Tests. My first two attempts at this code were incorrect and had unacceptable bugs, but the version shown here works. You need to use RegexOptions.SingleLine. The dot in a Regex matches all characters except a newline unless this is specified.
Tip To match multiline links, we require RegexOptions.Singleline. This is an important option.
C# program that scrapes with Regex
public struct LinkItem
public string Href;
public string Text;
public override string ToString()
return Href + "\n\t" + Text;
static class LinkFinder
public static List<LinkItem> Find(string file)
List<LinkItem> list = new List<LinkItem>();
// Find all matches in file.MatchCollection m1 = Regex.Matches(file, @"(<a.*?>.*?</a>)",
// Loop over each match.foreach (Match m in m1)
string value = m.Groups.Value;
LinkItem i = new LinkItem();
// Get href attribute.
Match m2 = Regex.Match(value, @"href=\""(.*?)\""",
i.Href = m2.Groups.Value;
// Remove inner tags from text.
string t = Regex.Replace(value, @"\s*<.*?>\s*", "",
i.Text = t;
Test the program on your website. It prints out matches to the console. Here we see part of the current results for the Wikipedia home page. The original HTML shows where the links were extracted. They are contained in a LI tag.
Note You will see my program successfully extracted the anchor text and also the HREF value.
anyone can edit
<li><a href="/wiki/Portal:Arts" title="Portal:Arts">Arts</a></li>
<li><a href="/wiki/Portal:Biography" title="Portal:Biography">Biography</a></li>
<li><a href="/wiki/Portal:Geography" title="Portal:Geography">Geography</a></li>
SingleLine. SingleLine is an important option. Microsoft states that SingleLine "Specifies single-line mode. Changes the meaning of the dot so it matches every character (instead of every character except \n)."
Performance. You can improve performance of the regular expressions by specifying RegexOptions.Compiled, and also use instance Regex objects, not the static methods I show. Normally, your Internet connection will be the bottleneck.
Summary. We scraped HTML content from the Internet. The code is more flexible than some other approaches. Using three regular expressions, you can extract HTML links into objects with a fair degree of accuracy.