NEED MOAR PROXIES?!?

Look no more! This automatic proxy scraper takes in a file populated with proxy sites, scrapes them, and pumps out a list of usable proxies.

Using a simple regular expressions (searches for ip:port) we can scrape all the proxy sites from a given URL. Example output:

./proxyScraper.pl proxy_list 213.85.92.10:80 221.176.14.72:80 202.108.50.75:81 120.192.249.74:80 202.106.169.142:80 111.206.81.248:80 ...[snip]...

We can easily feed this list into any proxy accepting software; Usually, I use the output to feed into proxychains, and use that to push any scripts/software through the newly acquired proxies.

The code is simple and uses basic Perl modules to crawl the given pages from a file, and prints out a list of proxies:

#!/usr/bin/perl # Usage: proxyScraper.pl [proxy_file] use strict; use warnings; use WWW::Mechanize; use Try::Tiny; my $source_file=shift; open (INPUT_FILE, "<$source_file") || die "Can't open $source_file: $!

"; my @sources = ; my $crawler = WWW::Mechanize->new(); foreach (@sources) { try { $crawler->get($_); # hunt for IP:PORT combination my @ips= $crawler->text() =~ /(\d{1,3}\.\d{1,3}\.\d{1,3}\.\d{1,3}\:\d{1,5})/g; foreach (@ips){ print "$_

"; } } catch { warn "[!] Error, who cares

"; } }

To start people off, here is subset of the proxy_list file I use with the script:

http://aliveproxy.com/fastest-proxies/ http://multiproxy.org/anon_proxy.htm http://www.ultraproxies.com/high-anonymous.html http://www.atomintersoft.com/high_anonymity_elite_proxy_list http://www.proxylists.net/ http://www.ip-adress.com/proxy_list/ http://www.dailiip.com/ http://www.samair.ru/proxy/ip-address-01.htm

Enjoy and happy crawling!