22from project .http .requests .parsers .proxyforeuParser import proxyforeuParser
33from project .http .requests .parsers .rebroweeblyParser import rebroweeblyParser
44from project .http .requests .parsers .samairproxyParser import semairproxyParser
5-
6- __author__ = 'pgaref'
7-
85import requests
96from requests .exceptions import ConnectionError
107import random
11- import os
128import time
13- from bs4 import BeautifulSoup
149from requests .exceptions import ReadTimeout
1510
11+ __author__ = 'pgaref'
1612
1713class RequestProxy :
1814 agent_file = '../data/user_agents.txt'
1915
2016 def __init__ (self , web_proxy_list = []):
2117 self .useragents = self .load_user_agents (RequestProxy .agent_file )
18+
2219 #####
2320 # Each of the classes below implements a specific URL Parser
24- # http://<USERNAME>:<PASSWORD>@<IP-ADDR>:<PORT>
2521 #####
2622 parsers = []
2723 parsers .append (freeproxyParser ('http://free-proxy-list.net' ))
@@ -83,24 +79,24 @@ def generate_proxied_request(self, url, params={}, req_timeout=30):
8379 request = None
8480 try :
8581 rand_proxy = random .choice (self .proxy_list )
86- print "Using proxy: " + str (rand_proxy )
82+ print "Using proxy: {0}" . format ( str (rand_proxy ) )
8783 request = requests .get (test_url , proxies = {"http" : rand_proxy },
8884 headers = req_headers , timeout = req_timeout )
8985 except ConnectionError :
9086 self .proxy_list .remove (rand_proxy )
91- print "Proxy unreachable - Removed Straggling proxy :" , rand_proxy , " PL Size = " , len (self .proxy_list )
87+ print "Proxy unreachable - Removed Straggling proxy: {0} PL Size = {1}" . format ( rand_proxy , len (self .proxy_list ) )
9288 pass
9389 except ReadTimeout :
9490 self .proxy_list .remove (rand_proxy )
95- print "Read timed out - Removed Straggling proxy :" , rand_proxy , " PL Size = " , len (self .proxy_list )
91+ print "Read timed out - Removed Straggling proxy: {0} PL Size = {1}" . format ( rand_proxy , len (self .proxy_list ) )
9692 pass
9793 return request
9894
9995if __name__ == '__main__' :
10096
10197 start = time .time ()
10298 req_proxy = RequestProxy ()
103- print "Initialization took: " , ( time .time ()- start )
99+ print "Initialization took: {0} sec" . format (( time .time ()- start ) )
104100 print "Size : " , len (req_proxy .get_proxy_list ())
105101 print " ALL = " , req_proxy .get_proxy_list ()
106102
@@ -109,7 +105,7 @@ def generate_proxied_request(self, url, params={}, req_timeout=30):
109105 while True :
110106 start = time .time ()
111107 request = req_proxy .generate_proxied_request (test_url )
112- print "Proxied Request Took: " , ( time .time ()- start ), " => Status: " , request .__str__ ()
108+ print "Proxied Request Took: {0} sec => Status: {1}" . format (( time .time ()- start ), request .__str__ () )
113109 print "Proxy List Size: " , len (req_proxy .get_proxy_list ())
114110
115111 print "-> Going to sleep.."
0 commit comments