Simple Python program stuck
Yesterday I wrote a simple Python program (really simple as shown below) to check HTTP status responses for about 5000 URLs. The thing is, the program seems to be stuck for every 400-500 URLs. Since I am really new to programming, I have no idea how to track down the problem.
I added the "a = a + 1" snippet to keep track of how much URL was processed when it got stuck.
How can I find the problem? Thank you very much!
I am using Ubuntu 11.10 and Python 2.7
#!/usr/bin/env python # -*- coding: utf-8 -*- import httplib raw_url_list = open ('url.txt', 'r') url_list = raw_url_list.readlines() result_file = open('result.txt', 'w') a = 0 for url in url_list: url = url.strip()[23:] conn = httplib.HTTPConnection('www.123456789.cn') conn.request('HEAD', url) res = conn.getresponse() result_file.write('http://www.123456789.cn%s, %s, %s \n' % (url, res.status, res.reason)) a = a + 1 print a raw_url_list.close() result_file.close()
source to share