signal.signal(signal.SIGALRM, alarmalarm)
signal.alarm(300)
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(control_url + "/max_req_no", headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=control_url + "/max_req_no", headers=headers)
+ f = urllib2.urlopen(req)
count = int(string.strip(f.readline()))
signal.alarm(0)
except Exception, e:
socket.setdefaulttimeout(240)
signal.alarm(300)
try:
- headers = urllib.urlencode( { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' } )
- f = urllib2.urlopen(control_url + "/queue.gz", headers)
+ headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache' }
+ req = urllib2.Request(url=control_url + "/queue.gz", headers=headers)
+ f = urllib2.urlopen(req)
signal.alarm(0)
except Exception, e:
signal.alarm(0)
q.lock(0)
q.read()
for r in reqs:
- if r.kind != 'group':
+ if r.kind != 'group':
raise Exception, 'handle_reqs: fatal: huh? %s' % r.kind
need_it = 0
for b in r.batches:
if lck == None:
sys.exit(1)
init_conf()
-
+ acl.try_reload()
+
status.push("fetching requests")
if has_new(config.control_url):
q = fetch_queue(config.control_url)
max_no = 0
q_new = []
for r in q:
- if r.no > max_no:
+ if r.no > max_no:
max_no = r.no
if r.no > last_count:
q_new.append(r)
f.close()
status.pop()
lck.close()
-
+
if __name__ == '__main__':
- # http connection is established (and few bytes transferred through it)
+ # http connection is established (and few bytes transferred through it)
# each $secs seconds.
loop.run_loop(main, secs = 10)