# coding=utf-8
from mod_python import apache
import hashlib, json, os, re, time, urllib, urllib2, urlparse
from bs4 import BeautifulSoup as bs4
# value for element, path of application relative to server root
BASE_PATH = '/'
# relative path to cache folder (relative to this file)
CACHE_PATH = '../cache'
# Cezar base URL
CEZAR_URL = 'http://msc.com.pl/cezar/'
# cache expiry (oldest valid cache timestamp)
CACHE_EXPIRY_LIMIT = int(time.time()) - 24*60*60
__dir__ = os.path.dirname(__file__)
# retrieves remote URL content, forwarding browser's UAS
def fetch_with_user_agent_spoof(cache_path, remote_url, user_agent):
opener = urllib2.build_opener()
opener.addheaders = [('User-Agent', user_agent)]
egg = None
for trigger, egg_actions in eggs.iteritems():
if cache_path.split('/')[-1].decode('utf-8').lower() == trigger.decode('utf-8'):
egg = egg_actions
break
if egg is not None:
remote_url = urlparse.urljoin(CEZAR_URL,
'?' + urllib.urlencode({'pid_search': egg['pid'], 'p': 21}))
content = opener.open(remote_url).read()
if egg is not None:
if 'replacements' in egg:
for replacement in egg['replacements']:
content = re.sub(replacement['from'], replacement['to'], content)
open(cache_path, 'w+').write(content)
# returns content of cached file, refreshing cache if necessary
def get_cache_content(cache_key, remote_url, force_refresh=False, user_agent=''):
cache_path = os.path.join(__dir__, CACHE_PATH, cache_key)
if force_refresh or not (os.path.exists(cache_path) and (int(os.path.getmtime(cache_path)) > CACHE_EXPIRY_LIMIT)):
fetch_with_user_agent_spoof(cache_path, remote_url, user_agent)
return open(cache_path, 'r').read()
def handler(req):
# MIME type fix for error messages
req.content_type = 'text/plain'
# we need to recover original request path, from before rewrite
orig_req = req
while True:
if orig_req.prev:
orig_req = orig_req.prev
else:
break
path = filter(None, re.sub('index\.py$', '', re.sub('^' + BASE_PATH, '', orig_req.uri)).split('/'))
if path:
# /[ANYTHING]/refresh forces cache refresh
no_cache = len(path) > 1 and path[1] == 'refresh'
user_agent = orig_req.headers_in['User-Agent']
# compiling remote original URI
search_url = urlparse.urljoin(CEZAR_URL,
'?' + urllib.urlencode(
{
'pid_search': path[0],
'p': 21
}
))
# fetching content
remote_content = get_cache_content(cache_key=path[0],
remote_url=search_url,
force_refresh=no_cache,
user_agent=user_agent).split('\n')
# slicing only interesting part of remote content
delimiter_regex = re.compile('---- page content ')
delimiters = [i for i, line in enumerate(remote_content) if re.search(delimiter_regex, line)]
if len(delimiters) < 2:
req.write('Malformed (even more than usually) content :(')
return apache.OK
# we need to sanitize line breaks and double-closed anchors manually
remote_content = bs4('
' + ' '.join(remote_content[delimiters[0]+1:delimiters[1]]).replace('
', '
').replace('', '') + '
', 'html.parser')
# stub template for output page
page_content = bs4('''
''', 'html.parser')
page_content.html.body.append(remote_content.div)
# internal links being rewritten to application links
for link in page_content.select('a[href^="?p=21&pid="]'):
link['href'] = urlparse.parse_qs(link['href'])['pid']
# garbage elements removed
garbage_selectors = ['script', 'table.msc_noprint', 'center > p']
for selector in garbage_selectors:
for garbage in page_content.select(selector):
garbage.extract()
# unnecessary tables removed
for table in page_content.select('table > tr > td > table')[4:]:
table.extract()
page_content.select('table > tr > td')[-1].extract()
# internal link targets (team, WK breakdown, etc.) removed
for internal_link in page_content.select('table > tr > td > table a'):
internal_link.attrs = {}
# internal link icons removed
for link_icon in page_content.select('img[src*="ico_link_8.gif"]'):
link_icon.extract()
# fetching all external pictures (not pic/* images) to local cache
for external_image in [image for image in page_content.select('img') if not image['src'].startswith('pic/')]:
image_url = urlparse.urljoin(CEZAR_URL, external_image['src'])
image_cache_path = 'foto/' + hashlib.md5(image_url).hexdigest() + '.' + image_url.split('.')[-1]
if not os.path.exists(image_cache_path) or no_cache:
fetch_with_user_agent_spoof(os.path.join(__dir__, image_cache_path), image_url, user_agent)
external_image['src'] = image_cache_path
# linking to original page
original_link_selectors = ['img[src^="foto/"]', 'span[style*=":28px"]']
for selector in original_link_selectors:
for element in page_content.select(selector):
link = element.wrap(page_content.new_tag('a'))
link['href'] = search_url
# credits info
credits_div = bs4('Pomysł Ivana,
Ivan jest zajebisty.
', 'html.parser', from_encoding='utf-8')
page_content.html.body.div.append(credits_div.div)
req.content_type = 'text/html'
req.write(page_content.prettify('utf-8'))
else:
req.write('Nothing to see here, move along.')
return apache.OK
eggs = {
'wąsłowicz': {
'pid': 13650,
'replacements': [
{ 'from': re.compile(r'src="\.\./cezar1/fots.*?"'), 'to': 'src="pic/egg.jpg"' },
{ 'from': 'asłowicz', 'to': 'ąsłowicz' },
{ 'from': 'ważną licencją', 'to': 'gwiazdką' }
]
},
'bubu': {
'pid': 1318
}
}