2015-06-01 07:15:50 +00:00
|
|
|
"""
|
2015-06-01 10:30:07 +00:00
|
|
|
Swisscows (Web, Images)
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
@website https://swisscows.ch
|
|
|
|
@provide-api no
|
|
|
|
|
|
|
|
@using-api no
|
|
|
|
@results HTML (using search portal)
|
|
|
|
@stable no (HTML can change)
|
|
|
|
@parse url, title, content
|
|
|
|
"""
|
|
|
|
|
|
|
|
from json import loads
|
|
|
|
import re
|
2016-11-06 02:51:38 +00:00
|
|
|
from lxml.html import fromstring
|
2016-11-30 17:43:03 +00:00
|
|
|
from searx.url_utils import unquote, urlencode
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
# engine dependent config
|
2015-06-01 10:30:07 +00:00
|
|
|
categories = ['general', 'images']
|
2015-06-01 07:15:50 +00:00
|
|
|
paging = True
|
|
|
|
language_support = True
|
|
|
|
|
|
|
|
# search-url
|
|
|
|
base_url = 'https://swisscows.ch/'
|
|
|
|
search_string = '?{query}&page={page}'
|
|
|
|
|
2016-12-15 06:34:43 +00:00
|
|
|
supported_languages_url = base_url
|
|
|
|
|
2015-06-01 07:15:50 +00:00
|
|
|
# regex
|
2016-11-30 17:43:03 +00:00
|
|
|
regex_json = re.compile(b'initialData: {"Request":(.|\n)*},\s*environment')
|
|
|
|
regex_json_remove_start = re.compile(b'^initialData:\s*')
|
|
|
|
regex_json_remove_end = re.compile(b',\s*environment$')
|
|
|
|
regex_img_url_remove_start = re.compile(b'^https?://i\.swisscows\.ch/\?link=')
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
|
|
|
|
# do search-request
|
|
|
|
def request(query, params):
|
2017-07-20 20:47:20 +00:00
|
|
|
if params['language'].split('-')[0] == 'no':
|
2016-10-30 02:04:01 +00:00
|
|
|
region = 'nb-NO'
|
2015-06-01 07:15:50 +00:00
|
|
|
else:
|
2016-08-06 04:34:56 +00:00
|
|
|
region = params['language']
|
|
|
|
ui_language = params['language'].split('-')[0]
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
search_path = search_string.format(
|
2016-11-30 17:43:03 +00:00
|
|
|
query=urlencode({'query': query, 'uiLanguage': ui_language, 'region': region}),
|
|
|
|
page=params['pageno']
|
|
|
|
)
|
2015-06-01 07:15:50 +00:00
|
|
|
|
2015-06-01 10:30:07 +00:00
|
|
|
# image search query is something like 'image?{query}&page={page}'
|
|
|
|
if params['category'] == 'images':
|
|
|
|
search_path = 'image' + search_path
|
|
|
|
|
2015-06-01 07:15:50 +00:00
|
|
|
params['url'] = base_url + search_path
|
|
|
|
|
|
|
|
return params
|
|
|
|
|
|
|
|
|
|
|
|
# get response from search-request
|
|
|
|
def response(resp):
|
|
|
|
results = []
|
|
|
|
|
2016-11-30 17:43:03 +00:00
|
|
|
json_regex = regex_json.search(resp.text)
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
# check if results are returned
|
|
|
|
if not json_regex:
|
|
|
|
return []
|
|
|
|
|
2016-11-30 17:43:03 +00:00
|
|
|
json_raw = regex_json_remove_end.sub(b'', regex_json_remove_start.sub(b'', json_regex.group()))
|
|
|
|
json = loads(json_raw.decode('utf-8'))
|
2015-06-01 07:15:50 +00:00
|
|
|
|
2015-06-01 10:30:07 +00:00
|
|
|
# parse results
|
2015-06-01 07:15:50 +00:00
|
|
|
for result in json['Results'].get('items', []):
|
2015-06-01 10:30:07 +00:00
|
|
|
result_title = result['Title'].replace(u'\uE000', '').replace(u'\uE001', '')
|
|
|
|
|
|
|
|
# parse image results
|
|
|
|
if result.get('ContentType', '').startswith('image'):
|
2016-11-30 17:43:03 +00:00
|
|
|
img_url = unquote(regex_img_url_remove_start.sub(b'', result['Url'].encode('utf-8')).decode('utf-8'))
|
2015-06-01 10:30:07 +00:00
|
|
|
|
|
|
|
# append result
|
|
|
|
results.append({'url': result['SourceUrl'],
|
2016-12-09 10:44:24 +00:00
|
|
|
'title': result['Title'],
|
2015-06-01 10:30:07 +00:00
|
|
|
'content': '',
|
|
|
|
'img_src': img_url,
|
|
|
|
'template': 'images.html'})
|
|
|
|
|
|
|
|
# parse general results
|
|
|
|
else:
|
|
|
|
result_url = result['Url'].replace(u'\uE000', '').replace(u'\uE001', '')
|
|
|
|
result_content = result['Description'].replace(u'\uE000', '').replace(u'\uE001', '')
|
|
|
|
|
|
|
|
# append result
|
|
|
|
results.append({'url': result_url,
|
2016-12-09 10:44:24 +00:00
|
|
|
'title': result_title,
|
|
|
|
'content': result_content})
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
# parse images
|
|
|
|
for result in json.get('Images', []):
|
|
|
|
# decode image url
|
2016-11-30 17:43:03 +00:00
|
|
|
img_url = unquote(regex_img_url_remove_start.sub(b'', result['Url'].encode('utf-8')).decode('utf-8'))
|
2015-06-01 07:15:50 +00:00
|
|
|
|
|
|
|
# append result
|
|
|
|
results.append({'url': result['SourceUrl'],
|
2016-12-09 10:44:24 +00:00
|
|
|
'title': result['Title'],
|
2015-06-01 07:15:50 +00:00
|
|
|
'content': '',
|
|
|
|
'img_src': img_url,
|
|
|
|
'template': 'images.html'})
|
|
|
|
|
|
|
|
# return results
|
|
|
|
return results
|
2016-11-06 02:51:38 +00:00
|
|
|
|
|
|
|
|
|
|
|
# get supported languages from their site
|
2016-12-15 06:34:43 +00:00
|
|
|
def _fetch_supported_languages(resp):
|
2016-11-06 02:51:38 +00:00
|
|
|
supported_languages = []
|
2016-12-15 06:34:43 +00:00
|
|
|
dom = fromstring(resp.text)
|
2016-11-06 02:51:38 +00:00
|
|
|
options = dom.xpath('//div[@id="regions-popup"]//ul/li/a')
|
|
|
|
for option in options:
|
2017-10-10 21:52:41 +00:00
|
|
|
code = option.xpath('./@data-search-language')[0]
|
2017-03-01 23:11:51 +00:00
|
|
|
if code.startswith('nb-'):
|
|
|
|
code = code.replace('nb', 'no', 1)
|
2016-11-06 02:51:38 +00:00
|
|
|
supported_languages.append(code)
|
|
|
|
|
|
|
|
return supported_languages
|