mirror of https://github.com/searxng/searxng.git
[fix] wikidata engine: faster processing, remove one HTTP redirection.
* Search URL is https://www.wikidata.org/w/index.php?{query}&ns0=1 (with ns0=1 at the end to avoid an HTTP redirection) * url_detail: remove the disabletidy=1 deprecated parameter * Add eval_xpath function: compile once for all xpath. * Add get_id_cache: retrieve all HTML with an id, avoid the slow to procress dynamic xpath '//div[@id="{propertyid}"]'.replace('{propertyid}') * Create an etree.HTMLParser() instead of using the global one (see #1575)
This commit is contained in:
parent
cbc5e13275
commit
6e0285b2db
|
@ -20,6 +20,7 @@ from searx.utils import match_language
|
||||||
|
|
||||||
from json import loads
|
from json import loads
|
||||||
from lxml.html import fromstring
|
from lxml.html import fromstring
|
||||||
|
from lxml import etree
|
||||||
|
|
||||||
logger = logger.getChild('wikidata')
|
logger = logger.getChild('wikidata')
|
||||||
result_count = 1
|
result_count = 1
|
||||||
|
@ -27,23 +28,23 @@ result_count = 1
|
||||||
# urls
|
# urls
|
||||||
wikidata_host = 'https://www.wikidata.org'
|
wikidata_host = 'https://www.wikidata.org'
|
||||||
url_search = wikidata_host \
|
url_search = wikidata_host \
|
||||||
+ '/w/index.php?{query}'
|
+ '/w/index.php?{query}&ns0=1'
|
||||||
|
|
||||||
wikidata_api = wikidata_host + '/w/api.php'
|
wikidata_api = wikidata_host + '/w/api.php'
|
||||||
url_detail = wikidata_api\
|
url_detail = wikidata_api\
|
||||||
+ '?action=parse&format=json&{query}'\
|
+ '?action=parse&format=json&{query}'\
|
||||||
+ '&redirects=1&prop=text%7Cdisplaytitle%7Clanglinks%7Crevid'\
|
+ '&redirects=1&prop=text%7Cdisplaytitle%7Cparsewarnings'\
|
||||||
+ '&disableeditsection=1&disabletidy=1&preview=1§ionpreview=1&disabletoc=1&utf8=1&formatversion=2'
|
+ '&disableeditsection=1&preview=1§ionpreview=1&disabletoc=1&utf8=1&formatversion=2'
|
||||||
|
|
||||||
url_map = 'https://www.openstreetmap.org/'\
|
url_map = 'https://www.openstreetmap.org/'\
|
||||||
+ '?lat={latitude}&lon={longitude}&zoom={zoom}&layers=M'
|
+ '?lat={latitude}&lon={longitude}&zoom={zoom}&layers=M'
|
||||||
url_image = 'https://commons.wikimedia.org/wiki/Special:FilePath/{filename}?width=500&height=400'
|
url_image = 'https://commons.wikimedia.org/wiki/Special:FilePath/{filename}?width=500&height=400'
|
||||||
|
|
||||||
# xpaths
|
# xpaths
|
||||||
|
div_ids_xpath = '//div[@id]'
|
||||||
wikidata_ids_xpath = '//ul[@class="mw-search-results"]/li//a/@href'
|
wikidata_ids_xpath = '//ul[@class="mw-search-results"]/li//a/@href'
|
||||||
title_xpath = '//*[contains(@class,"wikibase-title-label")]'
|
title_xpath = '//*[contains(@class,"wikibase-title-label")]'
|
||||||
description_xpath = '//div[contains(@class,"wikibase-entitytermsview-heading-description")]'
|
description_xpath = '//div[contains(@class,"wikibase-entitytermsview-heading-description")]'
|
||||||
property_xpath = '//div[@id="{propertyid}"]'
|
|
||||||
label_xpath = './/div[contains(@class,"wikibase-statementgroupview-property-label")]/a'
|
label_xpath = './/div[contains(@class,"wikibase-statementgroupview-property-label")]/a'
|
||||||
url_xpath = './/a[contains(@class,"external free") or contains(@class, "wb-external-id")]'
|
url_xpath = './/a[contains(@class,"external free") or contains(@class, "wb-external-id")]'
|
||||||
wikilink_xpath = './/ul[contains(@class,"wikibase-sitelinklistview-listview")]'\
|
wikilink_xpath = './/ul[contains(@class,"wikibase-sitelinklistview-listview")]'\
|
||||||
|
@ -56,6 +57,31 @@ language_fallback_xpath = '//sup[contains(@class,"wb-language-fallback-indicator
|
||||||
calendar_name_xpath = './/sup[contains(@class,"wb-calendar-name")]'
|
calendar_name_xpath = './/sup[contains(@class,"wb-calendar-name")]'
|
||||||
media_xpath = value_xpath + '//div[contains(@class,"commons-media-caption")]//a'
|
media_xpath = value_xpath + '//div[contains(@class,"commons-media-caption")]//a'
|
||||||
|
|
||||||
|
# xpath_cache
|
||||||
|
xpath_cache = {}
|
||||||
|
|
||||||
|
|
||||||
|
def get_xpath(xpath_str):
|
||||||
|
result = xpath_cache.get(xpath_str, None)
|
||||||
|
if not result:
|
||||||
|
result = etree.XPath(xpath_str)
|
||||||
|
xpath_cache[xpath_str] = result
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
def eval_xpath(element, xpath_str):
|
||||||
|
xpath = get_xpath(xpath_str)
|
||||||
|
return xpath(element)
|
||||||
|
|
||||||
|
|
||||||
|
def get_id_cache(result):
|
||||||
|
id_cache = {}
|
||||||
|
for e in eval_xpath(result, div_ids_xpath):
|
||||||
|
id = e.get('id')
|
||||||
|
if id.startswith('P'):
|
||||||
|
id_cache[id] = e
|
||||||
|
return id_cache
|
||||||
|
|
||||||
|
|
||||||
def request(query, params):
|
def request(query, params):
|
||||||
params['url'] = url_search.format(
|
params['url'] = url_search.format(
|
||||||
|
@ -65,8 +91,9 @@ def request(query, params):
|
||||||
|
|
||||||
def response(resp):
|
def response(resp):
|
||||||
results = []
|
results = []
|
||||||
html = fromstring(resp.text)
|
htmlparser = etree.HTMLParser()
|
||||||
search_results = html.xpath(wikidata_ids_xpath)
|
html = fromstring(resp.content.decode("utf-8"), parser=htmlparser)
|
||||||
|
search_results = eval_xpath(html, wikidata_ids_xpath)
|
||||||
|
|
||||||
if resp.search_params['language'].split('-')[0] == 'all':
|
if resp.search_params['language'].split('-')[0] == 'all':
|
||||||
language = 'en'
|
language = 'en'
|
||||||
|
@ -78,13 +105,13 @@ def response(resp):
|
||||||
wikidata_id = search_result.split('/')[-1]
|
wikidata_id = search_result.split('/')[-1]
|
||||||
url = url_detail.format(query=urlencode({'page': wikidata_id, 'uselang': language}))
|
url = url_detail.format(query=urlencode({'page': wikidata_id, 'uselang': language}))
|
||||||
htmlresponse = get(url)
|
htmlresponse = get(url)
|
||||||
jsonresponse = loads(htmlresponse.text)
|
jsonresponse = loads(htmlresponse.content.decode("utf-8"))
|
||||||
results += getDetail(jsonresponse, wikidata_id, language, resp.search_params['language'])
|
results += getDetail(jsonresponse, wikidata_id, language, resp.search_params['language'], htmlparser)
|
||||||
|
|
||||||
return results
|
return results
|
||||||
|
|
||||||
|
|
||||||
def getDetail(jsonresponse, wikidata_id, language, locale):
|
def getDetail(jsonresponse, wikidata_id, language, locale, htmlparser):
|
||||||
results = []
|
results = []
|
||||||
urls = []
|
urls = []
|
||||||
attributes = []
|
attributes = []
|
||||||
|
@ -95,21 +122,23 @@ def getDetail(jsonresponse, wikidata_id, language, locale):
|
||||||
if not title or not result:
|
if not title or not result:
|
||||||
return results
|
return results
|
||||||
|
|
||||||
title = fromstring(title)
|
title = fromstring(title, parser=htmlparser)
|
||||||
for elem in title.xpath(language_fallback_xpath):
|
for elem in eval_xpath(title, language_fallback_xpath):
|
||||||
elem.getparent().remove(elem)
|
elem.getparent().remove(elem)
|
||||||
title = extract_text(title.xpath(title_xpath))
|
title = extract_text(eval_xpath(title, title_xpath))
|
||||||
|
|
||||||
result = fromstring(result)
|
result = fromstring(result, parser=htmlparser)
|
||||||
for elem in result.xpath(language_fallback_xpath):
|
for elem in eval_xpath(result, language_fallback_xpath):
|
||||||
elem.getparent().remove(elem)
|
elem.getparent().remove(elem)
|
||||||
|
|
||||||
description = extract_text(result.xpath(description_xpath))
|
description = extract_text(eval_xpath(result, description_xpath))
|
||||||
|
|
||||||
|
id_cache = get_id_cache(result)
|
||||||
|
|
||||||
# URLS
|
# URLS
|
||||||
|
|
||||||
# official website
|
# official website
|
||||||
add_url(urls, result, 'P856', results=results)
|
add_url(urls, result, id_cache, 'P856', results=results)
|
||||||
|
|
||||||
# wikipedia
|
# wikipedia
|
||||||
wikipedia_link_count = 0
|
wikipedia_link_count = 0
|
||||||
|
@ -130,30 +159,30 @@ def getDetail(jsonresponse, wikidata_id, language, locale):
|
||||||
# if wikipedia_link_count == 0:
|
# if wikipedia_link_count == 0:
|
||||||
|
|
||||||
# more wikis
|
# more wikis
|
||||||
add_url(urls, result, default_label='Wikivoyage (' + language + ')', link_type=language + 'wikivoyage')
|
add_url(urls, result, id_cache, default_label='Wikivoyage (' + language + ')', link_type=language + 'wikivoyage')
|
||||||
add_url(urls, result, default_label='Wikiquote (' + language + ')', link_type=language + 'wikiquote')
|
add_url(urls, result, id_cache, default_label='Wikiquote (' + language + ')', link_type=language + 'wikiquote')
|
||||||
add_url(urls, result, default_label='Wikimedia Commons', link_type='commonswiki')
|
add_url(urls, result, id_cache, default_label='Wikimedia Commons', link_type='commonswiki')
|
||||||
|
|
||||||
add_url(urls, result, 'P625', 'OpenStreetMap', link_type='geo')
|
add_url(urls, result, id_cache, 'P625', 'OpenStreetMap', link_type='geo')
|
||||||
|
|
||||||
# musicbrainz
|
# musicbrainz
|
||||||
add_url(urls, result, 'P434', 'MusicBrainz', 'http://musicbrainz.org/artist/')
|
add_url(urls, result, id_cache, 'P434', 'MusicBrainz', 'http://musicbrainz.org/artist/')
|
||||||
add_url(urls, result, 'P435', 'MusicBrainz', 'http://musicbrainz.org/work/')
|
add_url(urls, result, id_cache, 'P435', 'MusicBrainz', 'http://musicbrainz.org/work/')
|
||||||
add_url(urls, result, 'P436', 'MusicBrainz', 'http://musicbrainz.org/release-group/')
|
add_url(urls, result, id_cache, 'P436', 'MusicBrainz', 'http://musicbrainz.org/release-group/')
|
||||||
add_url(urls, result, 'P966', 'MusicBrainz', 'http://musicbrainz.org/label/')
|
add_url(urls, result, id_cache, 'P966', 'MusicBrainz', 'http://musicbrainz.org/label/')
|
||||||
|
|
||||||
# IMDb
|
# IMDb
|
||||||
add_url(urls, result, 'P345', 'IMDb', 'https://www.imdb.com/', link_type='imdb')
|
add_url(urls, result, id_cache, 'P345', 'IMDb', 'https://www.imdb.com/', link_type='imdb')
|
||||||
# source code repository
|
# source code repository
|
||||||
add_url(urls, result, 'P1324')
|
add_url(urls, result, id_cache, 'P1324')
|
||||||
# blog
|
# blog
|
||||||
add_url(urls, result, 'P1581')
|
add_url(urls, result, id_cache, 'P1581')
|
||||||
# social media links
|
# social media links
|
||||||
add_url(urls, result, 'P2397', 'YouTube', 'https://www.youtube.com/channel/')
|
add_url(urls, result, id_cache, 'P2397', 'YouTube', 'https://www.youtube.com/channel/')
|
||||||
add_url(urls, result, 'P1651', 'YouTube', 'https://www.youtube.com/watch?v=')
|
add_url(urls, result, id_cache, 'P1651', 'YouTube', 'https://www.youtube.com/watch?v=')
|
||||||
add_url(urls, result, 'P2002', 'Twitter', 'https://twitter.com/')
|
add_url(urls, result, id_cache, 'P2002', 'Twitter', 'https://twitter.com/')
|
||||||
add_url(urls, result, 'P2013', 'Facebook', 'https://facebook.com/')
|
add_url(urls, result, id_cache, 'P2013', 'Facebook', 'https://facebook.com/')
|
||||||
add_url(urls, result, 'P2003', 'Instagram', 'https://instagram.com/')
|
add_url(urls, result, id_cache, 'P2003', 'Instagram', 'https://instagram.com/')
|
||||||
|
|
||||||
urls.append({'title': 'Wikidata',
|
urls.append({'title': 'Wikidata',
|
||||||
'url': 'https://www.wikidata.org/wiki/'
|
'url': 'https://www.wikidata.org/wiki/'
|
||||||
|
@ -163,132 +192,132 @@ def getDetail(jsonresponse, wikidata_id, language, locale):
|
||||||
|
|
||||||
# DATES
|
# DATES
|
||||||
# inception date
|
# inception date
|
||||||
add_attribute(attributes, result, 'P571', date=True)
|
add_attribute(attributes, id_cache, 'P571', date=True)
|
||||||
# dissolution date
|
# dissolution date
|
||||||
add_attribute(attributes, result, 'P576', date=True)
|
add_attribute(attributes, id_cache, 'P576', date=True)
|
||||||
# start date
|
# start date
|
||||||
add_attribute(attributes, result, 'P580', date=True)
|
add_attribute(attributes, id_cache, 'P580', date=True)
|
||||||
# end date
|
# end date
|
||||||
add_attribute(attributes, result, 'P582', date=True)
|
add_attribute(attributes, id_cache, 'P582', date=True)
|
||||||
# date of birth
|
# date of birth
|
||||||
add_attribute(attributes, result, 'P569', date=True)
|
add_attribute(attributes, id_cache, 'P569', date=True)
|
||||||
# date of death
|
# date of death
|
||||||
add_attribute(attributes, result, 'P570', date=True)
|
add_attribute(attributes, id_cache, 'P570', date=True)
|
||||||
# date of spacecraft launch
|
# date of spacecraft launch
|
||||||
add_attribute(attributes, result, 'P619', date=True)
|
add_attribute(attributes, id_cache, 'P619', date=True)
|
||||||
# date of spacecraft landing
|
# date of spacecraft landing
|
||||||
add_attribute(attributes, result, 'P620', date=True)
|
add_attribute(attributes, id_cache, 'P620', date=True)
|
||||||
|
|
||||||
# nationality
|
# nationality
|
||||||
add_attribute(attributes, result, 'P27')
|
add_attribute(attributes, id_cache, 'P27')
|
||||||
# country of origin
|
# country of origin
|
||||||
add_attribute(attributes, result, 'P495')
|
add_attribute(attributes, id_cache, 'P495')
|
||||||
# country
|
# country
|
||||||
add_attribute(attributes, result, 'P17')
|
add_attribute(attributes, id_cache, 'P17')
|
||||||
# headquarters
|
# headquarters
|
||||||
add_attribute(attributes, result, 'Q180')
|
add_attribute(attributes, id_cache, 'Q180')
|
||||||
|
|
||||||
# PLACES
|
# PLACES
|
||||||
# capital
|
# capital
|
||||||
add_attribute(attributes, result, 'P36', trim=True)
|
add_attribute(attributes, id_cache, 'P36', trim=True)
|
||||||
# head of state
|
# head of state
|
||||||
add_attribute(attributes, result, 'P35', trim=True)
|
add_attribute(attributes, id_cache, 'P35', trim=True)
|
||||||
# head of government
|
# head of government
|
||||||
add_attribute(attributes, result, 'P6', trim=True)
|
add_attribute(attributes, id_cache, 'P6', trim=True)
|
||||||
# type of government
|
# type of government
|
||||||
add_attribute(attributes, result, 'P122')
|
add_attribute(attributes, id_cache, 'P122')
|
||||||
# official language
|
# official language
|
||||||
add_attribute(attributes, result, 'P37')
|
add_attribute(attributes, id_cache, 'P37')
|
||||||
# population
|
# population
|
||||||
add_attribute(attributes, result, 'P1082', trim=True)
|
add_attribute(attributes, id_cache, 'P1082', trim=True)
|
||||||
# area
|
# area
|
||||||
add_attribute(attributes, result, 'P2046')
|
add_attribute(attributes, id_cache, 'P2046')
|
||||||
# currency
|
# currency
|
||||||
add_attribute(attributes, result, 'P38', trim=True)
|
add_attribute(attributes, id_cache, 'P38', trim=True)
|
||||||
# heigth (building)
|
# heigth (building)
|
||||||
add_attribute(attributes, result, 'P2048')
|
add_attribute(attributes, id_cache, 'P2048')
|
||||||
|
|
||||||
# MEDIA
|
# MEDIA
|
||||||
# platform (videogames)
|
# platform (videogames)
|
||||||
add_attribute(attributes, result, 'P400')
|
add_attribute(attributes, id_cache, 'P400')
|
||||||
# author
|
# author
|
||||||
add_attribute(attributes, result, 'P50')
|
add_attribute(attributes, id_cache, 'P50')
|
||||||
# creator
|
# creator
|
||||||
add_attribute(attributes, result, 'P170')
|
add_attribute(attributes, id_cache, 'P170')
|
||||||
# director
|
# director
|
||||||
add_attribute(attributes, result, 'P57')
|
add_attribute(attributes, id_cache, 'P57')
|
||||||
# performer
|
# performer
|
||||||
add_attribute(attributes, result, 'P175')
|
add_attribute(attributes, id_cache, 'P175')
|
||||||
# developer
|
# developer
|
||||||
add_attribute(attributes, result, 'P178')
|
add_attribute(attributes, id_cache, 'P178')
|
||||||
# producer
|
# producer
|
||||||
add_attribute(attributes, result, 'P162')
|
add_attribute(attributes, id_cache, 'P162')
|
||||||
# manufacturer
|
# manufacturer
|
||||||
add_attribute(attributes, result, 'P176')
|
add_attribute(attributes, id_cache, 'P176')
|
||||||
# screenwriter
|
# screenwriter
|
||||||
add_attribute(attributes, result, 'P58')
|
add_attribute(attributes, id_cache, 'P58')
|
||||||
# production company
|
# production company
|
||||||
add_attribute(attributes, result, 'P272')
|
add_attribute(attributes, id_cache, 'P272')
|
||||||
# record label
|
# record label
|
||||||
add_attribute(attributes, result, 'P264')
|
add_attribute(attributes, id_cache, 'P264')
|
||||||
# publisher
|
# publisher
|
||||||
add_attribute(attributes, result, 'P123')
|
add_attribute(attributes, id_cache, 'P123')
|
||||||
# original network
|
# original network
|
||||||
add_attribute(attributes, result, 'P449')
|
add_attribute(attributes, id_cache, 'P449')
|
||||||
# distributor
|
# distributor
|
||||||
add_attribute(attributes, result, 'P750')
|
add_attribute(attributes, id_cache, 'P750')
|
||||||
# composer
|
# composer
|
||||||
add_attribute(attributes, result, 'P86')
|
add_attribute(attributes, id_cache, 'P86')
|
||||||
# publication date
|
# publication date
|
||||||
add_attribute(attributes, result, 'P577', date=True)
|
add_attribute(attributes, id_cache, 'P577', date=True)
|
||||||
# genre
|
# genre
|
||||||
add_attribute(attributes, result, 'P136')
|
add_attribute(attributes, id_cache, 'P136')
|
||||||
# original language
|
# original language
|
||||||
add_attribute(attributes, result, 'P364')
|
add_attribute(attributes, id_cache, 'P364')
|
||||||
# isbn
|
# isbn
|
||||||
add_attribute(attributes, result, 'Q33057')
|
add_attribute(attributes, id_cache, 'Q33057')
|
||||||
# software license
|
# software license
|
||||||
add_attribute(attributes, result, 'P275')
|
add_attribute(attributes, id_cache, 'P275')
|
||||||
# programming language
|
# programming language
|
||||||
add_attribute(attributes, result, 'P277')
|
add_attribute(attributes, id_cache, 'P277')
|
||||||
# version
|
# version
|
||||||
add_attribute(attributes, result, 'P348', trim=True)
|
add_attribute(attributes, id_cache, 'P348', trim=True)
|
||||||
# narrative location
|
# narrative location
|
||||||
add_attribute(attributes, result, 'P840')
|
add_attribute(attributes, id_cache, 'P840')
|
||||||
|
|
||||||
# LANGUAGES
|
# LANGUAGES
|
||||||
# number of speakers
|
# number of speakers
|
||||||
add_attribute(attributes, result, 'P1098')
|
add_attribute(attributes, id_cache, 'P1098')
|
||||||
# writing system
|
# writing system
|
||||||
add_attribute(attributes, result, 'P282')
|
add_attribute(attributes, id_cache, 'P282')
|
||||||
# regulatory body
|
# regulatory body
|
||||||
add_attribute(attributes, result, 'P1018')
|
add_attribute(attributes, id_cache, 'P1018')
|
||||||
# language code
|
# language code
|
||||||
add_attribute(attributes, result, 'P218')
|
add_attribute(attributes, id_cache, 'P218')
|
||||||
|
|
||||||
# OTHER
|
# OTHER
|
||||||
# ceo
|
# ceo
|
||||||
add_attribute(attributes, result, 'P169', trim=True)
|
add_attribute(attributes, id_cache, 'P169', trim=True)
|
||||||
# founder
|
# founder
|
||||||
add_attribute(attributes, result, 'P112')
|
add_attribute(attributes, id_cache, 'P112')
|
||||||
# legal form (company/organization)
|
# legal form (company/organization)
|
||||||
add_attribute(attributes, result, 'P1454')
|
add_attribute(attributes, id_cache, 'P1454')
|
||||||
# operator
|
# operator
|
||||||
add_attribute(attributes, result, 'P137')
|
add_attribute(attributes, id_cache, 'P137')
|
||||||
# crew members (tripulation)
|
# crew members (tripulation)
|
||||||
add_attribute(attributes, result, 'P1029')
|
add_attribute(attributes, id_cache, 'P1029')
|
||||||
# taxon
|
# taxon
|
||||||
add_attribute(attributes, result, 'P225')
|
add_attribute(attributes, id_cache, 'P225')
|
||||||
# chemical formula
|
# chemical formula
|
||||||
add_attribute(attributes, result, 'P274')
|
add_attribute(attributes, id_cache, 'P274')
|
||||||
# winner (sports/contests)
|
# winner (sports/contests)
|
||||||
add_attribute(attributes, result, 'P1346')
|
add_attribute(attributes, id_cache, 'P1346')
|
||||||
# number of deaths
|
# number of deaths
|
||||||
add_attribute(attributes, result, 'P1120')
|
add_attribute(attributes, id_cache, 'P1120')
|
||||||
# currency code
|
# currency code
|
||||||
add_attribute(attributes, result, 'P498')
|
add_attribute(attributes, id_cache, 'P498')
|
||||||
|
|
||||||
image = add_image(result)
|
image = add_image(id_cache)
|
||||||
|
|
||||||
if len(attributes) == 0 and len(urls) == 2 and len(description) == 0:
|
if len(attributes) == 0 and len(urls) == 2 and len(description) == 0:
|
||||||
results.append({
|
results.append({
|
||||||
|
@ -310,43 +339,42 @@ def getDetail(jsonresponse, wikidata_id, language, locale):
|
||||||
|
|
||||||
|
|
||||||
# only returns first match
|
# only returns first match
|
||||||
def add_image(result):
|
def add_image(id_cache):
|
||||||
# P15: route map, P242: locator map, P154: logo, P18: image, P242: map, P41: flag, P2716: collage, P2910: icon
|
# P15: route map, P242: locator map, P154: logo, P18: image, P242: map, P41: flag, P2716: collage, P2910: icon
|
||||||
property_ids = ['P15', 'P242', 'P154', 'P18', 'P242', 'P41', 'P2716', 'P2910']
|
property_ids = ['P15', 'P242', 'P154', 'P18', 'P242', 'P41', 'P2716', 'P2910']
|
||||||
|
|
||||||
for property_id in property_ids:
|
for property_id in property_ids:
|
||||||
image = result.xpath(property_xpath.replace('{propertyid}', property_id))
|
image = id_cache.get(property_id, None)
|
||||||
if image:
|
if image is not None:
|
||||||
image_name = image[0].xpath(media_xpath)
|
image_name = eval_xpath(image, media_xpath)
|
||||||
image_src = url_image.replace('{filename}', extract_text(image_name[0]))
|
image_src = url_image.replace('{filename}', extract_text(image_name[0]))
|
||||||
return image_src
|
return image_src
|
||||||
|
|
||||||
|
|
||||||
# setting trim will only returned high ranked rows OR the first row
|
# setting trim will only returned high ranked rows OR the first row
|
||||||
def add_attribute(attributes, result, property_id, default_label=None, date=False, trim=False):
|
def add_attribute(attributes, id_cache, property_id, default_label=None, date=False, trim=False):
|
||||||
attribute = result.xpath(property_xpath.replace('{propertyid}', property_id))
|
attribute = id_cache.get(property_id, None)
|
||||||
if attribute:
|
if attribute is not None:
|
||||||
|
|
||||||
if default_label:
|
if default_label:
|
||||||
label = default_label
|
label = default_label
|
||||||
else:
|
else:
|
||||||
label = extract_text(attribute[0].xpath(label_xpath))
|
label = extract_text(eval_xpath(attribute, label_xpath))
|
||||||
label = label[0].upper() + label[1:]
|
label = label[0].upper() + label[1:]
|
||||||
|
|
||||||
if date:
|
if date:
|
||||||
trim = True
|
trim = True
|
||||||
# remove calendar name
|
# remove calendar name
|
||||||
calendar_name = attribute[0].xpath(calendar_name_xpath)
|
calendar_name = eval_xpath(attribute, calendar_name_xpath)
|
||||||
for calendar in calendar_name:
|
for calendar in calendar_name:
|
||||||
calendar.getparent().remove(calendar)
|
calendar.getparent().remove(calendar)
|
||||||
|
|
||||||
concat_values = ""
|
concat_values = ""
|
||||||
values = []
|
values = []
|
||||||
first_value = None
|
first_value = None
|
||||||
for row in attribute[0].xpath(property_row_xpath):
|
for row in eval_xpath(attribute, property_row_xpath):
|
||||||
if not first_value or not trim or row.xpath(preferred_rank_xpath):
|
if not first_value or not trim or eval_xpath(row, preferred_rank_xpath):
|
||||||
|
value = eval_xpath(row, value_xpath)
|
||||||
value = row.xpath(value_xpath)
|
|
||||||
if not value:
|
if not value:
|
||||||
continue
|
continue
|
||||||
value = extract_text(value)
|
value = extract_text(value)
|
||||||
|
@ -369,18 +397,18 @@ def add_attribute(attributes, result, property_id, default_label=None, date=Fals
|
||||||
|
|
||||||
|
|
||||||
# requires property_id unless it's a wiki link (defined in link_type)
|
# requires property_id unless it's a wiki link (defined in link_type)
|
||||||
def add_url(urls, result, property_id=None, default_label=None, url_prefix=None, results=None, link_type=None):
|
def add_url(urls, result, id_cache, property_id=None, default_label=None, url_prefix=None, results=None,
|
||||||
|
link_type=None):
|
||||||
links = []
|
links = []
|
||||||
|
|
||||||
# wiki links don't have property in wikidata page
|
# wiki links don't have property in wikidata page
|
||||||
if link_type and 'wiki' in link_type:
|
if link_type and 'wiki' in link_type:
|
||||||
links.append(get_wikilink(result, link_type))
|
links.append(get_wikilink(result, link_type))
|
||||||
else:
|
else:
|
||||||
dom_element = result.xpath(property_xpath.replace('{propertyid}', property_id))
|
dom_element = id_cache.get(property_id, None)
|
||||||
if dom_element:
|
if dom_element is not None:
|
||||||
dom_element = dom_element[0]
|
|
||||||
if not default_label:
|
if not default_label:
|
||||||
label = extract_text(dom_element.xpath(label_xpath))
|
label = extract_text(eval_xpath(dom_element, label_xpath))
|
||||||
label = label[0].upper() + label[1:]
|
label = label[0].upper() + label[1:]
|
||||||
|
|
||||||
if link_type == 'geo':
|
if link_type == 'geo':
|
||||||
|
@ -390,7 +418,7 @@ def add_url(urls, result, property_id=None, default_label=None, url_prefix=None,
|
||||||
links.append(get_imdblink(dom_element, url_prefix))
|
links.append(get_imdblink(dom_element, url_prefix))
|
||||||
|
|
||||||
else:
|
else:
|
||||||
url_results = dom_element.xpath(url_xpath)
|
url_results = eval_xpath(dom_element, url_xpath)
|
||||||
for link in url_results:
|
for link in url_results:
|
||||||
if link is not None:
|
if link is not None:
|
||||||
if url_prefix:
|
if url_prefix:
|
||||||
|
@ -410,7 +438,7 @@ def add_url(urls, result, property_id=None, default_label=None, url_prefix=None,
|
||||||
|
|
||||||
|
|
||||||
def get_imdblink(result, url_prefix):
|
def get_imdblink(result, url_prefix):
|
||||||
imdb_id = result.xpath(value_xpath)
|
imdb_id = eval_xpath(result, value_xpath)
|
||||||
if imdb_id:
|
if imdb_id:
|
||||||
imdb_id = extract_text(imdb_id)
|
imdb_id = extract_text(imdb_id)
|
||||||
id_prefix = imdb_id[:2]
|
id_prefix = imdb_id[:2]
|
||||||
|
@ -430,7 +458,7 @@ def get_imdblink(result, url_prefix):
|
||||||
|
|
||||||
|
|
||||||
def get_geolink(result):
|
def get_geolink(result):
|
||||||
coordinates = result.xpath(value_xpath)
|
coordinates = eval_xpath(result, value_xpath)
|
||||||
if not coordinates:
|
if not coordinates:
|
||||||
return None
|
return None
|
||||||
coordinates = extract_text(coordinates[0])
|
coordinates = extract_text(coordinates[0])
|
||||||
|
@ -477,7 +505,7 @@ def get_geolink(result):
|
||||||
|
|
||||||
|
|
||||||
def get_wikilink(result, wikiid):
|
def get_wikilink(result, wikiid):
|
||||||
url = result.xpath(wikilink_xpath.replace('{wikiid}', wikiid))
|
url = eval_xpath(result, wikilink_xpath.replace('{wikiid}', wikiid))
|
||||||
if not url:
|
if not url:
|
||||||
return None
|
return None
|
||||||
url = url[0]
|
url = url[0]
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
from lxml.html import fromstring
|
from lxml.html import fromstring
|
||||||
|
from lxml import etree
|
||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
import mock
|
import mock
|
||||||
from searx.engines import wikidata
|
from searx.engines import wikidata
|
||||||
|
@ -30,12 +31,12 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
|
|
||||||
wikidata.supported_languages = ['en', 'es']
|
wikidata.supported_languages = ['en', 'es']
|
||||||
wikidata.language_aliases = {}
|
wikidata.language_aliases = {}
|
||||||
response = mock.Mock(text='<html></html>', search_params={"language": "en"})
|
response = mock.Mock(content='<html></html>'.encode("utf-8"), search_params={"language": "en"})
|
||||||
self.assertEqual(wikidata.response(response), [])
|
self.assertEqual(wikidata.response(response), [])
|
||||||
|
|
||||||
def test_getDetail(self):
|
def test_getDetail(self):
|
||||||
response = {}
|
response = {}
|
||||||
results = wikidata.getDetail(response, "Q123", "en", "en-US")
|
results = wikidata.getDetail(response, "Q123", "en", "en-US", etree.HTMLParser())
|
||||||
self.assertEqual(results, [])
|
self.assertEqual(results, [])
|
||||||
|
|
||||||
title_html = '<div><div class="wikibase-title-label">Test</div></div>'
|
title_html = '<div><div class="wikibase-title-label">Test</div></div>'
|
||||||
|
@ -52,7 +53,7 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
response = {"parse": {"displaytitle": title_html, "text": html}}
|
response = {"parse": {"displaytitle": title_html, "text": html}}
|
||||||
|
|
||||||
results = wikidata.getDetail(response, "Q123", "en", "en-US")
|
results = wikidata.getDetail(response, "Q123", "en", "en-US", etree.HTMLParser())
|
||||||
self.assertEqual(len(results), 1)
|
self.assertEqual(len(results), 1)
|
||||||
self.assertEqual(results[0]['url'], 'https://en.wikipedia.org/wiki/Test')
|
self.assertEqual(results[0]['url'], 'https://en.wikipedia.org/wiki/Test')
|
||||||
|
|
||||||
|
@ -92,7 +93,7 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
response = {"parse": {"displaytitle": title_html, "text": html}}
|
response = {"parse": {"displaytitle": title_html, "text": html}}
|
||||||
|
|
||||||
results = wikidata.getDetail(response, "Q123", "yua", "yua_MX")
|
results = wikidata.getDetail(response, "Q123", "yua", "yua_MX", etree.HTMLParser())
|
||||||
self.assertEqual(len(results), 2)
|
self.assertEqual(len(results), 2)
|
||||||
self.assertEqual(results[0]['title'], 'Official website')
|
self.assertEqual(results[0]['title'], 'Official website')
|
||||||
self.assertEqual(results[0]['url'], 'https://officialsite.com')
|
self.assertEqual(results[0]['url'], 'https://officialsite.com')
|
||||||
|
@ -139,8 +140,8 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
</div>
|
</div>
|
||||||
"""
|
"""
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
image_src = wikidata.add_image(html_etree)
|
image_src = wikidata.add_image(id_cache)
|
||||||
self.assertEqual(image_src,
|
self.assertEqual(image_src,
|
||||||
"https://commons.wikimedia.org/wiki/Special:FilePath/image.png?width=500&height=400")
|
"https://commons.wikimedia.org/wiki/Special:FilePath/image.png?width=500&height=400")
|
||||||
|
|
||||||
|
@ -197,8 +198,9 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
</div>
|
</div>
|
||||||
"""
|
"""
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
|
||||||
image_src = wikidata.add_image(html_etree)
|
image_src = wikidata.add_image(id_cache)
|
||||||
self.assertEqual(image_src,
|
self.assertEqual(image_src,
|
||||||
"https://commons.wikimedia.org/wiki/Special:FilePath/logo.png?width=500&height=400")
|
"https://commons.wikimedia.org/wiki/Special:FilePath/logo.png?width=500&height=400")
|
||||||
|
|
||||||
|
@ -232,11 +234,12 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
attributes = []
|
attributes = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
|
||||||
wikidata.add_attribute(attributes, html_etree, "Fail")
|
wikidata.add_attribute(attributes, id_cache, "Fail")
|
||||||
self.assertEqual(attributes, [])
|
self.assertEqual(attributes, [])
|
||||||
|
|
||||||
wikidata.add_attribute(attributes, html_etree, "P27")
|
wikidata.add_attribute(attributes, id_cache, "P27")
|
||||||
self.assertEqual(len(attributes), 1)
|
self.assertEqual(len(attributes), 1)
|
||||||
self.assertEqual(attributes[0]["label"], "Country of citizenship")
|
self.assertEqual(attributes[0]["label"], "Country of citizenship")
|
||||||
self.assertEqual(attributes[0]["value"], "United Kingdom")
|
self.assertEqual(attributes[0]["value"], "United Kingdom")
|
||||||
|
@ -271,7 +274,8 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
attributes = []
|
attributes = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
wikidata.add_attribute(attributes, html_etree, "P569", date=True)
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
wikidata.add_attribute(attributes, id_cache, "P569", date=True)
|
||||||
self.assertEqual(len(attributes), 1)
|
self.assertEqual(len(attributes), 1)
|
||||||
self.assertEqual(attributes[0]["label"], "Date of birth")
|
self.assertEqual(attributes[0]["label"], "Date of birth")
|
||||||
self.assertEqual(attributes[0]["value"], "27 January 1832")
|
self.assertEqual(attributes[0]["value"], "27 January 1832")
|
||||||
|
@ -319,14 +323,16 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
attributes = []
|
attributes = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
wikidata.add_attribute(attributes, html_etree, "P6")
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
wikidata.add_attribute(attributes, id_cache, "P6")
|
||||||
self.assertEqual(len(attributes), 1)
|
self.assertEqual(len(attributes), 1)
|
||||||
self.assertEqual(attributes[0]["label"], "Head of government")
|
self.assertEqual(attributes[0]["label"], "Head of government")
|
||||||
self.assertEqual(attributes[0]["value"], "Old Prime Minister, Actual Prime Minister")
|
self.assertEqual(attributes[0]["value"], "Old Prime Minister, Actual Prime Minister")
|
||||||
|
|
||||||
attributes = []
|
attributes = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
wikidata.add_attribute(attributes, html_etree, "P6", trim=True)
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
wikidata.add_attribute(attributes, id_cache, "P6", trim=True)
|
||||||
self.assertEqual(len(attributes), 1)
|
self.assertEqual(len(attributes), 1)
|
||||||
self.assertEqual(attributes[0]["value"], "Actual Prime Minister")
|
self.assertEqual(attributes[0]["value"], "Actual Prime Minister")
|
||||||
|
|
||||||
|
@ -357,12 +363,13 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
urls = []
|
urls = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
wikidata.add_url(urls, html_etree, 'P856')
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
wikidata.add_url(urls, html_etree, id_cache, 'P856')
|
||||||
self.assertEquals(len(urls), 1)
|
self.assertEquals(len(urls), 1)
|
||||||
self.assertIn({'title': 'Official website', 'url': 'https://searx.me/'}, urls)
|
self.assertIn({'title': 'Official website', 'url': 'https://searx.me/'}, urls)
|
||||||
urls = []
|
urls = []
|
||||||
results = []
|
results = []
|
||||||
wikidata.add_url(urls, html_etree, 'P856', 'custom label', results=results)
|
wikidata.add_url(urls, html_etree, id_cache, 'P856', 'custom label', results=results)
|
||||||
self.assertEquals(len(urls), 1)
|
self.assertEquals(len(urls), 1)
|
||||||
self.assertEquals(len(results), 1)
|
self.assertEquals(len(results), 1)
|
||||||
self.assertIn({'title': 'custom label', 'url': 'https://searx.me/'}, urls)
|
self.assertIn({'title': 'custom label', 'url': 'https://searx.me/'}, urls)
|
||||||
|
@ -405,7 +412,8 @@ class TestWikidataEngine(SearxTestCase):
|
||||||
"""
|
"""
|
||||||
urls = []
|
urls = []
|
||||||
html_etree = fromstring(html)
|
html_etree = fromstring(html)
|
||||||
wikidata.add_url(urls, html_etree, 'P856')
|
id_cache = wikidata.get_id_cache(html_etree)
|
||||||
|
wikidata.add_url(urls, html_etree, id_cache, 'P856')
|
||||||
self.assertEquals(len(urls), 2)
|
self.assertEquals(len(urls), 2)
|
||||||
self.assertIn({'title': 'Official website', 'url': 'http://www.worldofwarcraft.com'}, urls)
|
self.assertIn({'title': 'Official website', 'url': 'http://www.worldofwarcraft.com'}, urls)
|
||||||
self.assertIn({'title': 'Official website', 'url': 'http://eu.battle.net/wow/en/'}, urls)
|
self.assertIn({'title': 'Official website', 'url': 'http://eu.battle.net/wow/en/'}, urls)
|
||||||
|
|
Loading…
Reference in New Issue