mirror of https://github.com/searxng/searxng.git
134 lines
4.1 KiB
Python
134 lines
4.1 KiB
Python
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||
|
"""Yandex (Web, images)"""
|
||
|
|
||
|
from json import loads
|
||
|
from urllib.parse import urlencode
|
||
|
from html import unescape
|
||
|
from lxml import html
|
||
|
from searx.exceptions import SearxEngineCaptchaException
|
||
|
from searx.utils import humanize_bytes, eval_xpath, eval_xpath_list, extract_text, extr
|
||
|
|
||
|
|
||
|
# Engine metadata
|
||
|
about = {
|
||
|
"website": 'https://yandex.com/',
|
||
|
"wikidata_id": 'Q5281',
|
||
|
"official_api_documentation": "?",
|
||
|
"use_official_api": False,
|
||
|
"require_api_key": False,
|
||
|
"results": 'HTML',
|
||
|
}
|
||
|
|
||
|
# Engine configuration
|
||
|
categories = []
|
||
|
paging = True
|
||
|
search_type = ""
|
||
|
|
||
|
# Search URL
|
||
|
base_url_web = 'https://yandex.com/search/site/'
|
||
|
base_url_images = 'https://yandex.com/images/search'
|
||
|
|
||
|
results_xpath = '//li[contains(@class, "serp-item")]'
|
||
|
url_xpath = './/a[@class="b-serp-item__title-link"]/@href'
|
||
|
title_xpath = './/h3[@class="b-serp-item__title"]/a[@class="b-serp-item__title-link"]/span'
|
||
|
content_xpath = './/div[@class="b-serp-item__content"]//div[@class="b-serp-item__text"]'
|
||
|
|
||
|
|
||
|
def catch_bad_response(resp):
|
||
|
if resp.url.path.startswith('/showcaptcha'):
|
||
|
raise SearxEngineCaptchaException()
|
||
|
|
||
|
|
||
|
def request(query, params):
|
||
|
query_params_web = {
|
||
|
"tmpl_version": "releases",
|
||
|
"text": query,
|
||
|
"web": "1",
|
||
|
"frame": "1",
|
||
|
"searchid": "3131712",
|
||
|
}
|
||
|
|
||
|
query_params_images = {
|
||
|
"text": query,
|
||
|
"uinfo": "sw-1920-sh-1080-ww-1125-wh-999",
|
||
|
}
|
||
|
|
||
|
if params['pageno'] > 1:
|
||
|
query_params_web.update({"p": params["pageno"] - 1})
|
||
|
query_params_images.update({"p": params["pageno"] - 1})
|
||
|
|
||
|
params["cookies"] = {'cookie': "yp=1716337604.sp.family%3A0#1685406411.szm.1:1920x1080:1920x999"}
|
||
|
|
||
|
if search_type == 'web':
|
||
|
params['url'] = f"{base_url_web}?{urlencode(query_params_web)}"
|
||
|
elif search_type == 'images':
|
||
|
params['url'] = f"{base_url_images}?{urlencode(query_params_images)}"
|
||
|
|
||
|
return params
|
||
|
|
||
|
|
||
|
def response(resp):
|
||
|
if search_type == 'web':
|
||
|
|
||
|
catch_bad_response(resp)
|
||
|
|
||
|
dom = html.fromstring(resp.text)
|
||
|
|
||
|
results = []
|
||
|
|
||
|
for result in eval_xpath_list(dom, results_xpath):
|
||
|
results.append(
|
||
|
{
|
||
|
'url': extract_text(eval_xpath(result, url_xpath)),
|
||
|
'title': extract_text(eval_xpath(result, title_xpath)),
|
||
|
'content': extract_text(eval_xpath(result, content_xpath)),
|
||
|
}
|
||
|
)
|
||
|
|
||
|
return results
|
||
|
|
||
|
if search_type == 'images':
|
||
|
|
||
|
catch_bad_response(resp)
|
||
|
|
||
|
html_data = html.fromstring(resp.text)
|
||
|
html_sample = unescape(html.tostring(html_data, encoding='unicode'))
|
||
|
|
||
|
content_between_tags = extr(
|
||
|
html_sample, '{"location":"/images/search/', 'advRsyaSearchColumn":null}}', default="fail"
|
||
|
)
|
||
|
json_data = '{"location":"/images/search/' + content_between_tags + 'advRsyaSearchColumn":null}}'
|
||
|
|
||
|
if content_between_tags == "fail":
|
||
|
content_between_tags = extr(html_sample, '{"location":"/images/search/', 'false}}}')
|
||
|
json_data = '{"location":"/images/search/' + content_between_tags + 'false}}}'
|
||
|
|
||
|
json_resp = loads(json_data)
|
||
|
|
||
|
results = []
|
||
|
for _, item_data in json_resp['initialState']['serpList']['items']['entities'].items():
|
||
|
title = item_data['snippet']['title']
|
||
|
source = item_data['snippet']['url']
|
||
|
thumb = item_data['image']
|
||
|
fullsize_image = item_data['viewerData']['dups'][0]['url']
|
||
|
height = item_data['viewerData']['dups'][0]['h']
|
||
|
width = item_data['viewerData']['dups'][0]['w']
|
||
|
filesize = item_data['viewerData']['dups'][0]['fileSizeInBytes']
|
||
|
humanized_filesize = humanize_bytes(filesize)
|
||
|
|
||
|
results.append(
|
||
|
{
|
||
|
'title': title,
|
||
|
'url': source,
|
||
|
'img_src': fullsize_image,
|
||
|
'filesize': humanized_filesize,
|
||
|
'thumbnail_src': thumb,
|
||
|
'template': 'images.html',
|
||
|
'resolution': f'{width} x {height}',
|
||
|
}
|
||
|
)
|
||
|
|
||
|
return results
|
||
|
|
||
|
return []
|