123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134 |
- # SPDX-License-Identifier: AGPL-3.0-or-later
- """Yandex (Web, images)"""
- from json import loads
- from urllib.parse import urlencode
- from html import unescape
- from lxml import html
- from searx.exceptions import SearxEngineCaptchaException
- from searx.utils import humanize_bytes, eval_xpath, eval_xpath_list, extract_text, extr
- # Engine metadata
- about = {
- "website": 'https://yandex.com/',
- "wikidata_id": 'Q5281',
- "official_api_documentation": "?",
- "use_official_api": False,
- "require_api_key": False,
- "results": 'HTML',
- }
- # Engine configuration
- categories = []
- paging = True
- search_type = ""
- # Search URL
- base_url_web = 'https://yandex.com/search/site/'
- base_url_images = 'https://yandex.com/images/search'
- results_xpath = '//li[contains(@class, "serp-item")]'
- url_xpath = './/a[@class="b-serp-item__title-link"]/@href'
- title_xpath = './/h3[@class="b-serp-item__title"]/a[@class="b-serp-item__title-link"]/span'
- content_xpath = './/div[@class="b-serp-item__content"]//div[@class="b-serp-item__text"]'
- def catch_bad_response(resp):
- if resp.url.path.startswith('/showcaptcha'):
- raise SearxEngineCaptchaException()
- def request(query, params):
- query_params_web = {
- "tmpl_version": "releases",
- "text": query,
- "web": "1",
- "frame": "1",
- "searchid": "3131712",
- }
- query_params_images = {
- "text": query,
- "uinfo": "sw-1920-sh-1080-ww-1125-wh-999",
- }
- if params['pageno'] > 1:
- query_params_web.update({"p": params["pageno"] - 1})
- query_params_images.update({"p": params["pageno"] - 1})
- params["cookies"] = {'cookie': "yp=1716337604.sp.family%3A0#1685406411.szm.1:1920x1080:1920x999"}
- if search_type == 'web':
- params['url'] = f"{base_url_web}?{urlencode(query_params_web)}"
- elif search_type == 'images':
- params['url'] = f"{base_url_images}?{urlencode(query_params_images)}"
- return params
- def response(resp):
- if search_type == 'web':
- catch_bad_response(resp)
- dom = html.fromstring(resp.text)
- results = []
- for result in eval_xpath_list(dom, results_xpath):
- results.append(
- {
- 'url': extract_text(eval_xpath(result, url_xpath)),
- 'title': extract_text(eval_xpath(result, title_xpath)),
- 'content': extract_text(eval_xpath(result, content_xpath)),
- }
- )
- return results
- if search_type == 'images':
- catch_bad_response(resp)
- html_data = html.fromstring(resp.text)
- html_sample = unescape(html.tostring(html_data, encoding='unicode'))
- content_between_tags = extr(
- html_sample, '{"location":"/images/search/', 'advRsyaSearchColumn":null}}', default="fail"
- )
- json_data = '{"location":"/images/search/' + content_between_tags + 'advRsyaSearchColumn":null}}'
- if content_between_tags == "fail":
- content_between_tags = extr(html_sample, '{"location":"/images/search/', 'false}}}')
- json_data = '{"location":"/images/search/' + content_between_tags + 'false}}}'
- json_resp = loads(json_data)
- results = []
- for _, item_data in json_resp['initialState']['serpList']['items']['entities'].items():
- title = item_data['snippet']['title']
- source = item_data['snippet']['url']
- thumb = item_data['image']
- fullsize_image = item_data['viewerData']['dups'][0]['url']
- height = item_data['viewerData']['dups'][0]['h']
- width = item_data['viewerData']['dups'][0]['w']
- filesize = item_data['viewerData']['dups'][0]['fileSizeInBytes']
- humanized_filesize = humanize_bytes(filesize)
- results.append(
- {
- 'title': title,
- 'url': source,
- 'img_src': fullsize_image,
- 'filesize': humanized_filesize,
- 'thumbnail_src': thumb,
- 'template': 'images.html',
- 'resolution': f'{width} x {height}',
- }
- )
- return results
- return []
|