1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677 |
- # SPDX-License-Identifier: AGPL-3.0-or-later
- """Ipernity (images)"""
- from datetime import datetime
- from json import loads, JSONDecodeError
- from urllib.parse import quote_plus
- from lxml import html
- from searx.utils import extr, extract_text, eval_xpath, eval_xpath_list
- about = {
- 'website': 'https://www.ipernity.com',
- 'official_api_documentation': 'https://www.ipernity.com/help/api',
- 'use_official_api': False,
- 'require_api_key': False,
- 'results': 'HTML',
- }
- paging = True
- categories = ['images']
- base_url = 'https://www.ipernity.com'
- page_size = 10
- def request(query, params):
- params['url'] = f"{base_url}/search/photo/@/page:{params['pageno']}:{page_size}?q={quote_plus(query)}"
- return params
- def response(resp):
- results = []
- doc = html.fromstring(resp.text)
- images = eval_xpath_list(doc, '//a[starts-with(@href, "/doc")]//img')
- result_index = 0
- for result in eval_xpath_list(doc, '//script[@type="text/javascript"]'):
- info_js = extr(extract_text(result), '] = ', '};') + '}'
- if not info_js:
- continue
- try:
- info_item = loads(info_js)
- if not info_item.get('mediakey'):
- continue
- thumbnail_src = extract_text(eval_xpath(images[result_index], './@src'))
- img_src = thumbnail_src.replace('240.jpg', '640.jpg')
- resolution = None
- if info_item.get("width") and info_item.get("height"):
- resolution = f'{info_item["width"]}x{info_item["height"]}'
- item = {
- 'template': 'images.html',
- 'url': f"{base_url}/doc/{info_item['user_id']}/{info_item['doc_id']}",
- 'title': info_item.get('title'),
- 'content': info_item.get('content', ''),
- 'resolution': resolution,
- 'publishedDate': datetime.fromtimestamp(int(info_item['posted_at'])),
- 'thumbnail_src': thumbnail_src,
- 'img_src': img_src,
- }
- results.append(item)
- result_index += 1
- except JSONDecodeError:
- continue
- return results
|