123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143 |
- # SPDX-License-Identifier: AGPL-3.0-or-later
- """Flickr (Images)
- """
- from typing import TYPE_CHECKING
- import json
- from time import time
- import re
- from urllib.parse import urlencode
- from searx.utils import ecma_unescape, html_to_text
- if TYPE_CHECKING:
- import logging
- logger: logging.Logger
- # about
- about = {
- "website": 'https://www.flickr.com',
- "wikidata_id": 'Q103204',
- "official_api_documentation": 'https://secure.flickr.com/services/api/flickr.photos.search.html',
- "use_official_api": False,
- "require_api_key": False,
- "results": 'HTML',
- }
- # engine dependent config
- categories = ['images']
- paging = True
- time_range_support = True
- safesearch = False
- time_range_dict = {
- 'day': 60 * 60 * 24,
- 'week': 60 * 60 * 24 * 7,
- 'month': 60 * 60 * 24 * 7 * 4,
- 'year': 60 * 60 * 24 * 7 * 52,
- }
- image_sizes = ('o', 'k', 'h', 'b', 'c', 'z', 'm', 'n', 't', 'q', 's')
- search_url = 'https://www.flickr.com/search?{query}&page={page}'
- time_range_url = '&min_upload_date={start}&max_upload_date={end}'
- photo_url = 'https://www.flickr.com/photos/{userid}/{photoid}'
- modelexport_re = re.compile(r"^\s*modelExport:\s*({.*}),$", re.M)
- def build_flickr_url(user_id, photo_id):
- return photo_url.format(userid=user_id, photoid=photo_id)
- def _get_time_range_url(time_range):
- if time_range in time_range_dict:
- return time_range_url.format(start=time(), end=str(int(time()) - time_range_dict[time_range]))
- return ''
- def request(query, params):
- params['url'] = search_url.format(query=urlencode({'text': query}), page=params['pageno']) + _get_time_range_url(
- params['time_range']
- )
- return params
- def response(resp): # pylint: disable=too-many-branches
- results = []
- matches = modelexport_re.search(resp.text)
- if matches is None:
- return results
- match = matches.group(1)
- model_export = json.loads(match)
- if 'legend' not in model_export:
- return results
- legend = model_export['legend']
- # handle empty page
- if not legend or not legend[0]:
- return results
- for x, index in enumerate(legend):
- if len(index) != 8:
- logger.debug("skip legend enty %s : %s", x, index)
- continue
- photo = model_export['main'][index[0]][int(index[1])][index[2]][index[3]][index[4]][index[5]][int(index[6])][
- index[7]
- ]
- author = ecma_unescape(photo.get('realname', ''))
- source = ecma_unescape(photo.get('username', ''))
- if source:
- source += ' @ Flickr'
- title = ecma_unescape(photo.get('title', ''))
- content = html_to_text(ecma_unescape(photo.get('description', '')))
- img_src = None
- # From the biggest to the lowest format
- size_data = None
- for image_size in image_sizes:
- if image_size in photo['sizes']['data']:
- size_data = photo['sizes']['data'][image_size]['data']
- break
- if not size_data:
- logger.debug('cannot find valid image size: {0}'.format(repr(photo['sizes']['data'])))
- continue
- img_src = size_data['url']
- resolution = f"{size_data['width']} x {size_data['height']}"
- # For a bigger thumbnail, keep only the url_z, not the url_n
- if 'n' in photo['sizes']['data']:
- thumbnail_src = photo['sizes']['data']['n']['data']['url']
- elif 'z' in photo['sizes']['data']:
- thumbnail_src = photo['sizes']['data']['z']['data']['url']
- else:
- thumbnail_src = img_src
- if 'ownerNsid' not in photo:
- # should not happen, disowned photo? Show it anyway
- url = img_src
- else:
- url = build_flickr_url(photo['ownerNsid'], photo['id'])
- result = {
- 'url': url,
- 'img_src': img_src,
- 'thumbnail_src': thumbnail_src,
- 'source': source,
- 'resolution': resolution,
- 'template': 'images.html',
- }
- result['author'] = author.encode(errors='ignore').decode()
- result['source'] = source.encode(errors='ignore').decode()
- result['title'] = title.encode(errors='ignore').decode()
- result['content'] = content.encode(errors='ignore').decode()
- results.append(result)
- return results
|