diff options
Diffstat (limited to 'searx/engines/google_images.py')
-rw-r--r-- | searx/engines/google_images.py | 95 |
1 files changed, 95 insertions, 0 deletions
diff --git a/searx/engines/google_images.py b/searx/engines/google_images.py new file mode 100644 index 0000000..9692f4b --- /dev/null +++ b/searx/engines/google_images.py @@ -0,0 +1,95 @@ +""" + Google (Images) + + @website https://www.google.com + @provide-api yes (https://developers.google.com/custom-search/) + + @using-api no + @results HTML chunks with JSON inside + @stable no + @parse url, title, img_src +""" + +from datetime import date, timedelta +from json import loads +from lxml import html +from searx.url_utils import urlencode + + +# engine dependent config +categories = ['images'] +paging = True +safesearch = True +time_range_support = True +number_of_results = 100 + +search_url = 'https://www.google.com/search'\ + '?{query}'\ + '&asearch=ichunk'\ + '&async=_id:rg_s,_pms:s'\ + '&tbm=isch'\ + '&yv=2'\ + '&{search_options}' +time_range_attr = "qdr:{range}" +time_range_custom_attr = "cdr:1,cd_min:{start},cd_max{end}" +time_range_dict = {'day': 'd', + 'week': 'w', + 'month': 'm'} + + +# do search-request +def request(query, params): + search_options = { + 'ijn': params['pageno'] - 1, + 'start': (params['pageno'] - 1) * number_of_results + } + + if params['time_range'] in time_range_dict: + search_options['tbs'] = time_range_attr.format(range=time_range_dict[params['time_range']]) + elif params['time_range'] == 'year': + now = date.today() + then = now - timedelta(days=365) + start = then.strftime('%m/%d/%Y') + end = now.strftime('%m/%d/%Y') + search_options['tbs'] = time_range_custom_attr.format(start=start, end=end) + + if safesearch and params['safesearch']: + search_options['safe'] = 'on' + + params['url'] = search_url.format(query=urlencode({'q': query}), + search_options=urlencode(search_options)) + + return params + + +# get response from search-request +def response(resp): + results = [] + + g_result = loads(resp.text) + + dom = html.fromstring(g_result[1][1]) + + # parse results + for result in dom.xpath('//div[@data-ved]'): + + try: + metadata = loads(''.join(result.xpath('./div[@class="rg_meta"]/text()'))) + except: + continue + + thumbnail_src = metadata['tu'] + + # http to https + thumbnail_src = thumbnail_src.replace("http://", "https://") + + # append result + results.append({'url': metadata['ru'], + 'title': metadata['pt'], + 'content': metadata['s'], + 'thumbnail_src': thumbnail_src, + 'img_src': metadata['ou'], + 'template': 'images.html'}) + + # return results + return results |