summaryrefslogtreecommitdiff
path: root/searx/engines/google_images.py
diff options
context:
space:
mode:
Diffstat (limited to 'searx/engines/google_images.py')
-rw-r--r--searx/engines/google_images.py95
1 files changed, 95 insertions, 0 deletions
diff --git a/searx/engines/google_images.py b/searx/engines/google_images.py
new file mode 100644
index 0000000..9692f4b
--- /dev/null
+++ b/searx/engines/google_images.py
@@ -0,0 +1,95 @@
+"""
+ Google (Images)
+
+ @website https://www.google.com
+ @provide-api yes (https://developers.google.com/custom-search/)
+
+ @using-api no
+ @results HTML chunks with JSON inside
+ @stable no
+ @parse url, title, img_src
+"""
+
+from datetime import date, timedelta
+from json import loads
+from lxml import html
+from searx.url_utils import urlencode
+
+
+# engine dependent config
+categories = ['images']
+paging = True
+safesearch = True
+time_range_support = True
+number_of_results = 100
+
+search_url = 'https://www.google.com/search'\
+ '?{query}'\
+ '&asearch=ichunk'\
+ '&async=_id:rg_s,_pms:s'\
+ '&tbm=isch'\
+ '&yv=2'\
+ '&{search_options}'
+time_range_attr = "qdr:{range}"
+time_range_custom_attr = "cdr:1,cd_min:{start},cd_max{end}"
+time_range_dict = {'day': 'd',
+ 'week': 'w',
+ 'month': 'm'}
+
+
+# do search-request
+def request(query, params):
+ search_options = {
+ 'ijn': params['pageno'] - 1,
+ 'start': (params['pageno'] - 1) * number_of_results
+ }
+
+ if params['time_range'] in time_range_dict:
+ search_options['tbs'] = time_range_attr.format(range=time_range_dict[params['time_range']])
+ elif params['time_range'] == 'year':
+ now = date.today()
+ then = now - timedelta(days=365)
+ start = then.strftime('%m/%d/%Y')
+ end = now.strftime('%m/%d/%Y')
+ search_options['tbs'] = time_range_custom_attr.format(start=start, end=end)
+
+ if safesearch and params['safesearch']:
+ search_options['safe'] = 'on'
+
+ params['url'] = search_url.format(query=urlencode({'q': query}),
+ search_options=urlencode(search_options))
+
+ return params
+
+
+# get response from search-request
+def response(resp):
+ results = []
+
+ g_result = loads(resp.text)
+
+ dom = html.fromstring(g_result[1][1])
+
+ # parse results
+ for result in dom.xpath('//div[@data-ved]'):
+
+ try:
+ metadata = loads(''.join(result.xpath('./div[@class="rg_meta"]/text()')))
+ except:
+ continue
+
+ thumbnail_src = metadata['tu']
+
+ # http to https
+ thumbnail_src = thumbnail_src.replace("http://", "https://")
+
+ # append result
+ results.append({'url': metadata['ru'],
+ 'title': metadata['pt'],
+ 'content': metadata['s'],
+ 'thumbnail_src': thumbnail_src,
+ 'img_src': metadata['ou'],
+ 'template': 'images.html'})
+
+ # return results
+ return results