Merge branch 'master' into patch-1
This commit is contained in:
		
						commit
						9012ad498c
					
				@ -35,9 +35,12 @@ site_url = 'https://duckduckgo.com/?{query}&iar=images&iax=1&ia=images'
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
# run query in site to get vqd number needed for requesting images
 | 
					# run query in site to get vqd number needed for requesting images
 | 
				
			||||||
# TODO: find a way to get this number without an extra request (is it a hash of the query?)
 | 
					# TODO: find a way to get this number without an extra request (is it a hash of the query?)
 | 
				
			||||||
def get_vqd(query):
 | 
					def get_vqd(query, headers):
 | 
				
			||||||
    res = get(site_url.format(query=urlencode({'q': query})))
 | 
					    query_url = site_url.format(query=urlencode({'q': query}))
 | 
				
			||||||
 | 
					    res = get(query_url, headers=headers)
 | 
				
			||||||
    content = res.text
 | 
					    content = res.text
 | 
				
			||||||
 | 
					    if content.find('vqd=\'') == -1:
 | 
				
			||||||
 | 
					        raise Exception('Request failed')
 | 
				
			||||||
    vqd = content[content.find('vqd=\'') + 5:]
 | 
					    vqd = content[content.find('vqd=\'') + 5:]
 | 
				
			||||||
    vqd = vqd[:vqd.find('\'')]
 | 
					    vqd = vqd[:vqd.find('\'')]
 | 
				
			||||||
    return vqd
 | 
					    return vqd
 | 
				
			||||||
@ -47,7 +50,7 @@ def get_vqd(query):
 | 
				
			|||||||
def request(query, params):
 | 
					def request(query, params):
 | 
				
			||||||
    # to avoid running actual external requests when testing
 | 
					    # to avoid running actual external requests when testing
 | 
				
			||||||
    if 'is_test' not in params:
 | 
					    if 'is_test' not in params:
 | 
				
			||||||
        vqd = get_vqd(query)
 | 
					        vqd = get_vqd(query, params['headers'])
 | 
				
			||||||
    else:
 | 
					    else:
 | 
				
			||||||
        vqd = '12345'
 | 
					        vqd = '12345'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -74,7 +77,7 @@ def response(resp):
 | 
				
			|||||||
    try:
 | 
					    try:
 | 
				
			||||||
        res_json = loads(content)
 | 
					        res_json = loads(content)
 | 
				
			||||||
    except:
 | 
					    except:
 | 
				
			||||||
        return []
 | 
					        raise Exception('Cannot parse results')
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    # parse results
 | 
					    # parse results
 | 
				
			||||||
    for result in res_json['results']:
 | 
					    for result in res_json['results']:
 | 
				
			||||||
 | 
				
			|||||||
@ -11,7 +11,6 @@
 | 
				
			|||||||
"""
 | 
					"""
 | 
				
			||||||
 | 
					
 | 
				
			||||||
from datetime import date, timedelta
 | 
					from datetime import date, timedelta
 | 
				
			||||||
from json import loads
 | 
					 | 
				
			||||||
from lxml import html
 | 
					from lxml import html
 | 
				
			||||||
from searx.url_utils import urlencode, urlparse, parse_qs
 | 
					from searx.url_utils import urlencode, urlparse, parse_qs
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -39,7 +38,6 @@ time_range_dict = {'day': 'd',
 | 
				
			|||||||
# do search-request
 | 
					# do search-request
 | 
				
			||||||
def request(query, params):
 | 
					def request(query, params):
 | 
				
			||||||
    search_options = {
 | 
					    search_options = {
 | 
				
			||||||
        'ijn': params['pageno'] - 1,
 | 
					 | 
				
			||||||
        'start': (params['pageno'] - 1) * number_of_results
 | 
					        'start': (params['pageno'] - 1) * number_of_results
 | 
				
			||||||
    }
 | 
					    }
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@ -53,7 +51,7 @@ def request(query, params):
 | 
				
			|||||||
        search_options['tbs'] = time_range_custom_attr.format(start=start, end=end)
 | 
					        search_options['tbs'] = time_range_custom_attr.format(start=start, end=end)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if safesearch and params['safesearch']:
 | 
					    if safesearch and params['safesearch']:
 | 
				
			||||||
        search_options['safe'] = 'on'
 | 
					        search_options['safe'] = 'active'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    params['url'] = search_url.format(query=urlencode({'q': query}),
 | 
					    params['url'] = search_url.format(query=urlencode({'q': query}),
 | 
				
			||||||
                                      search_options=urlencode(search_options))
 | 
					                                      search_options=urlencode(search_options))
 | 
				
			||||||
@ -63,24 +61,30 @@ def request(query, params):
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
# get response from search-request
 | 
					# get response from search-request
 | 
				
			||||||
def response(resp):
 | 
					def response(resp):
 | 
				
			||||||
    results = []
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
    dom = html.fromstring(resp.text)
 | 
					    dom = html.fromstring(resp.text)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    # parse results
 | 
					    results = []
 | 
				
			||||||
    for img in dom.xpath('//a'):
 | 
					    for element in dom.xpath('//div[@id="search"] //td'):
 | 
				
			||||||
        r = {
 | 
					        link = element.xpath('./a')[0]
 | 
				
			||||||
            'title': u' '.join(img.xpath('.//div[class="rg_ilmbg"]//text()')),
 | 
					
 | 
				
			||||||
 | 
					        google_url = urlparse(link.xpath('.//@href')[0])
 | 
				
			||||||
 | 
					        query = parse_qs(google_url.query)
 | 
				
			||||||
 | 
					        source_url = next(iter(query.get('q', [])), None)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        title_parts = element.xpath('./cite//following-sibling::*/text()')
 | 
				
			||||||
 | 
					        title_parts.extend(element.xpath('./cite//following-sibling::text()')[:-1])
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        result = {
 | 
				
			||||||
 | 
					            'title': ''.join(title_parts),
 | 
				
			||||||
            'content': '',
 | 
					            'content': '',
 | 
				
			||||||
            'template': 'images.html',
 | 
					            'template': 'images.html',
 | 
				
			||||||
 | 
					            'url': source_url,
 | 
				
			||||||
 | 
					            'img_src': source_url,
 | 
				
			||||||
 | 
					            'thumbnail_src': next(iter(link.xpath('.//img //@src')), None)
 | 
				
			||||||
        }
 | 
					        }
 | 
				
			||||||
        url = urlparse(img.xpath('.//@href')[0])
 | 
					 | 
				
			||||||
        query = parse_qs(url.query)
 | 
					 | 
				
			||||||
        r['url'] = query['imgrefurl'][0]
 | 
					 | 
				
			||||||
        r['img_src'] = query['imgurl'][0]
 | 
					 | 
				
			||||||
        r['thumbnail_src'] = r['img_src']
 | 
					 | 
				
			||||||
        # append result
 | 
					 | 
				
			||||||
        results.append(r)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
    # return results
 | 
					        if not source_url or not result['thumbnail_src']:
 | 
				
			||||||
 | 
					            continue
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        results.append(result)
 | 
				
			||||||
    return results
 | 
					    return results
 | 
				
			||||||
 | 
				
			|||||||
@ -41,7 +41,7 @@ class TestDuckduckgoImagesEngine(SearxTestCase):
 | 
				
			|||||||
        self.assertRaises(AttributeError, duckduckgo_images.response, '[]')
 | 
					        self.assertRaises(AttributeError, duckduckgo_images.response, '[]')
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        response = mock.Mock(text='If this error persists, please let us know: ops@duckduckgo.com')
 | 
					        response = mock.Mock(text='If this error persists, please let us know: ops@duckduckgo.com')
 | 
				
			||||||
        self.assertEqual(duckduckgo_images.response(response), [])
 | 
					        self.assertRaises(Exception, duckduckgo_images.response, response)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
        json = u"""
 | 
					        json = u"""
 | 
				
			||||||
        {
 | 
					        {
 | 
				
			||||||
 | 
				
			|||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user