diff --git a/searx/engines/google.py b/searx/engines/google.py index 4468558a..1c38ba34 100644 --- a/searx/engines/google.py +++ b/searx/engines/google.py @@ -287,7 +287,6 @@ def request(query, params): 'oe': "utf8", 'start': offset, 'filter': '0', - 'ucbcb': 1, **additional_parameters, } ) @@ -299,6 +298,7 @@ def request(query, params): query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]}) params['url'] = query_url + params['cookies']['CONSENT'] = "YES+" params['headers'].update(lang_info['headers']) if use_mobile_ui: params['headers']['Accept'] = '*/*' diff --git a/searx/engines/google_images.py b/searx/engines/google_images.py index b5c7b530..a65c0ce3 100644 --- a/searx/engines/google_images.py +++ b/searx/engines/google_images.py @@ -132,9 +132,7 @@ def request(query, params): + lang_info['subdomain'] + '/search' + "?" - + urlencode( - {'q': query, 'tbm': "isch", **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'num': 30, 'ucbcb': 1} - ) + + urlencode({'q': query, 'tbm': "isch", **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'num': 30}) ) if params['time_range'] in time_range_dict: @@ -143,6 +141,7 @@ def request(query, params): query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]}) params['url'] = query_url + params['cookies']['CONSENT'] = "YES+" params['headers'].update(lang_info['headers']) params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' return params diff --git a/searx/engines/google_news.py b/searx/engines/google_news.py index 2aef517f..0f97f928 100644 --- a/searx/engines/google_news.py +++ b/searx/engines/google_news.py @@ -97,13 +97,12 @@ def request(query, params): + lang_info['subdomain'] + '/search' + "?" - + urlencode( - {'q': query, **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'gl': lang_info['country'], 'ucbcb': 1} - ) + + urlencode({'q': query, **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'gl': lang_info['country']}) + ('&ceid=%s' % ceid) ) # ceid includes a ':' character which must not be urlencoded params['url'] = query_url + params['cookies']['CONSENT'] = "YES+" params['headers'].update(lang_info['headers']) params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' diff --git a/searx/engines/google_play_apps.py b/searx/engines/google_play_apps.py index c80a1f27..226e48da 100644 --- a/searx/engines/google_play_apps.py +++ b/searx/engines/google_play_apps.py @@ -22,11 +22,12 @@ about = { } categories = ["files", "apps"] -search_url = "https://play.google.com/store/search?{query}&c=apps&ucbcb=1" +search_url = "https://play.google.com/store/search?{query}&c=apps" def request(query, params): params["url"] = search_url.format(query=urlencode({"q": query})) + params['cookies']['CONSENT'] = "YES+" return params diff --git a/searx/engines/google_scholar.py b/searx/engines/google_scholar.py index e4c2d727..f9c73097 100644 --- a/searx/engines/google_scholar.py +++ b/searx/engines/google_scholar.py @@ -85,12 +85,13 @@ def request(query, params): + lang_info['subdomain'] + '/scholar' + "?" - + urlencode({'q': query, **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'start': offset, 'ucbcb': 1}) + + urlencode({'q': query, **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'start': offset}) ) query_url += time_range_url(params) params['url'] = query_url + params['cookies']['CONSENT'] = "YES+" params['headers'].update(lang_info['headers']) params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' diff --git a/searx/engines/google_videos.py b/searx/engines/google_videos.py index 3d3f6d93..6eb051e0 100644 --- a/searx/engines/google_videos.py +++ b/searx/engines/google_videos.py @@ -118,7 +118,7 @@ def request(query, params): + lang_info['subdomain'] + '/search' + "?" - + urlencode({'q': query, 'tbm': "vid", **lang_info['params'], 'ie': "utf8", 'oe': "utf8", 'ucbcb': 1}) + + urlencode({'q': query, 'tbm': "vid", **lang_info['params'], 'ie': "utf8", 'oe': "utf8"}) ) if params['time_range'] in time_range_dict: @@ -127,6 +127,7 @@ def request(query, params): query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]}) params['url'] = query_url + params['cookies']['CONSENT'] = "YES+" params['headers'].update(lang_info['headers']) params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' return params diff --git a/searx/engines/youtube_noapi.py b/searx/engines/youtube_noapi.py index 30f89d72..7992adf8 100644 --- a/searx/engines/youtube_noapi.py +++ b/searx/engines/youtube_noapi.py @@ -25,7 +25,7 @@ time_range_support = True # search-url base_url = 'https://www.youtube.com/results' -search_url = base_url + '?search_query={query}&page={page}&ucbcb=1' +search_url = base_url + '?search_query={query}&page={page}' time_range_url = '&sp=EgII{time_range}%253D%253D' # the key seems to be constant next_page_url = 'https://www.youtube.com/youtubei/v1/search?key=AIzaSyAO_FJ2SlqU8Q4STEHLGCilw_Y9_11qcW8' @@ -36,6 +36,7 @@ base_youtube_url = 'https://www.youtube.com/watch?v=' # do search-request def request(query, params): + params['cookies']['CONSENT'] = "YES+" if not params['engine_data'].get('next_page_token'): params['url'] = search_url.format(query=quote_plus(query), page=params['pageno']) if params['time_range'] in time_range_dict: