mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 19:17:07 -05:00 
			
		
		
		
	This patch was generated by black [1]::
    make format.python
[1] https://github.com/psf/black
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
		
	
			
		
			
				
	
	
		
			64 lines
		
	
	
		
			1.6 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			64 lines
		
	
	
		
			1.6 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
						|
"""
 | 
						|
 Mixcloud (Music)
 | 
						|
"""
 | 
						|
 | 
						|
from json import loads
 | 
						|
from dateutil import parser
 | 
						|
from urllib.parse import urlencode
 | 
						|
 | 
						|
# about
 | 
						|
about = {
 | 
						|
    "website": 'https://www.mixcloud.com/',
 | 
						|
    "wikidata_id": 'Q6883832',
 | 
						|
    "official_api_documentation": 'http://www.mixcloud.com/developers/',
 | 
						|
    "use_official_api": True,
 | 
						|
    "require_api_key": False,
 | 
						|
    "results": 'JSON',
 | 
						|
}
 | 
						|
 | 
						|
# engine dependent config
 | 
						|
categories = ['music']
 | 
						|
paging = True
 | 
						|
 | 
						|
# search-url
 | 
						|
url = 'https://api.mixcloud.com/'
 | 
						|
search_url = url + 'search/?{query}&type=cloudcast&limit=10&offset={offset}'
 | 
						|
 | 
						|
embedded_url = (
 | 
						|
    '<iframe scrolling="no" frameborder="0" allowTransparency="true" '
 | 
						|
    + 'data-src="https://www.mixcloud.com/widget/iframe/?feed={url}" width="300" height="300"></iframe>'
 | 
						|
)
 | 
						|
 | 
						|
 | 
						|
# do search-request
 | 
						|
def request(query, params):
 | 
						|
    offset = (params['pageno'] - 1) * 10
 | 
						|
 | 
						|
    params['url'] = search_url.format(query=urlencode({'q': query}), offset=offset)
 | 
						|
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
# get response from search-request
 | 
						|
def response(resp):
 | 
						|
    results = []
 | 
						|
 | 
						|
    search_res = loads(resp.text)
 | 
						|
 | 
						|
    # parse results
 | 
						|
    for result in search_res.get('data', []):
 | 
						|
        title = result['name']
 | 
						|
        url = result['url']
 | 
						|
        content = result['user']['name']
 | 
						|
        embedded = embedded_url.format(url=url)
 | 
						|
        publishedDate = parser.parse(result['created_time'])
 | 
						|
 | 
						|
        # append result
 | 
						|
        results.append(
 | 
						|
            {'url': url, 'title': title, 'embedded': embedded, 'publishedDate': publishedDate, 'content': content}
 | 
						|
        )
 | 
						|
 | 
						|
    # return results
 | 
						|
    return results
 |