mirror of
				https://github.com/searxng/searxng.git
				synced 2025-10-31 10:37:06 -04:00 
			
		
		
		
	
		
			
				
	
	
		
			138 lines
		
	
	
		
			4.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			138 lines
		
	
	
		
			4.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| """
 | |
| Jisho (the Japanese-English dictionary)
 | |
| """
 | |
| 
 | |
| from urllib.parse import urlencode, urljoin
 | |
| 
 | |
| # about
 | |
| about = {
 | |
|     "website": 'https://jisho.org',
 | |
|     "wikidata_id": 'Q24568389',
 | |
|     "official_api_documentation": "https://jisho.org/forum/54fefc1f6e73340b1f160000-is-there-any-kind-of-search-api",
 | |
|     "use_official_api": True,
 | |
|     "require_api_key": False,
 | |
|     "results": 'JSON',
 | |
|     "language": 'ja',
 | |
| }
 | |
| 
 | |
| categories = ['dictionaries']
 | |
| paging = False
 | |
| 
 | |
| URL = 'https://jisho.org'
 | |
| BASE_URL = 'https://jisho.org/word/'
 | |
| SEARCH_URL = URL + '/api/v1/search/words?{query}'
 | |
| 
 | |
| 
 | |
| def request(query, params):
 | |
|     query = urlencode({'keyword': query})
 | |
|     params['url'] = SEARCH_URL.format(query=query)
 | |
|     logger.debug(f"query_url --> {params['url']}")
 | |
|     return params
 | |
| 
 | |
| 
 | |
| def response(resp):
 | |
|     results = []
 | |
|     first_result = True
 | |
| 
 | |
|     search_results = resp.json()
 | |
| 
 | |
|     for page in search_results.get('data', []):
 | |
|         # Entries that are purely from Wikipedia are excluded.
 | |
|         parts_of_speech = page.get('senses') and page['senses'][0].get('parts_of_speech')
 | |
|         if parts_of_speech and parts_of_speech[0] == 'Wikipedia definition':
 | |
|             pass
 | |
| 
 | |
|         # Process alternative forms
 | |
|         alt_forms = []
 | |
|         for title_raw in page['japanese']:
 | |
|             if 'word' not in title_raw:
 | |
|                 alt_forms.append(title_raw['reading'])
 | |
|             else:
 | |
|                 title = title_raw['word']
 | |
|                 if 'reading' in title_raw:
 | |
|                     title += ' (' + title_raw['reading'] + ')'
 | |
|                 alt_forms.append(title)
 | |
| 
 | |
|         result_url = urljoin(BASE_URL, page['slug'])
 | |
|         definitions = get_definitions(page)
 | |
| 
 | |
|         # For results, we'll return the URL, all alternative forms (as title),
 | |
|         # and all definitions (as description) truncated to 300 characters.
 | |
|         content = " ".join(f"{engdef}." for _, engdef, _ in definitions)
 | |
|         results.append(
 | |
|             {'url': result_url, 'title': ", ".join(alt_forms), 'content': content[:300] + (content[300:] and '...')}
 | |
|         )
 | |
| 
 | |
|         # Like Wordnik, we'll return the first result in an infobox too.
 | |
|         if first_result:
 | |
|             first_result = False
 | |
|             results.append(get_infobox(alt_forms, result_url, definitions))
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def get_definitions(page):
 | |
|     # Process definitions
 | |
|     definitions = []
 | |
|     for defn_raw in page['senses']:
 | |
|         extra = []
 | |
|         # Extra data. Since they're not documented, this implementation is based solely by the author's assumptions.
 | |
|         if defn_raw.get('tags'):
 | |
|             if defn_raw.get('info'):
 | |
|                 # "usually written as kana: <kana>"
 | |
|                 extra.append(defn_raw['tags'][0] + ', ' + defn_raw['info'][0] + '. ')
 | |
|             else:
 | |
|                 # abbreviation, archaism, etc.
 | |
|                 extra.append(', '.join(defn_raw['tags']) + '. ')
 | |
|         elif defn_raw.get('info'):
 | |
|             # inconsistent
 | |
|             extra.append(', '.join(defn_raw['info']).capitalize() + '. ')
 | |
|         if defn_raw.get('restrictions'):
 | |
|             extra.append('Only applies to: ' + ', '.join(defn_raw['restrictions']) + '. ')
 | |
|         definitions.append(
 | |
|             (
 | |
|                 ', '.join(defn_raw['parts_of_speech']),
 | |
|                 '; '.join(defn_raw['english_definitions']),
 | |
|                 ''.join(extra)[:-1],
 | |
|             )
 | |
|         )
 | |
|     return definitions
 | |
| 
 | |
| 
 | |
| def get_infobox(alt_forms, result_url, definitions):
 | |
|     infobox_content = []
 | |
|     # title & alt_forms
 | |
|     infobox_title = alt_forms[0]
 | |
|     if len(alt_forms) > 1:
 | |
|         infobox_content.append(f'<p><i>Other forms:</i> {", ".join(alt_forms[1:])}</p>')
 | |
| 
 | |
|     # definitions
 | |
|     infobox_content.append(
 | |
|         '''
 | |
|         <small><a href="https://www.edrdg.org/wiki/index.php/JMdict-EDICT_Dictionary_Project">JMdict</a> 
 | |
|         and <a href="https://www.edrdg.org/enamdict/enamdict_doc.html">JMnedict</a> 
 | |
|         by <a href="https://www.edrdg.org/edrdg/licence.html">EDRDG</a>, CC BY-SA 3.0.</small>
 | |
|         <ul>
 | |
|     '''
 | |
|     )
 | |
|     for pos, engdef, extra in definitions:
 | |
|         if pos == 'Wikipedia definition':
 | |
|             infobox_content.append('</ul><small>Wikipedia, CC BY-SA 3.0.</small><ul>')
 | |
|         pos = f'<i>{pos}</i>: ' if pos else ''
 | |
|         extra = f' ({extra})' if extra else ''
 | |
|         infobox_content.append(f'<li>{pos}{engdef}{extra}</li>')
 | |
|     infobox_content.append('</ul>')
 | |
| 
 | |
|     #
 | |
|     return {
 | |
|         'infobox': infobox_title,
 | |
|         'content': ''.join(infobox_content),
 | |
|         'urls': [
 | |
|             {
 | |
|                 'title': 'Jisho.org',
 | |
|                 'url': result_url,
 | |
|             }
 | |
|         ],
 | |
|     }
 |