mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 19:17:07 -05:00 
			
		
		
		
	[feat] engine: implementation of Anna's Archive
Anna's Archive [1] is a free non-profit online shadow library metasearch engine providing access to a variety of book resources (also via IPFS), created by a team of anonymous archivists [2]. [1] https://annas-archive.org/ [2] https://annas-software.org/AnnaArchivist/annas-archive
This commit is contained in:
		
							parent
							
								
									fd26f37073
								
							
						
					
					
						commit
						e5637fe7b9
					
				
							
								
								
									
										63
									
								
								searx/engines/annas-archive.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										63
									
								
								searx/engines/annas-archive.py
									
									
									
									
									
										Normal file
									
								
							@ -0,0 +1,63 @@
 | 
				
			|||||||
 | 
					# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
				
			||||||
 | 
					# lint: pylint
 | 
				
			||||||
 | 
					"""Anna's Archive
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					"""
 | 
				
			||||||
 | 
					from typing import List, Dict, Any, Optional
 | 
				
			||||||
 | 
					from urllib.parse import quote
 | 
				
			||||||
 | 
					from lxml import html
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					from searx.utils import extract_text, eval_xpath
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# about
 | 
				
			||||||
 | 
					about: Dict[str, Any] = {
 | 
				
			||||||
 | 
					    "website": "https://annas-archive.org/",
 | 
				
			||||||
 | 
					    "wikidata_id": "Q115288326",
 | 
				
			||||||
 | 
					    "official_api_documentation": None,
 | 
				
			||||||
 | 
					    "use_official_api": False,
 | 
				
			||||||
 | 
					    "require_api_key": False,
 | 
				
			||||||
 | 
					    "results": "HTML",
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# engine dependent config
 | 
				
			||||||
 | 
					categories: List[str] = ["files"]
 | 
				
			||||||
 | 
					paging: bool = False
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# search-url
 | 
				
			||||||
 | 
					base_url: str = "https://annas-archive.org"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					# xpath queries
 | 
				
			||||||
 | 
					xpath_results: str = '//main//a[starts-with(@href,"/md5")]'
 | 
				
			||||||
 | 
					xpath_url: str = ".//@href"
 | 
				
			||||||
 | 
					xpath_title: str = ".//h3/text()[1]"
 | 
				
			||||||
 | 
					xpath_authors: str = './/div[contains(@class, "italic")]'
 | 
				
			||||||
 | 
					xpath_publisher: str = './/div[contains(@class, "text-sm")]'
 | 
				
			||||||
 | 
					xpath_file_info: str = './/div[contains(@class, "text-xs")]'
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					def request(query, params: Dict[str, Any]) -> Dict[str, Any]:
 | 
				
			||||||
 | 
					    search_url: str = base_url + "/search?q={search_query}"
 | 
				
			||||||
 | 
					    params["url"] = search_url.format(search_query=quote(query))
 | 
				
			||||||
 | 
					    return params
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					def response(resp) -> List[Dict[str, Optional[str]]]:
 | 
				
			||||||
 | 
					    results: List[Dict[str, Optional[str]]] = []
 | 
				
			||||||
 | 
					    dom = html.fromstring(resp.text)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    for item in dom.xpath(xpath_results):
 | 
				
			||||||
 | 
					        result: Dict[str, Optional[str]] = {}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        result["url"] = base_url + item.xpath(xpath_url)[0]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        result["title"] = extract_text(eval_xpath(item, xpath_title))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        result["content"] = "{publisher}. {authors}. {file_info}".format(
 | 
				
			||||||
 | 
					            authors=extract_text(eval_xpath(item, xpath_authors)),
 | 
				
			||||||
 | 
					            publisher=extract_text(eval_xpath(item, xpath_publisher)),
 | 
				
			||||||
 | 
					            file_info=extract_text(eval_xpath(item, xpath_file_info)),
 | 
				
			||||||
 | 
					        )
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					        results.append(result)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					    return results
 | 
				
			||||||
@ -297,6 +297,13 @@ engines:
 | 
				
			|||||||
    shortcut: 9g
 | 
					    shortcut: 9g
 | 
				
			||||||
    disabled: true
 | 
					    disabled: true
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					  - name: anna's archive
 | 
				
			||||||
 | 
					    engine: annas-archive
 | 
				
			||||||
 | 
					    paging: False
 | 
				
			||||||
 | 
					    categories: files
 | 
				
			||||||
 | 
					    disabled: true
 | 
				
			||||||
 | 
					    shortcut: aa
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  - name: apk mirror
 | 
					  - name: apk mirror
 | 
				
			||||||
    engine: apkmirror
 | 
					    engine: apkmirror
 | 
				
			||||||
    timeout: 4.0
 | 
					    timeout: 4.0
 | 
				
			||||||
 | 
				
			|||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user