2018-05-16 16:21:39 +02:00
|
|
|
from argparse import ArgumentParser
|
2018-05-15 21:57:54 +02:00
|
|
|
from bs4 import BeautifulSoup
|
2018-05-17 16:20:28 +02:00
|
|
|
from enum import Enum
|
|
|
|
|
from sys import stderr
|
|
|
|
|
from typing import Iterator, NewType
|
2018-05-15 21:57:54 +02:00
|
|
|
import re
|
2018-05-16 16:22:33 +02:00
|
|
|
import requests
|
2018-05-15 21:57:54 +02:00
|
|
|
|
2018-05-17 16:20:28 +02:00
|
|
|
Url = NewType("Url", str)
|
2018-05-16 17:08:58 +02:00
|
|
|
class Language(Enum):
|
|
|
|
|
EN = "en"
|
|
|
|
|
DE = "de"
|
2018-05-15 21:57:54 +02:00
|
|
|
|
2018-05-16 17:08:58 +02:00
|
|
|
DEFAULT_LANG = Language.DE
|
|
|
|
|
|
|
|
|
|
def cook_soup(url: Url) -> BeautifulSoup:
|
2018-05-15 21:57:54 +02:00
|
|
|
return BeautifulSoup(requests.get(url).text, "lxml")
|
|
|
|
|
|
2018-05-17 10:02:24 +02:00
|
|
|
def base_url(lang: Language = DEFAULT_LANG) -> Url:
|
|
|
|
|
return Url("https://{}.wiktionary.org".format(lang.value))
|
2018-05-15 21:57:54 +02:00
|
|
|
|
2018-05-17 10:02:24 +02:00
|
|
|
def entry_url(word: str, lang: Language = DEFAULT_LANG) -> Url:
|
|
|
|
|
return Url(base_url(lang) + "/wiki/" + word)
|
2018-05-15 21:57:54 +02:00
|
|
|
|
2018-05-17 10:02:24 +02:00
|
|
|
def rhymes_url(entry_url: Url, lang: Language = DEFAULT_LANG) -> Url:
|
2018-05-16 16:23:49 +02:00
|
|
|
soup = cook_soup(entry_url)
|
2018-05-15 21:57:54 +02:00
|
|
|
result_url = base_url(lang)
|
2018-05-17 16:20:28 +02:00
|
|
|
if lang == Language.DE:
|
|
|
|
|
pattern = re.compile(r"/wiki/Reim:Deutsch:.*")
|
|
|
|
|
elif lang == Language.EN:
|
|
|
|
|
pattern = re.compile(r"/wiki/Rhymes:English/.*")
|
|
|
|
|
anchor = soup.find("a", href=pattern)
|
|
|
|
|
if anchor is not None:
|
|
|
|
|
return result_url + anchor["href"]
|
|
|
|
|
else:
|
2018-05-16 16:22:33 +02:00
|
|
|
raise ValueError("Entry at {} not found.".format(entry_url))
|
|
|
|
|
|
2018-05-16 17:08:58 +02:00
|
|
|
def find_rhymes(rhymes_url: Url) -> Iterator[str]:
|
2018-05-16 16:23:49 +02:00
|
|
|
soup = cook_soup(rhymes_url)
|
2018-05-15 21:57:54 +02:00
|
|
|
for li in soup.select("div#content ul > li > a"):
|
2018-05-17 16:20:28 +02:00
|
|
|
yield li.text
|
2018-05-15 21:57:54 +02:00
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
2018-05-16 16:21:39 +02:00
|
|
|
parser = ArgumentParser(description="Find your rhyme in no time")
|
|
|
|
|
parser.add_argument("entry", metavar="WORD", type=str, help="the word to rhyme")
|
|
|
|
|
parser.add_argument("-l", "--lang", help="the language", choices=["de", "en"], nargs="?", default="de")
|
|
|
|
|
args = parser.parse_args()
|
2018-05-17 16:20:28 +02:00
|
|
|
try:
|
|
|
|
|
entry = entry_url(args.entry, lang=Language(args.lang))
|
|
|
|
|
rhymes = rhymes_url(entry, lang=Language(args.lang))
|
|
|
|
|
for rhyme in find_rhymes(rhymes):
|
|
|
|
|
print(rhyme)
|
|
|
|
|
except ValueError as error:
|
|
|
|
|
stderr.write("{}\n".format(error))
|