New URL with robots.txt
This commit is contained in:
parent
6df9e32d8f
commit
1d5283d382
4
utils.py
4
utils.py
@ -125,13 +125,13 @@ def get_fileio(date: DTDate, card_type: List[str] = CARD_VISA) -> BinaryIO: # py
|
||||
# ua = 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:64.0) Gecko/20100101 Firefox/64.0'
|
||||
# b.set_header('User-Agent', ua)
|
||||
# Ignore robots.txt
|
||||
# b.set_handle_robots(False)
|
||||
b.set_handle_robots(False)
|
||||
# Debugging flags
|
||||
# b.set_debug_http(True)
|
||||
# b.set_debug_redirects(True)
|
||||
# b.set_debug_responses(True)
|
||||
# PDF URL
|
||||
b.open('https://misc.firstdata.eu/CurrencyCalculator/fremdwaehrungskurse/pdf')
|
||||
b.open('https://online.firstdata.com/CurrencyCalculator/fremdwaehrungskurse/pdf')
|
||||
fm = b.forms()[0]
|
||||
# This must be done because I can't change the options otherwise
|
||||
fm.set_all_readonly(False)
|
||||
|
Loading…
Reference in New Issue
Block a user