New URL with robots.txt
This commit is contained in:
		
							
								
								
									
										4
									
								
								utils.py
									
									
									
									
									
								
							
							
						
						
									
										4
									
								
								utils.py
									
									
									
									
									
								
							@@ -125,13 +125,13 @@ def get_fileio(date: DTDate, card_type: List[str] = CARD_VISA) -> BinaryIO: # py
 | 
			
		||||
    # ua = 'Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:64.0) Gecko/20100101 Firefox/64.0'
 | 
			
		||||
    # b.set_header('User-Agent', ua)
 | 
			
		||||
    # Ignore robots.txt
 | 
			
		||||
    # b.set_handle_robots(False)
 | 
			
		||||
    b.set_handle_robots(False)
 | 
			
		||||
    # Debugging flags
 | 
			
		||||
    # b.set_debug_http(True)
 | 
			
		||||
    # b.set_debug_redirects(True)
 | 
			
		||||
    # b.set_debug_responses(True)
 | 
			
		||||
    # PDF URL
 | 
			
		||||
    b.open('https://misc.firstdata.eu/CurrencyCalculator/fremdwaehrungskurse/pdf')
 | 
			
		||||
    b.open('https://online.firstdata.com/CurrencyCalculator/fremdwaehrungskurse/pdf')
 | 
			
		||||
    fm = b.forms()[0]
 | 
			
		||||
    # This must be done because I can't change the options otherwise
 | 
			
		||||
    fm.set_all_readonly(False)
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user