@@ -53,7 +53,7 @@ class Driver:
5353 vulnerable_requests = []
5454
5555 def __init__ (self , uri , verify_exploit , use_crawler , quit_if_vulnerable ):
56- Logging .info ("Started scan" );
56+ Logging .info ("Started scan" )
5757
5858 self .input_uri = uri
5959 self .input_verify_exploit = verify_exploit
@@ -74,7 +74,7 @@ def __init__(self, uri, verify_exploit, use_crawler, quit_if_vulnerable):
7474 self .crawler_options .performance .max_threads = 12
7575
7676 try :
77- self .website_details = Scraper .get_details (self .input_uri );
77+ self .website_details = Scraper .get_details (self .input_uri )
7878 except Exception as e :
7979 Logging .red ("Error while scraping URL '{}': {}" .format (self .input_uri , str (e )))
8080 sys .exit (1 )
@@ -92,7 +92,7 @@ def __init__(self, uri, verify_exploit, use_crawler, quit_if_vulnerable):
9292 sys .exit (1 )
9393
9494 def cb_crawler_before_start (self ):
95- Logging .info ("Started crawler" );
95+ Logging .info ("Started crawler" )
9696
9797 def cb_crawler_after_finish (self , queue ):
9898 Logging .info ("Found {} vulnerable URI(s)" .format (len (self .vulnerable_requests )))
@@ -103,7 +103,7 @@ def cb_request_before_start(self, queue, queue_item):
103103 result = Exploit .is_vulnerable (queue_item , self .website_details ["angular_version" ], self .input_verify_exploit )
104104
105105 if result is not False :
106- self .vulnerable_requests .append (result );
106+ self .vulnerable_requests .append (result )
107107 Logging .red ("Request is vulnerable [" + result .request .method + "] " + result .request .url + " (PostData: " + str (result .request .data ) + ")" )
108108
109109 if self .input_quit_if_vulnerable :
@@ -112,4 +112,4 @@ def cb_request_before_start(self, queue, queue_item):
112112 return CrawlerActions .DO_CONTINUE_CRAWLING
113113
114114 def cb_request_after_finish (self , queue , queue_item , new_queue_items ):
115- return CrawlerActions .DO_CONTINUE_CRAWLING
115+ return CrawlerActions .DO_CONTINUE_CRAWLING
0 commit comments