@@ -126,7 +126,7 @@ def socli(query):
126126 query = urlencode (query )
127127 try :
128128 randomheaders ()
129- search_res = requests .get (soqurl + query , verify = False , headers = header )
129+ search_res = requests .get (soqurl + query , headers = header )
130130 soup = BeautifulSoup (search_res .text , 'html.parser' )
131131 try :
132132 res_url = sourl + (soup .find_all ("div" , class_ = "question-summary" )[0 ].a .get ('href' ))
@@ -207,7 +207,7 @@ def get_questions_for_query(query):
207207 """
208208 questions = []
209209 randomheaders ()
210- search_res = requests .get (soqurl + query , verify = False , headers = header )
210+ search_res = requests .get (soqurl + query , headers = header )
211211 soup = BeautifulSoup (search_res .text , 'html.parser' )
212212 try :
213213 soup .find_all ("div" , class_ = "question-summary" )[0 ] # For explicitly raising exception
@@ -235,7 +235,7 @@ def get_question_stats_and_answer(url):
235235 :return: tuple of ( question_title, question_desc, question_stats, answers )
236236 """
237237 randomheaders ()
238- res_page = requests .get (url , verify = False , headers = header )
238+ res_page = requests .get (url , headers = header )
239239 soup = BeautifulSoup (res_page .text , 'html.parser' )
240240 question_title , question_desc , question_stats = get_stats (soup )
241241 answers = [s .get_text () for s in soup .find_all ("div" , class_ = "post-text" )][1 :] # first post is question, discard it.
@@ -249,7 +249,7 @@ def socli_interactive_windows(query):
249249 :return:
250250 """
251251 try :
252- search_res = requests .get (soqurl + query , verify = False )
252+ search_res = requests .get (soqurl + query )
253253 soup = BeautifulSoup (search_res .text , 'html.parser' )
254254 try :
255255 soup .find_all ("div" , class_ = "question-summary" )[0 ] # For explictly raising exception
@@ -588,7 +588,7 @@ def socl_manusearch(query, rn):
588588 query = urlencode (query )
589589 try :
590590 randomheaders ()
591- search_res = requests .get (soqurl + query , verify = False , headers = header )
591+ search_res = requests .get (soqurl + query , headers = header )
592592 soup = BeautifulSoup (search_res .text , 'html.parser' )
593593 try :
594594 res_url = sourl + (soup .find_all ("div" , class_ = "question-summary" )[rn - 1 ].a .get ('href' ))
@@ -816,7 +816,7 @@ def dispres(url):
816816 :return:
817817 """
818818 randomheaders ()
819- res_page = requests .get (url + query , verify = False , headers = header )
819+ res_page = requests .get (url + query , headers = header )
820820 soup = BeautifulSoup (res_page .text , 'html.parser' )
821821 question_title , question_desc , question_stats = get_stats (soup )
822822
0 commit comments