Panda's Web table with Python scratching

My problem is that I need to create a web scraping of the table included in the following link: https://drive.google.com/drive/folders/1NzgjsD-fIPkVLjFMfu0bBpruwFoZd-LM
You have to download the file.

I could only print a small part of the table
with this code: import pandas as pd

def Read_html ():
Try:
table = pd.read_html ("file: /// C: /Users/ronye/Downloads/fuenteDeDatos.html")
for df in dfs:
print (df)
for tr in df:
Pressure (tr)
except:
print ("Error: no internet, connection loss")

Also, I need an idea of ​​how to store the value of each part of the table. Example: Initial_value = 24000000, saving = 400000, years = 5

Bing scratching without a substitute

Hello!

I could use some advice …

I would like to talk about the possibility of bing scratching without a substitute. To be honest, I have never been able to use the proxy scraper, and I can not justify $ 10 / million for shared proxies, but on the other hand, deathbycaptcha is very affordable. Would not it be better to just let Deathbycaptcha solve a few captchas when they pop up, then buy proxies … especially for someone like me who does not do this every month?

What would be an acceptable delay setting for Bing's scaping with such a setting? Or are deputies an absolute must? I mean … I can call bing from my browser and manually download a serial list without any problems. So it must be safe to use the Scrapebox without a proxy server, right? Or can bing say that I use software?

Many Thanks!

Python – Is there a way to speed up the scratching process?

The web scraping process is going on for quite a while now, and I'm wondering if I can structure or improve the code differently.

Code looks like this:

import numpy as np
Import pandas as pd
import requests
Import Json
from the sklearn import preprocessing
From sklearn.preprocessing import OneHotEncoder

results_2017 = []
results_2018 = []
for game_id in range (2017020001, 2017021271, 1):
url = & # 39; https: //statsapi.web.nhl.com/api/v1/game/ {} /boxscore'.format (game_id)
r_2017 = request.get (URL)
game_data_2017 = r_2017.json ()

for home away in ['home','away']:

game_dict_2017 = game_data_2017.get (& # 39; teams & # 39;). get (home theater) .get (& # 39; teamStats & # 39;). get (& # 39; teamSkaterStats & # 39;)
game_dict_2017['team'] = game_data_2017.get (& # 39; teams & # 39;). get (home theater) .get (& # 39; team & # 39;). get (& # 39; name & # 39;)
game_dict_2017['homeaway'] = Return home
game_dict_2017['game_id'] = game_id
results_2017.append (game_dict_2017)

df_2017 = pd.DataFrame (results_2017)

for game_id in range (2018020001, 2018020667, 1):
url = & # 39; https: //statsapi.web.nhl.com/api/v1/game/ {} /boxscore'.format (game_id)
r_2018 = request.get (URL)
game_data_2018 = r_2018.json ()

for home away in ['home','away']:

game_dict_2018 = game_data_2018.get (& # 39; teams & # 39;). get (home theater) .get (& # 39; teamStats & # 39;). get (& # 39; teamSkaterStats & # 39;)
game_dict_2018['team'] = game_data_2018.get (& # 39; teams & # 39;). get (home theater) .get (& # 39; team & # 39;). get (& # 39; name & # 39;)
game_dict_2018['homeaway'] = Return home
game_dict_2018['game_id'] = game_id
results_2018.append (game_dict_2018)


df_2018 = pd.DataFrame (results_2018)

df = df_2017.append (df_2018)

Scratching porn links?

Search for the KW porn in the title in the forum, but it seems to be no topic about it.

In another forum, I read a post from 2013 where someone used scrapbox links from porn sites using certain KWs.

I searched the internet for the last hour, but I can not really figure out how to do it.

Any experienced user can help me?

Someone

Many Thanks

Some sites scratching proxies

Some sources for proxies .. Some can already be added in SER 

http://www.moneyfanclub.com/proxy-forum-free-proxies/*
http://www.pr0xies.org/
http://www.pr0xies.org/2013/01/*
http://www.proxyfire.net/forum/showthread.php?p=
http://www.proxz.com/proxy_list_anonymous_us_0.html
http://www.proxz.com/proxy_list_anonymous_us_0_ext.html
http://www.proxz.com/proxy_list_fr_0.html
http://www.proxz.com/proxy_list_high_anonymous_0_ext.html
http://www.proxz.com/proxy_list_transparent_0.html
http://www.proxz.com/proxy_list_transparent_0_ext.html
http://www.proxz.com/proxy_list_transparent_175.html
http://www.proxz.com/proxy_list_uk_0.html
http://www.proxz.com/proxy_list_uk_0_ext.html
http://www.robotproxy.com/2013/01/*
http://www.ruzgarforum.com/guncel-scrapebox-proxy-servers/*
http://www.scrapeboxproxies.net/
lawofmoney.net/high-anonymous-09-11.html
seoblackhat.us/category/proxy-lists/
reviewseotools.com/
get-proxy.ru/tag/anonymous-proxy
yozgatfrm.com/anonymous-proxy/*
getfreeproxy.com/tag/elile-proxy
cool-proxy.ru/category/anonymous-proxy
freeproxy.org.ua/
world34.blogspot.com/
freedom.2surf.org/
4proxy.org.ua/tag/anonymous-proxy
elite-proxy.ru/
2proxy.org/tag/anonymous-proxy
proxy-hunter.blogspot.com/
thebotnet.com/freebies/
anon-proxy.ru/tag/elile-proxy
good-proxy.ru/
freeproxilist.info/
yahooaction.com/forum/http-proxy/*
alwinclores.com/
checkerproxy.net/ru/*
best-proxy.ru/
thebigproxylist.com/blog/
scribd.com/doc/72131887/Proxy-List
proxyvadi.net/category/irc-proxy
anonymousblog.info/
passwordscastle.net/showthread.php?124878-09-11-Fresh-proxy-List&goto=newpost
proxy-heaven.blogspot.com/
a1-warez.com/showthread.php?p=30983
elite-proxies.blogspot.com/

Scratching porn links?

Look for the KW porn in the title in the forum, but there seems to be no topic.

In another forum I read a post from 2013 where someone uses scrapbox links from porn sites that use certain KW.

I searched the internet for the last hour, but I can not really figure out how to do it.

Any experienced user can help me?

Someone

Many Thanks