pvm*_*lad 2 beautifulsoup web-crawler web-scraping python-3.x etherscan
我正在构建一个网络爬虫,每 30 秒不断刷新一些 etherscan URL,如果发生了任何未考虑到的新传输,它会向我发送一封电子邮件通知和一个指向 etherscan 上相关地址的链接,以便我可以手动检查他们出去。
我想要密切关注的地址之一在这里:
到目前为止我所做的:
from urllib.request import Request, urlopen
url = 'https://etherscan.io/token/0xd6a55c63865affd67e2fb9f284f87b7a9e5ff3bd?a=0x94f52b6520804eced0accad7ccb93c73523af089'
req = Request(url, headers={'User-Agent': 'XYZ/3.0'}) # I got this line from another post since "uClient = uReq(URL)" and "page_html = uClient.read()" would not work (I beleive that etherscan is attemption to block webscraping or something?)
response = urlopen(req, timeout=20).read()
response_close = urlopen(req, timeout=20).close()
page_soup = soup(response, "html.parser")
Transfers_info_table_1 = page_soup.find("div", {"class": "table-responsive"})
print(Transfers_info_table_1)
Run Code Online (Sandbox Code Playgroud)
有趣的是,当我运行它时,我得到以下输出:
<div class="table-responsive" style="visibility:hidden;">
<iframe frameborder="0" id="tokentxnsiframe" scrolling="no" src="" style="width: 100px; height: 600px; min-width: 100%;"></iframe>
</div>
Run Code Online (Sandbox Code Playgroud)
我期望获得整个转账表的输出。我在这里做错了什么?
由于该表存在于iframe.Copy srciframe 的值,然后使用 request 获取该 url 的内容。
from urllib.request import Request, urlopen
from bs4 import BeautifulSoup as soup
import pandas as pd
url = 'https://etherscan.io/token/generic-tokentxns2?m=normal&contractAddress=0xd6a55c63865affd67e2fb9f284f87b7a9e5ff3bd&a=0xd071f6e384cf271282fc37eb40456332307bb8af'
req = Request(url, headers={'User-Agent':'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.106 Safari/537.36'}) # I got this line from another post since "uClient = uReq(URL)" and "page_html = uClient.read()" would not work (I beleive that etherscan is attemption to block webscraping or something?)
response = urlopen(req, timeout=20).read()
response_close = urlopen(req, timeout=20).close()
page_soup = soup(response, "html.parser")
Transfers_info_table_1 = page_soup.find("table", {"class": "table table-md-text-normal table-hover mb-4"})
df=pd.read_html(str(Transfers_info_table_1))[0]
df.to_csv("TransferTable.csv",index=False)
Run Code Online (Sandbox Code Playgroud)
生成的 csv。
| 归档时间: |
|
| 查看次数: |
4466 次 |
| 最近记录: |