问题:
list_links_docs = []
soup = get_link_decode(URL)
for link in range(10):
link = soup.find_all("a",text="{}".format(link))
list_links_docs.append(link)
for link in list_links_docs:
lis = link.get("href")
print(lis)
执行这个程序出现错误
Traceback (most recent call last):
File "D:\APP\PyCharm 2018.1.4\helpers\pydev\pydevd.py", line 1664, in <module>
main()
File "D:\APP\PyCharm 2018.1.4\helpers\pydev\pydevd.py", line 1658, in main
globals = debugger.run(setup['file'], None, None, is_module)
File "D:\APP\PyCharm 2018.1.4\helpers\pydev\pydevd.py", line 1068, in run
pydev_imports.execfile(file, globals, locals) # execute the script
File "D:\APP\PyCharm 2018.1.4\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile
exec(compile(contents+"\n", file, 'exec'), glob, loc)
File "D:/PyDate/Climb_CTOLib.com/Android_Climb2.py", line 73, in <module>
lis = link.get("href")
File "D:\APP\Anaconda3\lib\site-packages\bs4\element.py", line 1884, in __getattr__
"ResultSet object has no attribute '%s'. You're probably treating a list of items like a single item. Did you call find_all() when you meant to call find()?" % key
AttributeError: ResultSet object has no attribute 'get'. You're probably treating a list of items like a single item. Did you call find_all() when you meant to call find()?
原因:
在后面get的时候,因为前面取a标签的时用了find_all方法,返回了一个ResultSet(结果集),导致出错。
下面是获取到的列表
[[], [<a href="/android/docs/android-pg-1.html">1</a>], [<a href="/android/docs/android-pg-2.html">2</a>], [], [], [], [], [], [], []]
解决:
不用find_all()选择器,换成find()选择器
更改后:
list_links_docs = []
soup = get_link_decode(URL)
for link in range(10):
link = soup.find("a",text="{}".format(link))
list_links_docs.append(link)
for link in list_links_docs:
lis = link.get("href")
print(lis)
07022004改
list_links_docs = []
soup = get_link_decode(URL)
for link in range(10):
link = soup.find("a",text="{}".format(link))
lis = link.get("href")
list_links_docs.append(lis)

本文详细解析了使用Python的BeautifulSoup库抓取网页链接时出现的常见错误——将find_all误用为find,导致从单个元素的属性获取变为从列表中获取。通过示例代码展示了如何正确地使用find方法定位单个元素并提取其href属性。
4420

被折叠的 条评论
为什么被折叠?



