PS C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project> scrapy crawl maoyan -s LOG_LEVEL=INFO
>>
2025-11-14 19:14:53 [scrapy.utils.log] INFO: Scrapy 2.11.0 started (bot: movie_analysis_project)
2025-11-14 19:14:53 [scrapy.utils.log] INFO: Versions: lxml 5.2.1.0, libxml2 2.13.1, cssselect 1.2.0, parsel 1.8.1, w3lib 2.1.2, Twisted 22.10.0, Python 3.12.7 | packaged by Anaconda, Inc. | (main, Oct 4 2024, 13:17:27) [MSC v.1929 64 bit (AMD64)], pyOpenSSL 24.2.1 (OpenSSL 3.0.15 3 Sep 2024), cryptography 43.0.0, Platform Windows-11-10.0.26100-SP0
2025-11-14 19:14:53 [scrapy.addons] INFO: Enabled addons:
[]
2025-11-14 19:14:53 [py.warnings] WARNING: D:\Anaconda\Lib\site-packages\scrapy\utils\request.py:254: ScrapyDeprecationWarning: '2.6' is a deprecated value for the 'REQUEST_FINGERPRINTER_IMPLEMENTATION' setting.
It is also the default value. In other words, it is normal to get this warning if you have not defined a value for the 'REQUEST_FINGERPRINTER_IMPLEMENTATION' setting. This is so for backward compatibility reasons, but it will change in a future version of Scrapy.
See the documentation of the 'REQUEST_FINGERPRINTER_IMPLEMENTATION' setting for information on how to handle this deprecation.
return cls(crawler)
2025-11-14 19:14:53 [scrapy.extensions.telnet] INFO: Telnet Password: b9dbfee686c0aa3d
2025-11-14 19:14:54 [scrapy.middleware] INFO: Enabled extensions:
['scrapy.extensions.corestats.CoreStats',
'scrapy.extensions.telnet.TelnetConsole',
'scrapy.extensions.logstats.LogStats']
2025-11-14 19:14:54 [scrapy.crawler] INFO: Overridden settings:
{'BOT_NAME': 'movie_analysis_project',
'CONCURRENT_REQUESTS': 1,
'DOWNLOAD_DELAY': 2,
'LOG_LEVEL': 'INFO',
'NEWSPIDER_MODULE': 'movie_analysis_project.spiders',
'RETRY_TIMES': 3,
'SPIDER_MODULES': ['movie_analysis_project.spiders'],
'USER_AGENT': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36'}
2025-11-14 19:14:55 [scrapy.middleware] INFO: Enabled downloader middlewares:
['scrapy.downloadermiddlewares.httpauth.HttpAuthMiddleware',
'scrapy.downloadermiddlewares.downloadtimeout.DownloadTimeoutMiddleware',
'scrapy.downloadermiddlewares.defaultheaders.DefaultHeadersMiddleware',
'scrapy.downloadermiddlewares.useragent.UserAgentMiddleware',
'scrapy.downloadermiddlewares.retry.RetryMiddleware',
'scrapy.downloadermiddlewares.redirect.MetaRefreshMiddleware',
'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware',
'scrapy.downloadermiddlewares.redirect.RedirectMiddleware',
'scrapy.downloadermiddlewares.cookies.CookiesMiddleware',
'scrapy.downloadermiddlewares.httpproxy.HttpProxyMiddleware',
'scrapy.downloadermiddlewares.stats.DownloaderStats']
2025-11-14 19:14:55 [scrapy.middleware] INFO: Enabled spider middlewares:
['scrapy.spidermiddlewares.httperror.HttpErrorMiddleware',
'scrapy.spidermiddlewares.offsite.OffsiteMiddleware',
'scrapy.spidermiddlewares.referer.RefererMiddleware',
'scrapy.spidermiddlewares.urllength.UrlLengthMiddleware',
'scrapy.spidermiddlewares.depth.DepthMiddleware']
Unhandled error in Deferred:
2025-11-14 19:14:58 [twisted] CRITICAL: Unhandled error in Deferred:
Traceback (most recent call last):
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 265, in crawl
return self._crawl(crawler, *args, **kwargs)
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 269, in _crawl
d = crawler.crawl(*args, **kwargs)
File "D:\Anaconda\Lib\site-packages\twisted\internet\defer.py", line 1947, in unwindGenerator
return _cancellableInlineCallbacks(gen)
File "D:\Anaconda\Lib\site-packages\twisted\internet\defer.py", line 1857, in _cancellableInlineCallbacks
_inlineCallbacks(None, gen, status, _copy_context())
--- <exception caught here> ---
File "D:\Anaconda\Lib\site-packages\twisted\internet\defer.py", line 1697, in _inlineCallbacks
result = context.run(gen.send, result)
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 158, in crawl
self.engine = self._create_engine()
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 172, in _create_engine
return ExecutionEngine(self, lambda _: self.stop())
File "D:\Anaconda\Lib\site-packages\scrapy\core\engine.py", line 100, in __init__
self.scraper = Scraper(crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\core\scraper.py", line 109, in __init__
self.itemproc: ItemPipelineManager = itemproc_cls.from_crawler(crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\middleware.py", line 90, in from_crawler
return cls.from_settings(crawler.settings, crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\middleware.py", line 66, in from_settings
mwcls = load_object(clspath)
File "D:\Anaconda\Lib\site-packages\scrapy\utils\misc.py", line 79, in load_object
mod = import_module(module)
File "D:\Anaconda\Lib\importlib\__init__.py", line 90, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1387, in _gcd_import
File "<frozen importlib._bootstrap>", line 1360, in _find_and_load
File "<frozen importlib._bootstrap>", line 1331, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 935, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 995, in exec_module
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
File "C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project\movie_analysis_project\pipelines.py", line 8, in <module>
from .utils.data_cleaner import clean_text, convert_date, extract_gender
builtins.ImportError: cannot import name 'extract_gender' from 'movie_analysis_project.utils.data_cleaner' (C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project\movie_analysis_project\utils\data_cleaner.py)
2025-11-14 19:14:58 [twisted] CRITICAL:
Traceback (most recent call last):
File "D:\Anaconda\Lib\site-packages\twisted\internet\defer.py", line 1697, in _inlineCallbacks
result = context.run(gen.send, result)
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 158, in crawl
self.engine = self._create_engine()
File "D:\Anaconda\Lib\site-packages\scrapy\crawler.py", line 172, in _create_engine
return ExecutionEngine(self, lambda _: self.stop())
File "D:\Anaconda\Lib\site-packages\scrapy\core\engine.py", line 100, in __init__
self.scraper = Scraper(crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\core\scraper.py", line 109, in __init__
self.itemproc: ItemPipelineManager = itemproc_cls.from_crawler(crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\middleware.py", line 90, in from_crawler
return cls.from_settings(crawler.settings, crawler)
File "D:\Anaconda\Lib\site-packages\scrapy\middleware.py", line 66, in from_settings
mwcls = load_object(clspath)
File "D:\Anaconda\Lib\site-packages\scrapy\utils\misc.py", line 79, in load_object
mod = import_module(module)
File "D:\Anaconda\Lib\importlib\__init__.py", line 90, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1387, in _gcd_import
File "<frozen importlib._bootstrap>", line 1360, in _find_and_load
File "<frozen importlib._bootstrap>", line 1331, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 935, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 995, in exec_module
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
File "C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project\movie_analysis_project\pipelines.py", line 8, in <module>
from .utils.data_cleaner import clean_text, convert_date, extract_gender
ImportError: cannot import name 'extract_gender' from 'movie_analysis_project.utils.data_cleaner' (C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project\movie_analysis_project\utils\data_cleaner.py)
PS C:\Users\童琪琪\Desktop\bishe.6\biyesheji.6\movie_analysis_project>
最新发布