2026-05-13 08:03:14 [scrapy.utils.log] INFO: Scrapy 2.14.1 started (bot: priceWatchBot) 2026-05-13 08:03:15 [scrapy.utils.log] INFO: Versions: {'lxml': '6.0.2', 'libxml2': '2.14.6', 'cssselect': '1.4.0', 'parsel': '1.11.0', 'w3lib': '2.4.0', 'Twisted': '25.5.0', 'Python': '3.13.1 (main, Jan 8 2025, 18:18:10) [GCC 14.2.0]', 'pyOpenSSL': '25.3.0 (OpenSSL 3.5.5 27 Jan 2026)', 'cryptography': '46.0.5', 'Platform': 'Linux-6.1.0-44-cloud-amd64-x86_64-with'} 2026-05-13 08:03:15 [priceWatchBotSpider] INFO: ID de l'URL : 1275 2026-05-13 08:03:15 [scrapy.addons] INFO: Enabled addons: [] 2026-05-13 08:03:15 [scrapy.extensions.telnet] INFO: Telnet Password: 7712f0b2f8722806 2026-05-13 08:03:15 [scrapy.middleware] INFO: Enabled extensions: ['scrapy.extensions.corestats.CoreStats', 'scrapy.extensions.logcount.LogCount', 'scrapy.extensions.telnet.TelnetConsole', 'scrapy.extensions.memusage.MemoryUsage', 'scrapy.extensions.logstats.LogStats'] 2026-05-13 08:03:15 [scrapy.crawler] INFO: Overridden settings: {'BOT_NAME': 'priceWatchBot', 'COOKIES_ENABLED': False, 'DOWNLOAD_DELAY': 3, 'FEED_EXPORT_ENCODING': 'utf-8', 'LOG_FILE': '/logs/priceWatchBot/priceWatchBotSpider/2d09a8c44ea211f1ad920242ac120003.log', 'LOG_LEVEL': 'INFO', 'NEWSPIDER_MODULE': 'priceWatchBot.spiders', 'SPIDER_MODULES': ['priceWatchBot.spiders']} 2026-05-13 08:03:15 [scrapy.middleware] INFO: Enabled downloader middlewares: ['scrapy.downloadermiddlewares.offsite.OffsiteMiddleware', 'scrapy.downloadermiddlewares.httpauth.HttpAuthMiddleware', 'scrapy.downloadermiddlewares.downloadtimeout.DownloadTimeoutMiddleware', 'scrapy.downloadermiddlewares.defaultheaders.DefaultHeadersMiddleware', 'scrapy.downloadermiddlewares.useragent.UserAgentMiddleware', 'scrapy.downloadermiddlewares.retry.RetryMiddleware', 'scrapy.downloadermiddlewares.redirect.MetaRefreshMiddleware', 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware', 'scrapy.downloadermiddlewares.redirect.RedirectMiddleware', 'scrapy.downloadermiddlewares.httpproxy.HttpProxyMiddleware', 'scrapy.downloadermiddlewares.stats.DownloaderStats'] 2026-05-13 08:03:15 [scrapy.middleware] INFO: Enabled spider middlewares: ['scrapy.spidermiddlewares.start.StartSpiderMiddleware', 'scrapy.spidermiddlewares.httperror.HttpErrorMiddleware', 'scrapy.spidermiddlewares.referer.RefererMiddleware', 'scrapy.spidermiddlewares.urllength.UrlLengthMiddleware', 'scrapy.spidermiddlewares.depth.DepthMiddleware'] 2026-05-13 08:03:15 [scrapy.middleware] INFO: Enabled item pipelines: ['priceWatchBot.pipelines.PricewatchbotPipeline'] 2026-05-13 08:03:15 [py.warnings] WARNING: /usr/local/lib/python3.13/site-packages/scrapy/pipelines/__init__.py:47: ScrapyDeprecationWarning: PricewatchbotPipeline.process_item() requires a spider argument, this is deprecated and the argument will not be passed in future Scrapy versions. If you need to access the spider instance you can save the crawler instance passed to from_crawler() and use its spider attribute. self._check_mw_method_spider_arg(pipe.process_item) 2026-05-13 08:03:15 [scrapy.core.engine] INFO: Spider opened 2026-05-13 08:03:15 [py.warnings] WARNING: /usr/local/lib/python3.13/site-packages/scrapy/core/spidermw.py:490: ScrapyDeprecationWarning: priceWatchBot.spiders.priceWatchBotSpider.PricewatchbotspiderSpider defines the deprecated start_requests() method. start_requests() has been deprecated in favor of a new method, start(), to support asynchronous code execution. start_requests() will stop being called in a future version of Scrapy. If you use Scrapy 2.13 or higher only, replace start_requests() with start(); note that start() is a coroutine (async def). If you need to maintain compatibility with lower Scrapy versions, when overriding start_requests() in a spider class, override start() as well; you can use super() to reuse the inherited start() implementation without copy-pasting. See the release notes of Scrapy 2.13 for details: https://docs.scrapy.org/en/2.13/news.html warn( 2026-05-13 08:03:15 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2026-05-13 08:03:15 [scrapy.extensions.telnet] INFO: Telnet console listening on 127.0.0.1:6042 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Processing https://www.easypiscine.fr/fr/accessoires-spa/1657-spot-lumineux-pour-spa-netspa-3700691407542.html 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Name: Spot lumineux NetSpa LED light pour spa 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Price: 8000 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Stock: unavailable 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Price without discount: 8000 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Discount: None 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: SKU: SP-N1407542 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: MPN: None 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: EAN: None 2026-05-13 08:03:20 [priceWatchBotSpider] INFO: Brand: None 2026-05-13 08:03:20 [priceWatchBot.pipelines] INFO: Traitement de l'item Spot lumineux NetSpa LED light pour spa dans la pipeline 2026-05-13 08:03:20 [priceWatchBot.pipelines] INFO: Prix actuel: 8000, dernier prix connu: 8000 2026-05-13 08:03:20 [priceWatchBot.pipelines] INFO: Le prix n'a pas changé, passage au suivant. 2026-05-13 08:03:20 [scrapy.core.engine] INFO: Closing spider (finished) 2026-05-13 08:03:20 [scrapy.statscollectors] INFO: Dumping Scrapy stats: {'downloader/request_bytes': 605, 'downloader/request_count': 2, 'downloader/request_method_count/GET': 2, 'downloader/response_bytes': 63191, 'downloader/response_count': 2, 'downloader/response_status_count/200': 1, 'downloader/response_status_count/301': 1, 'elapsed_time_seconds': 4.776363, 'finish_reason': 'finished', 'finish_time': datetime.datetime(2026, 5, 13, 8, 3, 20, 429736, tzinfo=datetime.timezone.utc), 'httpcompression/response_bytes': 372062, 'httpcompression/response_count': 1, 'item_scraped_count': 1, 'items_per_minute': 15.0, 'log_count/INFO': 16, 'memusage/max': 72847360, 'memusage/startup': 72847360, 'response_received_count': 1, 'responses_per_minute': 15.0, 'scheduler/dequeued': 2, 'scheduler/dequeued/memory': 2, 'scheduler/enqueued': 2, 'scheduler/enqueued/memory': 2, 'start_time': datetime.datetime(2026, 5, 13, 8, 3, 15, 653373, tzinfo=datetime.timezone.utc)} 2026-05-13 08:03:20 [scrapy.core.engine] INFO: Spider closed (finished)