From 4702ff0c227c0e0bec64a4278c2959fbbaba8462 Mon Sep 17 00:00:00 2001 From: Oli Passey Date: Wed, 2 Jul 2025 14:26:43 +0100 Subject: [PATCH] cron --- Dockerfile | 91 +++++++++++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 90 insertions(+), 1 deletion(-) diff --git a/Dockerfile b/Dockerfile index 4ebd1c8..216891f 100644 --- a/Dockerfile +++ b/Dockerfile @@ -66,6 +66,8 @@ async def main():\n\ from src.config import Config\n\ from src.database import DatabaseManager\n\ from src.scraper_manager import ScraperManager\n\ + from src.notification import NotificationManager\n\ + from src.shopping_list import AutoShoppingListGenerator\n\ \n\ logger.info("Starting scheduled price scraping")\n\ \n\ @@ -76,6 +78,8 @@ async def main():\n\ \n\ db_manager = DatabaseManager(config.database_path)\n\ scraper_manager = ScraperManager(config)\n\ + notification_manager = NotificationManager(config)\n\ + shopping_list_generator = AutoShoppingListGenerator(db_manager)\n\ \n\ products = db_manager.get_all_products()\n\ if not products:\n\ @@ -87,13 +91,17 @@ async def main():\n\ \n\ total = sum(len(sites) for sites in results.values())\n\ successful = sum(1 for sites in results.values() for result in sites.values() if result["success"])\n\ + failed = total - successful\n\ \n\ logger.info(f"Scraping complete: {successful}/{total} successful")\n\ \n\ - # Save results to database\n\ + # Save results and collect price alerts\n\ + price_alerts = []\n\ for product_id, site_results in results.items():\n\ + product = db_manager.get_product(product_id)\n\ for site_name, result in site_results.items():\n\ if result["success"]:\n\ + # Save to database\n\ db_manager.save_price_history(\n\ product_id=product_id,\n\ site_name=site_name,\n\ @@ -102,8 +110,89 @@ async def main():\n\ timestamp=datetime.now()\n\ )\n\ \n\ + # Check for price alerts\n\ + if product and product.get("target_price") and result["price"] <= product["target_price"]:\n\ + price_alerts.append({\n\ + "product": product,\n\ + "site": site_name,\n\ + "current_price": result["price"],\n\ + "target_price": product["target_price"],\n\ + "url": result.get("url", "")\n\ + })\n\ + \n\ + # Send price alerts if any\n\ + if price_alerts:\n\ + alert_message = "Price Alerts:\\n\\n"\n\ + for alert in price_alerts:\n\ + alert_message += f"🎯 {alert[\"product\"][\"name\"]}\\n"\n\ + alert_message += f" Store: {alert[\"site\"]}\\n"\n\ + alert_message += f" Price: £{alert[\"current_price\"]} (Target: £{alert[\"target_price\"]})\\n"\n\ + alert_message += f" URL: {alert[\"url\"]}\\n\\n"\n\ + \n\ + await notification_manager.send_notification(\n\ + subject=f"Price Alert: {len(price_alerts)} item(s) on sale!",\n\ + message=alert_message\n\ + )\n\ + logger.info(f"Sent price alerts for {len(price_alerts)} items")\n\ + \n\ + # Generate and send daily shopping list\n\ + try:\n\ + shopping_lists = shopping_list_generator.generate_all_shopping_lists()\n\ + if shopping_lists:\n\ + shopping_message = "Daily Shopping List (Best Prices):\\n\\n"\n\ + total_savings = 0\n\ + \n\ + for store_name, store_list in shopping_lists.items():\n\ + if store_list.items:\n\ + shopping_message += f"🏪 {store_name.upper()}:\\n"\n\ + store_total = 0\n\ + for item in store_list.items:\n\ + shopping_message += f" • {item.product_name} - £{item.current_price}\\n"\n\ + store_total += item.current_price\n\ + if item.savings_amount > 0:\n\ + total_savings += item.savings_amount\n\ + shopping_message += f" Subtotal: £{store_total:.2f}\\n\\n"\n\ + \n\ + if total_savings > 0:\n\ + shopping_message += f"💰 Total Savings: £{total_savings:.2f}\\n"\n\ + \n\ + await notification_manager.send_notification(\n\ + subject="Daily Shopping List - Best Prices",\n\ + message=shopping_message\n\ + )\n\ + logger.info("Sent daily shopping list")\n\ + except Exception as e:\n\ + logger.error(f"Failed to generate shopping list: {e}")\n\ + \n\ + # Send scraping summary\n\ + summary_message = f"Daily Price Scraping Summary:\\n\\n"\n\ + summary_message += f"📊 Products scraped: {len(products)}\\n"\n\ + summary_message += f"✅ Successful: {successful}\\n"\n\ + summary_message += f"❌ Failed: {failed}\\n"\n\ + summary_message += f"🎯 Price alerts: {len(price_alerts)}\\n"\n\ + summary_message += f"🕐 Completed at: {datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")}\\n"\n\ + \n\ + await notification_manager.send_notification(\n\ + subject="Daily Price Scraping Complete",\n\ + message=summary_message\n\ + )\n\ + logger.info("Sent scraping summary")\n\ + \n\ except Exception as e:\n\ logger.error(f"Scheduled scraping failed: {str(e)}", exc_info=True)\n\ + \n\ + # Send error notification\n\ + try:\n\ + from src.config import Config\n\ + from src.notification import NotificationManager\n\ + config = Config()\n\ + notification_manager = NotificationManager(config)\n\ + await notification_manager.send_notification(\n\ + subject="Price Scraping Failed",\n\ + message=f"Daily price scraping failed with error:\\n\\n{str(e)}"\n\ + )\n\ + except:\n\ + pass # If notification also fails, just log\n\ \n\ if __name__ == "__main__":\n\ asyncio.run(main())\n\