feat(fotograf-de-scraper): initial setup with backend and frontend scaffold [32788f42]
This commit is contained in:
45
fotograf-de-scraper/backend/Dockerfile
Normal file
45
fotograf-de-scraper/backend/Dockerfile
Normal file
@@ -0,0 +1,45 @@
|
||||
# Use an official Python runtime as a parent image
|
||||
FROM python:3.11-slim-buster
|
||||
|
||||
# Set the working directory in the container
|
||||
WORKDIR /app
|
||||
|
||||
# Install system dependencies for Chrome and other tools
|
||||
# Using a multi-stage build or a more specific base image could optimize this
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
chromium-driver \
|
||||
chromium \
|
||||
wget \
|
||||
unzip \
|
||||
fonts-liberation \
|
||||
libappindicator3-1 \
|
||||
libasound2 \
|
||||
libatk-bridge2.0-0 \
|
||||
libcups2 \
|
||||
libdrm-dev \
|
||||
libgbm-dev \
|
||||
libglvnd0 \
|
||||
libgtk-3-0 \
|
||||
libnspr4 \
|
||||
libnss3 \
|
||||
libxkbcommon0 \
|
||||
libxshmfence-dev \
|
||||
xdg-utils \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Set Chromium as default browser for Selenium
|
||||
ENV CHROME_BIN /usr/bin/chromium
|
||||
ENV CHROME_PATH /usr/bin/chromium
|
||||
|
||||
# Copy the requirements file and install Python dependencies
|
||||
COPY requirements.txt ./requirements.txt
|
||||
RUN pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
# Copy the application code
|
||||
COPY . .
|
||||
|
||||
# Expose the port FastAPI will run on
|
||||
EXPOSE 8000
|
||||
|
||||
# Command to run the application
|
||||
CMD ["uvicorn", "main:app", "--host", "0.0.0.0", "--port", "8000"]
|
||||
185
fotograf-de-scraper/backend/main.py
Normal file
185
fotograf-de-scraper/backend/main.py
Normal file
@@ -0,0 +1,185 @@
|
||||
import os
|
||||
from dotenv import load_dotenv
|
||||
from fastapi import FastAPI, HTTPException, BackgroundTasks
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
from typing import List, Dict, Any, Optional
|
||||
import time
|
||||
from datetime import datetime
|
||||
from selenium import webdriver
|
||||
from selenium.webdriver.chrome.options import Options
|
||||
from selenium.webdriver.common.by import By
|
||||
from selenium.webdriver.support.ui import WebDriverWait
|
||||
from selenium.webdriver.support import expected_conditions as EC
|
||||
from selenium.common.exceptions import TimeoutException, NoSuchElementException, StaleElementReferenceException, InvalidArgumentException
|
||||
import re
|
||||
import asyncio
|
||||
|
||||
# Load environment variables
|
||||
load_dotenv()
|
||||
|
||||
app = FastAPI(title="Fotograf.de Scraper API")
|
||||
|
||||
# Configure CORS
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["*"], # Adjust this to your frontend origin in production
|
||||
allow_credentials=True,
|
||||
allow_methods=["*"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
# --- Configuration & Constants ---
|
||||
LOGIN_URL = 'https://app.fotograf.de/login/login'
|
||||
|
||||
# --- Selectors from original scraper, expanded for dashboard jobs ---
|
||||
SELECTORS = {
|
||||
"cookie_accept_button": "#CybotCookiebotDialogBodyLevelButtonLevelOptinAllowAll",
|
||||
"login_user": "#login-email",
|
||||
"login_pass": "#login-password",
|
||||
"login_button": "#login-submit",
|
||||
"dashboard_jobs_table_rows": "//table[contains(@class, 'table-legacy')]/tbody/tr", # Assuming there's a table for jobs
|
||||
"job_row_name_link": ".//td[contains(@class, 'table-col-jobname')]//a",
|
||||
"job_row_status": ".//td[contains(@class, 'table-col-status')]//span",
|
||||
"job_row_date": ".//td[contains(@class, 'table-col-shootingDate')]",
|
||||
"job_row_shooting_type": ".//td[contains(@class, 'table-col-shootingType')]",
|
||||
}
|
||||
|
||||
# --- Utility functions from original scraper ---
|
||||
# (setup_driver, login, etc. will be adapted or moved into this file)
|
||||
|
||||
def setup_driver():
|
||||
print("Initialisiere Chrome WebDriver...")
|
||||
options = Options()
|
||||
options.add_argument('--headless')
|
||||
options.add_argument('--no-sandbox')
|
||||
options.add_argument('--disable-dev-shm-usage')
|
||||
options.add_argument('--window-size=1920,1200')
|
||||
options.binary_location = '/usr/bin/google-chrome' # Path to Chrome in Docker
|
||||
try:
|
||||
driver = webdriver.Chrome(options=options)
|
||||
return driver
|
||||
except Exception as e:
|
||||
print(f"Fehler bei der Initialisierung des WebDrivers: {e}")
|
||||
return None
|
||||
|
||||
def login(driver, username, password):
|
||||
print("Starte Login-Vorgang...")
|
||||
try:
|
||||
driver.get(LOGIN_URL)
|
||||
wait = WebDriverWait(driver, 10)
|
||||
try:
|
||||
print("Suche nach Cookie-Banner...")
|
||||
wait.until(EC.element_to_be_clickable((By.CSS_SELECTOR, SELECTORS["cookie_accept_button"]))).click()
|
||||
print("Cookie-Banner akzeptiert.")
|
||||
time.sleep(1)
|
||||
except TimeoutException:
|
||||
print("Kein Cookie-Banner gefunden, fahre fort.")
|
||||
print("Fülle Anmeldeformular aus...")
|
||||
wait.until(EC.visibility_of_element_located((By.CSS_SELECTOR, SELECTORS["login_user"]))).send_keys(username)
|
||||
driver.find_element(By.CSS_SELECTOR, SELECTORS["login_pass"]).send_keys(password)
|
||||
print("Klicke auf Login...")
|
||||
driver.find_element(By.CSS_SELECTOR, SELECTORS["login_button"]).click()
|
||||
print("Warte auf die nächste Seite...")
|
||||
wait.until(EC.url_contains('/config_dashboard/index'))
|
||||
print("Login erfolgreich!")
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Login fehlgeschlagen. Grund: {e}")
|
||||
# take_error_screenshot(driver, "login_error") # Removed for now, will re-add later if needed
|
||||
return False
|
||||
|
||||
# --- New function to get jobs from dashboard ---
|
||||
def get_jobs_from_dashboard(driver) -> List[Dict[str, Any]]:
|
||||
print("Navigiere zum Dashboard, um Aufträge abzurufen...")
|
||||
dashboard_url = "https://app.fotograf.de/config_dashboard/index"
|
||||
driver.get(dashboard_url)
|
||||
wait = WebDriverWait(driver, 20) # Increased timeout for dashboard load
|
||||
|
||||
jobs = []
|
||||
try:
|
||||
# Wait for the table rows to be present
|
||||
job_rows = wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["dashboard_jobs_table_rows"])))
|
||||
print(f"[{len(job_rows)}] Auftragszeilen auf dem Dashboard gefunden.")
|
||||
|
||||
for row in job_rows:
|
||||
try:
|
||||
name_element = row.find_element(By.XPATH, SELECTORS["job_row_name_link"])
|
||||
job_name = name_element.text.strip()
|
||||
job_url = name_element.get_attribute('href')
|
||||
|
||||
# Extract Job ID from URL
|
||||
job_id_match = re.search(r'/(\d+)$', job_url)
|
||||
job_id = job_id_match.group(1) if job_id_match else None
|
||||
|
||||
status_element = row.find_element(By.XPATH, SELECTORS["job_row_status"])
|
||||
job_status = status_element.text.strip()
|
||||
|
||||
date_element = row.find_element(By.XPATH, SELECTORS["job_row_date"])
|
||||
job_date = date_element.text.strip()
|
||||
|
||||
type_element = row.find_element(By.XPATH, SELECTORS["job_row_shooting_type"])
|
||||
shooting_type = type_element.text.strip()
|
||||
|
||||
jobs.append({
|
||||
"id": job_id,
|
||||
"name": job_name,
|
||||
"url": job_url,
|
||||
"status": job_status,
|
||||
"date": job_date,
|
||||
"shooting_type": shooting_type,
|
||||
})
|
||||
except NoSuchElementException as e:
|
||||
print(f"Warnung: Konnte nicht alle Elemente in einer Auftragszeile finden. Fehler: {e}")
|
||||
except Exception as e:
|
||||
print(f"Ein unerwarteter Fehler beim Parsen einer Auftragszeile: {e}")
|
||||
|
||||
except TimeoutException:
|
||||
print("Timeout: Keine Auftrags-Tabelle oder -Zeilen auf dem Dashboard gefunden.")
|
||||
except Exception as e:
|
||||
print(f"Ein Fehler ist aufgetreten beim Abrufen der Aufträge vom Dashboard: {e}")
|
||||
|
||||
return jobs
|
||||
|
||||
@app.get("/health")
|
||||
async def health_check():
|
||||
return {"status": "ok"}
|
||||
|
||||
@app.get("/api/jobs", response_model=List[Dict[str, Any]])
|
||||
async def get_jobs(account_type: str):
|
||||
username_env_var = f"{account_type.upper()}_USER"
|
||||
password_env_var = f"{account_type.upper()}_PW"
|
||||
|
||||
username = os.getenv(username_env_var)
|
||||
password = os.getenv(password_env_var)
|
||||
|
||||
if not username or not password:
|
||||
raise HTTPException(status_code=400, detail=f"Credentials for {account_type} not found. Please set {username_env_var} and {password_env_var} in your .env file.")
|
||||
|
||||
driver = None
|
||||
try:
|
||||
driver = setup_driver()
|
||||
if not driver:
|
||||
raise HTTPException(status_code=500, detail="Failed to initialize WebDriver.")
|
||||
|
||||
if not login(driver, username, password):
|
||||
raise HTTPException(status_code=401, detail="Login failed. Please check credentials.")
|
||||
|
||||
jobs = get_jobs_from_dashboard(driver)
|
||||
if not jobs:
|
||||
print("Keine Aufträge gefunden oder Fehler beim Abrufen vom Dashboard.")
|
||||
# Depending on desired behavior, might raise HTTPException or return empty list
|
||||
# For now, returning empty list if no jobs found but login was successful.
|
||||
|
||||
return jobs
|
||||
|
||||
except HTTPException as e:
|
||||
raise e # Re-raise HTTP exceptions
|
||||
except Exception as e:
|
||||
print(f"Ein unerwarteter Serverfehler ist aufgetreten: {e}")
|
||||
raise HTTPException(status_code=500, detail=f"Internal server error: {str(e)}")
|
||||
finally:
|
||||
if driver:
|
||||
print("Schließe WebDriver.")
|
||||
driver.quit()
|
||||
|
||||
# Integrate other scraper functions (process_reminder_mode, process_statistics_mode) as new API endpoints later
|
||||
6
fotograf-de-scraper/backend/requirements.txt
Normal file
6
fotograf-de-scraper/backend/requirements.txt
Normal file
@@ -0,0 +1,6 @@
|
||||
fastapi==0.111.0
|
||||
uvicorn==0.30.1
|
||||
python-dotenv==1.0.1
|
||||
selenium==4.22.0
|
||||
webdriver-manager==4.0.1
|
||||
pandas==2.2.2
|
||||
427
fotograf-de-scraper/backend/scraper.py
Normal file
427
fotograf-de-scraper/backend/scraper.py
Normal file
@@ -0,0 +1,427 @@
|
||||
import json
|
||||
import os
|
||||
import time
|
||||
import csv
|
||||
import math
|
||||
import re
|
||||
from datetime import datetime
|
||||
from selenium import webdriver
|
||||
from selenium.webdriver.chrome.options import Options
|
||||
from selenium.webdriver.common.by import By
|
||||
from selenium.webdriver.support.ui import WebDriverWait
|
||||
from selenium.webdriver.support import expected_conditions as EC
|
||||
from selenium.common.exceptions import TimeoutException, NoSuchElementException, StaleElementReferenceException, InvalidArgumentException
|
||||
|
||||
# --- Konfiguration & Konstanten ---
|
||||
CREDENTIALS_FILE = 'fotograf_credentials.json'
|
||||
OUTPUT_DIR = 'output'
|
||||
LOGIN_URL = 'https://app.fotograf.de/login/login'
|
||||
|
||||
# --- Selektoren ---
|
||||
SELECTORS = {
|
||||
"cookie_accept_button": "#CybotCookiebotDialogBodyLevelButtonLevelOptinAllowAll",
|
||||
"login_user": "#login-email",
|
||||
"login_pass": "#login-password",
|
||||
"login_button": "#login-submit",
|
||||
"job_name": "h1",
|
||||
"album_overview_rows": "//table/tbody/tr",
|
||||
"album_overview_link": ".//td[2]//a",
|
||||
"access_code_count": "//span[text()='Zugangscodes']/following-sibling::strong",
|
||||
"person_rows": "//div[contains(@class, 'border-legacy-silver-550') and .//span[text()='Logins']]",
|
||||
"person_vorname": ".//span[text()='Vorname']/following-sibling::strong",
|
||||
"person_logins": ".//span[text()='Logins']/following-sibling::strong",
|
||||
"person_access_code_link": ".//a[contains(@data-qa-id, 'guest-access-banner-access-code')]",
|
||||
# Selektoren für die Statistik-Zählung
|
||||
"person_all_photos": ".//div[@data-key]",
|
||||
"person_purchased_photos": ".//div[@data-key and .//img[@alt='Bestellungen mit diesem Foto']]",
|
||||
"person_access_card_photo": ".//div[@data-key and contains(@class, 'opacity-50')]", # NEU: Identifiziert die Zugangskarte
|
||||
"potential_buyer_link": "//a[contains(@href, '/config_customers/view_customer')]",
|
||||
"quick_login_url": "//a[@id='quick-login-url']",
|
||||
"buyer_email": "//span[contains(., '@')]"
|
||||
}
|
||||
|
||||
def take_error_screenshot(driver, error_name):
|
||||
os.makedirs(OUTPUT_DIR, exist_ok=True)
|
||||
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
filename = f"error_{error_name}_{timestamp}.png"
|
||||
filepath = os.path.join(OUTPUT_DIR, filename)
|
||||
try:
|
||||
driver.save_screenshot(filepath)
|
||||
print(f"!!! Fehler aufgetreten. Screenshot gespeichert unter: {filepath}")
|
||||
except Exception as e:
|
||||
print(f"!!! Konnte keinen Screenshot speichern: {e}")
|
||||
|
||||
def setup_driver():
|
||||
print("Initialisiere Chrome WebDriver...")
|
||||
options = Options()
|
||||
options.add_argument('--headless')
|
||||
options.add_argument('--no-sandbox')
|
||||
options.add_argument('--disable-dev-shm-usage')
|
||||
options.add_argument('--window-size=1920,1200')
|
||||
options.binary_location = '/usr/bin/google-chrome'
|
||||
try:
|
||||
driver = webdriver.Chrome(options=options)
|
||||
return driver
|
||||
except Exception as e:
|
||||
print(f"Fehler bei der Initialisierung des WebDrivers: {e}")
|
||||
return None
|
||||
|
||||
def load_all_credentials():
|
||||
try:
|
||||
with open(CREDENTIALS_FILE, 'r') as f:
|
||||
return json.load(f)
|
||||
except (FileNotFoundError, json.JSONDecodeError):
|
||||
return None
|
||||
|
||||
def login(driver, username, password):
|
||||
print("Starte Login-Vorgang...")
|
||||
try:
|
||||
driver.get(LOGIN_URL)
|
||||
wait = WebDriverWait(driver, 10)
|
||||
try:
|
||||
print("Suche nach Cookie-Banner...")
|
||||
wait.until(EC.element_to_be_clickable((By.CSS_SELECTOR, SELECTORS["cookie_accept_button"]))).click()
|
||||
print("Cookie-Banner akzeptiert.")
|
||||
time.sleep(1)
|
||||
except TimeoutException:
|
||||
print("Kein Cookie-Banner gefunden, fahre fort.")
|
||||
print("Fülle Anmeldeformular aus...")
|
||||
wait.until(EC.visibility_of_element_located((By.CSS_SELECTOR, SELECTORS["login_user"]))).send_keys(username)
|
||||
driver.find_element(By.CSS_SELECTOR, SELECTORS["login_pass"]).send_keys(password)
|
||||
print("Klicke auf Login...")
|
||||
driver.find_element(By.CSS_SELECTOR, SELECTORS["login_button"]).click()
|
||||
print("Warte auf die nächste Seite...")
|
||||
wait.until(EC.url_contains('/config_dashboard/index'))
|
||||
print("Login erfolgreich!")
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"Login fehlgeschlagen. Grund: {e}")
|
||||
take_error_screenshot(driver, "login_error")
|
||||
return False
|
||||
|
||||
def process_reminder_mode(driver, job_url):
|
||||
wait = WebDriverWait(driver, 15)
|
||||
|
||||
try:
|
||||
job_id_match = re.search(r'/(\d+)', job_url)
|
||||
if not job_id_match: raise ValueError("Konnte keine numerische Job-ID finden.")
|
||||
job_id = job_id_match.group(1)
|
||||
settings_url = f"https://app.fotograf.de/config_jobs_settings/index/{job_id}"
|
||||
except Exception as e:
|
||||
print(f"!!! FEHLER: Konnte keine Job-ID aus der URL '{job_url}' extrahieren. Grund: {e}")
|
||||
return []
|
||||
|
||||
print(f"\nVerarbeite Job-ID: {job_id}")
|
||||
driver.get(settings_url)
|
||||
try:
|
||||
job_name = wait.until(EC.visibility_of_element_located((By.CSS_SELECTOR, SELECTORS["job_name"]))).text
|
||||
print(f"Auftragsname: '{job_name}'")
|
||||
except TimeoutException:
|
||||
print("Konnte den Auftragsnamen nicht finden.")
|
||||
return []
|
||||
|
||||
albums_overview_url = f"https://app.fotograf.de/config_jobs_photos/index/{job_id}"
|
||||
print(f"Navigiere zur Alben-Übersicht: {albums_overview_url}")
|
||||
driver.get(albums_overview_url)
|
||||
|
||||
albums_to_visit = []
|
||||
try:
|
||||
album_rows = wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["album_overview_rows"])))
|
||||
print(f"{len(album_rows)} Alben in der Übersicht gefunden.")
|
||||
for row in album_rows:
|
||||
try:
|
||||
album_link = row.find_element(By.XPATH, SELECTORS["album_overview_link"])
|
||||
albums_to_visit.append({"name": album_link.text, "url": album_link.get_attribute('href')})
|
||||
except NoSuchElementException:
|
||||
continue
|
||||
print(f"{len(albums_to_visit)} gültige Album-Links gesammelt.")
|
||||
except TimeoutException:
|
||||
print("Konnte die Album-Liste nicht finden.")
|
||||
return []
|
||||
|
||||
final_results = []
|
||||
for album in albums_to_visit:
|
||||
print(f"\n--- Betrete Album: {album['name']} ---")
|
||||
driver.get(album['url'])
|
||||
try:
|
||||
total_codes_text = wait.until(EC.visibility_of_element_located((By.XPATH, SELECTORS["access_code_count"]))).text
|
||||
num_pages = math.ceil(int(total_codes_text) / 20)
|
||||
print(f"Album hat {total_codes_text} Zugangscodes auf {num_pages} Seite(n).")
|
||||
|
||||
for page_num in range(1, num_pages + 1):
|
||||
current_page_url = album['url']
|
||||
if page_num > 1: current_page_url += f"?page_guest_accesses={page_num}"
|
||||
|
||||
print(f" Verarbeite Seite {page_num}...")
|
||||
driver.get(current_page_url)
|
||||
|
||||
num_persons = len(wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["person_rows"]))))
|
||||
print(f" {num_persons} Personen auf dieser Seite gefunden.")
|
||||
|
||||
for i in range(num_persons):
|
||||
person_rows = wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["person_rows"])))
|
||||
person_row = person_rows[i]
|
||||
|
||||
login_count_text = person_row.find_element(By.XPATH, SELECTORS["person_logins"]).text
|
||||
|
||||
if int(login_count_text) <= 1:
|
||||
vorname = person_row.find_element(By.XPATH, SELECTORS["person_vorname"]).text
|
||||
|
||||
try:
|
||||
photo_container = person_row.find_element(By.XPATH, "./following-sibling::div[1]")
|
||||
purchase_icons = photo_container.find_elements(By.XPATH, SELECTORS["person_purchased_photos"])
|
||||
if len(purchase_icons) > 0:
|
||||
print(f" --> INFO: '{vorname}' hat bereits gekauft. Überspringe.")
|
||||
continue
|
||||
except NoSuchElementException:
|
||||
pass
|
||||
|
||||
print(f" --> ERFOLG: '{vorname}' mit {login_count_text} Login(s) gefunden (und kein Kauf).")
|
||||
|
||||
access_code_page_url = person_row.find_element(By.XPATH, SELECTORS["person_access_code_link"]).get_attribute('href')
|
||||
driver.get(access_code_page_url)
|
||||
print(f" Navigiere zur Kommunikations-Seite für '{vorname}'...")
|
||||
|
||||
for attempt in range(3):
|
||||
try:
|
||||
wait.until(EC.visibility_of_element_located((By.XPATH, SELECTORS["quick_login_url"])))
|
||||
schnell_login_url = driver.find_element(By.XPATH, SELECTORS["quick_login_url"]).get_attribute('href')
|
||||
potential_buyer_element = driver.find_element(By.XPATH, SELECTORS["potential_buyer_link"])
|
||||
kaeufer_name = potential_buyer_element.text
|
||||
|
||||
print(f" Käufer: '{kaeufer_name}', Schnell-Login: GEFUNDEN")
|
||||
potential_buyer_element.click()
|
||||
|
||||
print(f" Navigiere zur Käufer-Detailseite...")
|
||||
email = wait.until(EC.visibility_of_element_located((By.XPATH, SELECTORS["buyer_email"]))).text
|
||||
print(f" FINALE ERFOLG: E-Mail gefunden: {email}")
|
||||
|
||||
final_results.append({
|
||||
"Name des Kindes": vorname,
|
||||
"Name Käufer": kaeufer_name,
|
||||
"E-Mail-Adresse Käufer": email,
|
||||
"Schnell Login URL": schnell_login_url
|
||||
})
|
||||
break
|
||||
|
||||
except StaleElementReferenceException:
|
||||
print(f" Timing-Fehler, Versuch {attempt + 1}/3...")
|
||||
time.sleep(1)
|
||||
if attempt == 2: raise
|
||||
except TimeoutException:
|
||||
print(f" Timeout beim Warten auf Details für '{vorname}'.")
|
||||
take_error_screenshot(driver, f"timeout_error_{vorname}")
|
||||
break
|
||||
|
||||
print(f" Kehre zurück zur Album-Seite {page_num}...")
|
||||
driver.get(current_page_url)
|
||||
wait.until(EC.presence_of_element_located((By.XPATH, SELECTORS["person_rows"])))
|
||||
except TimeoutException:
|
||||
print(f" Keine Personen-Daten im Album '{album['name']}' gefunden. Überspringe.")
|
||||
continue
|
||||
|
||||
return final_results
|
||||
|
||||
def aggregate_results_by_email(results):
|
||||
print("\nBeginne mit der Aggregation der Ergebnisse pro E-Mail-Adresse...")
|
||||
aggregated_data = {}
|
||||
for result in results:
|
||||
email = result['E-Mail-Adresse Käufer']
|
||||
child_name = "Familienbilder" if result['Name des Kindes'] == "Familie" else result['Name des Kindes']
|
||||
html_link = f'<a href="{result["Schnell Login URL"]}">Fotos von {child_name}</a>'
|
||||
if email not in aggregated_data:
|
||||
aggregated_data[email] = {
|
||||
'Name Käufer': result['Name Käufer'].split(' ')[0],
|
||||
'E-Mail-Adresse Käufer': email,
|
||||
'Kindernamen_list': [child_name],
|
||||
'LinksHTML_list': [html_link]
|
||||
}
|
||||
else:
|
||||
aggregated_data[email]['Kindernamen_list'].append(child_name)
|
||||
aggregated_data[email]['LinksHTML_list'].append(html_link)
|
||||
|
||||
final_list = []
|
||||
for email, data in aggregated_data.items():
|
||||
names_list = data['Kindernamen_list']
|
||||
if len(names_list) > 2:
|
||||
kindernamen_str = ', '.join(names_list[:-1]) + ' und ' + names_list[-1]
|
||||
else:
|
||||
kindernamen_str = ' und '.join(names_list)
|
||||
final_list.append({
|
||||
'Name Käufer': data['Name Käufer'],
|
||||
'E-Mail-Adresse Käufer': email,
|
||||
'Kindernamen': kindernamen_str,
|
||||
'LinksHTML': '<br><br>'.join(data['LinksHTML_list'])
|
||||
})
|
||||
print(f"Aggregation abgeschlossen. {len(results)} Roh-Einträge zu {len(final_list)} einzigartigen E-Mails zusammengefasst.")
|
||||
return final_list
|
||||
|
||||
def save_aggregated_results_to_csv(results):
|
||||
if not results:
|
||||
print("\nKeine Daten zum Speichern vorhanden.")
|
||||
return
|
||||
|
||||
output_file = os.path.join(OUTPUT_DIR, 'supermailer_fertige_liste.csv')
|
||||
os.makedirs(OUTPUT_DIR, exist_ok=True)
|
||||
fieldnames = ["Name Käufer", "E-Mail-Adresse Käufer", "Kindernamen", "LinksHTML"]
|
||||
print(f"\nSpeichere {len(results)} aggregierte Ergebnisse in '{output_file}'...")
|
||||
with open(output_file, 'w', newline='', encoding='utf-8') as f:
|
||||
writer = csv.DictWriter(f, fieldnames=fieldnames)
|
||||
writer.writeheader()
|
||||
writer.writerows(results)
|
||||
print("Speichern erfolgreich!")
|
||||
|
||||
# --- Modus 2: Statistik-Auswertung ---
|
||||
def process_statistics_mode(driver, job_url):
|
||||
wait = WebDriverWait(driver, 15)
|
||||
|
||||
try:
|
||||
job_id = re.search(r'/(\d+)', job_url).group(1)
|
||||
except Exception:
|
||||
print(f"!!! FEHLER: Konnte keine Job-ID aus der URL '{job_url}' extrahieren.")
|
||||
return []
|
||||
|
||||
albums_overview_url = f"https://app.fotograf.de/config_jobs_photos/index/{job_id}"
|
||||
print(f"Navigiere zur Alben-Übersicht: {albums_overview_url}")
|
||||
driver.get(albums_overview_url)
|
||||
|
||||
albums_to_visit = []
|
||||
try:
|
||||
album_rows = wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["album_overview_rows"])))
|
||||
for row in album_rows:
|
||||
try:
|
||||
album_link = row.find_element(By.XPATH, SELECTORS["album_overview_link"])
|
||||
albums_to_visit.append({"name": album_link.text, "url": album_link.get_attribute('href')})
|
||||
except NoSuchElementException: continue
|
||||
except TimeoutException:
|
||||
print("Konnte die Album-Liste nicht finden.")
|
||||
return []
|
||||
|
||||
statistics = []
|
||||
print("\n--- STATISTIK-AUSWERTUNG ---")
|
||||
for album in albums_to_visit:
|
||||
print(f"\nAlbum: {album['name']}")
|
||||
driver.get(album['url'])
|
||||
try:
|
||||
total_codes_text = wait.until(EC.visibility_of_element_located((By.XPATH, SELECTORS["access_code_count"]))).text
|
||||
num_pages = math.ceil(int(total_codes_text) / 20)
|
||||
|
||||
total_children_in_album = 0
|
||||
children_with_purchase = 0
|
||||
children_with_all_purchased = 0
|
||||
|
||||
for page_num in range(1, num_pages + 1):
|
||||
if page_num > 1: driver.get(album['url'] + f"?page_guest_accesses={page_num}")
|
||||
|
||||
person_rows = wait.until(EC.presence_of_all_elements_located((By.XPATH, SELECTORS["person_rows"])))
|
||||
|
||||
for person_row in person_rows:
|
||||
total_children_in_album += 1
|
||||
try:
|
||||
photo_container = person_row.find_element(By.XPATH, "./following-sibling::div[1]")
|
||||
|
||||
# GEÄNDERTE ZÄHLLOGIK
|
||||
num_total_photos = len(photo_container.find_elements(By.XPATH, SELECTORS["person_all_photos"]))
|
||||
num_purchased_photos = len(photo_container.find_elements(By.XPATH, SELECTORS["person_purchased_photos"]))
|
||||
num_access_cards = len(photo_container.find_elements(By.XPATH, SELECTORS["person_access_card_photo"]))
|
||||
|
||||
buyable_photos = num_total_photos - num_access_cards
|
||||
|
||||
if num_purchased_photos > 0:
|
||||
children_with_purchase += 1
|
||||
|
||||
if buyable_photos > 0 and buyable_photos == num_purchased_photos:
|
||||
children_with_all_purchased += 1
|
||||
except NoSuchElementException:
|
||||
continue
|
||||
|
||||
print(f" - Kinder insgesamt: {total_children_in_album}")
|
||||
print(f" - Kinder mit (mind. 1) Kauf: {children_with_purchase}")
|
||||
print(f" - Kinder (Alle Bilder gekauft): {children_with_all_purchased}")
|
||||
statistics.append({
|
||||
"Album": album['name'],
|
||||
"Kinder insgesamt": total_children_in_album,
|
||||
"Kinder mit Käufen": children_with_purchase,
|
||||
"Kinder (Alle Bilder gekauft)": children_with_all_purchased
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
print(f" Fehler bei der Auswertung dieses Albums: {e}")
|
||||
continue
|
||||
|
||||
return statistics
|
||||
|
||||
def save_statistics_to_csv(results):
|
||||
if not results:
|
||||
print("\nKeine Statistikdaten zum Speichern vorhanden.")
|
||||
return
|
||||
|
||||
output_file = os.path.join(OUTPUT_DIR, 'job_statistik.csv')
|
||||
os.makedirs(OUTPUT_DIR, exist_ok=True)
|
||||
fieldnames = ["Album", "Kinder insgesamt", "Kinder mit Käufen", "Kinder (Alle Bilder gekauft)"]
|
||||
print(f"\nSpeichere Statistik für {len(results)} Alben in '{output_file}'...")
|
||||
with open(output_file, 'w', newline='', encoding='utf-8') as f:
|
||||
writer = csv.DictWriter(f, fieldnames=fieldnames)
|
||||
writer.writeheader()
|
||||
writer.writerows(results)
|
||||
print("Speichern erfolgreich!")
|
||||
|
||||
def get_profile_choice():
|
||||
all_credentials = load_all_credentials()
|
||||
if not all_credentials: return None
|
||||
profiles = list(all_credentials.keys())
|
||||
print("\nBitte wähle das zu verwendende Profil:")
|
||||
for i, p in enumerate(profiles): print(f" {i + 1}) {p}")
|
||||
while True:
|
||||
try:
|
||||
c = int(input(f"Gib eine Zahl zwischen 1 und {len(profiles)} ein: "))
|
||||
if 1 <= c <= len(profiles):
|
||||
p_name = profiles[c - 1]
|
||||
print(f"Profil '{p_name}' ausgewählt.")
|
||||
return all_credentials[p_name]
|
||||
else: print("Ungültige Auswahl.")
|
||||
except ValueError: print("Ungültige Eingabe.")
|
||||
|
||||
def main():
|
||||
print("--- Fotograf.de Scraper (v3.2 - The Master Analyst) ---")
|
||||
|
||||
while True:
|
||||
mode = input("Bitte Modus wählen:\n 1) E-Mail-Liste erstellen\n 2) Statistik auswerten\nWahl: ")
|
||||
if mode in ['1', '2']: break
|
||||
else: print("Ungültige Eingabe.")
|
||||
|
||||
credentials = get_profile_choice()
|
||||
if not credentials: return
|
||||
|
||||
job_url_raw = input("Bitte eine beliebige URL des zu bearbeitenden Fotoauftrags ein: ")
|
||||
|
||||
match = re.search(r'(https?://[^\s]+)', job_url_raw)
|
||||
if not match:
|
||||
print("Keine gültige URL in der Eingabe gefunden.")
|
||||
return
|
||||
job_url = match.group(1).strip()
|
||||
|
||||
if "fotograf.de/config_jobs_" not in job_url or not re.search(r'/\d+', job_url):
|
||||
print("Dies scheint keine gültige URL für einen Fotoauftrag zu sein.")
|
||||
return
|
||||
|
||||
driver = setup_driver()
|
||||
if not driver: return
|
||||
|
||||
try:
|
||||
if login(driver, credentials['username'], credentials['password']):
|
||||
if mode == '1':
|
||||
raw_results = process_reminder_mode(driver, job_url)
|
||||
aggregated_results = aggregate_results_by_email(raw_results)
|
||||
save_aggregated_results_to_csv(aggregated_results)
|
||||
elif mode == '2':
|
||||
stats_results = process_statistics_mode(driver, job_url)
|
||||
save_statistics_to_csv(stats_results)
|
||||
else:
|
||||
print("Skript wird beendet, da der Login fehlgeschlagen ist.")
|
||||
finally:
|
||||
print("\nSkript beendet. Schließe WebDriver.")
|
||||
if driver: driver.quit()
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
29
fotograf-de-scraper/docker-compose.yml
Normal file
29
fotograf-de-scraper/docker-compose.yml
Normal file
@@ -0,0 +1,29 @@
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
fotograf-de-scraper-backend:
|
||||
build:
|
||||
context: ./backend
|
||||
dockerfile: Dockerfile
|
||||
container_name: fotograf-de-scraper-backend
|
||||
env_file:
|
||||
- ./.env
|
||||
ports:
|
||||
- "8001:8000" # Map internal 8000 to external 8001 to avoid conflicts
|
||||
volumes:
|
||||
- ./backend:/app # Mount the backend code for easier development
|
||||
restart: unless-stopped
|
||||
|
||||
fotograf-de-scraper-frontend:
|
||||
build:
|
||||
context: ./frontend
|
||||
dockerfile: Dockerfile
|
||||
container_name: fotograf-de-scraper-frontend
|
||||
ports:
|
||||
- "3009:80" # Map internal 80 to external 3009
|
||||
depends_on:
|
||||
- fotograf-de-scraper-backend
|
||||
volumes:
|
||||
- ./frontend:/app # Mount the frontend code for easier development
|
||||
restart: unless-stopped
|
||||
|
||||
18
fotograf-de-scraper/frontend/Dockerfile
Normal file
18
fotograf-de-scraper/frontend/Dockerfile
Normal file
@@ -0,0 +1,18 @@
|
||||
# Use a lightweight Nginx image to serve the React app
|
||||
FROM nginx:alpine
|
||||
|
||||
# Set working directory to nginx's default static file directory
|
||||
WORKDIR /usr/share/nginx/html
|
||||
|
||||
# Remove default Nginx static assets
|
||||
RUN rm -rf ./*
|
||||
|
||||
# Copy the built React app from the builder stage
|
||||
# The React app is built using `npm run build` which creates a `dist` directory
|
||||
COPY ./dist .
|
||||
|
||||
# Expose port 80
|
||||
EXPOSE 80
|
||||
|
||||
# Command to start Nginx (default command of the base image)
|
||||
CMD ["nginx", "-g", "daemon off;"]
|
||||
109
fotograf-de-scraper/frontend/src/App.tsx
Normal file
109
fotograf-de-scraper/frontend/src/App.tsx
Normal file
@@ -0,0 +1,109 @@
|
||||
import { useState, useEffect } from 'react';
|
||||
import './App.css';
|
||||
|
||||
interface Job {
|
||||
id: string;
|
||||
name: string;
|
||||
url: string;
|
||||
status: string;
|
||||
date: string;
|
||||
shooting_type: string;
|
||||
}
|
||||
|
||||
function App() {
|
||||
const [accountType, setAccountType] = useState('kiga'); // Default to kindergarten
|
||||
const [jobs, setJobs] = useState<Job[]>([]);
|
||||
const [isLoading, setIsLoading] = useState(false);
|
||||
const [error, setError] = useState<string | null>(null);
|
||||
|
||||
const API_BASE_URL = import.meta.env.VITE_API_BASE_URL || 'http://localhost:8001';
|
||||
|
||||
const fetchJobs = async () => {
|
||||
setIsLoading(true);
|
||||
setError(null);
|
||||
try {
|
||||
const response = await fetch(`${API_BASE_URL}/api/jobs?account_type=${accountType}`);
|
||||
if (!response.ok) {
|
||||
const errData = await response.json();
|
||||
throw new Error(errData.detail || 'Fehler beim Abrufen der Aufträge');
|
||||
}
|
||||
const data: Job[] = await response.json();
|
||||
setJobs(data);
|
||||
} catch (err: any) {
|
||||
setError(err.message);
|
||||
console.error("Failed to fetch jobs:", err);
|
||||
} finally {
|
||||
setIsLoading(false);
|
||||
}
|
||||
};
|
||||
|
||||
useEffect(() => {
|
||||
fetchJobs();
|
||||
}, [accountType]); // Refetch when accountType changes
|
||||
|
||||
return (
|
||||
<div className="min-h-screen bg-gray-100 p-4">
|
||||
<div className="max-w-4xl mx-auto bg-white p-6 rounded-lg shadow-md">
|
||||
<h1 className="text-2xl font-bold mb-4">Fotograf.de Auftragsübersicht</h1>
|
||||
|
||||
<div className="mb-4">
|
||||
<label htmlFor="accountType" className="block text-sm font-medium text-gray-700">Account auswählen:</label>
|
||||
<select
|
||||
id="accountType"
|
||||
value={accountType}
|
||||
onChange={(e) => setAccountType(e.target.value)}
|
||||
className="mt-1 block w-full pl-3 pr-10 py-2 text-base border-gray-300 focus:outline-none focus:ring-indigo-500 focus:border-indigo-500 sm:text-sm rounded-md"
|
||||
>
|
||||
<option value="kiga">Kindergarten</option>
|
||||
<option value="schule">Schule</option>
|
||||
</select>
|
||||
</div>
|
||||
|
||||
<button
|
||||
onClick={fetchJobs}
|
||||
disabled={isLoading}
|
||||
className="mb-4 inline-flex items-center px-4 py-2 border border-transparent text-sm font-medium rounded-md shadow-sm text-white bg-indigo-600 hover:bg-indigo-700 focus:outline-none focus:ring-2 focus:ring-offset-2 focus:ring-indigo-500 disabled:opacity-50"
|
||||
>
|
||||
{isLoading ? 'Lade Aufträge...' : 'Aufträge neu laden'}
|
||||
</button>
|
||||
|
||||
{error && <p className="text-red-600 mb-4">Fehler: {error}</p>}
|
||||
|
||||
{jobs.length === 0 && !isLoading && !error && (
|
||||
<p className="text-gray-500">Keine Aufträge gefunden.</p>
|
||||
)}
|
||||
|
||||
{jobs.length > 0 && (
|
||||
<div className="overflow-x-auto">
|
||||
<table className="min-w-full divide-y divide-gray-200">
|
||||
<thead className="bg-gray-50">
|
||||
<tr>
|
||||
<th scope="col" className="px-6 py-3 text-left text-xs font-medium text-gray-500 uppercase tracking-wider">Name</th>
|
||||
<th scope="col" className="px-6 py-3 text-left text-xs font-medium text-gray-500 uppercase tracking-wider">Status</th>
|
||||
<th scope="col" className="px-6 py-3 text-left text-xs font-medium text-gray-500 uppercase tracking-wider">Datum</th>
|
||||
<th scope="col" className="px-6 py-3 text-left text-xs font-medium text-gray-500 uppercase tracking-wider">Typ</th>
|
||||
<th scope="col" className="relative px-6 py-3"><span className="sr-only">Aktionen</span></th>
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody className="bg-white divide-y divide-gray-200">
|
||||
{jobs.map((job) => (
|
||||
<tr key={job.id}>
|
||||
<td className="px-6 py-4 whitespace-nowrap text-sm font-medium text-gray-900">{job.name}</td>
|
||||
<td className="px-6 py-4 whitespace-nowrap text-sm text-gray-500">{job.status}</td>
|
||||
<td className="px-6 py-4 whitespace-nowrap text-sm text-gray-500">{job.date}</td>
|
||||
<td className="px-6 py-4 whitespace-nowrap text-sm text-gray-500">{job.shooting_type}</td>
|
||||
<td className="px-6 py-4 whitespace-nowrap text-right text-sm font-medium">
|
||||
<a href={job.url} target="_blank" rel="noopener noreferrer" className="text-indigo-600 hover:text-indigo-900">Details</a>
|
||||
</td>
|
||||
</tr>
|
||||
))}
|
||||
</tbody>
|
||||
</table>
|
||||
</div>
|
||||
)}
|
||||
</div>
|
||||
</div>
|
||||
);
|
||||
}
|
||||
|
||||
export default App;
|
||||
Reference in New Issue
Block a user