Imported RoyalRoad chapter checker, changed it to use FastAPI with metrics and uses their RSS feed instead of scraping with the FanFicFare module
This commit is contained in:
parent
4ebf3c1df8
commit
36a8767349
21
requirements.txt
Normal file
21
requirements.txt
Normal file
@ -0,0 +1,21 @@
|
||||
annotated-types==0.7.0
|
||||
anyio==4.11.0
|
||||
certifi==2025.10.5
|
||||
charset-normalizer==3.4.3
|
||||
click==8.3.0
|
||||
fastapi==0.118.2
|
||||
feedparser==6.0.12
|
||||
h11==0.16.0
|
||||
idna==3.10
|
||||
mysql-connector-python==9.4.0
|
||||
prometheus_client==0.23.1
|
||||
pydantic==2.12.0
|
||||
pydantic_core==2.41.1
|
||||
requests==2.32.5
|
||||
sgmllib3k==1.0.0
|
||||
sniffio==1.3.1
|
||||
starlette==0.48.0
|
||||
typing-inspection==0.4.2
|
||||
typing_extensions==4.15.0
|
||||
urllib3==2.5.0
|
||||
uvicorn==0.37.0
|
||||
79
src/db.py
Normal file
79
src/db.py
Normal file
@ -0,0 +1,79 @@
|
||||
import mysql.connector
|
||||
import threading
|
||||
from secret_handler import return_credentials
|
||||
import os
|
||||
import time
|
||||
import sys
|
||||
|
||||
db_username = return_credentials("/etc/secrets/db_username")
|
||||
db_password = return_credentials("/etc/secrets/db_password")
|
||||
db_host = os.getenv("SERVICE_RR_DB_HOST","localhost")
|
||||
db_database = os.getenv("SERVICE_RR_DB_HOST_DATABASE","app")
|
||||
|
||||
MAX_RETRIES = 5
|
||||
RETRY_DELAY = 5
|
||||
|
||||
MYSQL_CONFIG = {
|
||||
"host": db_host,
|
||||
"user": db_username,
|
||||
"password": db_password,
|
||||
"database": db_database
|
||||
}
|
||||
|
||||
_pool_lock = threading.Lock()
|
||||
_connection_pool = None
|
||||
|
||||
|
||||
def create_connection_pool():
|
||||
global _connection_pool
|
||||
for attempt in range(1, MAX_RETRIES+1):
|
||||
try:
|
||||
print(f"[MySQL] Attempt {attempt} to connect...")
|
||||
pool = mysql.connector.pooling.MySQLConnectionPool(
|
||||
pool_name="mypool",
|
||||
pool_size=5,
|
||||
pool_reset_session=True,
|
||||
**MYSQL_CONFIG
|
||||
)
|
||||
with _pool_lock:
|
||||
_connection_pool = pool
|
||||
print("[MySQL] Connection pool created successfully.")
|
||||
return
|
||||
except mysql.connector.Error as e:
|
||||
print(f"[MySQL] Attempt {attempt} failed: {e}")
|
||||
if attempt < MAX_RETRIES:
|
||||
time.sleep(RETRY_DELAY)
|
||||
print(f"[MySQL] Failed to connect after {MAX_RETRIES} attempts — exiting.")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
def close_connection_pool():
|
||||
global _connection_pool
|
||||
with _pool_lock:
|
||||
if _connection_pool:
|
||||
_connection_pool = None
|
||||
print("[MySQL] Connection pool closed.")
|
||||
|
||||
|
||||
def get_connection_pool():
|
||||
global _connection_pool
|
||||
with _pool_lock:
|
||||
if _connection_pool is None:
|
||||
create_connection_pool()
|
||||
return _connection_pool
|
||||
|
||||
|
||||
def get_db():
|
||||
pool = get_connection_pool()
|
||||
try:
|
||||
conn = pool.get_connection()
|
||||
if not conn.is_connected():
|
||||
conn.reconnect(attempts=MAX_RETRIES, delay=RETRY_DELAY)
|
||||
except Exception:
|
||||
create_connection_pool()
|
||||
pool = get_connection_pool()
|
||||
conn = pool.get_connection()
|
||||
try:
|
||||
yield conn
|
||||
finally:
|
||||
conn.close()
|
||||
26
src/feed_handler.py
Normal file
26
src/feed_handler.py
Normal file
@ -0,0 +1,26 @@
|
||||
import feedparser
|
||||
import re
|
||||
|
||||
def extract_number(s: str)->int | None:
|
||||
match = re.search(r"\d+", s)
|
||||
if match:
|
||||
return int(match.group())
|
||||
else:
|
||||
return None
|
||||
|
||||
def grab_latest_chapter_information(id:str)->tuple[int | None, str, str]:
|
||||
url = f"https://www.royalroad.com/fiction/syndication/{id}"
|
||||
feed = feedparser.parse(url)
|
||||
|
||||
if not feed.entries:
|
||||
raise ValueError(f"No entries found for feed {id}")
|
||||
|
||||
latest_chapter_data = feed["entries"][0]
|
||||
chapter_number = extract_number(latest_chapter_data["title"])
|
||||
chapter_link = latest_chapter_data["link"]
|
||||
title = feed["title"]
|
||||
|
||||
return chapter_number,chapter_link,title
|
||||
|
||||
if __name__ == "__main__":
|
||||
print(grab_latest_chapter_information("118891"))
|
||||
80
src/main.py
Normal file
80
src/main.py
Normal file
@ -0,0 +1,80 @@
|
||||
from fastapi import FastAPI, Depends, HTTPException, Response, Request
|
||||
import uvicorn
|
||||
from contextlib import asynccontextmanager
|
||||
from db import get_db, create_connection_pool, close_connection_pool
|
||||
import logging
|
||||
from feed_handler import grab_latest_chapter_information
|
||||
from send_notification import send_notification
|
||||
from metrics_server import REQUEST_COUNTER
|
||||
import asyncio
|
||||
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
logger.info("Starting application...")
|
||||
|
||||
logger.info("Creating MySQL connection pool...")
|
||||
create_connection_pool()
|
||||
|
||||
yield
|
||||
logger.info("Closing MySQL connection pool...")
|
||||
close_connection_pool()
|
||||
|
||||
api = FastAPI(
|
||||
title="Docker Repository Query",
|
||||
description="Queries Dockerhub and GHCR for new docker images",
|
||||
version="1.0.0",
|
||||
lifespan=lifespan
|
||||
)
|
||||
|
||||
@api.middleware("http")
|
||||
async def prometheus_middleware(request: Request, call_next):
|
||||
status = 500
|
||||
try:
|
||||
response = await call_next(request)
|
||||
status = response.status_code
|
||||
except Exception:
|
||||
raise
|
||||
finally:
|
||||
REQUEST_COUNTER.labels(request.method, request.url.path, status).inc()
|
||||
return response
|
||||
|
||||
|
||||
@api.get("/health")
|
||||
def return_health():
|
||||
return Response(status_code=200)
|
||||
|
||||
@api.get("/royalroad")
|
||||
def get_chapters(
|
||||
request: Request,
|
||||
db = Depends(get_db)
|
||||
):
|
||||
try:
|
||||
logger.info("Checking for new Royalroad chapters")
|
||||
|
||||
cursor = db.cursor()
|
||||
cursor.execute("SELECT id,royalroadId,lastChapter FROM stories where active=1")
|
||||
for id,royalroadId,last_chapter_db in cursor.fetchall():
|
||||
chapter_number,chapter_link,title = grab_latest_chapter_information(royalroadId)
|
||||
if chapter_number > last_chapter_db:
|
||||
cursor.execute("UPDATE stories SET lastChapter = %s WHERE id = %s",
|
||||
(chapter_number, id))
|
||||
db.commit()
|
||||
send_notification(title,chapter_number,chapter_link)
|
||||
return {"status": "checked"}
|
||||
except Exception as e:
|
||||
raise HTTPException(status_code=500, detail=str(e))
|
||||
|
||||
async def start_servers():
|
||||
config_main = uvicorn.Config("main:api", host="0.0.0.0", port=5000, log_level="info")
|
||||
config_metrics = uvicorn.Config("metrics_server:metrics_api", host="0.0.0.0", port=9000, log_level="info")
|
||||
|
||||
server_main = uvicorn.Server(config_main)
|
||||
server_metrics = uvicorn.Server(config_metrics)
|
||||
|
||||
await asyncio.gather(server_main.serve(), server_metrics.serve())
|
||||
|
||||
if __name__ == "__main__":
|
||||
asyncio.run(start_servers())
|
||||
10
src/metrics_server.py
Normal file
10
src/metrics_server.py
Normal file
@ -0,0 +1,10 @@
|
||||
from fastapi import FastAPI, Response
|
||||
from prometheus_client import generate_latest, CONTENT_TYPE_LATEST, Counter
|
||||
|
||||
metrics_api = FastAPI(title="Metrics Server", description="Prometheus metrics endpoint")
|
||||
|
||||
REQUEST_COUNTER = Counter("http_requests_total", "Total HTTP Requests", ["method", "endpoint", "status"])
|
||||
|
||||
@metrics_api.get("/metrics")
|
||||
async def metrics():
|
||||
return Response(generate_latest(), media_type=CONTENT_TYPE_LATEST)
|
||||
12
src/secret_handler.py
Normal file
12
src/secret_handler.py
Normal file
@ -0,0 +1,12 @@
|
||||
import sys
|
||||
|
||||
def return_credentials(path: str)->str:
|
||||
try:
|
||||
with open (path) as file:
|
||||
return file.read().strip()
|
||||
except FileNotFoundError:
|
||||
print(f"[FATAL] Secret file not found: {path}")
|
||||
sys.exit(1)
|
||||
except Exception as e:
|
||||
print(f"[FATAL] Failed to read secret file {path}: {e}")
|
||||
sys.exit(1)
|
||||
35
src/send_notification.py
Normal file
35
src/send_notification.py
Normal file
@ -0,0 +1,35 @@
|
||||
import requests
|
||||
from requests.exceptions import RequestException, Timeout, ConnectionError, HTTPError
|
||||
from secret_handler import return_credentials
|
||||
import os
|
||||
|
||||
backend_api_url=os.getenv("BACKEND_API_URL","localhost:8101/internal/receive-notifications")
|
||||
api_key= return_credentials("/etc/secrets/api_key")
|
||||
|
||||
def send_notification(title:str,chapter:int, link:str):
|
||||
headers = {
|
||||
"X-API-Key-Internal": backend_api_url,
|
||||
"Content-Type": "application/json"
|
||||
}
|
||||
|
||||
data = {
|
||||
"receipent_user_id": 1,
|
||||
"message": {
|
||||
"title": title,
|
||||
"info": f"Chapter {chapter} has been released",
|
||||
"link": link
|
||||
}}
|
||||
|
||||
try:
|
||||
response = requests.post(backend_api_url, headers=headers, json=data)
|
||||
response.raise_for_status()
|
||||
print("Success: Notification sent")
|
||||
|
||||
except Timeout:
|
||||
print("Error: request timed out")
|
||||
except ConnectionError:
|
||||
print("Error: connection failed")
|
||||
except HTTPError as e:
|
||||
print(f"HTTP error: {e.response.status_code} - {e.response.text}")
|
||||
except RequestException as e:
|
||||
print("Request failed:", str(e))
|
||||
Loading…
x
Reference in New Issue
Block a user