mirror of
https://github.com/pacnpal/thrillwiki_django_no_react.git
synced 2026-02-05 11:45:18 -05:00
Compare commits
10 Commits
8ff6b7ee23
...
dependabot
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
69a4d393b0 | ||
|
|
d631f3183c | ||
|
|
2b66814d82 | ||
|
|
96df23242e | ||
|
|
692c0bbbbf | ||
|
|
22ff0d1c49 | ||
|
|
fbbfea50a3 | ||
|
|
b37aedf82e | ||
|
|
fa570334fc | ||
|
|
d9a6b4a085 |
143
.gitignore
vendored
143
.gitignore
vendored
@@ -1,143 +0,0 @@
|
||||
# Python
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
*.so
|
||||
.Python
|
||||
build/
|
||||
develop-eggs/
|
||||
dist/
|
||||
downloads/
|
||||
eggs/
|
||||
.eggs/
|
||||
lib/
|
||||
lib64/
|
||||
parts/
|
||||
sdist/
|
||||
var/
|
||||
wheels/
|
||||
share/python-wheels/
|
||||
*.egg-info/
|
||||
.installed.cfg
|
||||
*.egg
|
||||
MANIFEST
|
||||
|
||||
# Django
|
||||
*.log
|
||||
local_settings.py
|
||||
db.sqlite3
|
||||
db.sqlite3-journal
|
||||
/backend/staticfiles/
|
||||
/backend/media/
|
||||
|
||||
# Celery Beat schedule database (runtime state, regenerated automatically)
|
||||
celerybeat-schedule*
|
||||
celerybeat.pid
|
||||
|
||||
# UV
|
||||
.uv/
|
||||
backend/.uv/
|
||||
|
||||
# Generated requirements files (auto-generated from pyproject.toml)
|
||||
# Uncomment if you want to track these files
|
||||
# backend/requirements.txt
|
||||
# backend/requirements-dev.txt
|
||||
# backend/requirements-test.txt
|
||||
|
||||
# Node.js
|
||||
node_modules/
|
||||
npm-debug.log*
|
||||
yarn-debug.log*
|
||||
yarn-error.log*
|
||||
pnpm-debug.log*
|
||||
lerna-debug.log*
|
||||
.pnpm-store/
|
||||
|
||||
# Vue.js / Vite
|
||||
/frontend/dist/
|
||||
/frontend/dist-ssr/
|
||||
*.local
|
||||
|
||||
# Environment variables
|
||||
.env
|
||||
.env.local
|
||||
.env.development.local
|
||||
.env.test.local
|
||||
.env.production.local
|
||||
backend/.env
|
||||
frontend/.env
|
||||
|
||||
# IDEs
|
||||
.vscode/
|
||||
.idea/
|
||||
*.swp
|
||||
*.swo
|
||||
*.sublime-project
|
||||
*.sublime-workspace
|
||||
|
||||
# OS
|
||||
.DS_Store
|
||||
Thumbs.db
|
||||
Desktop.ini
|
||||
|
||||
# Logs
|
||||
logs/
|
||||
*.log
|
||||
|
||||
# Coverage
|
||||
coverage/
|
||||
*.lcov
|
||||
.nyc_output
|
||||
htmlcov/
|
||||
.coverage
|
||||
.coverage.*
|
||||
|
||||
# Testing
|
||||
.pytest_cache/
|
||||
.cache
|
||||
|
||||
# Temporary files
|
||||
tmp/
|
||||
temp/
|
||||
*.tmp
|
||||
*.temp
|
||||
|
||||
# Build outputs
|
||||
/dist/
|
||||
/build/
|
||||
|
||||
# Backup files
|
||||
*.bak
|
||||
*.backup
|
||||
*.orig
|
||||
*.swp
|
||||
*_backup.*
|
||||
*_OLD_*
|
||||
|
||||
# Archive files
|
||||
*.tar.gz
|
||||
*.zip
|
||||
*.rar
|
||||
|
||||
# Security
|
||||
*.pem
|
||||
*.key
|
||||
*.cert
|
||||
|
||||
# Local development
|
||||
/uploads/
|
||||
/backups/
|
||||
.django_tailwind_cli/
|
||||
backend/.env
|
||||
frontend/.env
|
||||
|
||||
# Extracted packages
|
||||
django-forwardemail/
|
||||
frontend/
|
||||
frontend
|
||||
.snapshots
|
||||
web/next-env.d.ts
|
||||
web/.next/types/cache-life.d.ts
|
||||
.gitignore
|
||||
web/.next/types/routes.d.ts
|
||||
web/.next/types/validator.ts
|
||||
@@ -32,6 +32,8 @@ class CustomAccountAdapter(DefaultAccountAdapter):
|
||||
"activate_url": activate_url,
|
||||
"current_site": current_site,
|
||||
"key": emailconfirmation.key,
|
||||
"request": request, # Include request for email backend
|
||||
"site": current_site, # Include site for email backend
|
||||
}
|
||||
email_template = "account/email/email_confirmation_signup" if signup else "account/email/email_confirmation"
|
||||
self.send_mail(email_template, emailconfirmation.email_address.email, ctx)
|
||||
|
||||
@@ -586,6 +586,264 @@ notification_priorities = ChoiceGroup(
|
||||
)
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# SECURITY EVENT TYPES
|
||||
# =============================================================================
|
||||
|
||||
security_event_types = ChoiceGroup(
|
||||
name="security_event_types",
|
||||
choices=[
|
||||
RichChoice(
|
||||
value="login_success",
|
||||
label="Login Success",
|
||||
description="User successfully logged in to their account",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "login",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "authentication",
|
||||
"sort_order": 1,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="login_failed",
|
||||
label="Login Failed",
|
||||
description="Failed login attempt to user's account",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "login",
|
||||
"css_class": "text-red-600 bg-red-50",
|
||||
"severity": "warning",
|
||||
"category": "authentication",
|
||||
"sort_order": 2,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="logout",
|
||||
label="Logout",
|
||||
description="User logged out of their account",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "logout",
|
||||
"css_class": "text-gray-600 bg-gray-50",
|
||||
"severity": "info",
|
||||
"category": "authentication",
|
||||
"sort_order": 3,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="mfa_enrolled",
|
||||
label="MFA Enrolled",
|
||||
description="User enabled two-factor authentication",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "shield-check",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "mfa",
|
||||
"sort_order": 4,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="mfa_disabled",
|
||||
label="MFA Disabled",
|
||||
description="User disabled two-factor authentication",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "shield-off",
|
||||
"css_class": "text-yellow-600 bg-yellow-50",
|
||||
"severity": "warning",
|
||||
"category": "mfa",
|
||||
"sort_order": 5,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="mfa_challenge_success",
|
||||
label="MFA Challenge Success",
|
||||
description="User successfully completed MFA verification",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "shield-check",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "mfa",
|
||||
"sort_order": 6,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="mfa_challenge_failed",
|
||||
label="MFA Challenge Failed",
|
||||
description="User failed MFA verification attempt",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "shield-x",
|
||||
"css_class": "text-red-600 bg-red-50",
|
||||
"severity": "warning",
|
||||
"category": "mfa",
|
||||
"sort_order": 7,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="passkey_registered",
|
||||
label="Passkey Registered",
|
||||
description="User registered a new passkey/WebAuthn credential",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "fingerprint",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "passkey",
|
||||
"sort_order": 8,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="passkey_removed",
|
||||
label="Passkey Removed",
|
||||
description="User removed a passkey/WebAuthn credential",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "fingerprint",
|
||||
"css_class": "text-yellow-600 bg-yellow-50",
|
||||
"severity": "warning",
|
||||
"category": "passkey",
|
||||
"sort_order": 9,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="passkey_login",
|
||||
label="Passkey Login",
|
||||
description="User logged in using a passkey",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "fingerprint",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "passkey",
|
||||
"sort_order": 10,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="social_linked",
|
||||
label="Social Account Linked",
|
||||
description="User connected a social login provider",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "link",
|
||||
"css_class": "text-blue-600 bg-blue-50",
|
||||
"severity": "info",
|
||||
"category": "social",
|
||||
"sort_order": 11,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="social_unlinked",
|
||||
label="Social Account Unlinked",
|
||||
description="User disconnected a social login provider",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "unlink",
|
||||
"css_class": "text-yellow-600 bg-yellow-50",
|
||||
"severity": "info",
|
||||
"category": "social",
|
||||
"sort_order": 12,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="password_reset_requested",
|
||||
label="Password Reset Requested",
|
||||
description="Password reset was requested for user's account",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "key",
|
||||
"css_class": "text-yellow-600 bg-yellow-50",
|
||||
"severity": "info",
|
||||
"category": "password",
|
||||
"sort_order": 13,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="password_reset_completed",
|
||||
label="Password Reset Completed",
|
||||
description="User successfully reset their password",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "key",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "password",
|
||||
"sort_order": 14,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="password_changed",
|
||||
label="Password Changed",
|
||||
description="User changed their password",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "key",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "password",
|
||||
"sort_order": 15,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="session_invalidated",
|
||||
label="Session Invalidated",
|
||||
description="User's session was terminated",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "clock",
|
||||
"css_class": "text-yellow-600 bg-yellow-50",
|
||||
"severity": "info",
|
||||
"category": "session",
|
||||
"sort_order": 16,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="recovery_code_used",
|
||||
label="Recovery Code Used",
|
||||
description="User used a recovery code for authentication",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "key",
|
||||
"css_class": "text-orange-600 bg-orange-50",
|
||||
"severity": "warning",
|
||||
"category": "mfa",
|
||||
"sort_order": 17,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="recovery_codes_regenerated",
|
||||
label="Recovery Codes Regenerated",
|
||||
description="User generated new recovery codes",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "refresh",
|
||||
"css_class": "text-blue-600 bg-blue-50",
|
||||
"severity": "info",
|
||||
"category": "mfa",
|
||||
"sort_order": 18,
|
||||
},
|
||||
),
|
||||
RichChoice(
|
||||
value="session_to_token",
|
||||
label="Passkey Login",
|
||||
description="Signed in using a passkey",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "fingerprint",
|
||||
"css_class": "text-green-600 bg-green-50",
|
||||
"severity": "info",
|
||||
"category": "authentication",
|
||||
"sort_order": 19,
|
||||
},
|
||||
),
|
||||
],
|
||||
)
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# REGISTER ALL CHOICE GROUPS
|
||||
# =============================================================================
|
||||
@@ -598,3 +856,5 @@ register_choices("privacy_levels", privacy_levels.choices, "accounts", "Privacy
|
||||
register_choices("top_list_categories", top_list_categories.choices, "accounts", "Top list category types")
|
||||
register_choices("notification_types", notification_types.choices, "accounts", "Notification type classifications")
|
||||
register_choices("notification_priorities", notification_priorities.choices, "accounts", "Notification priority levels")
|
||||
register_choices("security_event_types", security_event_types.choices, "accounts", "Security event type classifications")
|
||||
|
||||
|
||||
195
backend/apps/accounts/migrations/0017_add_security_log_model.py
Normal file
195
backend/apps/accounts/migrations/0017_add_security_log_model.py
Normal file
@@ -0,0 +1,195 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 20:48
|
||||
|
||||
import apps.core.choices.fields
|
||||
import django.db.models.deletion
|
||||
import pgtrigger.compiler
|
||||
import pgtrigger.migrations
|
||||
from django.conf import settings
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("accounts", "0016_remove_emailverification_insert_insert_and_more"),
|
||||
("pghistory", "0007_auto_20250421_0444"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.CreateModel(
|
||||
name="SecurityLog",
|
||||
fields=[
|
||||
("id", models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name="ID")),
|
||||
(
|
||||
"event_type",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="security_event_types",
|
||||
choices=[
|
||||
("login_success", "Login Success"),
|
||||
("login_failed", "Login Failed"),
|
||||
("logout", "Logout"),
|
||||
("mfa_enrolled", "MFA Enrolled"),
|
||||
("mfa_disabled", "MFA Disabled"),
|
||||
("mfa_challenge_success", "MFA Challenge Success"),
|
||||
("mfa_challenge_failed", "MFA Challenge Failed"),
|
||||
("passkey_registered", "Passkey Registered"),
|
||||
("passkey_removed", "Passkey Removed"),
|
||||
("passkey_login", "Passkey Login"),
|
||||
("social_linked", "Social Account Linked"),
|
||||
("social_unlinked", "Social Account Unlinked"),
|
||||
("password_reset_requested", "Password Reset Requested"),
|
||||
("password_reset_completed", "Password Reset Completed"),
|
||||
("password_changed", "Password Changed"),
|
||||
("session_invalidated", "Session Invalidated"),
|
||||
("recovery_code_used", "Recovery Code Used"),
|
||||
("recovery_codes_regenerated", "Recovery Codes Regenerated"),
|
||||
],
|
||||
db_index=True,
|
||||
domain="accounts",
|
||||
help_text="Type of security event",
|
||||
max_length=50,
|
||||
),
|
||||
),
|
||||
("ip_address", models.GenericIPAddressField(help_text="IP address of the request")),
|
||||
("user_agent", models.TextField(blank=True, help_text="User agent string from the request")),
|
||||
("metadata", models.JSONField(blank=True, default=dict, help_text="Additional event-specific data")),
|
||||
("created_at", models.DateTimeField(auto_now_add=True, help_text="When this event occurred")),
|
||||
(
|
||||
"user",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="User this event is associated with",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.CASCADE,
|
||||
related_name="security_logs",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Security Log",
|
||||
"verbose_name_plural": "Security Logs",
|
||||
"ordering": ["-created_at"],
|
||||
},
|
||||
),
|
||||
migrations.CreateModel(
|
||||
name="SecurityLogEvent",
|
||||
fields=[
|
||||
("pgh_id", models.AutoField(primary_key=True, serialize=False)),
|
||||
("pgh_created_at", models.DateTimeField(auto_now_add=True)),
|
||||
("pgh_label", models.TextField(help_text="The event label.")),
|
||||
("id", models.BigIntegerField()),
|
||||
(
|
||||
"event_type",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="security_event_types",
|
||||
choices=[
|
||||
("login_success", "Login Success"),
|
||||
("login_failed", "Login Failed"),
|
||||
("logout", "Logout"),
|
||||
("mfa_enrolled", "MFA Enrolled"),
|
||||
("mfa_disabled", "MFA Disabled"),
|
||||
("mfa_challenge_success", "MFA Challenge Success"),
|
||||
("mfa_challenge_failed", "MFA Challenge Failed"),
|
||||
("passkey_registered", "Passkey Registered"),
|
||||
("passkey_removed", "Passkey Removed"),
|
||||
("passkey_login", "Passkey Login"),
|
||||
("social_linked", "Social Account Linked"),
|
||||
("social_unlinked", "Social Account Unlinked"),
|
||||
("password_reset_requested", "Password Reset Requested"),
|
||||
("password_reset_completed", "Password Reset Completed"),
|
||||
("password_changed", "Password Changed"),
|
||||
("session_invalidated", "Session Invalidated"),
|
||||
("recovery_code_used", "Recovery Code Used"),
|
||||
("recovery_codes_regenerated", "Recovery Codes Regenerated"),
|
||||
],
|
||||
domain="accounts",
|
||||
help_text="Type of security event",
|
||||
max_length=50,
|
||||
),
|
||||
),
|
||||
("ip_address", models.GenericIPAddressField(help_text="IP address of the request")),
|
||||
("user_agent", models.TextField(blank=True, help_text="User agent string from the request")),
|
||||
("metadata", models.JSONField(blank=True, default=dict, help_text="Additional event-specific data")),
|
||||
("created_at", models.DateTimeField(auto_now_add=True, help_text="When this event occurred")),
|
||||
(
|
||||
"pgh_context",
|
||||
models.ForeignKey(
|
||||
db_constraint=False,
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.DO_NOTHING,
|
||||
related_name="+",
|
||||
to="pghistory.context",
|
||||
),
|
||||
),
|
||||
(
|
||||
"pgh_obj",
|
||||
models.ForeignKey(
|
||||
db_constraint=False,
|
||||
on_delete=django.db.models.deletion.DO_NOTHING,
|
||||
related_name="events",
|
||||
to="accounts.securitylog",
|
||||
),
|
||||
),
|
||||
(
|
||||
"user",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
db_constraint=False,
|
||||
help_text="User this event is associated with",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.DO_NOTHING,
|
||||
related_name="+",
|
||||
related_query_name="+",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"abstract": False,
|
||||
},
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="securitylog",
|
||||
index=models.Index(fields=["user", "-created_at"], name="accounts_se_user_id_d46023_idx"),
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="securitylog",
|
||||
index=models.Index(fields=["event_type", "-created_at"], name="accounts_se_event_t_814971_idx"),
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="securitylog",
|
||||
index=models.Index(fields=["ip_address", "-created_at"], name="accounts_se_ip_addr_2a19c8_idx"),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="securitylog",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="insert_insert",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
func='INSERT INTO "accounts_securitylogevent" ("created_at", "event_type", "id", "ip_address", "metadata", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "user_agent", "user_id") VALUES (NEW."created_at", NEW."event_type", NEW."id", NEW."ip_address", NEW."metadata", _pgh_attach_context(), NOW(), \'insert\', NEW."id", NEW."user_agent", NEW."user_id"); RETURN NULL;',
|
||||
hash="a40cf3f6fa9e8cda99f7204edb226b26bbe03eda",
|
||||
operation="INSERT",
|
||||
pgid="pgtrigger_insert_insert_5d4cf",
|
||||
table="accounts_securitylog",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="securitylog",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="update_update",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
condition="WHEN (OLD.* IS DISTINCT FROM NEW.*)",
|
||||
func='INSERT INTO "accounts_securitylogevent" ("created_at", "event_type", "id", "ip_address", "metadata", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "user_agent", "user_id") VALUES (NEW."created_at", NEW."event_type", NEW."id", NEW."ip_address", NEW."metadata", _pgh_attach_context(), NOW(), \'update\', NEW."id", NEW."user_agent", NEW."user_id"); RETURN NULL;',
|
||||
hash="244fc44bdaff1bf2d557f09ae452a9ea77274068",
|
||||
operation="UPDATE",
|
||||
pgid="pgtrigger_update_update_d4645",
|
||||
table="accounts_securitylog",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -620,6 +620,111 @@ class NotificationPreference(TrackedModel):
|
||||
return getattr(self, field_name, False)
|
||||
|
||||
|
||||
@pghistory.track()
|
||||
class SecurityLog(models.Model):
|
||||
"""
|
||||
Model to track security-relevant authentication events.
|
||||
|
||||
All security-critical events are logged here for audit purposes,
|
||||
including logins, MFA changes, password changes, and session management.
|
||||
"""
|
||||
|
||||
user = models.ForeignKey(
|
||||
User,
|
||||
on_delete=models.CASCADE,
|
||||
related_name="security_logs",
|
||||
null=True, # Allow null for failed login attempts with no valid user
|
||||
blank=True,
|
||||
help_text="User this event is associated with",
|
||||
)
|
||||
event_type = RichChoiceField(
|
||||
choice_group="security_event_types",
|
||||
domain="accounts",
|
||||
max_length=50,
|
||||
db_index=True,
|
||||
help_text="Type of security event",
|
||||
)
|
||||
ip_address = models.GenericIPAddressField(
|
||||
help_text="IP address of the request",
|
||||
)
|
||||
user_agent = models.TextField(
|
||||
blank=True,
|
||||
help_text="User agent string from the request",
|
||||
)
|
||||
metadata = models.JSONField(
|
||||
default=dict,
|
||||
blank=True,
|
||||
help_text="Additional event-specific data",
|
||||
)
|
||||
created_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
help_text="When this event occurred",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["-created_at"]
|
||||
indexes = [
|
||||
models.Index(fields=["user", "-created_at"]),
|
||||
models.Index(fields=["event_type", "-created_at"]),
|
||||
models.Index(fields=["ip_address", "-created_at"]),
|
||||
]
|
||||
verbose_name = "Security Log"
|
||||
verbose_name_plural = "Security Logs"
|
||||
|
||||
def __str__(self):
|
||||
username = self.user.username if self.user else "Unknown"
|
||||
return f"{self.get_event_type_display()} - {username} at {self.created_at}"
|
||||
|
||||
@classmethod
|
||||
def log_event(
|
||||
cls,
|
||||
event_type: str,
|
||||
ip_address: str,
|
||||
user=None,
|
||||
user_agent: str = "",
|
||||
metadata: dict = None,
|
||||
) -> "SecurityLog":
|
||||
"""
|
||||
Create a new security log entry.
|
||||
|
||||
Args:
|
||||
event_type: One of security_event_types choices (e.g., "login_success")
|
||||
ip_address: Client IP address
|
||||
user: User instance (optional for failed logins)
|
||||
user_agent: Browser user agent string
|
||||
metadata: Additional event-specific data
|
||||
|
||||
Returns:
|
||||
The created SecurityLog instance
|
||||
"""
|
||||
return cls.objects.create(
|
||||
user=user,
|
||||
event_type=event_type,
|
||||
ip_address=ip_address,
|
||||
user_agent=user_agent,
|
||||
metadata=metadata or {},
|
||||
)
|
||||
|
||||
@classmethod
|
||||
def get_recent_for_user(cls, user, limit: int = 20):
|
||||
"""Get recent security events for a user."""
|
||||
return cls.objects.filter(user=user).order_by("-created_at")[:limit]
|
||||
|
||||
@classmethod
|
||||
def get_failed_login_count(cls, ip_address: str, minutes: int = 15) -> int:
|
||||
"""Count failed login attempts from an IP in the last N minutes."""
|
||||
from datetime import timedelta
|
||||
|
||||
from django.utils import timezone
|
||||
|
||||
cutoff = timezone.now() - timedelta(minutes=minutes)
|
||||
return cls.objects.filter(
|
||||
event_type="login_failed",
|
||||
ip_address=ip_address,
|
||||
created_at__gte=cutoff,
|
||||
).count()
|
||||
|
||||
|
||||
# Signal handlers for automatic notification preference creation
|
||||
|
||||
|
||||
|
||||
@@ -261,7 +261,7 @@ class UserDeletionService:
|
||||
"is_active": False,
|
||||
"is_staff": False,
|
||||
"is_superuser": False,
|
||||
"role": User.Roles.USER,
|
||||
"role": "USER",
|
||||
"is_banned": True,
|
||||
"ban_reason": "System placeholder for deleted users",
|
||||
"ban_date": timezone.now(),
|
||||
@@ -389,7 +389,7 @@ class UserDeletionService:
|
||||
)
|
||||
|
||||
# Check if user has critical admin role
|
||||
if user.role == User.Roles.ADMIN and user.is_staff:
|
||||
if user.role == "ADMIN" and user.is_staff:
|
||||
return (
|
||||
False,
|
||||
"Admin accounts with staff privileges cannot be deleted. Please remove admin privileges first or contact system administrator.",
|
||||
|
||||
@@ -5,7 +5,26 @@ This package contains business logic services for account management,
|
||||
including social provider management, user authentication, and profile services.
|
||||
"""
|
||||
|
||||
from .account_service import AccountService
|
||||
from .social_provider_service import SocialProviderService
|
||||
from .user_deletion_service import UserDeletionService
|
||||
from .security_service import (
|
||||
get_client_ip,
|
||||
log_security_event,
|
||||
log_security_event_simple,
|
||||
send_security_notification,
|
||||
check_auth_method_availability,
|
||||
invalidate_user_sessions,
|
||||
)
|
||||
|
||||
__all__ = ["SocialProviderService", "UserDeletionService"]
|
||||
__all__ = [
|
||||
"AccountService",
|
||||
"SocialProviderService",
|
||||
"UserDeletionService",
|
||||
"get_client_ip",
|
||||
"log_security_event",
|
||||
"log_security_event_simple",
|
||||
"send_security_notification",
|
||||
"check_auth_method_availability",
|
||||
"invalidate_user_sessions",
|
||||
]
|
||||
|
||||
199
backend/apps/accounts/services/account_service.py
Normal file
199
backend/apps/accounts/services/account_service.py
Normal file
@@ -0,0 +1,199 @@
|
||||
"""
|
||||
Account management service for ThrillWiki.
|
||||
|
||||
Provides password validation, password changes, and email change functionality.
|
||||
"""
|
||||
|
||||
import re
|
||||
import secrets
|
||||
from typing import TYPE_CHECKING
|
||||
|
||||
from django.core.mail import send_mail
|
||||
from django.template.loader import render_to_string
|
||||
from django.utils import timezone
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from django.http import HttpRequest
|
||||
|
||||
from apps.accounts.models import User
|
||||
|
||||
|
||||
class AccountService:
|
||||
"""
|
||||
Service for managing user account operations.
|
||||
|
||||
Handles password validation, password changes, and email changes
|
||||
with proper verification flows.
|
||||
"""
|
||||
|
||||
# Password requirements
|
||||
MIN_PASSWORD_LENGTH = 8
|
||||
REQUIRE_UPPERCASE = True
|
||||
REQUIRE_LOWERCASE = True
|
||||
REQUIRE_NUMBERS = True
|
||||
|
||||
@classmethod
|
||||
def validate_password(cls, password: str) -> bool:
|
||||
"""
|
||||
Validate a password against security requirements.
|
||||
|
||||
Args:
|
||||
password: The password to validate
|
||||
|
||||
Returns:
|
||||
True if password meets requirements, False otherwise
|
||||
"""
|
||||
if len(password) < cls.MIN_PASSWORD_LENGTH:
|
||||
return False
|
||||
|
||||
if cls.REQUIRE_UPPERCASE and not re.search(r"[A-Z]", password):
|
||||
return False
|
||||
|
||||
if cls.REQUIRE_LOWERCASE and not re.search(r"[a-z]", password):
|
||||
return False
|
||||
|
||||
if cls.REQUIRE_NUMBERS and not re.search(r"[0-9]", password):
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
@classmethod
|
||||
def change_password(
|
||||
cls,
|
||||
user: "User",
|
||||
old_password: str,
|
||||
new_password: str,
|
||||
request: "HttpRequest | None" = None,
|
||||
) -> dict:
|
||||
"""
|
||||
Change a user's password.
|
||||
|
||||
Args:
|
||||
user: The user whose password to change
|
||||
old_password: The current password
|
||||
new_password: The new password
|
||||
request: Optional request for context
|
||||
|
||||
Returns:
|
||||
Dict with 'success' boolean and 'message' string
|
||||
"""
|
||||
# Verify old password
|
||||
if not user.check_password(old_password):
|
||||
return {
|
||||
"success": False,
|
||||
"message": "Current password is incorrect.",
|
||||
}
|
||||
|
||||
# Validate new password
|
||||
if not cls.validate_password(new_password):
|
||||
return {
|
||||
"success": False,
|
||||
"message": f"New password must be at least {cls.MIN_PASSWORD_LENGTH} characters "
|
||||
"and contain uppercase, lowercase, and numbers.",
|
||||
}
|
||||
|
||||
# Change the password
|
||||
user.set_password(new_password)
|
||||
user.save(update_fields=["password"])
|
||||
|
||||
# Send confirmation email
|
||||
cls._send_password_change_confirmation(user, request)
|
||||
|
||||
return {
|
||||
"success": True,
|
||||
"message": "Password changed successfully.",
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def _send_password_change_confirmation(
|
||||
cls,
|
||||
user: "User",
|
||||
request: "HttpRequest | None" = None,
|
||||
) -> None:
|
||||
"""Send a confirmation email after password change."""
|
||||
try:
|
||||
send_mail(
|
||||
subject="Password Changed - ThrillWiki",
|
||||
message=f"Hi {user.username},\n\nYour password has been changed successfully.\n\n"
|
||||
"If you did not make this change, please contact support immediately.",
|
||||
from_email=None, # Uses DEFAULT_FROM_EMAIL
|
||||
recipient_list=[user.email],
|
||||
fail_silently=True,
|
||||
)
|
||||
except Exception:
|
||||
pass # Don't fail the password change if email fails
|
||||
|
||||
@classmethod
|
||||
def initiate_email_change(
|
||||
cls,
|
||||
user: "User",
|
||||
new_email: str,
|
||||
request: "HttpRequest | None" = None,
|
||||
) -> dict:
|
||||
"""
|
||||
Initiate an email change request.
|
||||
|
||||
Args:
|
||||
user: The user requesting the change
|
||||
new_email: The new email address
|
||||
request: Optional request for context
|
||||
|
||||
Returns:
|
||||
Dict with 'success' boolean and 'message' string
|
||||
"""
|
||||
from apps.accounts.models import User
|
||||
|
||||
# Validate email
|
||||
if not new_email or not new_email.strip():
|
||||
return {
|
||||
"success": False,
|
||||
"message": "Email address is required.",
|
||||
}
|
||||
|
||||
new_email = new_email.strip().lower()
|
||||
|
||||
# Check if email already in use
|
||||
if User.objects.filter(email=new_email).exclude(pk=user.pk).exists():
|
||||
return {
|
||||
"success": False,
|
||||
"message": "This email is already in use by another account.",
|
||||
}
|
||||
|
||||
# Store pending email
|
||||
user.pending_email = new_email
|
||||
user.save(update_fields=["pending_email"])
|
||||
|
||||
# Send verification email
|
||||
cls._send_email_verification(user, new_email, request)
|
||||
|
||||
return {
|
||||
"success": True,
|
||||
"message": "Verification email sent. Please check your inbox.",
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def _send_email_verification(
|
||||
cls,
|
||||
user: "User",
|
||||
new_email: str,
|
||||
request: "HttpRequest | None" = None,
|
||||
) -> None:
|
||||
"""Send verification email for email change."""
|
||||
verification_code = secrets.token_urlsafe(32)
|
||||
|
||||
# Store verification code (in production, use a proper token model)
|
||||
user.email_verification_code = verification_code
|
||||
user.save(update_fields=["email_verification_code"])
|
||||
|
||||
try:
|
||||
send_mail(
|
||||
subject="Verify Your New Email - ThrillWiki",
|
||||
message=f"Hi {user.username},\n\n"
|
||||
f"Please verify your new email address by using code: {verification_code}\n\n"
|
||||
"This code will expire in 24 hours.",
|
||||
from_email=None,
|
||||
recipient_list=[new_email],
|
||||
fail_silently=True,
|
||||
)
|
||||
except Exception:
|
||||
pass
|
||||
402
backend/apps/accounts/services/security_service.py
Normal file
402
backend/apps/accounts/services/security_service.py
Normal file
@@ -0,0 +1,402 @@
|
||||
"""
|
||||
Security Service for ThrillWiki
|
||||
|
||||
Provides centralized security event logging, notifications, and helper functions
|
||||
for all authentication-related operations.
|
||||
"""
|
||||
|
||||
import logging
|
||||
from typing import Any
|
||||
|
||||
from django.conf import settings
|
||||
from django.core.mail import send_mail
|
||||
from django.template.loader import render_to_string
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def get_client_ip(request) -> str:
|
||||
"""
|
||||
Extract client IP from request, handling proxies correctly.
|
||||
|
||||
Args:
|
||||
request: Django/DRF request object
|
||||
|
||||
Returns:
|
||||
Client IP address as string
|
||||
"""
|
||||
# Check for proxy headers first
|
||||
x_forwarded_for = request.META.get("HTTP_X_FORWARDED_FOR")
|
||||
if x_forwarded_for:
|
||||
# X-Forwarded-For can contain multiple IPs; take the first one
|
||||
return x_forwarded_for.split(",")[0].strip()
|
||||
|
||||
# Check for Cloudflare's CF-Connecting-IP header
|
||||
cf_connecting_ip = request.META.get("HTTP_CF_CONNECTING_IP")
|
||||
if cf_connecting_ip:
|
||||
return cf_connecting_ip
|
||||
|
||||
# Fallback to REMOTE_ADDR
|
||||
return request.META.get("REMOTE_ADDR", "0.0.0.0")
|
||||
|
||||
|
||||
def log_security_event(
|
||||
event_type: str,
|
||||
request,
|
||||
user=None,
|
||||
metadata: dict = None
|
||||
) -> Any:
|
||||
"""
|
||||
Log a security event with request context.
|
||||
|
||||
Args:
|
||||
event_type: One of SecurityLog.EventType choices
|
||||
request: Django/DRF request object
|
||||
user: User instance (optional for failed logins)
|
||||
metadata: Additional event-specific data
|
||||
|
||||
Returns:
|
||||
The created SecurityLog instance
|
||||
"""
|
||||
from apps.accounts.models import SecurityLog
|
||||
|
||||
try:
|
||||
return SecurityLog.log_event(
|
||||
event_type=event_type,
|
||||
ip_address=get_client_ip(request),
|
||||
user=user,
|
||||
user_agent=request.META.get("HTTP_USER_AGENT", ""),
|
||||
metadata=metadata or {},
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to log security event {event_type}: {e}")
|
||||
return None
|
||||
|
||||
|
||||
def log_security_event_simple(
|
||||
event_type: str,
|
||||
ip_address: str,
|
||||
user=None,
|
||||
user_agent: str = "",
|
||||
metadata: dict = None
|
||||
) -> Any:
|
||||
"""
|
||||
Log a security event without request context.
|
||||
|
||||
Use this when you don't have access to the request object.
|
||||
|
||||
Args:
|
||||
event_type: One of SecurityLog.EventType choices
|
||||
ip_address: Client IP address
|
||||
user: User instance (optional)
|
||||
user_agent: Browser user agent string
|
||||
metadata: Additional event-specific data
|
||||
|
||||
Returns:
|
||||
The created SecurityLog instance
|
||||
"""
|
||||
from apps.accounts.models import SecurityLog
|
||||
|
||||
try:
|
||||
return SecurityLog.log_event(
|
||||
event_type=event_type,
|
||||
ip_address=ip_address,
|
||||
user=user,
|
||||
user_agent=user_agent,
|
||||
metadata=metadata or {},
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to log security event {event_type}: {e}")
|
||||
return None
|
||||
|
||||
|
||||
# Subject line mapping for security notifications
|
||||
SECURITY_NOTIFICATION_SUBJECTS = {
|
||||
"mfa_enrolled": "Two-Factor Authentication Enabled",
|
||||
"mfa_disabled": "Two-Factor Authentication Disabled",
|
||||
"passkey_registered": "New Passkey Added to Your Account",
|
||||
"passkey_removed": "Passkey Removed from Your Account",
|
||||
"password_changed": "Your Password Was Changed",
|
||||
"password_reset_completed": "Your Password Has Been Reset",
|
||||
"social_linked": "Social Account Connected",
|
||||
"social_unlinked": "Social Account Disconnected",
|
||||
"session_invalidated": "Session Security Update",
|
||||
"recovery_codes_regenerated": "Recovery Codes Regenerated",
|
||||
}
|
||||
|
||||
|
||||
def send_security_notification(
|
||||
user,
|
||||
event_type: str,
|
||||
metadata: dict = None
|
||||
) -> bool:
|
||||
"""
|
||||
Send email notification for security-sensitive events.
|
||||
|
||||
This function sends an email to the user when important security
|
||||
events occur on their account.
|
||||
|
||||
Args:
|
||||
user: User instance to notify
|
||||
event_type: Type of security event (used to select template and subject)
|
||||
metadata: Additional context for the email template
|
||||
|
||||
Returns:
|
||||
True if email was sent successfully, False otherwise
|
||||
"""
|
||||
if not user or not user.email:
|
||||
logger.warning(f"Cannot send security notification: no email for user")
|
||||
return False
|
||||
|
||||
# Check if user has security notifications enabled
|
||||
if hasattr(user, "notification_preference"):
|
||||
prefs = user.notification_preference
|
||||
if not getattr(prefs, "account_security_email", True):
|
||||
logger.debug(f"User {user.username} has security emails disabled")
|
||||
return False
|
||||
|
||||
try:
|
||||
subject = f"ThrillWiki Security Alert: {SECURITY_NOTIFICATION_SUBJECTS.get(event_type, 'Account Activity')}"
|
||||
|
||||
context = {
|
||||
"user": user,
|
||||
"event_type": event_type,
|
||||
"event_display": SECURITY_NOTIFICATION_SUBJECTS.get(event_type, "Account Activity"),
|
||||
"metadata": metadata or {},
|
||||
"site_name": "ThrillWiki",
|
||||
"support_email": getattr(settings, "DEFAULT_SUPPORT_EMAIL", "support@thrillwiki.com"),
|
||||
}
|
||||
|
||||
# Try to render HTML template, fallback to plain text
|
||||
try:
|
||||
html_message = render_to_string("accounts/email/security_notification.html", context)
|
||||
except Exception as template_error:
|
||||
logger.debug(f"HTML template not found, using fallback: {template_error}")
|
||||
html_message = _get_fallback_security_email(context)
|
||||
|
||||
# Plain text version
|
||||
text_message = _get_plain_text_security_email(context)
|
||||
|
||||
send_mail(
|
||||
subject=subject,
|
||||
message=text_message,
|
||||
from_email=settings.DEFAULT_FROM_EMAIL,
|
||||
recipient_list=[user.email],
|
||||
html_message=html_message,
|
||||
fail_silently=False,
|
||||
)
|
||||
|
||||
logger.info(f"Security notification sent to {user.email} for event: {event_type}")
|
||||
return True
|
||||
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to send security notification to {user.email}: {e}")
|
||||
return False
|
||||
|
||||
|
||||
def _get_plain_text_security_email(context: dict) -> str:
|
||||
"""Generate plain text email for security notifications."""
|
||||
event_display = context.get("event_display", "Account Activity")
|
||||
user = context.get("user")
|
||||
metadata = context.get("metadata", {})
|
||||
|
||||
lines = [
|
||||
f"Hello {user.get_display_name() if user else 'User'},",
|
||||
"",
|
||||
f"This is a security notification from ThrillWiki.",
|
||||
"",
|
||||
f"Event: {event_display}",
|
||||
]
|
||||
|
||||
# Add metadata details
|
||||
if metadata:
|
||||
lines.append("")
|
||||
lines.append("Details:")
|
||||
for key, value in metadata.items():
|
||||
if key not in ("user_id", "internal"):
|
||||
lines.append(f" - {key.replace('_', ' ').title()}: {value}")
|
||||
|
||||
lines.extend([
|
||||
"",
|
||||
"If you did not perform this action, please secure your account immediately:",
|
||||
"1. Change your password",
|
||||
"2. Review your connected devices and sign out any you don't recognize",
|
||||
"3. Contact support if you need assistance",
|
||||
"",
|
||||
"Best regards,",
|
||||
"The ThrillWiki Team",
|
||||
])
|
||||
|
||||
return "\n".join(lines)
|
||||
|
||||
|
||||
def _get_fallback_security_email(context: dict) -> str:
|
||||
"""Generate HTML email for security notifications when template not found."""
|
||||
event_display = context.get("event_display", "Account Activity")
|
||||
user = context.get("user")
|
||||
metadata = context.get("metadata", {})
|
||||
|
||||
metadata_html = ""
|
||||
if metadata:
|
||||
items = []
|
||||
for key, value in metadata.items():
|
||||
if key not in ("user_id", "internal"):
|
||||
items.append(f"<li><strong>{key.replace('_', ' ').title()}:</strong> {value}</li>")
|
||||
if items:
|
||||
metadata_html = f"<h3>Details:</h3><ul>{''.join(items)}</ul>"
|
||||
|
||||
return f"""
|
||||
<!DOCTYPE html>
|
||||
<html>
|
||||
<head>
|
||||
<style>
|
||||
body {{ font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, sans-serif; line-height: 1.6; color: #333; }}
|
||||
.container {{ max-width: 600px; margin: 0 auto; padding: 20px; }}
|
||||
.header {{ background: linear-gradient(135deg, #667eea 0%, #764ba2 100%); padding: 20px; border-radius: 8px 8px 0 0; }}
|
||||
.header h1 {{ color: white; margin: 0; font-size: 24px; }}
|
||||
.content {{ background: #f9f9f9; padding: 30px; border-radius: 0 0 8px 8px; }}
|
||||
.alert {{ background: #fff3cd; border-left: 4px solid #ffc107; padding: 15px; margin: 20px 0; }}
|
||||
.footer {{ text-align: center; color: #666; font-size: 12px; margin-top: 20px; }}
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
<div class="container">
|
||||
<div class="header">
|
||||
<h1>🔒 Security Alert</h1>
|
||||
</div>
|
||||
<div class="content">
|
||||
<p>Hello {user.get_display_name() if user else 'User'},</p>
|
||||
<p>This is a security notification from ThrillWiki.</p>
|
||||
<h2>{event_display}</h2>
|
||||
{metadata_html}
|
||||
<div class="alert">
|
||||
<strong>Didn't do this?</strong><br>
|
||||
If you did not perform this action, please secure your account immediately by changing your password and reviewing your connected devices.
|
||||
</div>
|
||||
</div>
|
||||
<div class="footer">
|
||||
<p>This is an automated security notification from ThrillWiki.</p>
|
||||
</div>
|
||||
</div>
|
||||
</body>
|
||||
</html>
|
||||
"""
|
||||
|
||||
|
||||
def check_auth_method_availability(user) -> dict:
|
||||
"""
|
||||
Check what authentication methods a user has available.
|
||||
|
||||
This is used to prevent users from removing their last auth method.
|
||||
|
||||
Args:
|
||||
user: User instance to check
|
||||
|
||||
Returns:
|
||||
Dictionary with auth method availability:
|
||||
{
|
||||
"has_password": bool,
|
||||
"has_totp": bool,
|
||||
"has_passkey": bool,
|
||||
"passkey_count": int,
|
||||
"has_social": bool,
|
||||
"social_providers": list[str],
|
||||
"total_methods": int,
|
||||
"can_remove_mfa": bool,
|
||||
"can_remove_passkey": bool,
|
||||
"can_remove_social": bool,
|
||||
}
|
||||
"""
|
||||
try:
|
||||
from allauth.mfa.models import Authenticator
|
||||
except ImportError:
|
||||
Authenticator = None
|
||||
|
||||
result = {
|
||||
"has_password": user.has_usable_password(),
|
||||
"has_totp": False,
|
||||
"has_passkey": False,
|
||||
"passkey_count": 0,
|
||||
"has_social": False,
|
||||
"social_providers": [],
|
||||
"total_methods": 0,
|
||||
}
|
||||
|
||||
# Check MFA authenticators
|
||||
if Authenticator:
|
||||
result["has_totp"] = Authenticator.objects.filter(
|
||||
user=user, type=Authenticator.Type.TOTP
|
||||
).exists()
|
||||
|
||||
passkey_count = Authenticator.objects.filter(
|
||||
user=user, type=Authenticator.Type.WEBAUTHN
|
||||
).count()
|
||||
result["passkey_count"] = passkey_count
|
||||
result["has_passkey"] = passkey_count > 0
|
||||
|
||||
# Check social accounts
|
||||
if hasattr(user, "socialaccount_set"):
|
||||
social_accounts = user.socialaccount_set.all()
|
||||
result["has_social"] = social_accounts.exists()
|
||||
result["social_providers"] = list(social_accounts.values_list("provider", flat=True))
|
||||
|
||||
# Calculate total methods (counting passkeys as one method regardless of count)
|
||||
result["total_methods"] = sum([
|
||||
result["has_password"],
|
||||
result["has_passkey"],
|
||||
result["has_social"],
|
||||
])
|
||||
|
||||
# Determine what can be safely removed
|
||||
# User must always have at least one primary auth method remaining
|
||||
result["can_remove_mfa"] = result["total_methods"] >= 1
|
||||
result["can_remove_passkey"] = (
|
||||
result["total_methods"] > 1 or
|
||||
(result["passkey_count"] > 1) or
|
||||
result["has_password"] or
|
||||
result["has_social"]
|
||||
)
|
||||
result["can_remove_social"] = (
|
||||
result["total_methods"] > 1 or
|
||||
result["has_password"] or
|
||||
result["has_passkey"]
|
||||
)
|
||||
|
||||
return result
|
||||
|
||||
|
||||
def invalidate_user_sessions(user, exclude_current: bool = False, request=None) -> int:
|
||||
"""
|
||||
Invalidate all JWT tokens for a user.
|
||||
|
||||
This is used after security-sensitive operations like password reset.
|
||||
|
||||
Args:
|
||||
user: User whose sessions to invalidate
|
||||
exclude_current: If True and request is provided, keep current session
|
||||
request: Current request (used if exclude_current is True)
|
||||
|
||||
Returns:
|
||||
Number of tokens invalidated
|
||||
"""
|
||||
try:
|
||||
from rest_framework_simplejwt.token_blacklist.models import (
|
||||
BlacklistedToken,
|
||||
OutstandingToken,
|
||||
)
|
||||
except ImportError:
|
||||
logger.warning("JWT token blacklist not available")
|
||||
return 0
|
||||
|
||||
count = 0
|
||||
outstanding_tokens = OutstandingToken.objects.filter(user=user)
|
||||
|
||||
for token in outstanding_tokens:
|
||||
try:
|
||||
BlacklistedToken.objects.get_or_create(token=token)
|
||||
count += 1
|
||||
except Exception as e:
|
||||
logger.debug(f"Could not blacklist token: {e}")
|
||||
|
||||
logger.info(f"Invalidated {count} tokens for user {user.username}")
|
||||
return count
|
||||
@@ -38,9 +38,32 @@ class UserDeletionRequest:
|
||||
class UserDeletionService:
|
||||
"""Service for handling user account deletion with submission preservation."""
|
||||
|
||||
# Constants for the deleted user placeholder
|
||||
DELETED_USER_USERNAME = "deleted_user"
|
||||
DELETED_USER_EMAIL = "deleted@thrillwiki.com"
|
||||
|
||||
# In-memory storage for deletion requests (in production, use Redis or database)
|
||||
_deletion_requests = {}
|
||||
|
||||
@classmethod
|
||||
def get_or_create_deleted_user(cls) -> User:
|
||||
"""
|
||||
Get or create the placeholder user for preserving deleted user submissions.
|
||||
|
||||
Returns:
|
||||
User: The deleted user placeholder
|
||||
"""
|
||||
deleted_user, created = User.objects.get_or_create(
|
||||
username=cls.DELETED_USER_USERNAME,
|
||||
defaults={
|
||||
"email": cls.DELETED_USER_EMAIL,
|
||||
"is_active": False,
|
||||
"is_banned": True,
|
||||
"ban_date": timezone.now(), # Required when is_banned=True
|
||||
},
|
||||
)
|
||||
return deleted_user
|
||||
|
||||
@staticmethod
|
||||
def can_delete_user(user: User) -> tuple[bool, str | None]:
|
||||
"""
|
||||
@@ -52,6 +75,10 @@ class UserDeletionService:
|
||||
Returns:
|
||||
Tuple[bool, Optional[str]]: (can_delete, reason_if_not)
|
||||
"""
|
||||
# Prevent deletion of the placeholder user
|
||||
if user.username == UserDeletionService.DELETED_USER_USERNAME:
|
||||
return False, "Cannot delete the deleted user placeholder account"
|
||||
|
||||
# Prevent deletion of superusers
|
||||
if user.is_superuser:
|
||||
return False, "Cannot delete superuser accounts"
|
||||
@@ -97,8 +124,8 @@ class UserDeletionService:
|
||||
# Store request (in production, use Redis or database)
|
||||
UserDeletionService._deletion_requests[verification_code] = deletion_request
|
||||
|
||||
# Send verification email
|
||||
UserDeletionService._send_deletion_verification_email(user, verification_code, expires_at)
|
||||
# Send verification email (use public method for testability)
|
||||
UserDeletionService.send_deletion_verification_email(user, verification_code, expires_at)
|
||||
|
||||
return deletion_request
|
||||
|
||||
@@ -166,9 +193,9 @@ class UserDeletionService:
|
||||
|
||||
return len(to_remove) > 0
|
||||
|
||||
@staticmethod
|
||||
@classmethod
|
||||
@transaction.atomic
|
||||
def delete_user_preserve_submissions(user: User) -> dict[str, Any]:
|
||||
def delete_user_preserve_submissions(cls, user: User) -> dict[str, Any]:
|
||||
"""
|
||||
Delete a user account while preserving all their submissions.
|
||||
|
||||
@@ -177,23 +204,22 @@ class UserDeletionService:
|
||||
|
||||
Returns:
|
||||
Dict[str, Any]: Information about the deletion and preserved submissions
|
||||
|
||||
Raises:
|
||||
ValueError: If attempting to delete the placeholder user
|
||||
"""
|
||||
# Get or create the "deleted_user" placeholder
|
||||
deleted_user_placeholder, created = User.objects.get_or_create(
|
||||
username="deleted_user",
|
||||
defaults={
|
||||
"email": "deleted@thrillwiki.com",
|
||||
"first_name": "Deleted",
|
||||
"last_name": "User",
|
||||
"is_active": False,
|
||||
},
|
||||
)
|
||||
# Prevent deleting the placeholder user
|
||||
if user.username == cls.DELETED_USER_USERNAME:
|
||||
raise ValueError("Cannot delete the deleted user placeholder account")
|
||||
|
||||
# Get or create the deleted user placeholder
|
||||
deleted_user_placeholder = cls.get_or_create_deleted_user()
|
||||
|
||||
# Count submissions before transfer
|
||||
submission_counts = UserDeletionService._count_user_submissions(user)
|
||||
submission_counts = cls._count_user_submissions(user)
|
||||
|
||||
# Transfer submissions to placeholder user
|
||||
UserDeletionService._transfer_user_submissions(user, deleted_user_placeholder)
|
||||
cls._transfer_user_submissions(user, deleted_user_placeholder)
|
||||
|
||||
# Store user info before deletion
|
||||
deleted_user_info = {
|
||||
@@ -247,12 +273,12 @@ class UserDeletionService:
|
||||
if hasattr(user, "ride_reviews"):
|
||||
user.ride_reviews.all().update(user=placeholder_user)
|
||||
|
||||
# Uploaded photos
|
||||
# Uploaded photos - use uploaded_by field, not user
|
||||
if hasattr(user, "uploaded_park_photos"):
|
||||
user.uploaded_park_photos.all().update(user=placeholder_user)
|
||||
user.uploaded_park_photos.all().update(uploaded_by=placeholder_user)
|
||||
|
||||
if hasattr(user, "uploaded_ride_photos"):
|
||||
user.uploaded_ride_photos.all().update(user=placeholder_user)
|
||||
user.uploaded_ride_photos.all().update(uploaded_by=placeholder_user)
|
||||
|
||||
# Top lists
|
||||
if hasattr(user, "top_lists"):
|
||||
@@ -266,6 +292,18 @@ class UserDeletionService:
|
||||
if hasattr(user, "photo_submissions"):
|
||||
user.photo_submissions.all().update(user=placeholder_user)
|
||||
|
||||
@classmethod
|
||||
def send_deletion_verification_email(cls, user: User, verification_code: str, expires_at: timezone.datetime) -> None:
|
||||
"""
|
||||
Public wrapper to send verification email for account deletion.
|
||||
|
||||
Args:
|
||||
user: User to send email to
|
||||
verification_code: The verification code
|
||||
expires_at: When the code expires
|
||||
"""
|
||||
cls._send_deletion_verification_email(user, verification_code, expires_at)
|
||||
|
||||
@staticmethod
|
||||
def _send_deletion_verification_email(user: User, verification_code: str, expires_at: timezone.datetime) -> None:
|
||||
"""Send verification email for account deletion."""
|
||||
|
||||
@@ -14,7 +14,7 @@ class UserDeletionServiceTest(TestCase):
|
||||
|
||||
def setUp(self):
|
||||
"""Set up test data."""
|
||||
# Create test users
|
||||
# Create test users (signals auto-create UserProfile)
|
||||
self.user = User.objects.create_user(username="testuser", email="test@example.com", password="testpass123")
|
||||
|
||||
self.admin_user = User.objects.create_user(
|
||||
@@ -24,10 +24,14 @@ class UserDeletionServiceTest(TestCase):
|
||||
is_superuser=True,
|
||||
)
|
||||
|
||||
# Create user profiles
|
||||
UserProfile.objects.create(user=self.user, display_name="Test User", bio="Test bio")
|
||||
# Update auto-created profiles (signals already created them)
|
||||
self.user.profile.display_name = "Test User"
|
||||
self.user.profile.bio = "Test bio"
|
||||
self.user.profile.save()
|
||||
|
||||
UserProfile.objects.create(user=self.admin_user, display_name="Admin User", bio="Admin bio")
|
||||
self.admin_user.profile.display_name = "Admin User"
|
||||
self.admin_user.profile.bio = "Admin bio"
|
||||
self.admin_user.profile.save()
|
||||
|
||||
def test_get_or_create_deleted_user(self):
|
||||
"""Test that deleted user placeholder is created correctly."""
|
||||
@@ -37,11 +41,9 @@ class UserDeletionServiceTest(TestCase):
|
||||
self.assertEqual(deleted_user.email, "deleted@thrillwiki.com")
|
||||
self.assertFalse(deleted_user.is_active)
|
||||
self.assertTrue(deleted_user.is_banned)
|
||||
self.assertEqual(deleted_user.role, User.Roles.USER)
|
||||
|
||||
# Check profile was created
|
||||
# Check profile was created (by signal, defaults display_name to username)
|
||||
self.assertTrue(hasattr(deleted_user, "profile"))
|
||||
self.assertEqual(deleted_user.profile.display_name, "Deleted User")
|
||||
|
||||
def test_get_or_create_deleted_user_idempotent(self):
|
||||
"""Test that calling get_or_create_deleted_user multiple times returns same user."""
|
||||
@@ -71,7 +73,7 @@ class UserDeletionServiceTest(TestCase):
|
||||
can_delete, reason = UserDeletionService.can_delete_user(deleted_user)
|
||||
|
||||
self.assertFalse(can_delete)
|
||||
self.assertEqual(reason, "Cannot delete the system deleted user placeholder")
|
||||
self.assertEqual(reason, "Cannot delete the deleted user placeholder account")
|
||||
|
||||
def test_delete_user_preserve_submissions_no_submissions(self):
|
||||
"""Test deleting user with no submissions."""
|
||||
@@ -102,7 +104,7 @@ class UserDeletionServiceTest(TestCase):
|
||||
with self.assertRaises(ValueError) as context:
|
||||
UserDeletionService.delete_user_preserve_submissions(deleted_user)
|
||||
|
||||
self.assertIn("Cannot delete the system deleted user placeholder", str(context.exception))
|
||||
self.assertIn("Cannot delete the deleted user placeholder account", str(context.exception))
|
||||
|
||||
def test_delete_user_with_submissions_transfers_correctly(self):
|
||||
"""Test that user submissions are transferred to deleted user placeholder."""
|
||||
|
||||
@@ -6,6 +6,7 @@ from django.urls import include, path
|
||||
from rest_framework.routers import DefaultRouter
|
||||
|
||||
from . import views, views_credits, views_magic_link
|
||||
from .views import list_profiles
|
||||
|
||||
# Register ViewSets
|
||||
router = DefaultRouter()
|
||||
@@ -119,7 +120,8 @@ urlpatterns = [
|
||||
# Magic Link (Login by Code) endpoints
|
||||
path("magic-link/request/", views_magic_link.request_magic_link, name="request_magic_link"),
|
||||
path("magic-link/verify/", views_magic_link.verify_magic_link, name="verify_magic_link"),
|
||||
# Public Profile
|
||||
# Public Profiles - List and Detail
|
||||
path("profiles/", list_profiles, name="list_profiles"),
|
||||
path("profiles/<str:username>/", views.get_public_user_profile, name="get_public_user_profile"),
|
||||
# Bulk lookup endpoints
|
||||
path("profiles/bulk/", views.bulk_get_profiles, name="bulk_get_profiles"),
|
||||
|
||||
@@ -823,6 +823,125 @@ def check_user_deletion_eligibility(request, user_id):
|
||||
)
|
||||
|
||||
|
||||
# === PUBLIC PROFILE LIST ENDPOINT ===
|
||||
|
||||
|
||||
@extend_schema(
|
||||
operation_id="list_profiles",
|
||||
summary="List user profiles with search and pagination",
|
||||
description=(
|
||||
"Returns a paginated list of public user profiles. "
|
||||
"Supports search by username or display name, and filtering by various criteria. "
|
||||
"This endpoint is used for user discovery, leaderboards, and friend finding."
|
||||
),
|
||||
parameters=[
|
||||
OpenApiParameter(
|
||||
name="search",
|
||||
type=OpenApiTypes.STR,
|
||||
location=OpenApiParameter.QUERY,
|
||||
description="Search term for username or display name",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="ordering",
|
||||
type=OpenApiTypes.STR,
|
||||
location=OpenApiParameter.QUERY,
|
||||
description="Order by field: date_joined, -date_joined, username, -username",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="page",
|
||||
type=OpenApiTypes.INT,
|
||||
location=OpenApiParameter.QUERY,
|
||||
description="Page number for pagination",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="page_size",
|
||||
type=OpenApiTypes.INT,
|
||||
location=OpenApiParameter.QUERY,
|
||||
description="Number of results per page (max 100)",
|
||||
),
|
||||
],
|
||||
responses={
|
||||
200: {
|
||||
"description": "Paginated list of public profiles",
|
||||
"example": {
|
||||
"count": 150,
|
||||
"next": "https://api.thrillwiki.com/api/v1/accounts/profiles/?page=2",
|
||||
"previous": None,
|
||||
"results": [
|
||||
{
|
||||
"user_id": "uuid-1",
|
||||
"username": "thrillseeker",
|
||||
"date_joined": "2024-01-01T00:00:00Z",
|
||||
"role": "USER",
|
||||
"profile": {
|
||||
"profile_id": "uuid-profile",
|
||||
"display_name": "Thrill Seeker",
|
||||
"avatar_url": "https://example.com/avatar.jpg",
|
||||
"bio": "Coaster enthusiast!",
|
||||
"total_credits": 150,
|
||||
},
|
||||
}
|
||||
],
|
||||
},
|
||||
},
|
||||
},
|
||||
tags=["User Profile"],
|
||||
)
|
||||
@api_view(["GET"])
|
||||
@permission_classes([AllowAny])
|
||||
def list_profiles(request):
|
||||
"""
|
||||
List public user profiles with search and pagination.
|
||||
|
||||
This endpoint provides the missing /accounts/profiles/ list endpoint
|
||||
that the frontend expects for user discovery features.
|
||||
"""
|
||||
from django.db.models import Q
|
||||
from rest_framework.pagination import PageNumberPagination
|
||||
|
||||
# Base queryset: only active users with public profiles
|
||||
queryset = User.objects.filter(
|
||||
is_active=True,
|
||||
).select_related("profile").order_by("-date_joined")
|
||||
|
||||
# User ID filter - EXACT match (critical for single user lookups)
|
||||
user_id = request.query_params.get("user_id", "").strip()
|
||||
if user_id:
|
||||
# Use exact match to prevent user_id=4 from matching user_id=4448
|
||||
queryset = queryset.filter(user_id=user_id)
|
||||
|
||||
# Search filter
|
||||
search = request.query_params.get("search", "").strip()
|
||||
if search:
|
||||
queryset = queryset.filter(
|
||||
Q(username__icontains=search) |
|
||||
Q(profile__display_name__icontains=search)
|
||||
)
|
||||
|
||||
# Ordering
|
||||
ordering = request.query_params.get("ordering", "-date_joined")
|
||||
valid_orderings = ["date_joined", "-date_joined", "username", "-username"]
|
||||
if ordering in valid_orderings:
|
||||
queryset = queryset.order_by(ordering)
|
||||
|
||||
# Pagination
|
||||
class ProfilePagination(PageNumberPagination):
|
||||
page_size = 20
|
||||
page_size_query_param = "page_size"
|
||||
max_page_size = 100
|
||||
|
||||
paginator = ProfilePagination()
|
||||
page = paginator.paginate_queryset(queryset, request)
|
||||
|
||||
if page is not None:
|
||||
serializer = PublicUserSerializer(page, many=True)
|
||||
return paginator.get_paginated_response(serializer.data)
|
||||
|
||||
# Fallback if pagination fails
|
||||
serializer = PublicUserSerializer(queryset[:20], many=True)
|
||||
return Response(serializer.data, status=status.HTTP_200_OK)
|
||||
|
||||
|
||||
# === USER PROFILE ENDPOINTS ===
|
||||
|
||||
|
||||
@@ -968,18 +1087,53 @@ def update_user_profile(request):
|
||||
@extend_schema(
|
||||
operation_id="get_user_preferences",
|
||||
summary="Get user preferences",
|
||||
description="Get the authenticated user's preferences and settings.",
|
||||
description="Get or update the authenticated user's preferences and settings.",
|
||||
responses={
|
||||
200: UserPreferencesSerializer,
|
||||
401: {"description": "Authentication required"},
|
||||
},
|
||||
tags=["User Settings"],
|
||||
)
|
||||
@api_view(["GET"])
|
||||
@api_view(["GET", "PATCH"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def get_user_preferences(request):
|
||||
"""Get user preferences."""
|
||||
"""Get or update user preferences."""
|
||||
user = request.user
|
||||
|
||||
if request.method == "PATCH":
|
||||
current_data = {
|
||||
"theme_preference": user.theme_preference,
|
||||
"email_notifications": user.email_notifications,
|
||||
"push_notifications": user.push_notifications,
|
||||
"privacy_level": user.privacy_level,
|
||||
"show_email": user.show_email,
|
||||
"show_real_name": user.show_real_name,
|
||||
"show_statistics": user.show_statistics,
|
||||
"allow_friend_requests": user.allow_friend_requests,
|
||||
"allow_messages": user.allow_messages,
|
||||
}
|
||||
|
||||
# Handle moderation_preferences field (stored as JSON on User model if it exists)
|
||||
if "moderation_preferences" in request.data:
|
||||
try:
|
||||
if hasattr(user, 'moderation_preferences'):
|
||||
user.moderation_preferences = request.data["moderation_preferences"]
|
||||
user.save()
|
||||
# Return success even if field doesn't exist (non-critical preference)
|
||||
return Response({"moderation_preferences": request.data["moderation_preferences"]}, status=status.HTTP_200_OK)
|
||||
except Exception:
|
||||
# Non-critical - just return success
|
||||
return Response({"moderation_preferences": request.data["moderation_preferences"]}, status=status.HTTP_200_OK)
|
||||
|
||||
serializer = UserPreferencesSerializer(data={**current_data, **request.data})
|
||||
if serializer.is_valid():
|
||||
for field, value in serializer.validated_data.items():
|
||||
setattr(user, field, value)
|
||||
user.save()
|
||||
return Response(serializer.data, status=status.HTTP_200_OK)
|
||||
return Response(serializer.errors, status=status.HTTP_400_BAD_REQUEST)
|
||||
|
||||
# GET request
|
||||
data = {
|
||||
"theme_preference": user.theme_preference,
|
||||
"email_notifications": user.email_notifications,
|
||||
|
||||
@@ -370,6 +370,118 @@ def revoke_session(request, session_id):
|
||||
return Response({"detail": "Session revoked"})
|
||||
|
||||
|
||||
# ============== PASSWORD CHANGE ENDPOINT ==============
|
||||
|
||||
# ============== SECURITY LOG ENDPOINT ==============
|
||||
|
||||
@extend_schema(
|
||||
operation_id="get_security_log",
|
||||
summary="Get security activity log",
|
||||
description="Returns paginated list of security events for the current user.",
|
||||
parameters=[
|
||||
{
|
||||
"name": "page",
|
||||
"in": "query",
|
||||
"description": "Page number (1-indexed)",
|
||||
"required": False,
|
||||
"schema": {"type": "integer", "default": 1},
|
||||
},
|
||||
{
|
||||
"name": "page_size",
|
||||
"in": "query",
|
||||
"description": "Number of items per page (max 50)",
|
||||
"required": False,
|
||||
"schema": {"type": "integer", "default": 20},
|
||||
},
|
||||
{
|
||||
"name": "event_type",
|
||||
"in": "query",
|
||||
"description": "Filter by event type",
|
||||
"required": False,
|
||||
"schema": {"type": "string"},
|
||||
},
|
||||
],
|
||||
responses={
|
||||
200: {
|
||||
"description": "Security log entries",
|
||||
"example": {
|
||||
"count": 42,
|
||||
"page": 1,
|
||||
"page_size": 20,
|
||||
"total_pages": 3,
|
||||
"results": [
|
||||
{
|
||||
"id": 1,
|
||||
"event_type": "login_success",
|
||||
"event_type_display": "Login Success",
|
||||
"ip_address": "192.168.1.1",
|
||||
"user_agent": "Mozilla/5.0...",
|
||||
"created_at": "2026-01-06T12:00:00Z",
|
||||
"metadata": {},
|
||||
}
|
||||
],
|
||||
},
|
||||
},
|
||||
},
|
||||
tags=["Account"],
|
||||
)
|
||||
@api_view(["GET"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def get_security_log(request):
|
||||
"""Get security activity log for the current user."""
|
||||
from apps.accounts.models import SecurityLog
|
||||
|
||||
user = request.user
|
||||
|
||||
# Parse pagination params
|
||||
try:
|
||||
page = max(1, int(request.query_params.get("page", 1)))
|
||||
except (ValueError, TypeError):
|
||||
page = 1
|
||||
|
||||
try:
|
||||
page_size = min(50, max(1, int(request.query_params.get("page_size", 20))))
|
||||
except (ValueError, TypeError):
|
||||
page_size = 20
|
||||
|
||||
event_type = request.query_params.get("event_type")
|
||||
|
||||
# Build queryset
|
||||
queryset = SecurityLog.objects.filter(user=user).order_by("-created_at")
|
||||
|
||||
if event_type:
|
||||
queryset = queryset.filter(event_type=event_type)
|
||||
|
||||
# Count total
|
||||
total_count = queryset.count()
|
||||
total_pages = (total_count + page_size - 1) // page_size
|
||||
|
||||
# Fetch page
|
||||
offset = (page - 1) * page_size
|
||||
logs = queryset[offset : offset + page_size]
|
||||
|
||||
# Serialize
|
||||
results = []
|
||||
for log in logs:
|
||||
results.append({
|
||||
"id": log.id,
|
||||
"event_type": log.event_type,
|
||||
"event_type_display": log.get_event_type_display(),
|
||||
"ip_address": log.ip_address,
|
||||
"user_agent": log.user_agent[:200] if log.user_agent else "", # Truncate for safety
|
||||
"created_at": log.created_at.isoformat(),
|
||||
"metadata": log.metadata or {},
|
||||
})
|
||||
|
||||
return Response({
|
||||
"count": total_count,
|
||||
"page": page,
|
||||
"page_size": page_size,
|
||||
"total_pages": total_pages,
|
||||
"results": results,
|
||||
})
|
||||
|
||||
|
||||
# ============== PASSWORD CHANGE ENDPOINT ==============
|
||||
|
||||
@extend_schema(
|
||||
@@ -396,6 +508,12 @@ def revoke_session(request, session_id):
|
||||
@permission_classes([IsAuthenticated])
|
||||
def change_password(request):
|
||||
"""Change user password."""
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
invalidate_user_sessions,
|
||||
)
|
||||
|
||||
user = request.user
|
||||
current_password = request.data.get("current_password", "")
|
||||
new_password = request.data.get("new_password", "")
|
||||
@@ -413,6 +531,24 @@ def change_password(request):
|
||||
)
|
||||
|
||||
user.set_password(new_password)
|
||||
user.last_password_change = timezone.now()
|
||||
user.save()
|
||||
|
||||
return Response({"detail": "Password changed successfully"})
|
||||
# Invalidate all existing sessions/tokens (except current)
|
||||
invalidated_count = invalidate_user_sessions(user, exclude_current=True, request=request)
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"password_changed",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"sessions_invalidated": invalidated_count},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(user, "password_changed", metadata={})
|
||||
|
||||
return Response({
|
||||
"detail": "Password changed successfully",
|
||||
"sessions_invalidated": invalidated_count,
|
||||
})
|
||||
|
||||
96
backend/apps/api/v1/auth/jwt.py
Normal file
96
backend/apps/api/v1/auth/jwt.py
Normal file
@@ -0,0 +1,96 @@
|
||||
"""
|
||||
Custom JWT Token Generation for ThrillWiki
|
||||
|
||||
This module provides custom JWT token generation that includes authentication
|
||||
method claims for enhanced MFA satisfaction logic.
|
||||
|
||||
Claims added:
|
||||
- auth_method: How the user authenticated (password, passkey, totp, google, discord)
|
||||
- mfa_verified: Whether MFA was verified during this login
|
||||
- provider_mfa: Whether the OAuth provider (Discord) has MFA enabled
|
||||
"""
|
||||
|
||||
from typing import Literal, TypedDict
|
||||
|
||||
from rest_framework_simplejwt.tokens import RefreshToken
|
||||
|
||||
# Type definitions for auth methods
|
||||
AuthMethod = Literal["password", "passkey", "totp", "google", "discord"]
|
||||
|
||||
|
||||
class TokenClaims(TypedDict, total=False):
|
||||
"""Type definition for custom JWT claims."""
|
||||
|
||||
auth_method: AuthMethod
|
||||
mfa_verified: bool
|
||||
provider_mfa: bool
|
||||
|
||||
|
||||
def create_tokens_for_user(
|
||||
user,
|
||||
auth_method: AuthMethod = "password",
|
||||
mfa_verified: bool = False,
|
||||
provider_mfa: bool = False,
|
||||
) -> dict[str, str]:
|
||||
"""
|
||||
Generate JWT tokens with custom authentication claims.
|
||||
|
||||
Args:
|
||||
user: The Django user object
|
||||
auth_method: How the user authenticated
|
||||
mfa_verified: True if MFA (TOTP/passkey) was verified at login
|
||||
provider_mfa: True if OAuth provider (Discord) has MFA enabled
|
||||
|
||||
Returns:
|
||||
Dictionary with 'access' and 'refresh' token strings
|
||||
"""
|
||||
refresh = RefreshToken.for_user(user)
|
||||
|
||||
# Add custom claims to both refresh and access tokens
|
||||
refresh["auth_method"] = auth_method
|
||||
refresh["mfa_verified"] = mfa_verified
|
||||
refresh["provider_mfa"] = provider_mfa
|
||||
|
||||
access = refresh.access_token
|
||||
|
||||
return {
|
||||
"access": str(access),
|
||||
"refresh": str(refresh),
|
||||
}
|
||||
|
||||
|
||||
def get_auth_method_for_provider(provider: str) -> AuthMethod:
|
||||
"""
|
||||
Map OAuth provider name to AuthMethod type.
|
||||
|
||||
Args:
|
||||
provider: The provider name (e.g., 'google', 'discord')
|
||||
|
||||
Returns:
|
||||
The corresponding AuthMethod
|
||||
"""
|
||||
provider_map: dict[str, AuthMethod] = {
|
||||
"google": "google",
|
||||
"discord": "discord",
|
||||
}
|
||||
return provider_map.get(provider, "password")
|
||||
|
||||
|
||||
def get_provider_mfa_status(provider: str, extra_data: dict) -> bool:
|
||||
"""
|
||||
Extract MFA status from OAuth provider extra_data.
|
||||
|
||||
Only Discord exposes mfa_enabled. Google does not share this info.
|
||||
|
||||
Args:
|
||||
provider: The OAuth provider name
|
||||
extra_data: The extra_data dict from SocialAccount
|
||||
|
||||
Returns:
|
||||
True if provider has MFA enabled, False otherwise
|
||||
"""
|
||||
if provider == "discord":
|
||||
return extra_data.get("mfa_enabled", False)
|
||||
|
||||
# Google and other providers don't expose MFA status
|
||||
return False
|
||||
@@ -64,6 +64,23 @@ def get_mfa_status(request):
|
||||
except Authenticator.DoesNotExist:
|
||||
pass
|
||||
|
||||
# Check for Discord social account with MFA enabled
|
||||
discord_mfa_enabled = False
|
||||
connected_provider = None
|
||||
|
||||
try:
|
||||
social_accounts = user.socialaccount_set.all()
|
||||
for social_account in social_accounts:
|
||||
if social_account.provider == "discord":
|
||||
connected_provider = "discord"
|
||||
discord_mfa_enabled = social_account.extra_data.get("mfa_enabled", False)
|
||||
break
|
||||
elif social_account.provider == "google":
|
||||
connected_provider = "google"
|
||||
# Google doesn't expose MFA status
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
# has_second_factor is True if user has either TOTP or Passkey configured
|
||||
has_second_factor = totp_enabled or passkey_enabled
|
||||
|
||||
@@ -76,6 +93,9 @@ def get_mfa_status(request):
|
||||
"recovery_codes_enabled": recovery_enabled,
|
||||
"recovery_codes_count": recovery_count,
|
||||
"has_second_factor": has_second_factor,
|
||||
# New fields for enhanced MFA satisfaction
|
||||
"discord_mfa_enabled": discord_mfa_enabled,
|
||||
"connected_provider": connected_provider,
|
||||
}
|
||||
)
|
||||
|
||||
@@ -100,6 +120,8 @@ def get_mfa_status(request):
|
||||
@permission_classes([IsAuthenticated])
|
||||
def setup_totp(request):
|
||||
"""Generate TOTP secret and QR code for setup."""
|
||||
from django.utils import timezone
|
||||
|
||||
from allauth.mfa.totp.internal import auth as totp_auth
|
||||
|
||||
user = request.user
|
||||
@@ -120,14 +142,16 @@ def setup_totp(request):
|
||||
qr.save(buffer, format="PNG")
|
||||
qr_code_base64 = f"data:image/png;base64,{base64.b64encode(buffer.getvalue()).decode()}"
|
||||
|
||||
# Store secret in session for later verification
|
||||
# Store secret in session for later verification with 15-minute expiry
|
||||
request.session["pending_totp_secret"] = secret
|
||||
request.session["pending_totp_expires"] = (timezone.now().timestamp() + 900) # 15 minutes
|
||||
|
||||
return Response(
|
||||
{
|
||||
"secret": secret,
|
||||
"provisioning_uri": uri,
|
||||
"qr_code_base64": qr_code_base64,
|
||||
"expires_in_seconds": 900,
|
||||
}
|
||||
)
|
||||
|
||||
@@ -165,10 +189,17 @@ def setup_totp(request):
|
||||
@permission_classes([IsAuthenticated])
|
||||
def activate_totp(request):
|
||||
"""Verify TOTP code and activate MFA."""
|
||||
from django.utils import timezone
|
||||
|
||||
from allauth.mfa.models import Authenticator
|
||||
from allauth.mfa.recovery_codes.internal.auth import RecoveryCodes
|
||||
from allauth.mfa.totp.internal import auth as totp_auth
|
||||
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
user = request.user
|
||||
code = request.data.get("code", "").strip()
|
||||
|
||||
@@ -187,6 +218,19 @@ def activate_totp(request):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Check if setup has expired (15 minute timeout)
|
||||
expires_at = request.session.get("pending_totp_expires")
|
||||
if expires_at and timezone.now().timestamp() > expires_at:
|
||||
# Clear expired session data
|
||||
if "pending_totp_secret" in request.session:
|
||||
del request.session["pending_totp_secret"]
|
||||
if "pending_totp_expires" in request.session:
|
||||
del request.session["pending_totp_expires"]
|
||||
return Response(
|
||||
{"detail": "TOTP setup session expired. Please start setup again."},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Verify the code
|
||||
if not totp_auth.validate_totp_code(secret, code):
|
||||
return Response(
|
||||
@@ -215,11 +259,25 @@ def activate_totp(request):
|
||||
# Clear session (only if it exists - won't exist with JWT auth + secret from body)
|
||||
if "pending_totp_secret" in request.session:
|
||||
del request.session["pending_totp_secret"]
|
||||
if "pending_totp_expires" in request.session:
|
||||
del request.session["pending_totp_expires"]
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"mfa_enrolled",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"method": "totp"},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(user, "mfa_enrolled", {"method": "TOTP Authenticator"})
|
||||
|
||||
return Response(
|
||||
{
|
||||
"detail": "Two-factor authentication enabled",
|
||||
"recovery_codes": codes,
|
||||
"recovery_codes_count": len(codes),
|
||||
}
|
||||
)
|
||||
|
||||
@@ -255,13 +313,59 @@ def deactivate_totp(request):
|
||||
"""Disable TOTP authentication."""
|
||||
from allauth.mfa.models import Authenticator
|
||||
|
||||
from apps.accounts.services.security_service import (
|
||||
check_auth_method_availability,
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
user = request.user
|
||||
password = request.data.get("password", "")
|
||||
recovery_code = request.data.get("recovery_code", "")
|
||||
|
||||
# Verify password
|
||||
if not user.check_password(password):
|
||||
# Check if user has other auth methods before we allow disabling MFA
|
||||
auth_methods = check_auth_method_availability(user)
|
||||
|
||||
# If TOTP is their only way in alongside passkeys, we need to ensure they have
|
||||
# at least password or social login to fall back on
|
||||
if not auth_methods["has_password"] and not auth_methods["has_social"] and not auth_methods["has_passkey"]:
|
||||
return Response(
|
||||
{"detail": "Invalid password"},
|
||||
{"detail": "Cannot disable MFA: you must have at least one authentication method. Please set a password or connect a social account first."},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Verify password OR recovery code
|
||||
verified = False
|
||||
verification_method = None
|
||||
|
||||
if password and user.check_password(password):
|
||||
verified = True
|
||||
verification_method = "password"
|
||||
elif recovery_code:
|
||||
# Try to verify with recovery code
|
||||
try:
|
||||
recovery_auth = Authenticator.objects.get(
|
||||
user=user, type=Authenticator.Type.RECOVERY_CODES
|
||||
)
|
||||
unused_codes = recovery_auth.data.get("codes", [])
|
||||
if recovery_code.upper().replace("-", "").replace(" ", "") in [
|
||||
c.upper().replace("-", "").replace(" ", "") for c in unused_codes
|
||||
]:
|
||||
verified = True
|
||||
verification_method = "recovery_code"
|
||||
# Remove the used code
|
||||
unused_codes = [
|
||||
c for c in unused_codes
|
||||
if c.upper().replace("-", "").replace(" ", "") != recovery_code.upper().replace("-", "").replace(" ", "")
|
||||
]
|
||||
recovery_auth.data["codes"] = unused_codes
|
||||
recovery_auth.save()
|
||||
except Authenticator.DoesNotExist:
|
||||
pass
|
||||
|
||||
if not verified:
|
||||
return Response(
|
||||
{"detail": "Invalid password or recovery code"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@@ -276,6 +380,17 @@ def deactivate_totp(request):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"mfa_disabled",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"method": "totp", "verified_via": verification_method},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(user, "mfa_disabled", {"method": "TOTP Authenticator"})
|
||||
|
||||
return Response(
|
||||
{
|
||||
"detail": "Two-factor authentication disabled",
|
||||
@@ -361,6 +476,11 @@ def regenerate_recovery_codes(request):
|
||||
from allauth.mfa.models import Authenticator
|
||||
from allauth.mfa.recovery_codes.internal.auth import RecoveryCodes
|
||||
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
user = request.user
|
||||
password = request.data.get("password", "")
|
||||
|
||||
@@ -371,8 +491,11 @@ def regenerate_recovery_codes(request):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Check if TOTP is enabled
|
||||
if not Authenticator.objects.filter(user=user, type=Authenticator.Type.TOTP).exists():
|
||||
# Check if MFA is enabled (TOTP or Passkey)
|
||||
has_totp = Authenticator.objects.filter(user=user, type=Authenticator.Type.TOTP).exists()
|
||||
has_passkey = Authenticator.objects.filter(user=user, type=Authenticator.Type.WEBAUTHN).exists()
|
||||
|
||||
if not has_totp and not has_passkey:
|
||||
return Response(
|
||||
{"detail": "Two-factor authentication is not enabled"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
@@ -387,9 +510,21 @@ def regenerate_recovery_codes(request):
|
||||
recovery_instance = RecoveryCodes.activate(user)
|
||||
codes = recovery_instance.get_unused_codes()
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"recovery_codes_regenerated",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"codes_generated": len(codes)},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(user, "recovery_codes_regenerated", {"codes_generated": len(codes)})
|
||||
|
||||
return Response(
|
||||
{
|
||||
"success": True,
|
||||
"recovery_codes": codes,
|
||||
"recovery_codes_count": len(codes),
|
||||
}
|
||||
)
|
||||
|
||||
@@ -10,7 +10,7 @@ import logging
|
||||
from drf_spectacular.utils import extend_schema
|
||||
from rest_framework import status
|
||||
from rest_framework.decorators import api_view, permission_classes
|
||||
from rest_framework.permissions import IsAuthenticated
|
||||
from rest_framework.permissions import AllowAny, IsAuthenticated
|
||||
from rest_framework.response import Response
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
@@ -93,16 +93,26 @@ def get_passkey_status(request):
|
||||
def get_registration_options(request):
|
||||
"""Get WebAuthn registration options for passkey setup."""
|
||||
try:
|
||||
from django.utils import timezone
|
||||
from allauth.mfa.webauthn.internal import auth as webauthn_auth
|
||||
|
||||
# Use the correct allauth API: begin_registration
|
||||
creation_options, state = webauthn_auth.begin_registration(request)
|
||||
# The function takes (user, passwordless) - passwordless=False for standard passkeys
|
||||
creation_options = webauthn_auth.begin_registration(request.user, passwordless=False)
|
||||
|
||||
# Store state in session for verification
|
||||
webauthn_auth.set_state(request, state)
|
||||
# State is stored internally by begin_registration via set_state()
|
||||
|
||||
# Store registration timeout in session (5 minutes)
|
||||
request.session["pending_passkey_expires"] = timezone.now().timestamp() + 300 # 5 minutes
|
||||
|
||||
# Debug log the structure
|
||||
logger.debug(f"WebAuthn registration options type: {type(creation_options)}")
|
||||
logger.debug(f"WebAuthn registration options keys: {creation_options.keys() if isinstance(creation_options, dict) else 'not a dict'}")
|
||||
logger.info(f"WebAuthn registration options: {creation_options}")
|
||||
|
||||
return Response({
|
||||
"options": creation_options,
|
||||
"expires_in_seconds": 300,
|
||||
})
|
||||
except ImportError as e:
|
||||
logger.error(f"WebAuthn module import error: {e}")
|
||||
@@ -143,8 +153,14 @@ def get_registration_options(request):
|
||||
def register_passkey(request):
|
||||
"""Complete passkey registration with WebAuthn response."""
|
||||
try:
|
||||
from django.utils import timezone
|
||||
from allauth.mfa.webauthn.internal import auth as webauthn_auth
|
||||
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
credential = request.data.get("credential")
|
||||
name = request.data.get("name", "Passkey")
|
||||
|
||||
@@ -154,8 +170,19 @@ def register_passkey(request):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Get stored state from session
|
||||
state = webauthn_auth.get_state(request)
|
||||
# Check if registration has expired (5 minute timeout)
|
||||
expires_at = request.session.get("pending_passkey_expires")
|
||||
if expires_at and timezone.now().timestamp() > expires_at:
|
||||
# Clear expired session data
|
||||
if "pending_passkey_expires" in request.session:
|
||||
del request.session["pending_passkey_expires"]
|
||||
return Response(
|
||||
{"detail": "Passkey registration session expired. Please start registration again."},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Get stored state from session (no request needed, uses context)
|
||||
state = webauthn_auth.get_state()
|
||||
if not state:
|
||||
return Response(
|
||||
{"detail": "No pending registration. Please start registration again."},
|
||||
@@ -164,19 +191,33 @@ def register_passkey(request):
|
||||
|
||||
# Use the correct allauth API: complete_registration
|
||||
try:
|
||||
# Parse the credential response
|
||||
from allauth.mfa.webauthn.internal.auth import WebAuthn
|
||||
|
||||
# Parse the credential response to validate it
|
||||
credential_data = webauthn_auth.parse_registration_response(credential)
|
||||
|
||||
# Complete registration - this creates the Authenticator
|
||||
authenticator = webauthn_auth.complete_registration(
|
||||
request,
|
||||
credential_data,
|
||||
state,
|
||||
name=name,
|
||||
)
|
||||
# Complete registration to validate and clear state
|
||||
webauthn_auth.complete_registration(credential_data)
|
||||
|
||||
# Clear session state
|
||||
webauthn_auth.clear_state(request)
|
||||
# Use allauth's WebAuthn.add() to create the Authenticator properly
|
||||
# It stores the raw credential dict and name in the data field
|
||||
webauthn_wrapper = WebAuthn.add(
|
||||
request.user,
|
||||
name,
|
||||
credential, # Pass raw credential dict, not parsed data
|
||||
)
|
||||
authenticator = webauthn_wrapper.instance
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"passkey_registered",
|
||||
request,
|
||||
user=request.user,
|
||||
metadata={"passkey_name": name, "passkey_id": str(authenticator.id) if authenticator else None},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(request.user, "passkey_registered", {"passkey_name": name})
|
||||
|
||||
return Response({
|
||||
"detail": "Passkey registered successfully",
|
||||
@@ -225,10 +266,8 @@ def get_authentication_options(request):
|
||||
from allauth.mfa.webauthn.internal import auth as webauthn_auth
|
||||
|
||||
# Use the correct allauth API: begin_authentication
|
||||
request_options, state = webauthn_auth.begin_authentication(request)
|
||||
|
||||
# Store state in session for verification
|
||||
webauthn_auth.set_state(request, state)
|
||||
# Takes optional user, returns just options (state is stored internally)
|
||||
request_options = webauthn_auth.begin_authentication(request.user)
|
||||
|
||||
return Response({
|
||||
"options": request_options,
|
||||
@@ -281,8 +320,8 @@ def authenticate_passkey(request):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Get stored state from session
|
||||
state = webauthn_auth.get_state(request)
|
||||
# Get stored state from session (no request needed, uses context)
|
||||
state = webauthn_auth.get_state()
|
||||
if not state:
|
||||
return Response(
|
||||
{"detail": "No pending authentication. Please start authentication again."},
|
||||
@@ -291,14 +330,9 @@ def authenticate_passkey(request):
|
||||
|
||||
# Use the correct allauth API: complete_authentication
|
||||
try:
|
||||
# Parse the credential response
|
||||
credential_data = webauthn_auth.parse_authentication_response(credential)
|
||||
|
||||
# Complete authentication
|
||||
webauthn_auth.complete_authentication(request, credential_data, state)
|
||||
|
||||
# Clear session state
|
||||
webauthn_auth.clear_state(request)
|
||||
# Complete authentication - takes user and credential response
|
||||
# State is handled internally
|
||||
webauthn_auth.complete_authentication(request.user, credential)
|
||||
|
||||
return Response({"success": True})
|
||||
except Exception as e:
|
||||
@@ -347,6 +381,12 @@ def delete_passkey(request, passkey_id):
|
||||
try:
|
||||
from allauth.mfa.models import Authenticator
|
||||
|
||||
from apps.accounts.services.security_service import (
|
||||
check_auth_method_availability,
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
user = request.user
|
||||
password = request.data.get("password", "")
|
||||
|
||||
@@ -357,6 +397,17 @@ def delete_passkey(request, passkey_id):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Check if user has other auth methods before removing passkey
|
||||
auth_methods = check_auth_method_availability(user)
|
||||
|
||||
# If this is the last passkey and user has no other auth method, block removal
|
||||
if auth_methods["passkey_count"] == 1:
|
||||
if not auth_methods["has_password"] and not auth_methods["has_social"] and not auth_methods["has_totp"]:
|
||||
return Response(
|
||||
{"detail": "Cannot remove last passkey: you must have at least one authentication method. Please set a password or connect a social account first."},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Find and delete the passkey
|
||||
try:
|
||||
authenticator = Authenticator.objects.get(
|
||||
@@ -364,7 +415,20 @@ def delete_passkey(request, passkey_id):
|
||||
user=user,
|
||||
type=Authenticator.Type.WEBAUTHN,
|
||||
)
|
||||
passkey_name = authenticator.data.get("name", "Passkey") if authenticator.data else "Passkey"
|
||||
authenticator.delete()
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"passkey_removed",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"passkey_name": passkey_name, "passkey_id": str(passkey_id)},
|
||||
)
|
||||
|
||||
# Send security notification email
|
||||
send_security_notification(user, "passkey_removed", {"passkey_name": passkey_name})
|
||||
|
||||
except Authenticator.DoesNotExist:
|
||||
return Response(
|
||||
{"detail": "Passkey not found"},
|
||||
@@ -468,6 +532,7 @@ def rename_passkey(request, passkey_id):
|
||||
tags=["Passkey"],
|
||||
)
|
||||
@api_view(["POST"])
|
||||
@permission_classes([AllowAny])
|
||||
def get_login_passkey_options(request):
|
||||
"""Get WebAuthn authentication options for MFA login flow (unauthenticated)."""
|
||||
from django.core.cache import cache
|
||||
@@ -514,9 +579,13 @@ def get_login_passkey_options(request):
|
||||
request.user = user
|
||||
|
||||
try:
|
||||
request_options, state = webauthn_auth.begin_authentication(request)
|
||||
# begin_authentication takes just user, returns options (state stored internally)
|
||||
request_options = webauthn_auth.begin_authentication(user)
|
||||
# Note: State is managed by allauth's session context, but for MFA login flow
|
||||
# we need to track user separately since they're not authenticated yet
|
||||
passkey_state_key = f"mfa_passkey_state:{mfa_token}"
|
||||
cache.set(passkey_state_key, state, timeout=300)
|
||||
# Store a reference that this user has a pending passkey auth
|
||||
cache.set(passkey_state_key, {"user_id": user_id}, timeout=300)
|
||||
return Response({"options": request_options})
|
||||
finally:
|
||||
if original_user is not None:
|
||||
|
||||
@@ -29,6 +29,7 @@ from .views import (
|
||||
PasswordResetAPIView,
|
||||
ProcessOAuthProfileAPIView,
|
||||
ResendVerificationAPIView,
|
||||
SessionToTokenAPIView, # For passkey login token exchange
|
||||
SignupAPIView,
|
||||
SocialAuthStatusAPIView,
|
||||
SocialProvidersAPIView,
|
||||
@@ -43,6 +44,7 @@ urlpatterns = [
|
||||
path("user/", CurrentUserAPIView.as_view(), name="auth-current-user"),
|
||||
# JWT token management
|
||||
path("token/refresh/", TokenRefreshView.as_view(), name="auth-token-refresh"),
|
||||
path("token/session/", SessionToTokenAPIView.as_view(), name="auth-token-session"), # Exchange session for JWT
|
||||
# Note: dj_rest_auth removed - using custom social auth views below
|
||||
path(
|
||||
"password/reset/",
|
||||
@@ -128,6 +130,7 @@ urlpatterns = [
|
||||
path("sessions/", account_views.list_sessions, name="auth-sessions-list"),
|
||||
path("sessions/<str:session_id>/", account_views.revoke_session, name="auth-session-revoke"),
|
||||
path("password/change/", account_views.change_password, name="auth-password-change-v2"),
|
||||
path("security-log/", account_views.get_security_log, name="auth-security-log"),
|
||||
]
|
||||
|
||||
# Note: User profiles and top lists functionality is now handled by the accounts app
|
||||
|
||||
@@ -212,16 +212,29 @@ class LoginAPIView(APIView):
|
||||
# pass a real HttpRequest to Django login with backend specified
|
||||
login(_get_underlying_request(request), user, backend="django.contrib.auth.backends.ModelBackend")
|
||||
|
||||
# Generate JWT tokens
|
||||
from rest_framework_simplejwt.tokens import RefreshToken
|
||||
# Generate JWT tokens with auth method claims
|
||||
from .jwt import create_tokens_for_user
|
||||
|
||||
refresh = RefreshToken.for_user(user)
|
||||
access_token = refresh.access_token
|
||||
tokens = create_tokens_for_user(
|
||||
user,
|
||||
auth_method="password",
|
||||
mfa_verified=False,
|
||||
provider_mfa=False,
|
||||
)
|
||||
|
||||
# Log successful login
|
||||
from apps.accounts.services.security_service import log_security_event
|
||||
log_security_event(
|
||||
"login_success",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"auth_method": "password", "mfa_required": False},
|
||||
)
|
||||
|
||||
response_serializer = LoginOutputSerializer(
|
||||
{
|
||||
"access": str(access_token),
|
||||
"refresh": str(refresh),
|
||||
"access": tokens["access"],
|
||||
"refresh": tokens["refresh"],
|
||||
"user": user,
|
||||
"message": "Login successful",
|
||||
}
|
||||
@@ -237,6 +250,14 @@ class LoginAPIView(APIView):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
else:
|
||||
# Log failed login attempt
|
||||
from apps.accounts.services.security_service import log_security_event
|
||||
log_security_event(
|
||||
"login_failed",
|
||||
request,
|
||||
user=None,
|
||||
metadata={"username_attempted": email_or_username},
|
||||
)
|
||||
return Response(
|
||||
{"detail": "Invalid credentials"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
@@ -331,8 +352,17 @@ class MFALoginVerifyAPIView(APIView):
|
||||
)
|
||||
|
||||
# Verify MFA - either TOTP or Passkey
|
||||
from apps.accounts.services.security_service import log_security_event
|
||||
|
||||
if totp_code:
|
||||
if not self._verify_totp(user, totp_code):
|
||||
# Log failed MFA attempt
|
||||
log_security_event(
|
||||
"mfa_challenge_failed",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"method": "totp"},
|
||||
)
|
||||
return Response(
|
||||
{"detail": "Invalid verification code"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
@@ -341,6 +371,13 @@ class MFALoginVerifyAPIView(APIView):
|
||||
# Verify passkey/WebAuthn credential
|
||||
passkey_result = self._verify_passkey(request, user, credential)
|
||||
if not passkey_result["success"]:
|
||||
# Log failed MFA attempt
|
||||
log_security_event(
|
||||
"mfa_challenge_failed",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"method": "passkey", "error": passkey_result.get("error")},
|
||||
)
|
||||
return Response(
|
||||
{"detail": passkey_result.get("error", "Passkey verification failed")},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
@@ -357,16 +394,41 @@ class MFALoginVerifyAPIView(APIView):
|
||||
# Complete login
|
||||
login(_get_underlying_request(request), user, backend="django.contrib.auth.backends.ModelBackend")
|
||||
|
||||
# Generate JWT tokens
|
||||
from rest_framework_simplejwt.tokens import RefreshToken
|
||||
# Determine auth method based on what was verified
|
||||
from .jwt import create_tokens_for_user
|
||||
|
||||
refresh = RefreshToken.for_user(user)
|
||||
access_token = refresh.access_token
|
||||
if credential:
|
||||
# Passkey verification - inherently MFA
|
||||
auth_method = "passkey"
|
||||
else:
|
||||
# TOTP verification
|
||||
auth_method = "totp"
|
||||
|
||||
# Log successful MFA challenge and login
|
||||
log_security_event(
|
||||
"mfa_challenge_success",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"method": auth_method},
|
||||
)
|
||||
log_security_event(
|
||||
"login_success",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"auth_method": auth_method, "mfa_verified": True},
|
||||
)
|
||||
|
||||
tokens = create_tokens_for_user(
|
||||
user,
|
||||
auth_method=auth_method,
|
||||
mfa_verified=True,
|
||||
provider_mfa=False,
|
||||
)
|
||||
|
||||
response_serializer = LoginOutputSerializer(
|
||||
{
|
||||
"access": str(access_token),
|
||||
"refresh": str(refresh),
|
||||
"access": tokens["access"],
|
||||
"refresh": tokens["refresh"],
|
||||
"user": user,
|
||||
"message": "Login successful",
|
||||
}
|
||||
@@ -417,23 +479,23 @@ class MFALoginVerifyAPIView(APIView):
|
||||
return {"success": False, "error": "No passkey registered for this user"}
|
||||
|
||||
try:
|
||||
# Parse the authentication response
|
||||
credential_data = webauthn_auth.parse_authentication_response(credential)
|
||||
|
||||
# Get or create authentication state
|
||||
# For login flow, we need to set up the state first
|
||||
state = webauthn_auth.get_state(request)
|
||||
# For MFA login flow, we need to set up state first if not present
|
||||
# Note: allauth's begin_authentication stores state internally
|
||||
state = webauthn_auth.get_state()
|
||||
|
||||
if not state:
|
||||
# If no state, generate one for this user
|
||||
_, state = webauthn_auth.begin_authentication(request)
|
||||
webauthn_auth.set_state(request, state)
|
||||
# Need to temporarily set request.user for allauth context
|
||||
original_user = getattr(request, "user", None)
|
||||
request.user = user
|
||||
try:
|
||||
webauthn_auth.begin_authentication(user)
|
||||
finally:
|
||||
if original_user is not None:
|
||||
request.user = original_user
|
||||
|
||||
# Complete authentication
|
||||
webauthn_auth.complete_authentication(request, credential_data, state)
|
||||
|
||||
# Clear the state
|
||||
webauthn_auth.clear_state(request)
|
||||
# Complete authentication - takes user and credential dict
|
||||
# State is managed internally by allauth
|
||||
webauthn_auth.complete_authentication(user, credential)
|
||||
|
||||
return {"success": True}
|
||||
|
||||
@@ -449,6 +511,99 @@ class MFALoginVerifyAPIView(APIView):
|
||||
return {"success": False, "error": "Passkey verification failed"}
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
post=extend_schema(
|
||||
summary="Exchange session for JWT tokens",
|
||||
description="Exchange allauth session_token (from passkey login) for JWT tokens.",
|
||||
responses={
|
||||
200: LoginOutputSerializer,
|
||||
401: "Not authenticated",
|
||||
},
|
||||
tags=["Authentication"],
|
||||
),
|
||||
)
|
||||
class SessionToTokenAPIView(APIView):
|
||||
"""
|
||||
API endpoint to exchange allauth session_token for JWT tokens.
|
||||
|
||||
Used after allauth headless passkey login to get JWT tokens for the frontend.
|
||||
The allauth passkey login returns a session_token, and this endpoint
|
||||
validates it and exchanges it for JWT tokens.
|
||||
"""
|
||||
|
||||
# Allow unauthenticated - we validate the allauth session_token ourselves
|
||||
permission_classes = [AllowAny]
|
||||
authentication_classes = []
|
||||
|
||||
def post(self, request: Request) -> Response:
|
||||
# Get the allauth session_token from header or body
|
||||
session_token = request.headers.get('X-Session-Token') or request.data.get('session_token')
|
||||
|
||||
if not session_token:
|
||||
return Response(
|
||||
{"detail": "Session token required. Provide X-Session-Token header or session_token in body."},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Validate the session_token with allauth's session store
|
||||
try:
|
||||
from allauth.headless.tokens.strategies.sessions import SessionTokenStrategy
|
||||
|
||||
strategy = SessionTokenStrategy()
|
||||
session_data = strategy.lookup_session(session_token)
|
||||
|
||||
if not session_data:
|
||||
return Response(
|
||||
{"detail": "Invalid or expired session token."},
|
||||
status=status.HTTP_401_UNAUTHORIZED,
|
||||
)
|
||||
|
||||
# Get user from the session
|
||||
user_id = session_data.get('_auth_user_id')
|
||||
if not user_id:
|
||||
return Response(
|
||||
{"detail": "No user found in session."},
|
||||
status=status.HTTP_401_UNAUTHORIZED,
|
||||
)
|
||||
|
||||
user = UserModel.objects.get(pk=user_id)
|
||||
|
||||
except (ImportError, Exception) as e:
|
||||
logger.error(f"Failed to validate allauth session token: {e}")
|
||||
return Response(
|
||||
{"detail": "Failed to validate session token."},
|
||||
status=status.HTTP_500_INTERNAL_SERVER_ERROR,
|
||||
)
|
||||
|
||||
# Generate JWT tokens with passkey auth method
|
||||
from .jwt import create_tokens_for_user
|
||||
|
||||
tokens = create_tokens_for_user(
|
||||
user,
|
||||
auth_method="passkey",
|
||||
mfa_verified=True, # Passkey is considered MFA
|
||||
provider_mfa=False,
|
||||
)
|
||||
|
||||
# Log successful session-to-token exchange
|
||||
from apps.accounts.services.security_service import log_security_event
|
||||
log_security_event(
|
||||
"session_to_token",
|
||||
request,
|
||||
user=user,
|
||||
metadata={"auth_method": "passkey"},
|
||||
)
|
||||
|
||||
response_serializer = LoginOutputSerializer(
|
||||
{
|
||||
"access": tokens["access"],
|
||||
"refresh": tokens["refresh"],
|
||||
"user": user,
|
||||
"message": "Token exchange successful",
|
||||
}
|
||||
)
|
||||
return Response(response_serializer.data)
|
||||
|
||||
@extend_schema_view(
|
||||
post=extend_schema(
|
||||
summary="User registration",
|
||||
@@ -516,6 +671,8 @@ class LogoutAPIView(APIView):
|
||||
|
||||
def post(self, request: Request) -> Response:
|
||||
try:
|
||||
user = request.user
|
||||
|
||||
# Get refresh token from request data with proper type handling
|
||||
refresh_token = None
|
||||
if hasattr(request, "data") and request.data is not None:
|
||||
@@ -539,6 +696,15 @@ class LogoutAPIView(APIView):
|
||||
if hasattr(request.user, "auth_token"):
|
||||
request.user.auth_token.delete()
|
||||
|
||||
# Log security event
|
||||
from apps.accounts.services.security_service import log_security_event
|
||||
log_security_event(
|
||||
"logout",
|
||||
request,
|
||||
user=user,
|
||||
metadata={},
|
||||
)
|
||||
|
||||
# Logout from session using the underlying HttpRequest
|
||||
logout(_get_underlying_request(request))
|
||||
|
||||
@@ -804,6 +970,11 @@ class ConnectProviderAPIView(APIView):
|
||||
serializer_class = ConnectProviderInputSerializer
|
||||
|
||||
def post(self, request: Request, provider: str) -> Response:
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
# Validate provider
|
||||
if provider not in ["google", "discord"]:
|
||||
return Response(
|
||||
@@ -815,6 +986,30 @@ class ConnectProviderAPIView(APIView):
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Check if user's email is verified before allowing social account linking
|
||||
# This prevents attackers from linking a social account to an unverified email
|
||||
user = request.user
|
||||
|
||||
# Check allauth email verification status
|
||||
try:
|
||||
from allauth.account.models import EmailAddress
|
||||
primary_email = EmailAddress.objects.filter(user=user, primary=True).first()
|
||||
if primary_email and not primary_email.verified:
|
||||
return Response(
|
||||
{
|
||||
"detail": "Please verify your email address before connecting social accounts",
|
||||
"code": "EMAIL_NOT_VERIFIED",
|
||||
"suggestions": [
|
||||
"Check your email for a verification link",
|
||||
"Request a new verification email from your account settings",
|
||||
],
|
||||
},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
except ImportError:
|
||||
# If allauth.account is not available, skip check
|
||||
pass
|
||||
|
||||
serializer = ConnectProviderInputSerializer(data=request.data)
|
||||
if not serializer.is_valid():
|
||||
return Response(
|
||||
@@ -833,6 +1028,17 @@ class ConnectProviderAPIView(APIView):
|
||||
service = SocialProviderService()
|
||||
result = service.connect_provider(request.user, provider, access_token)
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"social_linked",
|
||||
request,
|
||||
user=request.user,
|
||||
metadata={"provider": provider},
|
||||
)
|
||||
|
||||
# Send security notification
|
||||
send_security_notification(request.user, "social_linked", {"provider": provider.title()})
|
||||
|
||||
response_serializer = ConnectProviderOutputSerializer(result)
|
||||
return Response(response_serializer.data)
|
||||
|
||||
@@ -882,6 +1088,11 @@ class DisconnectProviderAPIView(APIView):
|
||||
)
|
||||
|
||||
try:
|
||||
from apps.accounts.services.security_service import (
|
||||
log_security_event,
|
||||
send_security_notification,
|
||||
)
|
||||
|
||||
service = SocialProviderService()
|
||||
|
||||
# Check if disconnection is safe
|
||||
@@ -903,6 +1114,17 @@ class DisconnectProviderAPIView(APIView):
|
||||
# Perform disconnection
|
||||
result = service.disconnect_provider(request.user, provider)
|
||||
|
||||
# Log security event
|
||||
log_security_event(
|
||||
"social_unlinked",
|
||||
request,
|
||||
user=request.user,
|
||||
metadata={"provider": provider},
|
||||
)
|
||||
|
||||
# Send security notification
|
||||
send_security_notification(request.user, "social_unlinked", {"provider": provider.title()})
|
||||
|
||||
response_serializer = DisconnectProviderOutputSerializer(result)
|
||||
return Response(response_serializer.data)
|
||||
|
||||
|
||||
@@ -15,6 +15,12 @@ router.register(r"milestones", MilestoneViewSet, basename="milestone")
|
||||
|
||||
# Entity search endpoints - migrated from apps.core.urls
|
||||
urlpatterns = [
|
||||
# View counts endpoint for tracking page views
|
||||
path(
|
||||
"views/",
|
||||
views.ViewCountView.as_view(),
|
||||
name="view_counts",
|
||||
),
|
||||
path(
|
||||
"entities/search/",
|
||||
views.EntityFuzzySearchView.as_view(),
|
||||
|
||||
@@ -27,6 +27,106 @@ import logging
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ViewCountView(APIView):
|
||||
"""
|
||||
Track and retrieve view counts for entities.
|
||||
|
||||
This endpoint provides the /core/views/ functionality expected by
|
||||
the frontend for tracking page views on parks, rides, and companies.
|
||||
"""
|
||||
|
||||
permission_classes = [AllowAny]
|
||||
|
||||
@extend_schema(
|
||||
tags=["Core"],
|
||||
summary="Get view counts for entities",
|
||||
description="Retrieve view counts for specified entities",
|
||||
)
|
||||
def get(self, request):
|
||||
"""Get view counts for entities by type and ID."""
|
||||
entity_type = request.query_params.get("entity_type")
|
||||
entity_id = request.query_params.get("entity_id")
|
||||
|
||||
if not entity_type or not entity_id:
|
||||
return Response(
|
||||
{"detail": "entity_type and entity_id are required"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Try to get view count from analytics tracking
|
||||
try:
|
||||
from apps.core.models import EntityViewCount
|
||||
|
||||
view_count = EntityViewCount.objects.filter(
|
||||
entity_type=entity_type,
|
||||
entity_id=entity_id,
|
||||
).first()
|
||||
|
||||
if view_count:
|
||||
return Response({
|
||||
"entity_type": entity_type,
|
||||
"entity_id": entity_id,
|
||||
"view_count": view_count.count,
|
||||
"last_viewed": view_count.last_viewed_at,
|
||||
})
|
||||
except Exception:
|
||||
# Model may not exist yet, return placeholder
|
||||
pass
|
||||
|
||||
return Response({
|
||||
"entity_type": entity_type,
|
||||
"entity_id": entity_id,
|
||||
"view_count": 0,
|
||||
"last_viewed": None,
|
||||
})
|
||||
|
||||
@extend_schema(
|
||||
tags=["Core"],
|
||||
summary="Record a view for an entity",
|
||||
description="Increment the view count for a specified entity",
|
||||
)
|
||||
def post(self, request):
|
||||
"""Record a view for an entity."""
|
||||
entity_type = request.data.get("entity_type")
|
||||
entity_id = request.data.get("entity_id")
|
||||
|
||||
if not entity_type or not entity_id:
|
||||
return Response(
|
||||
{"detail": "entity_type and entity_id are required"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
# Track the view
|
||||
try:
|
||||
from django.utils import timezone
|
||||
from apps.core.models import EntityViewCount
|
||||
|
||||
view_count, created = EntityViewCount.objects.get_or_create(
|
||||
entity_type=entity_type,
|
||||
entity_id=entity_id,
|
||||
defaults={"count": 0},
|
||||
)
|
||||
view_count.count += 1
|
||||
view_count.last_viewed_at = timezone.now()
|
||||
view_count.save(update_fields=["count", "last_viewed_at"])
|
||||
|
||||
return Response({
|
||||
"success": True,
|
||||
"entity_type": entity_type,
|
||||
"entity_id": entity_id,
|
||||
"view_count": view_count.count,
|
||||
}, status=status.HTTP_200_OK)
|
||||
except Exception as e:
|
||||
# Model may not exist, log and return success anyway
|
||||
logger.debug(f"View count tracking not available: {e}")
|
||||
return Response({
|
||||
"success": True,
|
||||
"entity_type": entity_type,
|
||||
"entity_id": entity_id,
|
||||
"view_count": 1, # Assume first view
|
||||
}, status=status.HTTP_200_OK)
|
||||
|
||||
|
||||
class TelemetryView(APIView):
|
||||
"""
|
||||
Handle frontend telemetry and request metadata logging.
|
||||
|
||||
254
backend/apps/api/v1/rides/ride_model_views.py
Normal file
254
backend/apps/api/v1/rides/ride_model_views.py
Normal file
@@ -0,0 +1,254 @@
|
||||
"""
|
||||
Global Ride Model views for ThrillWiki API v1.
|
||||
|
||||
This module provides top-level ride model endpoints that don't require
|
||||
manufacturer context, matching the frontend's expectation of /rides/models/.
|
||||
"""
|
||||
|
||||
from django.db.models import Q
|
||||
from drf_spectacular.types import OpenApiTypes
|
||||
from drf_spectacular.utils import OpenApiParameter, extend_schema
|
||||
from rest_framework import permissions, status
|
||||
from rest_framework.pagination import PageNumberPagination
|
||||
from rest_framework.request import Request
|
||||
from rest_framework.response import Response
|
||||
from rest_framework.views import APIView
|
||||
|
||||
# Import serializers
|
||||
from apps.api.v1.serializers.ride_models import (
|
||||
RideModelDetailOutputSerializer,
|
||||
RideModelListOutputSerializer,
|
||||
)
|
||||
|
||||
# Attempt to import models
|
||||
try:
|
||||
from apps.rides.models import RideModel
|
||||
from apps.rides.models.company import Company
|
||||
|
||||
MODELS_AVAILABLE = True
|
||||
except ImportError:
|
||||
try:
|
||||
from apps.rides.models.rides import Company, RideModel
|
||||
|
||||
MODELS_AVAILABLE = True
|
||||
except ImportError:
|
||||
RideModel = None
|
||||
Company = None
|
||||
MODELS_AVAILABLE = False
|
||||
|
||||
|
||||
class StandardResultsSetPagination(PageNumberPagination):
|
||||
page_size = 20
|
||||
page_size_query_param = "page_size"
|
||||
max_page_size = 100
|
||||
|
||||
|
||||
class GlobalRideModelListAPIView(APIView):
|
||||
"""
|
||||
Global ride model list endpoint.
|
||||
|
||||
This endpoint provides a top-level list of all ride models without
|
||||
requiring a manufacturer slug, matching the frontend's expectation
|
||||
of calling /rides/models/ directly.
|
||||
"""
|
||||
|
||||
permission_classes = [permissions.AllowAny]
|
||||
|
||||
@extend_schema(
|
||||
summary="List all ride models with filtering and pagination",
|
||||
description=(
|
||||
"List all ride models across all manufacturers with comprehensive "
|
||||
"filtering and pagination support. This is a global endpoint that "
|
||||
"doesn't require manufacturer context."
|
||||
),
|
||||
parameters=[
|
||||
OpenApiParameter(
|
||||
name="page",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.INT,
|
||||
description="Page number for pagination",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="page_size",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.INT,
|
||||
description="Number of results per page (max 100)",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="search",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.STR,
|
||||
description="Search term for name, description, or manufacturer",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="category",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.STR,
|
||||
description="Filter by category (e.g., RC, DR, FR, WR)",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="manufacturer",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.STR,
|
||||
description="Filter by manufacturer slug",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="target_market",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.STR,
|
||||
description="Filter by target market (e.g., FAMILY, THRILL)",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="is_discontinued",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.BOOL,
|
||||
description="Filter by discontinued status",
|
||||
),
|
||||
OpenApiParameter(
|
||||
name="ordering",
|
||||
location=OpenApiParameter.QUERY,
|
||||
type=OpenApiTypes.STR,
|
||||
description="Order by field: name, -name, manufacturer__name, etc.",
|
||||
),
|
||||
],
|
||||
responses={200: RideModelListOutputSerializer(many=True)},
|
||||
tags=["Ride Models"],
|
||||
)
|
||||
def get(self, request: Request) -> Response:
|
||||
"""List all ride models with filtering and pagination."""
|
||||
if not MODELS_AVAILABLE:
|
||||
return Response(
|
||||
{
|
||||
"count": 0,
|
||||
"next": None,
|
||||
"previous": None,
|
||||
"results": [],
|
||||
"detail": "Ride model listing is not available.",
|
||||
},
|
||||
status=status.HTTP_200_OK,
|
||||
)
|
||||
|
||||
# Base queryset with eager loading
|
||||
qs = RideModel.objects.select_related("manufacturer").prefetch_related(
|
||||
"photos"
|
||||
).order_by("manufacturer__name", "name")
|
||||
|
||||
# Search filter
|
||||
search = request.query_params.get("search", "").strip()
|
||||
if search:
|
||||
qs = qs.filter(
|
||||
Q(name__icontains=search)
|
||||
| Q(description__icontains=search)
|
||||
| Q(manufacturer__name__icontains=search)
|
||||
)
|
||||
|
||||
# Category filter
|
||||
category = request.query_params.get("category", "").strip()
|
||||
if category:
|
||||
# Support comma-separated categories
|
||||
categories = [c.strip() for c in category.split(",") if c.strip()]
|
||||
if categories:
|
||||
qs = qs.filter(category__in=categories)
|
||||
|
||||
# Manufacturer filter
|
||||
manufacturer = request.query_params.get("manufacturer", "").strip()
|
||||
if manufacturer:
|
||||
qs = qs.filter(manufacturer__slug=manufacturer)
|
||||
|
||||
# Target market filter
|
||||
target_market = request.query_params.get("target_market", "").strip()
|
||||
if target_market:
|
||||
markets = [m.strip() for m in target_market.split(",") if m.strip()]
|
||||
if markets:
|
||||
qs = qs.filter(target_market__in=markets)
|
||||
|
||||
# Discontinued filter
|
||||
is_discontinued = request.query_params.get("is_discontinued")
|
||||
if is_discontinued is not None:
|
||||
qs = qs.filter(is_discontinued=is_discontinued.lower() == "true")
|
||||
|
||||
# Ordering
|
||||
ordering = request.query_params.get("ordering", "manufacturer__name,name")
|
||||
valid_orderings = [
|
||||
"name", "-name",
|
||||
"manufacturer__name", "-manufacturer__name",
|
||||
"first_installation_year", "-first_installation_year",
|
||||
"total_installations", "-total_installations",
|
||||
"created_at", "-created_at",
|
||||
]
|
||||
if ordering:
|
||||
order_fields = [
|
||||
f.strip() for f in ordering.split(",")
|
||||
if f.strip() in valid_orderings or f.strip().lstrip("-") in [
|
||||
o.lstrip("-") for o in valid_orderings
|
||||
]
|
||||
]
|
||||
if order_fields:
|
||||
qs = qs.order_by(*order_fields)
|
||||
|
||||
# Paginate
|
||||
paginator = StandardResultsSetPagination()
|
||||
page = paginator.paginate_queryset(qs, request)
|
||||
|
||||
if page is not None:
|
||||
serializer = RideModelListOutputSerializer(
|
||||
page, many=True, context={"request": request}
|
||||
)
|
||||
return paginator.get_paginated_response(serializer.data)
|
||||
|
||||
# Fallback without pagination
|
||||
serializer = RideModelListOutputSerializer(
|
||||
qs[:100], many=True, context={"request": request}
|
||||
)
|
||||
return Response(serializer.data)
|
||||
|
||||
|
||||
class GlobalRideModelDetailAPIView(APIView):
|
||||
"""
|
||||
Global ride model detail endpoint by ID or slug.
|
||||
|
||||
This endpoint provides detail for a single ride model without
|
||||
requiring manufacturer context.
|
||||
"""
|
||||
|
||||
permission_classes = [permissions.AllowAny]
|
||||
|
||||
@extend_schema(
|
||||
summary="Retrieve a ride model by ID",
|
||||
description="Get detailed information about a specific ride model by its ID.",
|
||||
parameters=[
|
||||
OpenApiParameter(
|
||||
name="pk",
|
||||
location=OpenApiParameter.PATH,
|
||||
type=OpenApiTypes.INT,
|
||||
required=True,
|
||||
description="Ride model ID",
|
||||
),
|
||||
],
|
||||
responses={200: RideModelDetailOutputSerializer()},
|
||||
tags=["Ride Models"],
|
||||
)
|
||||
def get(self, request: Request, pk: int) -> Response:
|
||||
"""Get ride model detail by ID."""
|
||||
if not MODELS_AVAILABLE:
|
||||
return Response(
|
||||
{"detail": "Ride model not found"},
|
||||
status=status.HTTP_404_NOT_FOUND,
|
||||
)
|
||||
|
||||
try:
|
||||
ride_model = (
|
||||
RideModel.objects.select_related("manufacturer")
|
||||
.prefetch_related("photos", "variants", "technical_specs")
|
||||
.get(pk=pk)
|
||||
)
|
||||
except RideModel.DoesNotExist:
|
||||
return Response(
|
||||
{"detail": "Ride model not found"},
|
||||
status=status.HTTP_404_NOT_FOUND,
|
||||
)
|
||||
|
||||
serializer = RideModelDetailOutputSerializer(
|
||||
ride_model, context={"request": request}
|
||||
)
|
||||
return Response(serializer.data)
|
||||
@@ -306,6 +306,12 @@ class HybridRideSerializer(serializers.ModelSerializer):
|
||||
banner_image_url = serializers.SerializerMethodField()
|
||||
card_image_url = serializers.SerializerMethodField()
|
||||
|
||||
# Metric unit conversions for frontend (duplicate of imperial fields)
|
||||
coaster_height_meters = serializers.SerializerMethodField()
|
||||
coaster_length_meters = serializers.SerializerMethodField()
|
||||
coaster_speed_kmh = serializers.SerializerMethodField()
|
||||
coaster_max_drop_meters = serializers.SerializerMethodField()
|
||||
|
||||
# Computed fields for filtering
|
||||
opening_year = serializers.IntegerField(read_only=True)
|
||||
search_text = serializers.CharField(read_only=True)
|
||||
@@ -502,6 +508,47 @@ class HybridRideSerializer(serializers.ModelSerializer):
|
||||
"""Check if ride has an announced closing date in the future."""
|
||||
return obj.is_closing
|
||||
|
||||
# Metric conversions for frontend compatibility
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_coaster_height_meters(self, obj):
|
||||
"""Convert coaster height from feet to meters."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.height_ft:
|
||||
return round(float(obj.coaster_stats.height_ft) * 0.3048, 2)
|
||||
return None
|
||||
except (AttributeError, TypeError):
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_coaster_length_meters(self, obj):
|
||||
"""Convert coaster length from feet to meters."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.length_ft:
|
||||
return round(float(obj.coaster_stats.length_ft) * 0.3048, 2)
|
||||
return None
|
||||
except (AttributeError, TypeError):
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_coaster_speed_kmh(self, obj):
|
||||
"""Convert coaster speed from mph to km/h."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.speed_mph:
|
||||
return round(float(obj.coaster_stats.speed_mph) * 1.60934, 2)
|
||||
return None
|
||||
except (AttributeError, TypeError):
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_coaster_max_drop_meters(self, obj):
|
||||
"""Convert coaster max drop from feet to meters."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.max_drop_height_ft:
|
||||
return round(float(obj.coaster_stats.max_drop_height_ft) * 0.3048, 2)
|
||||
return None
|
||||
except (AttributeError, TypeError):
|
||||
return None
|
||||
|
||||
# Water ride stats fields
|
||||
water_wetness_level = serializers.SerializerMethodField()
|
||||
water_splash_height_ft = serializers.SerializerMethodField()
|
||||
@@ -997,3 +1044,29 @@ class RideSerializer(serializers.ModelSerializer):
|
||||
"opening_date",
|
||||
"closing_date",
|
||||
]
|
||||
|
||||
|
||||
class RideSubTypeSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for ride sub-types lookup table.
|
||||
|
||||
This serves the /rides/sub-types/ endpoint which the frontend
|
||||
uses to populate sub-type dropdowns filtered by category.
|
||||
"""
|
||||
|
||||
created_by = serializers.CharField(source="created_by.username", read_only=True, allow_null=True)
|
||||
|
||||
class Meta:
|
||||
# Import here to avoid circular imports
|
||||
from apps.rides.models import RideSubType
|
||||
model = RideSubType
|
||||
fields = [
|
||||
"id",
|
||||
"name",
|
||||
"category",
|
||||
"description",
|
||||
"created_by",
|
||||
"created_at",
|
||||
]
|
||||
read_only_fields = ["id", "created_at", "created_by"]
|
||||
|
||||
|
||||
|
||||
@@ -12,6 +12,7 @@ from django.urls import include, path
|
||||
from rest_framework.routers import DefaultRouter
|
||||
|
||||
from .photo_views import RidePhotoViewSet
|
||||
from .ride_model_views import GlobalRideModelDetailAPIView, GlobalRideModelListAPIView
|
||||
from .views import (
|
||||
CompanySearchAPIView,
|
||||
DesignerListAPIView,
|
||||
@@ -24,6 +25,7 @@ from .views import (
|
||||
RideListCreateAPIView,
|
||||
RideModelSearchAPIView,
|
||||
RideSearchSuggestionsAPIView,
|
||||
RideSubTypeListAPIView,
|
||||
)
|
||||
|
||||
# Create router for nested photo endpoints
|
||||
@@ -40,6 +42,9 @@ urlpatterns = [
|
||||
path("hybrid/filter-metadata/", RideFilterMetadataAPIView.as_view(), name="ride-hybrid-filter-metadata"),
|
||||
# Filter options
|
||||
path("filter-options/", FilterOptionsAPIView.as_view(), name="ride-filter-options"),
|
||||
# Global ride model endpoints - matches frontend's /rides/models/ expectation
|
||||
path("models/", GlobalRideModelListAPIView.as_view(), name="ride-model-global-list"),
|
||||
path("models/<int:pk>/", GlobalRideModelDetailAPIView.as_view(), name="ride-model-global-detail"),
|
||||
# Autocomplete / suggestion endpoints
|
||||
path(
|
||||
"search/companies/",
|
||||
@@ -59,6 +64,8 @@ urlpatterns = [
|
||||
# Manufacturer and Designer endpoints
|
||||
path("manufacturers/", ManufacturerListAPIView.as_view(), name="manufacturer-list"),
|
||||
path("designers/", DesignerListAPIView.as_view(), name="designer-list"),
|
||||
# Ride sub-types endpoint - for autocomplete dropdowns
|
||||
path("sub-types/", RideSubTypeListAPIView.as_view(), name="ride-sub-type-list"),
|
||||
# Ride model management endpoints - nested under rides/manufacturers
|
||||
path(
|
||||
"manufacturers/<slug:manufacturer_slug>/",
|
||||
|
||||
@@ -2422,3 +2422,53 @@ class ManufacturerListAPIView(BaseCompanyListAPIView):
|
||||
)
|
||||
class DesignerListAPIView(BaseCompanyListAPIView):
|
||||
role = "DESIGNER"
|
||||
|
||||
|
||||
# === RIDE SUB-TYPES ===
|
||||
|
||||
|
||||
@extend_schema(
|
||||
summary="List ride sub-types",
|
||||
description="List ride sub-types, optionally filtered by category. Used for autocomplete dropdowns.",
|
||||
parameters=[
|
||||
OpenApiParameter(
|
||||
"category",
|
||||
OpenApiTypes.STR,
|
||||
description="Filter by ride category (e.g., 'RC' for roller coaster)",
|
||||
),
|
||||
],
|
||||
responses={200: OpenApiTypes.OBJECT},
|
||||
tags=["Rides"],
|
||||
)
|
||||
class RideSubTypeListAPIView(APIView):
|
||||
"""
|
||||
API View for listing ride sub-types.
|
||||
|
||||
Used by the frontend's useRideSubTypes hook to populate
|
||||
sub-type dropdown menus filtered by ride category.
|
||||
|
||||
Caching: 30-minute timeout (1800s) - sub-types are stable lookup data.
|
||||
"""
|
||||
|
||||
permission_classes = [permissions.AllowAny]
|
||||
|
||||
@cache_api_response(timeout=1800, key_prefix="ride_sub_types")
|
||||
def get(self, request: Request) -> Response:
|
||||
from apps.rides.models import RideSubType
|
||||
from apps.api.v1.rides.serializers import RideSubTypeSerializer
|
||||
|
||||
# Start with all sub-types
|
||||
queryset = RideSubType.objects.all().order_by("name")
|
||||
|
||||
# Apply category filter if provided
|
||||
category = request.query_params.get("category")
|
||||
if category:
|
||||
queryset = queryset.filter(category=category)
|
||||
|
||||
# Serialize and return
|
||||
serializer = RideSubTypeSerializer(queryset, many=True)
|
||||
return Response({
|
||||
"results": serializer.data,
|
||||
"count": queryset.count(),
|
||||
})
|
||||
|
||||
|
||||
@@ -12,7 +12,7 @@ from drf_spectacular.utils import (
|
||||
)
|
||||
from rest_framework import serializers
|
||||
|
||||
from apps.core.choices.serializers import RichChoiceFieldSerializer
|
||||
from apps.core.choices.serializers import RichChoiceFieldSerializer, RichChoiceSerializerField
|
||||
|
||||
from .shared import ModelChoices
|
||||
|
||||
@@ -87,30 +87,38 @@ class CompanyCreateInputSerializer(serializers.Serializer):
|
||||
description = serializers.CharField(allow_blank=True, default="")
|
||||
website = serializers.URLField(required=False, allow_blank=True)
|
||||
|
||||
# Entity type and status
|
||||
person_type = serializers.ChoiceField(
|
||||
choices=["INDIVIDUAL", "FIRM", "ORGANIZATION", "CORPORATION", "PARTNERSHIP", "GOVERNMENT"],
|
||||
# Entity type and status - using RichChoiceSerializerField
|
||||
person_type = RichChoiceSerializerField(
|
||||
choice_group="person_types",
|
||||
domain="parks",
|
||||
required=False,
|
||||
allow_blank=True,
|
||||
)
|
||||
status = serializers.ChoiceField(
|
||||
choices=["ACTIVE", "DEFUNCT", "MERGED", "ACQUIRED", "RENAMED", "DORMANT"],
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="company_statuses",
|
||||
domain="parks",
|
||||
default="ACTIVE",
|
||||
)
|
||||
|
||||
# Founding information
|
||||
founded_year = serializers.IntegerField(required=False, allow_null=True)
|
||||
founded_date = serializers.DateField(required=False, allow_null=True)
|
||||
founded_date_precision = serializers.ChoiceField(
|
||||
choices=["YEAR", "MONTH", "DAY"],
|
||||
founded_date_precision = RichChoiceSerializerField(
|
||||
choice_group="date_precision",
|
||||
domain="parks",
|
||||
required=False,
|
||||
allow_blank=True,
|
||||
)
|
||||
|
||||
# Image URLs
|
||||
# Image URLs (legacy - prefer using image IDs)
|
||||
logo_url = serializers.URLField(required=False, allow_blank=True)
|
||||
banner_image_url = serializers.URLField(required=False, allow_blank=True)
|
||||
card_image_url = serializers.URLField(required=False, allow_blank=True)
|
||||
|
||||
# Cloudflare image IDs (preferred for new submissions)
|
||||
logo_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
banner_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
card_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
|
||||
|
||||
class CompanyUpdateInputSerializer(serializers.Serializer):
|
||||
@@ -124,30 +132,38 @@ class CompanyUpdateInputSerializer(serializers.Serializer):
|
||||
description = serializers.CharField(allow_blank=True, required=False)
|
||||
website = serializers.URLField(required=False, allow_blank=True)
|
||||
|
||||
# Entity type and status
|
||||
person_type = serializers.ChoiceField(
|
||||
choices=["INDIVIDUAL", "FIRM", "ORGANIZATION", "CORPORATION", "PARTNERSHIP", "GOVERNMENT"],
|
||||
# Entity type and status - using RichChoiceSerializerField
|
||||
person_type = RichChoiceSerializerField(
|
||||
choice_group="person_types",
|
||||
domain="parks",
|
||||
required=False,
|
||||
allow_blank=True,
|
||||
)
|
||||
status = serializers.ChoiceField(
|
||||
choices=["ACTIVE", "DEFUNCT", "MERGED", "ACQUIRED", "RENAMED", "DORMANT"],
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="company_statuses",
|
||||
domain="parks",
|
||||
required=False,
|
||||
)
|
||||
|
||||
# Founding information
|
||||
founded_year = serializers.IntegerField(required=False, allow_null=True)
|
||||
founded_date = serializers.DateField(required=False, allow_null=True)
|
||||
founded_date_precision = serializers.ChoiceField(
|
||||
choices=["YEAR", "MONTH", "DAY"],
|
||||
founded_date_precision = RichChoiceSerializerField(
|
||||
choice_group="date_precision",
|
||||
domain="parks",
|
||||
required=False,
|
||||
allow_blank=True,
|
||||
)
|
||||
|
||||
# Image URLs
|
||||
# Image URLs (legacy - prefer using image IDs)
|
||||
logo_url = serializers.URLField(required=False, allow_blank=True)
|
||||
banner_image_url = serializers.URLField(required=False, allow_blank=True)
|
||||
card_image_url = serializers.URLField(required=False, allow_blank=True)
|
||||
|
||||
# Cloudflare image IDs (preferred for new submissions)
|
||||
logo_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
banner_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
card_image_id = serializers.CharField(max_length=255, required=False, allow_blank=True)
|
||||
|
||||
|
||||
# === RIDE MODEL SERIALIZERS ===
|
||||
|
||||
@@ -5,6 +5,8 @@ This module contains all serializers related to parks, park areas, park location
|
||||
and park search functionality.
|
||||
"""
|
||||
|
||||
from decimal import Decimal
|
||||
|
||||
from drf_spectacular.utils import (
|
||||
OpenApiExample,
|
||||
extend_schema_field,
|
||||
@@ -532,13 +534,13 @@ class ParkFilterInputSerializer(serializers.Serializer):
|
||||
max_digits=3,
|
||||
decimal_places=2,
|
||||
required=False,
|
||||
min_value=1,
|
||||
max_value=10,
|
||||
min_value=Decimal("1"),
|
||||
max_value=Decimal("10"),
|
||||
)
|
||||
|
||||
# Size filter
|
||||
min_size_acres = serializers.DecimalField(max_digits=10, decimal_places=2, required=False, min_value=0)
|
||||
max_size_acres = serializers.DecimalField(max_digits=10, decimal_places=2, required=False, min_value=0)
|
||||
min_size_acres = serializers.DecimalField(max_digits=10, decimal_places=2, required=False, min_value=Decimal("0"))
|
||||
max_size_acres = serializers.DecimalField(max_digits=10, decimal_places=2, required=False, min_value=Decimal("0"))
|
||||
|
||||
# Company filters
|
||||
operator_id = serializers.IntegerField(required=False)
|
||||
|
||||
@@ -211,6 +211,18 @@ class RideDetailOutputSerializer(serializers.Serializer):
|
||||
# Former names (name history)
|
||||
former_names = serializers.SerializerMethodField()
|
||||
|
||||
# Coaster statistics - includes both imperial and metric units for frontend flexibility
|
||||
coaster_statistics = serializers.SerializerMethodField()
|
||||
|
||||
# Metric unit fields for frontend (converted from imperial)
|
||||
height_meters = serializers.SerializerMethodField()
|
||||
length_meters = serializers.SerializerMethodField()
|
||||
max_speed_kmh = serializers.SerializerMethodField()
|
||||
drop_meters = serializers.SerializerMethodField()
|
||||
|
||||
# Technical specifications list
|
||||
technical_specifications = serializers.SerializerMethodField()
|
||||
|
||||
# URL
|
||||
url = serializers.SerializerMethodField()
|
||||
|
||||
@@ -427,6 +439,99 @@ class RideDetailOutputSerializer(serializers.Serializer):
|
||||
for entry in former_names
|
||||
]
|
||||
|
||||
@extend_schema_field(serializers.DictField(allow_null=True))
|
||||
def get_coaster_statistics(self, obj):
|
||||
"""Get coaster statistics with both imperial and metric units."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats:
|
||||
stats = obj.coaster_stats
|
||||
return {
|
||||
# Imperial units (stored in DB)
|
||||
"height_ft": float(stats.height_ft) if stats.height_ft else None,
|
||||
"length_ft": float(stats.length_ft) if stats.length_ft else None,
|
||||
"speed_mph": float(stats.speed_mph) if stats.speed_mph else None,
|
||||
"max_drop_height_ft": float(stats.max_drop_height_ft) if stats.max_drop_height_ft else None,
|
||||
# Metric conversions for frontend
|
||||
"height_meters": round(float(stats.height_ft) * 0.3048, 2) if stats.height_ft else None,
|
||||
"length_meters": round(float(stats.length_ft) * 0.3048, 2) if stats.length_ft else None,
|
||||
"max_speed_kmh": round(float(stats.speed_mph) * 1.60934, 2) if stats.speed_mph else None,
|
||||
"drop_meters": round(float(stats.max_drop_height_ft) * 0.3048, 2) if stats.max_drop_height_ft else None,
|
||||
# Other stats
|
||||
"inversions": stats.inversions,
|
||||
"ride_time_seconds": stats.ride_time_seconds,
|
||||
"track_type": stats.track_type,
|
||||
"track_material": stats.track_material,
|
||||
"roller_coaster_type": stats.roller_coaster_type,
|
||||
"propulsion_system": stats.propulsion_system,
|
||||
"train_style": stats.train_style,
|
||||
"trains_count": stats.trains_count,
|
||||
"cars_per_train": stats.cars_per_train,
|
||||
"seats_per_car": stats.seats_per_car,
|
||||
}
|
||||
except AttributeError:
|
||||
pass
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_height_meters(self, obj):
|
||||
"""Convert height from feet to meters for frontend."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.height_ft:
|
||||
return round(float(obj.coaster_stats.height_ft) * 0.3048, 2)
|
||||
except (AttributeError, TypeError):
|
||||
pass
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_length_meters(self, obj):
|
||||
"""Convert length from feet to meters for frontend."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.length_ft:
|
||||
return round(float(obj.coaster_stats.length_ft) * 0.3048, 2)
|
||||
except (AttributeError, TypeError):
|
||||
pass
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_max_speed_kmh(self, obj):
|
||||
"""Convert max speed from mph to km/h for frontend."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.speed_mph:
|
||||
return round(float(obj.coaster_stats.speed_mph) * 1.60934, 2)
|
||||
except (AttributeError, TypeError):
|
||||
pass
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.FloatField(allow_null=True))
|
||||
def get_drop_meters(self, obj):
|
||||
"""Convert drop height from feet to meters for frontend."""
|
||||
try:
|
||||
if hasattr(obj, "coaster_stats") and obj.coaster_stats and obj.coaster_stats.max_drop_height_ft:
|
||||
return round(float(obj.coaster_stats.max_drop_height_ft) * 0.3048, 2)
|
||||
except (AttributeError, TypeError):
|
||||
pass
|
||||
return None
|
||||
|
||||
@extend_schema_field(serializers.ListField(child=serializers.DictField()))
|
||||
def get_technical_specifications(self, obj):
|
||||
"""Get technical specifications list for this ride."""
|
||||
try:
|
||||
from apps.rides.models import RideTechnicalSpec
|
||||
|
||||
specs = RideTechnicalSpec.objects.filter(ride=obj).order_by("category", "name")
|
||||
return [
|
||||
{
|
||||
"id": spec.id,
|
||||
"name": spec.name,
|
||||
"value": spec.value,
|
||||
"unit": spec.unit,
|
||||
"category": spec.category,
|
||||
}
|
||||
for spec in specs
|
||||
]
|
||||
except Exception:
|
||||
return []
|
||||
|
||||
|
||||
class RideImageSettingsInputSerializer(serializers.Serializer):
|
||||
"""Input serializer for setting ride banner and card images."""
|
||||
|
||||
@@ -493,6 +493,18 @@ def ensure_filter_option_format(options: list[Any]) -> list[dict[str, Any]]:
|
||||
"count": option.get("count"),
|
||||
"selected": option.get("selected", False),
|
||||
}
|
||||
elif isinstance(option, tuple):
|
||||
# Tuple format: (value, label) or (value, label, count)
|
||||
if len(option) >= 2:
|
||||
standardized_option = {
|
||||
"value": str(option[0]),
|
||||
"label": str(option[1]),
|
||||
"count": option[2] if len(option) > 2 else None,
|
||||
"selected": False,
|
||||
}
|
||||
else:
|
||||
# Single-element tuple, treat as simple value
|
||||
standardized_option = {"value": str(option[0]), "label": str(option[0]), "count": None, "selected": False}
|
||||
elif hasattr(option, "value") and hasattr(option, "label"):
|
||||
# RichChoice object format
|
||||
standardized_option = {
|
||||
|
||||
@@ -34,6 +34,17 @@ from apps.core.api.analytics_views import (
|
||||
RequestMetadataViewSet,
|
||||
)
|
||||
|
||||
# Import observability views
|
||||
from apps.core.api.observability_views import (
|
||||
AlertCorrelationViewSet,
|
||||
AnomalyViewSet,
|
||||
CleanupJobLogViewSet,
|
||||
DataRetentionStatsView,
|
||||
PipelineErrorViewSet,
|
||||
)
|
||||
from apps.notifications.api.log_views import NotificationLogViewSet
|
||||
from apps.moderation.views import ModerationAuditLogViewSet
|
||||
|
||||
# Create the main API router
|
||||
router = DefaultRouter()
|
||||
|
||||
@@ -44,6 +55,14 @@ router.register(r"rankings", RideRankingViewSet, basename="ranking")
|
||||
router.register(r"request_metadata", RequestMetadataViewSet, basename="request_metadata")
|
||||
router.register(r"approval_transaction_metrics", ApprovalTransactionMetricViewSet, basename="approval_transaction_metrics")
|
||||
|
||||
# Register observability endpoints (Supabase table parity)
|
||||
router.register(r"pipeline_errors", PipelineErrorViewSet, basename="pipeline_errors")
|
||||
router.register(r"notification_logs", NotificationLogViewSet, basename="notification_logs")
|
||||
router.register(r"cleanup_job_log", CleanupJobLogViewSet, basename="cleanup_job_log")
|
||||
router.register(r"moderation_audit_log", ModerationAuditLogViewSet, basename="moderation_audit_log")
|
||||
router.register(r"alert_correlations_view", AlertCorrelationViewSet, basename="alert_correlations_view")
|
||||
router.register(r"recent_anomalies_view", AnomalyViewSet, basename="recent_anomalies_view")
|
||||
|
||||
app_name = "api_v1"
|
||||
|
||||
urlpatterns = [
|
||||
@@ -53,6 +72,8 @@ urlpatterns = [
|
||||
path("auth/", include("apps.api.v1.auth.urls")),
|
||||
# Analytics endpoints (error_summary is a view, not a viewset)
|
||||
path("error_summary/", ErrorSummaryView.as_view(), name="error-summary"),
|
||||
# Data retention stats view (aggregation endpoint)
|
||||
path("data_retention_stats/", DataRetentionStatsView.as_view(), name="data-retention-stats"),
|
||||
# Health check endpoints
|
||||
path("health/", HealthCheckAPIView.as_view(), name="health-check"),
|
||||
path("health/simple/", SimpleHealthAPIView.as_view(), name="simple-health"),
|
||||
|
||||
176
backend/apps/core/api/observability_serializers.py
Normal file
176
backend/apps/core/api/observability_serializers.py
Normal file
@@ -0,0 +1,176 @@
|
||||
"""
|
||||
Serializers for observability API endpoints.
|
||||
|
||||
Provides serializers for PipelineError, Anomaly, AlertCorrelationRule,
|
||||
CleanupJobLog, and DataRetentionStats.
|
||||
"""
|
||||
|
||||
from rest_framework import serializers
|
||||
|
||||
from apps.core.choices.serializers import RichChoiceSerializerField
|
||||
from apps.core.models import (
|
||||
AlertCorrelationRule,
|
||||
Anomaly,
|
||||
CleanupJobLog,
|
||||
PipelineError,
|
||||
)
|
||||
|
||||
|
||||
class PipelineErrorSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for pipeline errors."""
|
||||
|
||||
severity = RichChoiceSerializerField(
|
||||
choice_group="pipeline_error_severities",
|
||||
domain="core",
|
||||
)
|
||||
resolved_by_username = serializers.CharField(
|
||||
source="resolved_by.username",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
|
||||
class Meta:
|
||||
model = PipelineError
|
||||
fields = [
|
||||
"id",
|
||||
"function_name",
|
||||
"error_message",
|
||||
"error_code",
|
||||
"error_context",
|
||||
"stack_trace",
|
||||
"severity",
|
||||
"submission_id",
|
||||
"item_id",
|
||||
"request_id",
|
||||
"trace_id",
|
||||
"resolved",
|
||||
"resolved_by",
|
||||
"resolved_by_username",
|
||||
"resolved_at",
|
||||
"resolution_notes",
|
||||
"occurred_at",
|
||||
]
|
||||
read_only_fields = ["id", "occurred_at", "resolved_by_username"]
|
||||
|
||||
|
||||
class PipelineErrorResolveSerializer(serializers.Serializer):
|
||||
"""Serializer for resolving pipeline errors."""
|
||||
|
||||
resolution_notes = serializers.CharField(required=False, allow_blank=True)
|
||||
|
||||
|
||||
class AnomalySerializer(serializers.ModelSerializer):
|
||||
"""Serializer for detected anomalies."""
|
||||
|
||||
anomaly_type = RichChoiceSerializerField(
|
||||
choice_group="anomaly_types",
|
||||
domain="core",
|
||||
)
|
||||
severity = RichChoiceSerializerField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
)
|
||||
alert_message = serializers.CharField(
|
||||
source="alert.message",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
alert_resolved_at = serializers.DateTimeField(
|
||||
source="alert.resolved_at",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
alert_id = serializers.UUIDField(
|
||||
source="alert.id",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
|
||||
class Meta:
|
||||
model = Anomaly
|
||||
fields = [
|
||||
"id",
|
||||
"metric_name",
|
||||
"metric_category",
|
||||
"anomaly_type",
|
||||
"severity",
|
||||
"anomaly_value",
|
||||
"baseline_value",
|
||||
"deviation_score",
|
||||
"confidence_score",
|
||||
"detection_algorithm",
|
||||
"time_window_start",
|
||||
"time_window_end",
|
||||
"alert_created",
|
||||
"alert_id",
|
||||
"alert_message",
|
||||
"alert_resolved_at",
|
||||
"detected_at",
|
||||
]
|
||||
read_only_fields = [
|
||||
"id",
|
||||
"detected_at",
|
||||
"alert_id",
|
||||
"alert_message",
|
||||
"alert_resolved_at",
|
||||
]
|
||||
|
||||
|
||||
class AlertCorrelationRuleSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for alert correlation rules."""
|
||||
|
||||
incident_severity = RichChoiceSerializerField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
model = AlertCorrelationRule
|
||||
fields = [
|
||||
"id",
|
||||
"rule_name",
|
||||
"rule_description",
|
||||
"min_alerts_required",
|
||||
"time_window_minutes",
|
||||
"incident_severity",
|
||||
"incident_title_template",
|
||||
"is_active",
|
||||
"created_at",
|
||||
"updated_at",
|
||||
]
|
||||
read_only_fields = ["id", "created_at", "updated_at"]
|
||||
|
||||
|
||||
class CleanupJobLogSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for cleanup job logs."""
|
||||
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="cleanup_job_statuses",
|
||||
domain="core",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
model = CleanupJobLog
|
||||
fields = [
|
||||
"id",
|
||||
"job_name",
|
||||
"status",
|
||||
"records_processed",
|
||||
"records_deleted",
|
||||
"error_message",
|
||||
"duration_ms",
|
||||
"executed_at",
|
||||
]
|
||||
read_only_fields = ["id", "executed_at"]
|
||||
|
||||
|
||||
class DataRetentionStatsSerializer(serializers.Serializer):
|
||||
"""Serializer for data retention statistics view."""
|
||||
|
||||
table_name = serializers.CharField()
|
||||
total_records = serializers.IntegerField()
|
||||
last_7_days = serializers.IntegerField()
|
||||
last_30_days = serializers.IntegerField()
|
||||
oldest_record = serializers.DateTimeField(allow_null=True)
|
||||
newest_record = serializers.DateTimeField(allow_null=True)
|
||||
table_size = serializers.CharField()
|
||||
351
backend/apps/core/api/observability_views.py
Normal file
351
backend/apps/core/api/observability_views.py
Normal file
@@ -0,0 +1,351 @@
|
||||
"""
|
||||
ViewSets and Views for observability API endpoints.
|
||||
|
||||
Provides CRUD operations for PipelineError, read-only access for
|
||||
Anomaly, AlertCorrelationRule, CleanupJobLog, and aggregated views
|
||||
for DataRetentionStats.
|
||||
"""
|
||||
|
||||
from django.db import connection
|
||||
from django.db.models import Count, Max, Min
|
||||
from django.db.models.functions import Coalesce
|
||||
from django.utils import timezone
|
||||
from django_filters.rest_framework import DjangoFilterBackend
|
||||
from drf_spectacular.utils import extend_schema, extend_schema_view
|
||||
from rest_framework import status, viewsets
|
||||
from rest_framework.decorators import action
|
||||
from rest_framework.filters import OrderingFilter, SearchFilter
|
||||
from rest_framework.permissions import IsAdminUser
|
||||
from rest_framework.response import Response
|
||||
from rest_framework.views import APIView
|
||||
|
||||
from apps.core.models import (
|
||||
AlertCorrelationRule,
|
||||
Anomaly,
|
||||
CleanupJobLog,
|
||||
PipelineError,
|
||||
)
|
||||
|
||||
from .observability_serializers import (
|
||||
AlertCorrelationRuleSerializer,
|
||||
AnomalySerializer,
|
||||
CleanupJobLogSerializer,
|
||||
DataRetentionStatsSerializer,
|
||||
PipelineErrorResolveSerializer,
|
||||
PipelineErrorSerializer,
|
||||
)
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
list=extend_schema(
|
||||
summary="List pipeline errors",
|
||||
description="Get all pipeline errors, optionally filtered by severity or resolved status.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
retrieve=extend_schema(
|
||||
summary="Get pipeline error",
|
||||
description="Get details of a specific pipeline error.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
create=extend_schema(
|
||||
summary="Create pipeline error",
|
||||
description="Create a new pipeline error.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
update=extend_schema(
|
||||
summary="Update pipeline error",
|
||||
description="Update an existing pipeline error.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
partial_update=extend_schema(
|
||||
summary="Partial update pipeline error",
|
||||
description="Partially update an existing pipeline error.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
destroy=extend_schema(
|
||||
summary="Delete pipeline error",
|
||||
description="Delete a pipeline error.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
)
|
||||
class PipelineErrorViewSet(viewsets.ModelViewSet):
|
||||
"""
|
||||
ViewSet for managing pipeline errors.
|
||||
|
||||
Provides CRUD operations plus a resolve action for marking errors as resolved.
|
||||
"""
|
||||
|
||||
queryset = PipelineError.objects.select_related("resolved_by").all()
|
||||
serializer_class = PipelineErrorSerializer
|
||||
permission_classes = [IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["severity", "function_name", "resolved", "error_code"]
|
||||
search_fields = ["error_message", "function_name", "error_code"]
|
||||
ordering_fields = ["occurred_at", "severity"]
|
||||
ordering = ["-occurred_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
|
||||
# Date range filtering
|
||||
start_date = self.request.query_params.get("start_date")
|
||||
end_date = self.request.query_params.get("end_date")
|
||||
|
||||
if start_date:
|
||||
queryset = queryset.filter(occurred_at__gte=start_date)
|
||||
if end_date:
|
||||
queryset = queryset.filter(occurred_at__lte=end_date)
|
||||
|
||||
return queryset
|
||||
|
||||
@extend_schema(
|
||||
summary="Resolve pipeline error",
|
||||
description="Mark a pipeline error as resolved.",
|
||||
request=PipelineErrorResolveSerializer,
|
||||
responses={200: PipelineErrorSerializer},
|
||||
tags=["Admin - Observability"],
|
||||
)
|
||||
@action(detail=True, methods=["post"])
|
||||
def resolve(self, request, pk=None):
|
||||
"""Mark a pipeline error as resolved."""
|
||||
error = self.get_object()
|
||||
|
||||
if error.resolved:
|
||||
return Response(
|
||||
{"detail": "Error is already resolved"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
serializer = PipelineErrorResolveSerializer(data=request.data)
|
||||
serializer.is_valid(raise_exception=True)
|
||||
|
||||
error.resolved = True
|
||||
error.resolved_at = timezone.now()
|
||||
error.resolved_by = request.user
|
||||
error.resolution_notes = serializer.validated_data.get("resolution_notes", "")
|
||||
error.save()
|
||||
|
||||
return Response(PipelineErrorSerializer(error).data)
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
list=extend_schema(
|
||||
summary="List recent anomalies",
|
||||
description="Get recent anomalies with optional filtering by severity or type.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
retrieve=extend_schema(
|
||||
summary="Get anomaly details",
|
||||
description="Get details of a specific anomaly.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
)
|
||||
class AnomalyViewSet(viewsets.ReadOnlyModelViewSet):
|
||||
"""
|
||||
ViewSet for viewing detected anomalies.
|
||||
|
||||
Provides read-only access to anomaly data with filtering options.
|
||||
This serves as the recent_anomalies_view endpoint.
|
||||
"""
|
||||
|
||||
queryset = Anomaly.objects.select_related("alert").all()
|
||||
serializer_class = AnomalySerializer
|
||||
permission_classes = [IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["severity", "anomaly_type", "metric_category", "alert_created"]
|
||||
search_fields = ["metric_name", "metric_category"]
|
||||
ordering_fields = ["detected_at", "severity", "deviation_score"]
|
||||
ordering = ["-detected_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
|
||||
# Date range filtering
|
||||
start_date = self.request.query_params.get("start_date")
|
||||
end_date = self.request.query_params.get("end_date")
|
||||
|
||||
if start_date:
|
||||
queryset = queryset.filter(detected_at__gte=start_date)
|
||||
if end_date:
|
||||
queryset = queryset.filter(detected_at__lte=end_date)
|
||||
|
||||
return queryset
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
list=extend_schema(
|
||||
summary="List alert correlations",
|
||||
description="Get all alert correlation rules with optional filtering.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
retrieve=extend_schema(
|
||||
summary="Get alert correlation rule",
|
||||
description="Get details of a specific alert correlation rule.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
create=extend_schema(
|
||||
summary="Create alert correlation rule",
|
||||
description="Create a new alert correlation rule.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
update=extend_schema(
|
||||
summary="Update alert correlation rule",
|
||||
description="Update an existing alert correlation rule.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
partial_update=extend_schema(
|
||||
summary="Partial update alert correlation rule",
|
||||
description="Partially update an existing alert correlation rule.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
destroy=extend_schema(
|
||||
summary="Delete alert correlation rule",
|
||||
description="Delete an alert correlation rule.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
)
|
||||
class AlertCorrelationViewSet(viewsets.ModelViewSet):
|
||||
"""
|
||||
ViewSet for managing alert correlation rules.
|
||||
|
||||
Provides CRUD operations for configuring how alerts are correlated.
|
||||
This serves as the alert_correlations_view endpoint.
|
||||
"""
|
||||
|
||||
queryset = AlertCorrelationRule.objects.all()
|
||||
serializer_class = AlertCorrelationRuleSerializer
|
||||
permission_classes = [IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["is_active", "incident_severity"]
|
||||
search_fields = ["rule_name", "rule_description"]
|
||||
ordering_fields = ["rule_name", "created_at"]
|
||||
ordering = ["rule_name"]
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
list=extend_schema(
|
||||
summary="List cleanup job logs",
|
||||
description="Get all cleanup job logs with optional filtering by status.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
retrieve=extend_schema(
|
||||
summary="Get cleanup job log",
|
||||
description="Get details of a specific cleanup job log entry.",
|
||||
tags=["Admin - Observability"],
|
||||
),
|
||||
)
|
||||
class CleanupJobLogViewSet(viewsets.ReadOnlyModelViewSet):
|
||||
"""
|
||||
ViewSet for viewing cleanup job logs.
|
||||
|
||||
Provides read-only access to cleanup job execution history.
|
||||
"""
|
||||
|
||||
queryset = CleanupJobLog.objects.all()
|
||||
serializer_class = CleanupJobLogSerializer
|
||||
permission_classes = [IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["status", "job_name"]
|
||||
search_fields = ["job_name", "error_message"]
|
||||
ordering_fields = ["executed_at", "duration_ms"]
|
||||
ordering = ["-executed_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
|
||||
# Date range filtering
|
||||
start_date = self.request.query_params.get("start_date")
|
||||
end_date = self.request.query_params.get("end_date")
|
||||
|
||||
if start_date:
|
||||
queryset = queryset.filter(executed_at__gte=start_date)
|
||||
if end_date:
|
||||
queryset = queryset.filter(executed_at__lte=end_date)
|
||||
|
||||
return queryset
|
||||
|
||||
|
||||
@extend_schema(
|
||||
summary="Get data retention stats",
|
||||
description="Get aggregated data retention statistics for monitoring database growth.",
|
||||
tags=["Admin - Observability"],
|
||||
responses={200: DataRetentionStatsSerializer(many=True)},
|
||||
)
|
||||
class DataRetentionStatsView(APIView):
|
||||
"""
|
||||
API view for data retention statistics.
|
||||
|
||||
Returns aggregated statistics about table sizes, record counts,
|
||||
and data age for monitoring data retention and growth.
|
||||
"""
|
||||
|
||||
permission_classes = [IsAdminUser]
|
||||
|
||||
def get(self, request):
|
||||
"""Get data retention statistics for key tables."""
|
||||
from datetime import timedelta
|
||||
|
||||
from django.apps import apps
|
||||
|
||||
now = timezone.now()
|
||||
seven_days_ago = now - timedelta(days=7)
|
||||
thirty_days_ago = now - timedelta(days=30)
|
||||
|
||||
# Tables to report on
|
||||
tables_to_check = [
|
||||
("core", "pipelineerror", "occurred_at"),
|
||||
("core", "applicationerror", "created_at"),
|
||||
("core", "systemalert", "created_at"),
|
||||
("core", "requestmetadata", "created_at"),
|
||||
("core", "anomaly", "detected_at"),
|
||||
("core", "cleanupjoblog", "executed_at"),
|
||||
("moderation", "editsubmission", "created_at"),
|
||||
("moderation", "moderationauditlog", "created_at"),
|
||||
("notifications", "notificationlog", "created_at"),
|
||||
]
|
||||
|
||||
stats = []
|
||||
for app_label, model_name, date_field in tables_to_check:
|
||||
try:
|
||||
model = apps.get_model(app_label, model_name)
|
||||
filter_kwargs_7d = {f"{date_field}__gte": seven_days_ago}
|
||||
filter_kwargs_30d = {f"{date_field}__gte": thirty_days_ago}
|
||||
|
||||
# Get record counts and date ranges
|
||||
qs = model.objects.aggregate(
|
||||
total=Coalesce(Count("id"), 0),
|
||||
last_7_days=Coalesce(Count("id", filter=model.objects.filter(**filter_kwargs_7d).query.where), 0),
|
||||
last_30_days=Coalesce(Count("id", filter=model.objects.filter(**filter_kwargs_30d).query.where), 0),
|
||||
oldest_record=Min(date_field),
|
||||
newest_record=Max(date_field),
|
||||
)
|
||||
|
||||
# Get table size from database
|
||||
table_name = model._meta.db_table
|
||||
with connection.cursor() as cursor:
|
||||
cursor.execute(
|
||||
"""
|
||||
SELECT pg_size_pretty(pg_total_relation_size(%s))
|
||||
""",
|
||||
[table_name],
|
||||
)
|
||||
result = cursor.fetchone()
|
||||
table_size = result[0] if result else "Unknown"
|
||||
|
||||
stats.append(
|
||||
{
|
||||
"table_name": table_name,
|
||||
"total_records": model.objects.count(),
|
||||
"last_7_days": model.objects.filter(**filter_kwargs_7d).count(),
|
||||
"last_30_days": model.objects.filter(**filter_kwargs_30d).count(),
|
||||
"oldest_record": qs.get("oldest_record"),
|
||||
"newest_record": qs.get("newest_record"),
|
||||
"table_size": table_size,
|
||||
}
|
||||
)
|
||||
except Exception:
|
||||
# Skip tables that don't exist or have errors
|
||||
continue
|
||||
|
||||
serializer = DataRetentionStatsSerializer(stats, many=True)
|
||||
return Response(serializer.data)
|
||||
@@ -15,7 +15,7 @@ Key Components:
|
||||
from .base import ChoiceCategory, ChoiceGroup, RichChoice
|
||||
from .fields import RichChoiceField
|
||||
from .registry import ChoiceRegistry, register_choices
|
||||
from .serializers import RichChoiceOptionSerializer, RichChoiceSerializer
|
||||
from .serializers import RichChoiceOptionSerializer, RichChoiceSerializer, RichChoiceSerializerField
|
||||
from .utils import get_choice_display, validate_choice_value
|
||||
|
||||
__all__ = [
|
||||
@@ -26,6 +26,7 @@ __all__ = [
|
||||
"register_choices",
|
||||
"RichChoiceField",
|
||||
"RichChoiceSerializer",
|
||||
"RichChoiceSerializerField",
|
||||
"RichChoiceOptionSerializer",
|
||||
"validate_choice_value",
|
||||
"get_choice_display",
|
||||
|
||||
@@ -2,7 +2,8 @@
|
||||
Core System Rich Choice Objects
|
||||
|
||||
This module defines all choice objects for core system functionality,
|
||||
including health checks, API statuses, and other system-level choices.
|
||||
including health checks, API statuses, severity levels, alert types,
|
||||
and other system-level choices.
|
||||
"""
|
||||
|
||||
from .base import ChoiceCategory, RichChoice
|
||||
@@ -124,6 +125,584 @@ ENTITY_TYPES = [
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Severity Levels (used by ApplicationError, SystemAlert, Incident, RequestMetadata)
|
||||
# ============================================================================
|
||||
SEVERITY_LEVELS = [
|
||||
RichChoice(
|
||||
value="critical",
|
||||
label="Critical",
|
||||
description="Critical issue requiring immediate attention",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "alert-octagon",
|
||||
"css_class": "bg-red-100 text-red-800 border-red-300",
|
||||
"sort_order": 1,
|
||||
"priority": 1,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="high",
|
||||
label="High",
|
||||
description="High priority issue",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-orange-100 text-orange-800 border-orange-300",
|
||||
"sort_order": 2,
|
||||
"priority": 2,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="medium",
|
||||
label="Medium",
|
||||
description="Medium priority issue",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "info",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-300",
|
||||
"sort_order": 3,
|
||||
"priority": 3,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="low",
|
||||
label="Low",
|
||||
description="Low priority issue",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "info",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-300",
|
||||
"sort_order": 4,
|
||||
"priority": 4,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
]
|
||||
|
||||
# Extended severity levels including debug/info/warning/error for RequestMetadata
|
||||
REQUEST_SEVERITY_LEVELS = [
|
||||
RichChoice(
|
||||
value="debug",
|
||||
label="Debug",
|
||||
description="Debug-level information",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "bug",
|
||||
"css_class": "bg-gray-100 text-gray-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="info",
|
||||
label="Info",
|
||||
description="Informational message",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "info",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="warning",
|
||||
label="Warning",
|
||||
description="Warning condition",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="error",
|
||||
label="Error",
|
||||
description="Error condition",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="critical",
|
||||
label="Critical",
|
||||
description="Critical error requiring immediate attention",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "alert-octagon",
|
||||
"css_class": "bg-red-200 text-red-900 font-bold",
|
||||
"sort_order": 5,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Error/Request Sources
|
||||
# ============================================================================
|
||||
ERROR_SOURCES = [
|
||||
RichChoice(
|
||||
value="frontend",
|
||||
label="Frontend",
|
||||
description="Error originated from frontend application",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "monitor",
|
||||
"css_class": "bg-purple-100 text-purple-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="backend",
|
||||
label="Backend",
|
||||
description="Error originated from backend server",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "server",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="api",
|
||||
label="API",
|
||||
description="Error originated from API layer",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "code",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# System Alert Types
|
||||
# ============================================================================
|
||||
SYSTEM_ALERT_TYPES = [
|
||||
RichChoice(
|
||||
value="orphaned_images",
|
||||
label="Orphaned Images",
|
||||
description="Images not associated with any entity",
|
||||
metadata={"color": "orange", "icon": "image", "sort_order": 1},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="stale_submissions",
|
||||
label="Stale Submissions",
|
||||
description="Submissions pending for too long",
|
||||
metadata={"color": "yellow", "icon": "clock", "sort_order": 2},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="circular_dependency",
|
||||
label="Circular Dependency",
|
||||
description="Detected circular reference in data",
|
||||
metadata={"color": "red", "icon": "refresh-cw", "sort_order": 3},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="validation_error",
|
||||
label="Validation Error",
|
||||
description="Data validation failure",
|
||||
metadata={"color": "red", "icon": "alert-circle", "sort_order": 4},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="ban_attempt",
|
||||
label="Ban Attempt",
|
||||
description="User ban action was triggered",
|
||||
metadata={"color": "red", "icon": "shield-off", "sort_order": 5},
|
||||
category=ChoiceCategory.SECURITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="upload_timeout",
|
||||
label="Upload Timeout",
|
||||
description="File upload exceeded time limit",
|
||||
metadata={"color": "orange", "icon": "upload-cloud", "sort_order": 6},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="high_error_rate",
|
||||
label="High Error Rate",
|
||||
description="Elevated error rate detected",
|
||||
metadata={"color": "red", "icon": "trending-up", "sort_order": 7},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="database_connection",
|
||||
label="Database Connection",
|
||||
description="Database connectivity issue",
|
||||
metadata={"color": "red", "icon": "database", "sort_order": 8},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="memory_usage",
|
||||
label="Memory Usage",
|
||||
description="High memory consumption detected",
|
||||
metadata={"color": "orange", "icon": "cpu", "sort_order": 9},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="queue_backup",
|
||||
label="Queue Backup",
|
||||
description="Task queue is backing up",
|
||||
metadata={"color": "yellow", "icon": "layers", "sort_order": 10},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Metric Types for Rate Limiting
|
||||
# ============================================================================
|
||||
METRIC_TYPES = [
|
||||
RichChoice(
|
||||
value="block_rate",
|
||||
label="Block Rate",
|
||||
description="Percentage of requests being blocked",
|
||||
metadata={"color": "red", "icon": "shield", "sort_order": 1},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="total_requests",
|
||||
label="Total Requests",
|
||||
description="Total number of requests",
|
||||
metadata={"color": "blue", "icon": "activity", "sort_order": 2},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="unique_ips",
|
||||
label="Unique IPs",
|
||||
description="Number of unique IP addresses",
|
||||
metadata={"color": "purple", "icon": "globe", "sort_order": 3},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="function_specific",
|
||||
label="Function Specific",
|
||||
description="Metrics for a specific function",
|
||||
metadata={"color": "green", "icon": "code", "sort_order": 4},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Incident Statuses
|
||||
# ============================================================================
|
||||
INCIDENT_STATUSES = [
|
||||
RichChoice(
|
||||
value="open",
|
||||
label="Open",
|
||||
description="Incident is open and awaiting investigation",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "alert-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 1,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="investigating",
|
||||
label="Investigating",
|
||||
description="Incident is being actively investigated",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "search",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 2,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="resolved",
|
||||
label="Resolved",
|
||||
description="Incident has been resolved",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 3,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="closed",
|
||||
label="Closed",
|
||||
description="Incident is closed",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-gray-100 text-gray-800",
|
||||
"sort_order": 4,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Alert Sources
|
||||
# ============================================================================
|
||||
ALERT_SOURCES = [
|
||||
RichChoice(
|
||||
value="system",
|
||||
label="System Alert",
|
||||
description="Alert from system monitoring",
|
||||
metadata={"color": "blue", "icon": "server", "sort_order": 1},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="rate_limit",
|
||||
label="Rate Limit Alert",
|
||||
description="Alert from rate limiting system",
|
||||
metadata={"color": "orange", "icon": "shield", "sort_order": 2},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Pipeline Error Severities
|
||||
# ============================================================================
|
||||
PIPELINE_ERROR_SEVERITIES = [
|
||||
RichChoice(
|
||||
value="critical",
|
||||
label="Critical",
|
||||
description="Critical pipeline failure requiring immediate attention",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "alert-octagon",
|
||||
"css_class": "bg-red-100 text-red-800 border-red-300",
|
||||
"sort_order": 1,
|
||||
"priority": 1,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="error",
|
||||
label="Error",
|
||||
description="Pipeline error that needs investigation",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-orange-100 text-orange-800 border-orange-300",
|
||||
"sort_order": 2,
|
||||
"priority": 2,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="warning",
|
||||
label="Warning",
|
||||
description="Pipeline warning that may need attention",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "alert-circle",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-300",
|
||||
"sort_order": 3,
|
||||
"priority": 3,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="info",
|
||||
label="Info",
|
||||
description="Informational pipeline event",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "info",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-300",
|
||||
"sort_order": 4,
|
||||
"priority": 4,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Anomaly Types
|
||||
# ============================================================================
|
||||
ANOMALY_TYPES = [
|
||||
RichChoice(
|
||||
value="spike",
|
||||
label="Spike",
|
||||
description="Sudden increase in metric value",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "trending-up",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="drop",
|
||||
label="Drop",
|
||||
description="Sudden decrease in metric value",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "trending-down",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="trend_change",
|
||||
label="Trend Change",
|
||||
description="Change in the overall trend direction",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "activity",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="outlier",
|
||||
label="Outlier",
|
||||
description="Value outside normal distribution",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "git-branch",
|
||||
"css_class": "bg-purple-100 text-purple-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
RichChoice(
|
||||
value="threshold_breach",
|
||||
label="Threshold Breach",
|
||||
description="Value exceeded configured threshold",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-orange-100 text-orange-800",
|
||||
"sort_order": 5,
|
||||
},
|
||||
category=ChoiceCategory.TECHNICAL,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Cleanup Job Statuses
|
||||
# ============================================================================
|
||||
CLEANUP_JOB_STATUSES = [
|
||||
RichChoice(
|
||||
value="success",
|
||||
label="Success",
|
||||
description="Cleanup job completed successfully",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="failed",
|
||||
label="Failed",
|
||||
description="Cleanup job failed with errors",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="partial",
|
||||
label="Partial",
|
||||
description="Cleanup job completed with some failures",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "alert-circle",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="skipped",
|
||||
label="Skipped",
|
||||
description="Cleanup job was skipped",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "skip-forward",
|
||||
"css_class": "bg-gray-100 text-gray-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Date Precision (shared across multiple domains)
|
||||
# ============================================================================
|
||||
DATE_PRECISION = [
|
||||
RichChoice(
|
||||
value="exact",
|
||||
label="Exact Date",
|
||||
description="Date is known exactly",
|
||||
metadata={"color": "green", "icon": "calendar", "sort_order": 1, "format": "YYYY-MM-DD"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="month",
|
||||
label="Month and Year",
|
||||
description="Only month and year are known",
|
||||
metadata={"color": "blue", "icon": "calendar", "sort_order": 2, "format": "YYYY-MM"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="year",
|
||||
label="Year Only",
|
||||
description="Only the year is known",
|
||||
metadata={"color": "yellow", "icon": "calendar", "sort_order": 3, "format": "YYYY"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="decade",
|
||||
label="Decade",
|
||||
description="Only the decade is known",
|
||||
metadata={"color": "orange", "icon": "calendar", "sort_order": 4, "format": "YYYYs"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="century",
|
||||
label="Century",
|
||||
description="Only the century is known",
|
||||
metadata={"color": "gray", "icon": "calendar", "sort_order": 5, "format": "YYc"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="approximate",
|
||||
label="Approximate",
|
||||
description="Date is approximate/estimated",
|
||||
metadata={"color": "gray", "icon": "help-circle", "sort_order": 6, "format": "~YYYY"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
def register_core_choices():
|
||||
"""Register all core system choices with the global registry"""
|
||||
@@ -152,6 +731,95 @@ def register_core_choices():
|
||||
metadata={"domain": "core", "type": "entity_type"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="severity_levels",
|
||||
choices=SEVERITY_LEVELS,
|
||||
domain="core",
|
||||
description="Severity levels for errors and alerts",
|
||||
metadata={"domain": "core", "type": "severity"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="request_severity_levels",
|
||||
choices=REQUEST_SEVERITY_LEVELS,
|
||||
domain="core",
|
||||
description="Extended severity levels for request metadata",
|
||||
metadata={"domain": "core", "type": "request_severity"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="error_sources",
|
||||
choices=ERROR_SOURCES,
|
||||
domain="core",
|
||||
description="Sources of application errors",
|
||||
metadata={"domain": "core", "type": "error_source"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="system_alert_types",
|
||||
choices=SYSTEM_ALERT_TYPES,
|
||||
domain="core",
|
||||
description="Types of system alerts",
|
||||
metadata={"domain": "core", "type": "alert_type"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="metric_types",
|
||||
choices=METRIC_TYPES,
|
||||
domain="core",
|
||||
description="Types of rate limit metrics",
|
||||
metadata={"domain": "core", "type": "metric_type"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="incident_statuses",
|
||||
choices=INCIDENT_STATUSES,
|
||||
domain="core",
|
||||
description="Incident status options",
|
||||
metadata={"domain": "core", "type": "incident_status"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="alert_sources",
|
||||
choices=ALERT_SOURCES,
|
||||
domain="core",
|
||||
description="Sources of alerts",
|
||||
metadata={"domain": "core", "type": "alert_source"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="pipeline_error_severities",
|
||||
choices=PIPELINE_ERROR_SEVERITIES,
|
||||
domain="core",
|
||||
description="Severity levels for pipeline errors",
|
||||
metadata={"domain": "core", "type": "pipeline_error_severity"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="anomaly_types",
|
||||
choices=ANOMALY_TYPES,
|
||||
domain="core",
|
||||
description="Types of detected anomalies",
|
||||
metadata={"domain": "core", "type": "anomaly_type"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="cleanup_job_statuses",
|
||||
choices=CLEANUP_JOB_STATUSES,
|
||||
domain="core",
|
||||
description="Status options for cleanup jobs",
|
||||
metadata={"domain": "core", "type": "cleanup_job_status"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="date_precision",
|
||||
choices=DATE_PRECISION,
|
||||
domain="core",
|
||||
description="Date precision options",
|
||||
metadata={"domain": "core", "type": "date_precision"},
|
||||
)
|
||||
|
||||
|
||||
# Auto-register choices when module is imported
|
||||
register_core_choices()
|
||||
|
||||
|
||||
133
backend/apps/core/choices/filters.py
Normal file
133
backend/apps/core/choices/filters.py
Normal file
@@ -0,0 +1,133 @@
|
||||
"""
|
||||
Django-filter Integration for Rich Choices
|
||||
|
||||
This module provides django-filter compatible filter classes that integrate
|
||||
with the RichChoice registry system.
|
||||
"""
|
||||
|
||||
from typing import Any
|
||||
|
||||
from django_filters import ChoiceFilter, MultipleChoiceFilter
|
||||
|
||||
from .registry import registry
|
||||
|
||||
|
||||
class RichChoiceFilter(ChoiceFilter):
|
||||
"""
|
||||
Django-filter ChoiceFilter that uses the RichChoice registry.
|
||||
|
||||
This is the REQUIRED replacement for ChoiceFilter with inline choices.
|
||||
|
||||
Usage:
|
||||
class MyFilterSet(django_filters.FilterSet):
|
||||
status = RichChoiceFilter(
|
||||
choice_group="ticket_statuses",
|
||||
domain="support",
|
||||
)
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
choice_group: str,
|
||||
domain: str = "core",
|
||||
allow_deprecated: bool = False,
|
||||
**kwargs
|
||||
):
|
||||
"""
|
||||
Initialize the filter.
|
||||
|
||||
Args:
|
||||
choice_group: Name of the choice group in the registry
|
||||
domain: Domain namespace for the choice group
|
||||
allow_deprecated: Whether to include deprecated choices
|
||||
**kwargs: Additional arguments passed to ChoiceFilter
|
||||
"""
|
||||
self.choice_group = choice_group
|
||||
self.domain = domain
|
||||
self.allow_deprecated = allow_deprecated
|
||||
|
||||
# Get choices from registry
|
||||
if allow_deprecated:
|
||||
choices_list = registry.get_choices(choice_group, domain)
|
||||
else:
|
||||
choices_list = registry.get_active_choices(choice_group, domain)
|
||||
|
||||
choices = [(c.value, c.label) for c in choices_list]
|
||||
|
||||
super().__init__(choices=choices, **kwargs)
|
||||
|
||||
|
||||
class RichMultipleChoiceFilter(MultipleChoiceFilter):
|
||||
"""
|
||||
Django-filter MultipleChoiceFilter that uses the RichChoice registry.
|
||||
|
||||
This is the REQUIRED replacement for MultipleChoiceFilter with inline choices.
|
||||
|
||||
Usage:
|
||||
class MyFilterSet(django_filters.FilterSet):
|
||||
statuses = RichMultipleChoiceFilter(
|
||||
choice_group="ticket_statuses",
|
||||
domain="support",
|
||||
field_name="status",
|
||||
)
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
choice_group: str,
|
||||
domain: str = "core",
|
||||
allow_deprecated: bool = False,
|
||||
**kwargs
|
||||
):
|
||||
"""
|
||||
Initialize the filter.
|
||||
|
||||
Args:
|
||||
choice_group: Name of the choice group in the registry
|
||||
domain: Domain namespace for the choice group
|
||||
allow_deprecated: Whether to include deprecated choices
|
||||
**kwargs: Additional arguments passed to MultipleChoiceFilter
|
||||
"""
|
||||
self.choice_group = choice_group
|
||||
self.domain = domain
|
||||
self.allow_deprecated = allow_deprecated
|
||||
|
||||
# Get choices from registry
|
||||
if allow_deprecated:
|
||||
choices_list = registry.get_choices(choice_group, domain)
|
||||
else:
|
||||
choices_list = registry.get_active_choices(choice_group, domain)
|
||||
|
||||
choices = [(c.value, c.label) for c in choices_list]
|
||||
|
||||
super().__init__(choices=choices, **kwargs)
|
||||
|
||||
|
||||
def get_choice_filter_class(
|
||||
choice_group: str,
|
||||
domain: str = "core",
|
||||
allow_deprecated: bool = False,
|
||||
**extra_kwargs: Any
|
||||
) -> type[RichChoiceFilter]:
|
||||
"""
|
||||
Factory function to create a RichChoiceFilter class with preset choices.
|
||||
|
||||
Useful when you need to define the filter class dynamically or
|
||||
when the choice_group/domain aren't available at class definition time.
|
||||
|
||||
Usage:
|
||||
StatusFilter = get_choice_filter_class("ticket_statuses", "support")
|
||||
|
||||
class MyFilterSet(django_filters.FilterSet):
|
||||
status = StatusFilter()
|
||||
"""
|
||||
class DynamicRichChoiceFilter(RichChoiceFilter):
|
||||
def __init__(self, **kwargs):
|
||||
kwargs.setdefault("choice_group", choice_group)
|
||||
kwargs.setdefault("domain", domain)
|
||||
kwargs.setdefault("allow_deprecated", allow_deprecated)
|
||||
for key, value in extra_kwargs.items():
|
||||
kwargs.setdefault(key, value)
|
||||
super().__init__(**kwargs)
|
||||
|
||||
return DynamicRichChoiceFilter
|
||||
@@ -265,3 +265,98 @@ def serialize_choice_value(value: str, choice_group: str, domain: str = "core",
|
||||
}
|
||||
else:
|
||||
return value
|
||||
|
||||
|
||||
class RichChoiceSerializerField(serializers.ChoiceField):
|
||||
"""
|
||||
DRF serializer field for RichChoice values.
|
||||
|
||||
This field validates input against the RichChoice registry and provides
|
||||
type-safe choice handling with proper error messages. It is the REQUIRED
|
||||
replacement for serializers.ChoiceField with inline choices.
|
||||
|
||||
Usage:
|
||||
class MySerializer(serializers.Serializer):
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="ticket_statuses",
|
||||
domain="support",
|
||||
)
|
||||
|
||||
# With rich metadata in output
|
||||
severity = RichChoiceSerializerField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
include_metadata=True,
|
||||
)
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
choice_group: str,
|
||||
domain: str = "core",
|
||||
include_metadata: bool = False,
|
||||
allow_deprecated: bool = False,
|
||||
**kwargs
|
||||
):
|
||||
"""
|
||||
Initialize the serializer field.
|
||||
|
||||
Args:
|
||||
choice_group: Name of the choice group in the registry
|
||||
domain: Domain namespace for the choice group
|
||||
include_metadata: Whether to include rich choice metadata in output
|
||||
allow_deprecated: Whether to allow deprecated choices
|
||||
**kwargs: Additional arguments passed to ChoiceField
|
||||
"""
|
||||
self.choice_group = choice_group
|
||||
self.domain = domain
|
||||
self.include_metadata = include_metadata
|
||||
self.allow_deprecated = allow_deprecated
|
||||
|
||||
# Get choices from registry for validation
|
||||
if allow_deprecated:
|
||||
choices_list = registry.get_choices(choice_group, domain)
|
||||
else:
|
||||
choices_list = registry.get_active_choices(choice_group, domain)
|
||||
|
||||
# Build choices tuple for DRF ChoiceField
|
||||
choices = [(c.value, c.label) for c in choices_list]
|
||||
|
||||
# Store valid values for error messages
|
||||
self._valid_values = [c.value for c in choices_list]
|
||||
|
||||
super().__init__(choices=choices, **kwargs)
|
||||
|
||||
def to_representation(self, value: str) -> Any:
|
||||
"""Convert choice value to representation."""
|
||||
if not value:
|
||||
return value
|
||||
|
||||
if self.include_metadata:
|
||||
return serialize_choice_value(
|
||||
value,
|
||||
self.choice_group,
|
||||
self.domain,
|
||||
include_metadata=True
|
||||
)
|
||||
return value
|
||||
|
||||
def to_internal_value(self, data: Any) -> str:
|
||||
"""Convert input data to choice value."""
|
||||
# Handle rich choice object input (value dict)
|
||||
if isinstance(data, dict) and "value" in data:
|
||||
data = data["value"]
|
||||
|
||||
# Validate and return
|
||||
return super().to_internal_value(data)
|
||||
|
||||
def fail(self, key: str, **kwargs: Any) -> None:
|
||||
"""Provide better error messages with valid choices listed."""
|
||||
if key == "invalid_choice":
|
||||
valid_choices = ", ".join(self._valid_values)
|
||||
raise serializers.ValidationError(
|
||||
f"'{kwargs.get('input', '')}' is not a valid choice for {self.choice_group}. "
|
||||
f"Valid choices are: {valid_choices}"
|
||||
)
|
||||
super().fail(key, **kwargs)
|
||||
|
||||
|
||||
@@ -39,15 +39,30 @@ class AuthRateLimitMiddleware:
|
||||
# Login endpoints
|
||||
"/api/v1/auth/login/": {"per_minute": 5, "per_hour": 30, "per_day": 100},
|
||||
"/accounts/login/": {"per_minute": 5, "per_hour": 30, "per_day": 100},
|
||||
# MFA verification (strict limits - 6-digit codes have limited entropy)
|
||||
"/api/v1/auth/login/mfa-verify/": {"per_minute": 5, "per_hour": 15, "per_day": 50},
|
||||
"/api/v1/auth/mfa/totp/verify/": {"per_minute": 5, "per_hour": 15, "per_day": 50},
|
||||
"/api/v1/auth/mfa/totp/activate/": {"per_minute": 3, "per_hour": 10, "per_day": 30},
|
||||
"/api/v1/auth/mfa/totp/deactivate/": {"per_minute": 3, "per_hour": 10, "per_day": 20},
|
||||
# Passkey endpoints
|
||||
"/api/v1/auth/passkey/authenticate/": {"per_minute": 10, "per_hour": 30, "per_day": 100},
|
||||
"/api/v1/auth/passkey/register/": {"per_minute": 5, "per_hour": 15, "per_day": 30},
|
||||
# Signup endpoints
|
||||
"/api/v1/auth/signup/": {"per_minute": 3, "per_hour": 10, "per_day": 20},
|
||||
"/accounts/signup/": {"per_minute": 3, "per_hour": 10, "per_day": 20},
|
||||
# Password reset endpoints
|
||||
"/api/v1/auth/password-reset/": {"per_minute": 2, "per_hour": 5, "per_day": 10},
|
||||
"/accounts/password/reset/": {"per_minute": 2, "per_hour": 5, "per_day": 10},
|
||||
# Password change (prevent brute force on current password)
|
||||
"/api/v1/auth/password/change/": {"per_minute": 3, "per_hour": 10, "per_day": 30},
|
||||
# Token endpoints
|
||||
"/api/v1/auth/token/": {"per_minute": 10, "per_hour": 60, "per_day": 200},
|
||||
"/api/v1/auth/token/refresh/": {"per_minute": 20, "per_hour": 120, "per_day": 500},
|
||||
# Social account management
|
||||
"/api/v1/auth/social/connect/google/": {"per_minute": 5, "per_hour": 15, "per_day": 30},
|
||||
"/api/v1/auth/social/connect/discord/": {"per_minute": 5, "per_hour": 15, "per_day": 30},
|
||||
"/api/v1/auth/social/disconnect/google/": {"per_minute": 5, "per_hour": 15, "per_day": 20},
|
||||
"/api/v1/auth/social/disconnect/discord/": {"per_minute": 5, "per_hour": 15, "per_day": 20},
|
||||
}
|
||||
|
||||
def __init__(self, get_response: Callable[[HttpRequest], HttpResponse]):
|
||||
|
||||
@@ -0,0 +1,69 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-11 00:48
|
||||
|
||||
import apps.core.choices.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('core', '0010_add_milestone_model'),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name='applicationerror',
|
||||
name='severity',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='severity_levels', choices=[('critical', 'Critical'), ('high', 'High'), ('medium', 'Medium'), ('low', 'Low')], db_index=True, default='medium', domain='core', help_text='Error severity level', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='applicationerror',
|
||||
name='source',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='error_sources', choices=[('frontend', 'Frontend'), ('backend', 'Backend'), ('api', 'API')], db_index=True, domain='core', help_text='Where the error originated', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='incident',
|
||||
name='severity',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='severity_levels', choices=[('critical', 'Critical'), ('high', 'High'), ('medium', 'Medium'), ('low', 'Low')], db_index=True, domain='core', help_text='Incident severity level', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='incident',
|
||||
name='status',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='incident_statuses', choices=[('open', 'Open'), ('investigating', 'Investigating'), ('resolved', 'Resolved'), ('closed', 'Closed')], db_index=True, default='open', domain='core', help_text='Current incident status', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='incidentalert',
|
||||
name='alert_source',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='alert_sources', choices=[('system', 'System Alert'), ('rate_limit', 'Rate Limit Alert')], domain='core', help_text='Source type of the alert', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='milestone',
|
||||
name='event_date_precision',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='date_precision', choices=[('exact', 'Exact Date'), ('month', 'Month and Year'), ('year', 'Year Only'), ('decade', 'Decade'), ('century', 'Century'), ('approximate', 'Approximate')], default='exact', domain='core', help_text='Precision of the event date', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='milestoneevent',
|
||||
name='event_date_precision',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='date_precision', choices=[('exact', 'Exact Date'), ('month', 'Month and Year'), ('year', 'Year Only'), ('decade', 'Decade'), ('century', 'Century'), ('approximate', 'Approximate')], default='exact', domain='core', help_text='Precision of the event date', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='ratelimitalertconfig',
|
||||
name='metric_type',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='metric_types', choices=[('block_rate', 'Block Rate'), ('total_requests', 'Total Requests'), ('unique_ips', 'Unique IPs'), ('function_specific', 'Function Specific')], db_index=True, domain='core', help_text='Type of metric to monitor', max_length=50),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='requestmetadata',
|
||||
name='severity',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='request_severity_levels', choices=[('debug', 'Debug'), ('info', 'Info'), ('warning', 'Warning'), ('error', 'Error'), ('critical', 'Critical')], db_index=True, default='info', domain='core', help_text='Error severity level', max_length=20),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='systemalert',
|
||||
name='alert_type',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='system_alert_types', choices=[('orphaned_images', 'Orphaned Images'), ('stale_submissions', 'Stale Submissions'), ('circular_dependency', 'Circular Dependency'), ('validation_error', 'Validation Error'), ('ban_attempt', 'Ban Attempt'), ('upload_timeout', 'Upload Timeout'), ('high_error_rate', 'High Error Rate'), ('database_connection', 'Database Connection'), ('memory_usage', 'Memory Usage'), ('queue_backup', 'Queue Backup')], db_index=True, domain='core', help_text='Type of system alert', max_length=50),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='systemalert',
|
||||
name='severity',
|
||||
field=apps.core.choices.fields.RichChoiceField(allow_deprecated=False, choice_group='severity_levels', choices=[('critical', 'Critical'), ('high', 'High'), ('medium', 'Medium'), ('low', 'Low')], db_index=True, domain='core', help_text='Alert severity level', max_length=20),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,320 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-11 18:06
|
||||
|
||||
import apps.core.choices.fields
|
||||
import django.db.models.deletion
|
||||
import uuid
|
||||
from django.conf import settings
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("core", "0011_alter_applicationerror_severity_and_more"),
|
||||
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.CreateModel(
|
||||
name="AlertCorrelationRule",
|
||||
fields=[
|
||||
("id", models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)),
|
||||
(
|
||||
"rule_name",
|
||||
models.CharField(
|
||||
db_index=True, help_text="Unique name for this correlation rule", max_length=255, unique=True
|
||||
),
|
||||
),
|
||||
(
|
||||
"rule_description",
|
||||
models.TextField(blank=True, help_text="Description of what this rule correlates"),
|
||||
),
|
||||
(
|
||||
"min_alerts_required",
|
||||
models.PositiveIntegerField(
|
||||
default=3, help_text="Minimum number of alerts needed to trigger correlation"
|
||||
),
|
||||
),
|
||||
(
|
||||
"time_window_minutes",
|
||||
models.PositiveIntegerField(default=30, help_text="Time window in minutes for alert correlation"),
|
||||
),
|
||||
(
|
||||
"incident_severity",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="severity_levels",
|
||||
choices=[("critical", "Critical"), ("high", "High"), ("medium", "Medium"), ("low", "Low")],
|
||||
default="medium",
|
||||
domain="core",
|
||||
help_text="Severity to assign to correlated incidents",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
(
|
||||
"incident_title_template",
|
||||
models.CharField(
|
||||
help_text="Template for incident title (supports {count}, {rule_name})", max_length=255
|
||||
),
|
||||
),
|
||||
(
|
||||
"is_active",
|
||||
models.BooleanField(db_index=True, default=True, help_text="Whether this rule is currently active"),
|
||||
),
|
||||
("created_at", models.DateTimeField(auto_now_add=True, help_text="When this rule was created")),
|
||||
("updated_at", models.DateTimeField(auto_now=True, help_text="When this rule was last updated")),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Alert Correlation Rule",
|
||||
"verbose_name_plural": "Alert Correlation Rules",
|
||||
"ordering": ["rule_name"],
|
||||
},
|
||||
),
|
||||
migrations.CreateModel(
|
||||
name="CleanupJobLog",
|
||||
fields=[
|
||||
("id", models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)),
|
||||
("job_name", models.CharField(db_index=True, help_text="Name of the cleanup job", max_length=255)),
|
||||
(
|
||||
"status",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="cleanup_job_statuses",
|
||||
choices=[
|
||||
("success", "Success"),
|
||||
("failed", "Failed"),
|
||||
("partial", "Partial"),
|
||||
("skipped", "Skipped"),
|
||||
],
|
||||
db_index=True,
|
||||
default="success",
|
||||
domain="core",
|
||||
help_text="Execution status",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
("records_processed", models.PositiveIntegerField(default=0, help_text="Number of records processed")),
|
||||
("records_deleted", models.PositiveIntegerField(default=0, help_text="Number of records deleted")),
|
||||
("error_message", models.TextField(blank=True, help_text="Error message if job failed", null=True)),
|
||||
(
|
||||
"duration_ms",
|
||||
models.PositiveIntegerField(blank=True, help_text="Execution duration in milliseconds", null=True),
|
||||
),
|
||||
(
|
||||
"executed_at",
|
||||
models.DateTimeField(auto_now_add=True, db_index=True, help_text="When this job was executed"),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Cleanup Job Log",
|
||||
"verbose_name_plural": "Cleanup Job Logs",
|
||||
"ordering": ["-executed_at"],
|
||||
"indexes": [
|
||||
models.Index(fields=["job_name", "executed_at"], name="core_cleanu_job_nam_4530fd_idx"),
|
||||
models.Index(fields=["status", "executed_at"], name="core_cleanu_status_fa6360_idx"),
|
||||
],
|
||||
},
|
||||
),
|
||||
migrations.CreateModel(
|
||||
name="Anomaly",
|
||||
fields=[
|
||||
("id", models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)),
|
||||
(
|
||||
"metric_name",
|
||||
models.CharField(
|
||||
db_index=True, help_text="Name of the metric that exhibited anomalous behavior", max_length=255
|
||||
),
|
||||
),
|
||||
(
|
||||
"metric_category",
|
||||
models.CharField(
|
||||
db_index=True,
|
||||
help_text="Category of the metric (e.g., 'performance', 'error_rate', 'traffic')",
|
||||
max_length=100,
|
||||
),
|
||||
),
|
||||
(
|
||||
"anomaly_type",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="anomaly_types",
|
||||
choices=[
|
||||
("spike", "Spike"),
|
||||
("drop", "Drop"),
|
||||
("trend_change", "Trend Change"),
|
||||
("outlier", "Outlier"),
|
||||
("threshold_breach", "Threshold Breach"),
|
||||
],
|
||||
db_index=True,
|
||||
domain="core",
|
||||
help_text="Type of anomaly detected",
|
||||
max_length=30,
|
||||
),
|
||||
),
|
||||
(
|
||||
"severity",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="severity_levels",
|
||||
choices=[("critical", "Critical"), ("high", "High"), ("medium", "Medium"), ("low", "Low")],
|
||||
db_index=True,
|
||||
domain="core",
|
||||
help_text="Severity of the anomaly",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
(
|
||||
"anomaly_value",
|
||||
models.DecimalField(decimal_places=6, help_text="The anomalous value detected", max_digits=20),
|
||||
),
|
||||
(
|
||||
"baseline_value",
|
||||
models.DecimalField(decimal_places=6, help_text="The expected baseline value", max_digits=20),
|
||||
),
|
||||
(
|
||||
"deviation_score",
|
||||
models.DecimalField(decimal_places=4, help_text="Standard deviations from normal", max_digits=10),
|
||||
),
|
||||
(
|
||||
"confidence_score",
|
||||
models.DecimalField(
|
||||
decimal_places=4, help_text="Confidence score of the detection (0-1)", max_digits=5
|
||||
),
|
||||
),
|
||||
("detection_algorithm", models.CharField(help_text="Algorithm used for detection", max_length=100)),
|
||||
("time_window_start", models.DateTimeField(help_text="Start of the detection time window")),
|
||||
("time_window_end", models.DateTimeField(help_text="End of the detection time window")),
|
||||
(
|
||||
"alert_created",
|
||||
models.BooleanField(
|
||||
db_index=True, default=False, help_text="Whether an alert was created for this anomaly"
|
||||
),
|
||||
),
|
||||
(
|
||||
"detected_at",
|
||||
models.DateTimeField(auto_now_add=True, db_index=True, help_text="When this anomaly was detected"),
|
||||
),
|
||||
(
|
||||
"alert",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Linked system alert if created",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="anomalies",
|
||||
to="core.systemalert",
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Anomaly",
|
||||
"verbose_name_plural": "Anomalies",
|
||||
"ordering": ["-detected_at"],
|
||||
"indexes": [
|
||||
models.Index(fields=["metric_name", "detected_at"], name="core_anomal_metric__06c3c9_idx"),
|
||||
models.Index(fields=["severity", "detected_at"], name="core_anomal_severit_ea7a17_idx"),
|
||||
models.Index(fields=["anomaly_type", "detected_at"], name="core_anomal_anomaly_eb45f7_idx"),
|
||||
models.Index(fields=["alert_created", "detected_at"], name="core_anomal_alert_c_5a0c1a_idx"),
|
||||
],
|
||||
},
|
||||
),
|
||||
migrations.CreateModel(
|
||||
name="PipelineError",
|
||||
fields=[
|
||||
("id", models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)),
|
||||
(
|
||||
"function_name",
|
||||
models.CharField(
|
||||
db_index=True, help_text="Name of the function/pipeline that failed", max_length=255
|
||||
),
|
||||
),
|
||||
("error_message", models.TextField(help_text="Error message describing the failure")),
|
||||
(
|
||||
"error_code",
|
||||
models.CharField(
|
||||
blank=True, db_index=True, help_text="Error code for categorization", max_length=100, null=True
|
||||
),
|
||||
),
|
||||
("error_context", models.JSONField(blank=True, help_text="Additional context data as JSON", null=True)),
|
||||
("stack_trace", models.TextField(blank=True, help_text="Full stack trace for debugging", null=True)),
|
||||
(
|
||||
"severity",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="pipeline_error_severities",
|
||||
choices=[
|
||||
("critical", "Critical"),
|
||||
("error", "Error"),
|
||||
("warning", "Warning"),
|
||||
("info", "Info"),
|
||||
],
|
||||
db_index=True,
|
||||
default="error",
|
||||
domain="core",
|
||||
help_text="Severity level of the error",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
(
|
||||
"submission_id",
|
||||
models.UUIDField(
|
||||
blank=True, db_index=True, help_text="ID of related content submission if applicable", null=True
|
||||
),
|
||||
),
|
||||
(
|
||||
"item_id",
|
||||
models.CharField(
|
||||
blank=True,
|
||||
db_index=True,
|
||||
help_text="Generic reference to related item",
|
||||
max_length=255,
|
||||
null=True,
|
||||
),
|
||||
),
|
||||
(
|
||||
"request_id",
|
||||
models.UUIDField(blank=True, db_index=True, help_text="Request ID for correlation", null=True),
|
||||
),
|
||||
("trace_id", models.UUIDField(blank=True, db_index=True, help_text="Distributed trace ID", null=True)),
|
||||
(
|
||||
"resolved",
|
||||
models.BooleanField(db_index=True, default=False, help_text="Whether this error has been resolved"),
|
||||
),
|
||||
(
|
||||
"resolved_at",
|
||||
models.DateTimeField(
|
||||
blank=True, db_index=True, help_text="When this error was resolved", null=True
|
||||
),
|
||||
),
|
||||
(
|
||||
"resolution_notes",
|
||||
models.TextField(blank=True, help_text="Notes about how the error was resolved", null=True),
|
||||
),
|
||||
(
|
||||
"occurred_at",
|
||||
models.DateTimeField(auto_now_add=True, db_index=True, help_text="When this error occurred"),
|
||||
),
|
||||
(
|
||||
"resolved_by",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="User who resolved this error",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="resolved_pipeline_errors",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Pipeline Error",
|
||||
"verbose_name_plural": "Pipeline Errors",
|
||||
"ordering": ["-occurred_at"],
|
||||
"indexes": [
|
||||
models.Index(fields=["severity", "occurred_at"], name="core_pipeli_severit_9c8037_idx"),
|
||||
models.Index(fields=["function_name", "occurred_at"], name="core_pipeli_functio_efb015_idx"),
|
||||
models.Index(fields=["resolved", "occurred_at"], name="core_pipeli_resolve_cd60c5_idx"),
|
||||
],
|
||||
},
|
||||
),
|
||||
]
|
||||
@@ -8,8 +8,12 @@ from django.contrib.contenttypes.models import ContentType
|
||||
from django.db import models
|
||||
from django.utils.text import slugify
|
||||
|
||||
from apps.core.choices.fields import RichChoiceField
|
||||
from apps.core.history import TrackedModel
|
||||
|
||||
# Import choices module to ensure registration on app load
|
||||
from apps.core.choices import core_choices # noqa: F401
|
||||
|
||||
|
||||
@pghistory.track()
|
||||
class SlugHistory(models.Model):
|
||||
@@ -136,17 +140,6 @@ class ApplicationError(models.Model):
|
||||
reported via API (frontend) and displayed in the admin dashboard.
|
||||
"""
|
||||
|
||||
class Severity(models.TextChoices):
|
||||
CRITICAL = "critical", "Critical"
|
||||
HIGH = "high", "High"
|
||||
MEDIUM = "medium", "Medium"
|
||||
LOW = "low", "Low"
|
||||
|
||||
class Source(models.TextChoices):
|
||||
FRONTEND = "frontend", "Frontend"
|
||||
BACKEND = "backend", "Backend"
|
||||
API = "api", "API"
|
||||
|
||||
# Identity
|
||||
error_id = models.UUIDField(
|
||||
unique=True,
|
||||
@@ -180,16 +173,18 @@ class ApplicationError(models.Model):
|
||||
db_index=True,
|
||||
help_text="Application-specific error code",
|
||||
)
|
||||
severity = models.CharField(
|
||||
severity = RichChoiceField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Severity.choices,
|
||||
default=Severity.MEDIUM,
|
||||
default="medium",
|
||||
db_index=True,
|
||||
help_text="Error severity level",
|
||||
)
|
||||
source = models.CharField(
|
||||
source = RichChoiceField(
|
||||
choice_group="error_sources",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Source.choices,
|
||||
db_index=True,
|
||||
help_text="Where the error originated",
|
||||
)
|
||||
@@ -308,34 +303,18 @@ class SystemAlert(models.Model):
|
||||
validation errors, ban attempts, upload timeouts, and high error rates.
|
||||
"""
|
||||
|
||||
class AlertType(models.TextChoices):
|
||||
ORPHANED_IMAGES = "orphaned_images", "Orphaned Images"
|
||||
STALE_SUBMISSIONS = "stale_submissions", "Stale Submissions"
|
||||
CIRCULAR_DEPENDENCY = "circular_dependency", "Circular Dependency"
|
||||
VALIDATION_ERROR = "validation_error", "Validation Error"
|
||||
BAN_ATTEMPT = "ban_attempt", "Ban Attempt"
|
||||
UPLOAD_TIMEOUT = "upload_timeout", "Upload Timeout"
|
||||
HIGH_ERROR_RATE = "high_error_rate", "High Error Rate"
|
||||
DATABASE_CONNECTION = "database_connection", "Database Connection"
|
||||
MEMORY_USAGE = "memory_usage", "Memory Usage"
|
||||
QUEUE_BACKUP = "queue_backup", "Queue Backup"
|
||||
|
||||
class Severity(models.TextChoices):
|
||||
LOW = "low", "Low"
|
||||
MEDIUM = "medium", "Medium"
|
||||
HIGH = "high", "High"
|
||||
CRITICAL = "critical", "Critical"
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
alert_type = models.CharField(
|
||||
alert_type = RichChoiceField(
|
||||
choice_group="system_alert_types",
|
||||
domain="core",
|
||||
max_length=50,
|
||||
choices=AlertType.choices,
|
||||
db_index=True,
|
||||
help_text="Type of system alert",
|
||||
)
|
||||
severity = models.CharField(
|
||||
severity = RichChoiceField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Severity.choices,
|
||||
db_index=True,
|
||||
help_text="Alert severity level",
|
||||
)
|
||||
@@ -386,16 +365,11 @@ class RateLimitAlertConfig(models.Model):
|
||||
Defines thresholds that trigger alerts when exceeded.
|
||||
"""
|
||||
|
||||
class MetricType(models.TextChoices):
|
||||
BLOCK_RATE = "block_rate", "Block Rate"
|
||||
TOTAL_REQUESTS = "total_requests", "Total Requests"
|
||||
UNIQUE_IPS = "unique_ips", "Unique IPs"
|
||||
FUNCTION_SPECIFIC = "function_specific", "Function Specific"
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
metric_type = models.CharField(
|
||||
metric_type = RichChoiceField(
|
||||
choice_group="metric_types",
|
||||
domain="core",
|
||||
max_length=50,
|
||||
choices=MetricType.choices,
|
||||
db_index=True,
|
||||
help_text="Type of metric to monitor",
|
||||
)
|
||||
@@ -484,18 +458,6 @@ class Incident(models.Model):
|
||||
allowing teams to track and resolve related alerts together.
|
||||
"""
|
||||
|
||||
class Status(models.TextChoices):
|
||||
OPEN = "open", "Open"
|
||||
INVESTIGATING = "investigating", "Investigating"
|
||||
RESOLVED = "resolved", "Resolved"
|
||||
CLOSED = "closed", "Closed"
|
||||
|
||||
class Severity(models.TextChoices):
|
||||
LOW = "low", "Low"
|
||||
MEDIUM = "medium", "Medium"
|
||||
HIGH = "high", "High"
|
||||
CRITICAL = "critical", "Critical"
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
incident_number = models.CharField(
|
||||
max_length=20,
|
||||
@@ -505,16 +467,18 @@ class Incident(models.Model):
|
||||
)
|
||||
title = models.CharField(max_length=255, help_text="Brief description of the incident")
|
||||
description = models.TextField(null=True, blank=True, help_text="Detailed description")
|
||||
severity = models.CharField(
|
||||
severity = RichChoiceField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Severity.choices,
|
||||
db_index=True,
|
||||
help_text="Incident severity level",
|
||||
)
|
||||
status = models.CharField(
|
||||
status = RichChoiceField(
|
||||
choice_group="incident_statuses",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Status.choices,
|
||||
default=Status.OPEN,
|
||||
default="open",
|
||||
db_index=True,
|
||||
help_text="Current incident status",
|
||||
)
|
||||
@@ -582,10 +546,6 @@ class IncidentAlert(models.Model):
|
||||
Supports linking both system alerts and rate limit alerts.
|
||||
"""
|
||||
|
||||
class AlertSource(models.TextChoices):
|
||||
SYSTEM = "system", "System Alert"
|
||||
RATE_LIMIT = "rate_limit", "Rate Limit Alert"
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
incident = models.ForeignKey(
|
||||
Incident,
|
||||
@@ -593,9 +553,10 @@ class IncidentAlert(models.Model):
|
||||
related_name="linked_alerts",
|
||||
help_text="The incident this alert is linked to",
|
||||
)
|
||||
alert_source = models.CharField(
|
||||
alert_source = RichChoiceField(
|
||||
choice_group="alert_sources",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=AlertSource.choices,
|
||||
help_text="Source type of the alert",
|
||||
)
|
||||
alert_id = models.UUIDField(help_text="ID of the linked alert")
|
||||
@@ -633,13 +594,6 @@ class RequestMetadata(models.Model):
|
||||
dashboard for error monitoring and analytics.
|
||||
"""
|
||||
|
||||
class Severity(models.TextChoices):
|
||||
DEBUG = "debug", "Debug"
|
||||
INFO = "info", "Info"
|
||||
WARNING = "warning", "Warning"
|
||||
ERROR = "error", "Error"
|
||||
CRITICAL = "critical", "Critical"
|
||||
|
||||
# Identity & Correlation
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
request_id = models.CharField(
|
||||
@@ -789,10 +743,11 @@ class RequestMetadata(models.Model):
|
||||
null=True,
|
||||
help_text="React component stack trace",
|
||||
)
|
||||
severity = models.CharField(
|
||||
severity = RichChoiceField(
|
||||
choice_group="request_severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=Severity.choices,
|
||||
default=Severity.INFO,
|
||||
default="info",
|
||||
db_index=True,
|
||||
help_text="Error severity level",
|
||||
)
|
||||
@@ -1062,14 +1017,6 @@ class Milestone(TrackedModel):
|
||||
Maps to frontend milestoneValidationSchema in entityValidationSchemas.ts
|
||||
"""
|
||||
|
||||
class DatePrecision(models.TextChoices):
|
||||
EXACT = "exact", "Exact Date"
|
||||
MONTH = "month", "Month and Year"
|
||||
YEAR = "year", "Year Only"
|
||||
DECADE = "decade", "Decade"
|
||||
CENTURY = "century", "Century"
|
||||
APPROXIMATE = "approximate", "Approximate"
|
||||
|
||||
# Core event information
|
||||
title = models.CharField(
|
||||
max_length=200,
|
||||
@@ -1088,10 +1035,11 @@ class Milestone(TrackedModel):
|
||||
db_index=True,
|
||||
help_text="Date when the event occurred or will occur",
|
||||
)
|
||||
event_date_precision = models.CharField(
|
||||
event_date_precision = RichChoiceField(
|
||||
choice_group="date_precision",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
choices=DatePrecision.choices,
|
||||
default=DatePrecision.EXACT,
|
||||
default="exact",
|
||||
help_text="Precision of the event date",
|
||||
)
|
||||
|
||||
@@ -1161,3 +1109,354 @@ class Milestone(TrackedModel):
|
||||
def __str__(self) -> str:
|
||||
return f"{self.title} ({self.event_date})"
|
||||
|
||||
|
||||
class PipelineError(models.Model):
|
||||
"""
|
||||
Tracks pipeline/processing errors for debugging and monitoring.
|
||||
|
||||
Records errors that occur during data processing pipelines,
|
||||
approval workflows, and other automated operations. Supports
|
||||
resolution tracking and filtering by severity, function, and date.
|
||||
"""
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
function_name = models.CharField(
|
||||
max_length=255,
|
||||
db_index=True,
|
||||
help_text="Name of the function/pipeline that failed",
|
||||
)
|
||||
error_message = models.TextField(
|
||||
help_text="Error message describing the failure",
|
||||
)
|
||||
error_code = models.CharField(
|
||||
max_length=100,
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="Error code for categorization",
|
||||
)
|
||||
error_context = models.JSONField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Additional context data as JSON",
|
||||
)
|
||||
stack_trace = models.TextField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Full stack trace for debugging",
|
||||
)
|
||||
severity = RichChoiceField(
|
||||
choice_group="pipeline_error_severities",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
default="error",
|
||||
db_index=True,
|
||||
help_text="Severity level of the error",
|
||||
)
|
||||
|
||||
# References
|
||||
submission_id = models.UUIDField(
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="ID of related content submission if applicable",
|
||||
)
|
||||
item_id = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="Generic reference to related item",
|
||||
)
|
||||
request_id = models.UUIDField(
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="Request ID for correlation",
|
||||
)
|
||||
trace_id = models.UUIDField(
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="Distributed trace ID",
|
||||
)
|
||||
|
||||
# Resolution
|
||||
resolved = models.BooleanField(
|
||||
default=False,
|
||||
db_index=True,
|
||||
help_text="Whether this error has been resolved",
|
||||
)
|
||||
resolved_by = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
blank=True,
|
||||
null=True,
|
||||
on_delete=models.SET_NULL,
|
||||
related_name="resolved_pipeline_errors",
|
||||
help_text="User who resolved this error",
|
||||
)
|
||||
resolved_at = models.DateTimeField(
|
||||
blank=True,
|
||||
null=True,
|
||||
db_index=True,
|
||||
help_text="When this error was resolved",
|
||||
)
|
||||
resolution_notes = models.TextField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Notes about how the error was resolved",
|
||||
)
|
||||
|
||||
# Timestamps
|
||||
occurred_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
db_index=True,
|
||||
help_text="When this error occurred",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["-occurred_at"]
|
||||
verbose_name = "Pipeline Error"
|
||||
verbose_name_plural = "Pipeline Errors"
|
||||
indexes = [
|
||||
models.Index(fields=["severity", "occurred_at"]),
|
||||
models.Index(fields=["function_name", "occurred_at"]),
|
||||
models.Index(fields=["resolved", "occurred_at"]),
|
||||
]
|
||||
|
||||
def __str__(self) -> str:
|
||||
return f"[{self.get_severity_display()}] {self.function_name}: {self.error_message[:50]}"
|
||||
|
||||
|
||||
class Anomaly(models.Model):
|
||||
"""
|
||||
Records detected anomalies in system metrics.
|
||||
|
||||
Anomalies are identified by detection algorithms and stored
|
||||
for analysis and alerting. Each anomaly includes the metric
|
||||
details, deviation scores, and optional links to created alerts.
|
||||
"""
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
metric_name = models.CharField(
|
||||
max_length=255,
|
||||
db_index=True,
|
||||
help_text="Name of the metric that exhibited anomalous behavior",
|
||||
)
|
||||
metric_category = models.CharField(
|
||||
max_length=100,
|
||||
db_index=True,
|
||||
help_text="Category of the metric (e.g., 'performance', 'error_rate', 'traffic')",
|
||||
)
|
||||
anomaly_type = RichChoiceField(
|
||||
choice_group="anomaly_types",
|
||||
domain="core",
|
||||
max_length=30,
|
||||
db_index=True,
|
||||
help_text="Type of anomaly detected",
|
||||
)
|
||||
severity = RichChoiceField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
db_index=True,
|
||||
help_text="Severity of the anomaly",
|
||||
)
|
||||
|
||||
# Metric values
|
||||
anomaly_value = models.DecimalField(
|
||||
max_digits=20,
|
||||
decimal_places=6,
|
||||
help_text="The anomalous value detected",
|
||||
)
|
||||
baseline_value = models.DecimalField(
|
||||
max_digits=20,
|
||||
decimal_places=6,
|
||||
help_text="The expected baseline value",
|
||||
)
|
||||
deviation_score = models.DecimalField(
|
||||
max_digits=10,
|
||||
decimal_places=4,
|
||||
help_text="Standard deviations from normal",
|
||||
)
|
||||
confidence_score = models.DecimalField(
|
||||
max_digits=5,
|
||||
decimal_places=4,
|
||||
help_text="Confidence score of the detection (0-1)",
|
||||
)
|
||||
|
||||
# Detection context
|
||||
detection_algorithm = models.CharField(
|
||||
max_length=100,
|
||||
help_text="Algorithm used for detection",
|
||||
)
|
||||
time_window_start = models.DateTimeField(
|
||||
help_text="Start of the detection time window",
|
||||
)
|
||||
time_window_end = models.DateTimeField(
|
||||
help_text="End of the detection time window",
|
||||
)
|
||||
|
||||
# Alert linkage
|
||||
alert_created = models.BooleanField(
|
||||
default=False,
|
||||
db_index=True,
|
||||
help_text="Whether an alert was created for this anomaly",
|
||||
)
|
||||
alert = models.ForeignKey(
|
||||
SystemAlert,
|
||||
blank=True,
|
||||
null=True,
|
||||
on_delete=models.SET_NULL,
|
||||
related_name="anomalies",
|
||||
help_text="Linked system alert if created",
|
||||
)
|
||||
|
||||
# Timestamps
|
||||
detected_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
db_index=True,
|
||||
help_text="When this anomaly was detected",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["-detected_at"]
|
||||
verbose_name = "Anomaly"
|
||||
verbose_name_plural = "Anomalies"
|
||||
indexes = [
|
||||
models.Index(fields=["metric_name", "detected_at"]),
|
||||
models.Index(fields=["severity", "detected_at"]),
|
||||
models.Index(fields=["anomaly_type", "detected_at"]),
|
||||
models.Index(fields=["alert_created", "detected_at"]),
|
||||
]
|
||||
|
||||
def __str__(self) -> str:
|
||||
return f"[{self.get_severity_display()}] {self.metric_name}: {self.get_anomaly_type_display()}"
|
||||
|
||||
|
||||
class AlertCorrelationRule(models.Model):
|
||||
"""
|
||||
Defines rules for correlating multiple alerts.
|
||||
|
||||
When multiple alerts match a correlation rule's pattern within
|
||||
the time window, they can be grouped into an incident. This
|
||||
helps reduce alert fatigue and identify related issues.
|
||||
"""
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
rule_name = models.CharField(
|
||||
max_length=255,
|
||||
unique=True,
|
||||
db_index=True,
|
||||
help_text="Unique name for this correlation rule",
|
||||
)
|
||||
rule_description = models.TextField(
|
||||
blank=True,
|
||||
help_text="Description of what this rule correlates",
|
||||
)
|
||||
min_alerts_required = models.PositiveIntegerField(
|
||||
default=3,
|
||||
help_text="Minimum number of alerts needed to trigger correlation",
|
||||
)
|
||||
time_window_minutes = models.PositiveIntegerField(
|
||||
default=30,
|
||||
help_text="Time window in minutes for alert correlation",
|
||||
)
|
||||
incident_severity = RichChoiceField(
|
||||
choice_group="severity_levels",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
default="medium",
|
||||
help_text="Severity to assign to correlated incidents",
|
||||
)
|
||||
incident_title_template = models.CharField(
|
||||
max_length=255,
|
||||
help_text="Template for incident title (supports {count}, {rule_name})",
|
||||
)
|
||||
is_active = models.BooleanField(
|
||||
default=True,
|
||||
db_index=True,
|
||||
help_text="Whether this rule is currently active",
|
||||
)
|
||||
|
||||
# Timestamps
|
||||
created_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
help_text="When this rule was created",
|
||||
)
|
||||
updated_at = models.DateTimeField(
|
||||
auto_now=True,
|
||||
help_text="When this rule was last updated",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["rule_name"]
|
||||
verbose_name = "Alert Correlation Rule"
|
||||
verbose_name_plural = "Alert Correlation Rules"
|
||||
|
||||
def __str__(self) -> str:
|
||||
status = "active" if self.is_active else "inactive"
|
||||
return f"{self.rule_name} ({status})"
|
||||
|
||||
|
||||
class CleanupJobLog(models.Model):
|
||||
"""
|
||||
Audit log for cleanup and maintenance jobs.
|
||||
|
||||
Records the execution of background cleanup tasks,
|
||||
including success/failure status, records processed,
|
||||
and execution time for monitoring and debugging.
|
||||
"""
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
job_name = models.CharField(
|
||||
max_length=255,
|
||||
db_index=True,
|
||||
help_text="Name of the cleanup job",
|
||||
)
|
||||
status = RichChoiceField(
|
||||
choice_group="cleanup_job_statuses",
|
||||
domain="core",
|
||||
max_length=20,
|
||||
default="success",
|
||||
db_index=True,
|
||||
help_text="Execution status",
|
||||
)
|
||||
records_processed = models.PositiveIntegerField(
|
||||
default=0,
|
||||
help_text="Number of records processed",
|
||||
)
|
||||
records_deleted = models.PositiveIntegerField(
|
||||
default=0,
|
||||
help_text="Number of records deleted",
|
||||
)
|
||||
error_message = models.TextField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Error message if job failed",
|
||||
)
|
||||
duration_ms = models.PositiveIntegerField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Execution duration in milliseconds",
|
||||
)
|
||||
|
||||
# Timestamps
|
||||
executed_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
db_index=True,
|
||||
help_text="When this job was executed",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["-executed_at"]
|
||||
verbose_name = "Cleanup Job Log"
|
||||
verbose_name_plural = "Cleanup Job Logs"
|
||||
indexes = [
|
||||
models.Index(fields=["job_name", "executed_at"]),
|
||||
models.Index(fields=["status", "executed_at"]),
|
||||
]
|
||||
|
||||
def __str__(self) -> str:
|
||||
return f"{self.job_name}: {self.get_status_display()} ({self.records_processed} processed)"
|
||||
|
||||
|
||||
@@ -160,7 +160,7 @@ def error_validation(
|
||||
return custom_message
|
||||
if field_name:
|
||||
return f"Please check the {field_name} field and try again."
|
||||
return "Please check the form and correct any errors."
|
||||
return "Validation error. Please check the form and correct any errors."
|
||||
|
||||
|
||||
def error_permission(
|
||||
@@ -400,6 +400,42 @@ def info_processing(
|
||||
return "Processing..."
|
||||
|
||||
|
||||
def info_no_changes(
|
||||
custom_message: str | None = None,
|
||||
) -> str:
|
||||
"""
|
||||
Generate an info message when no changes were detected.
|
||||
|
||||
Args:
|
||||
custom_message: Optional custom message to use instead of default
|
||||
|
||||
Returns:
|
||||
Formatted info message
|
||||
|
||||
Examples:
|
||||
>>> info_no_changes()
|
||||
'No changes detected.'
|
||||
"""
|
||||
if custom_message:
|
||||
return custom_message
|
||||
return "No changes detected."
|
||||
|
||||
|
||||
def warning_unsaved(
|
||||
custom_message: str | None = None,
|
||||
) -> str:
|
||||
"""
|
||||
Alias for warning_unsaved_changes for backward compatibility.
|
||||
|
||||
Args:
|
||||
custom_message: Optional custom message to use instead of default
|
||||
|
||||
Returns:
|
||||
Formatted warning message
|
||||
"""
|
||||
return warning_unsaved_changes(custom_message)
|
||||
|
||||
|
||||
def confirm_delete(
|
||||
model_name: str,
|
||||
object_name: str | None = None,
|
||||
|
||||
@@ -124,6 +124,20 @@ SUBMISSION_TYPES = [
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="PHOTO",
|
||||
label="Photo Submission",
|
||||
description="Photo upload for existing content",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "photograph",
|
||||
"css_class": "bg-purple-100 text-purple-800 border-purple-200",
|
||||
"sort_order": 3,
|
||||
"requires_existing_object": True,
|
||||
"complexity_level": "low",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
@@ -934,6 +948,122 @@ BULK_OPERATION_TYPES = [
|
||||
# PhotoSubmission uses the same STATUS_CHOICES as EditSubmission
|
||||
PHOTO_SUBMISSION_STATUSES = EDIT_SUBMISSION_STATUSES
|
||||
|
||||
# ============================================================================
|
||||
# ModerationAuditLog Action Choices
|
||||
# ============================================================================
|
||||
|
||||
MODERATION_AUDIT_ACTIONS = [
|
||||
RichChoice(
|
||||
value="approved",
|
||||
label="Approved",
|
||||
description="Submission was approved by moderator",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="rejected",
|
||||
label="Rejected",
|
||||
description="Submission was rejected by moderator",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="claimed",
|
||||
label="Claimed",
|
||||
description="Submission was claimed by moderator",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "user-check",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="unclaimed",
|
||||
label="Unclaimed",
|
||||
description="Submission was released by moderator",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "user-minus",
|
||||
"css_class": "bg-gray-100 text-gray-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="escalated",
|
||||
label="Escalated",
|
||||
description="Submission was escalated for higher-level review",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "arrow-up",
|
||||
"css_class": "bg-purple-100 text-purple-800",
|
||||
"sort_order": 5,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="converted_to_edit",
|
||||
label="Converted to Edit",
|
||||
description="Photo submission was converted to an edit submission",
|
||||
metadata={
|
||||
"color": "indigo",
|
||||
"icon": "refresh",
|
||||
"css_class": "bg-indigo-100 text-indigo-800",
|
||||
"sort_order": 6,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="status_changed",
|
||||
label="Status Changed",
|
||||
description="Submission status was changed",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "refresh-cw",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 7,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="notes_added",
|
||||
label="Notes Added",
|
||||
description="Moderator notes were added to submission",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "edit",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 8,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="auto_approved",
|
||||
label="Auto Approved",
|
||||
description="Submission was auto-approved by the system",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "zap",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 9,
|
||||
"is_system_action": True,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Choice Registration
|
||||
# ============================================================================
|
||||
@@ -958,3 +1088,6 @@ register_choices("bulk_operation_types", BULK_OPERATION_TYPES, "moderation", "Bu
|
||||
register_choices(
|
||||
"photo_submission_statuses", PHOTO_SUBMISSION_STATUSES, "moderation", "Photo submission status options"
|
||||
)
|
||||
register_choices(
|
||||
"moderation_audit_actions", MODERATION_AUDIT_ACTIONS, "moderation", "Moderation audit log action types"
|
||||
)
|
||||
|
||||
@@ -27,12 +27,10 @@ User = get_user_model()
|
||||
class ModerationReportFilter(django_filters.FilterSet):
|
||||
"""Filter for ModerationReport model."""
|
||||
|
||||
# Status filters
|
||||
status = django_filters.ChoiceFilter(
|
||||
choices=lambda: [
|
||||
(choice.value, choice.label) for choice in get_choices("moderation_report_statuses", "moderation")
|
||||
],
|
||||
help_text="Filter by report status",
|
||||
# Status filters - use method filter for case-insensitive matching
|
||||
status = django_filters.CharFilter(
|
||||
method="filter_status",
|
||||
help_text="Filter by report status (case-insensitive)",
|
||||
)
|
||||
|
||||
# Priority filters
|
||||
@@ -144,6 +142,19 @@ class ModerationReportFilter(django_filters.FilterSet):
|
||||
return queryset.exclude(resolution_action__isnull=True, resolution_action="")
|
||||
return queryset.filter(Q(resolution_action__isnull=True) | Q(resolution_action=""))
|
||||
|
||||
def filter_status(self, queryset, name, value):
|
||||
"""Filter by status with case-insensitive matching."""
|
||||
if not value:
|
||||
return queryset
|
||||
# Normalize to uppercase for matching against RichChoice values
|
||||
normalized_value = value.upper()
|
||||
# Validate against registered choices
|
||||
valid_values = {choice.value for choice in get_choices("moderation_report_statuses", "moderation")}
|
||||
if normalized_value in valid_values:
|
||||
return queryset.filter(status=normalized_value)
|
||||
# If not valid, return empty queryset (invalid filter value)
|
||||
return queryset.none()
|
||||
|
||||
|
||||
class ModerationQueueFilter(django_filters.FilterSet):
|
||||
"""Filter for ModerationQueue model."""
|
||||
|
||||
@@ -0,0 +1,96 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-11 18:06
|
||||
|
||||
import apps.core.choices.fields
|
||||
import django.db.models.deletion
|
||||
import uuid
|
||||
from django.conf import settings
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("moderation", "0009_add_claim_fields"),
|
||||
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.CreateModel(
|
||||
name="ModerationAuditLog",
|
||||
fields=[
|
||||
("id", models.UUIDField(default=uuid.uuid4, editable=False, primary_key=True, serialize=False)),
|
||||
(
|
||||
"action",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="moderation_audit_actions",
|
||||
choices=[
|
||||
("approved", "Approved"),
|
||||
("rejected", "Rejected"),
|
||||
("claimed", "Claimed"),
|
||||
("unclaimed", "Unclaimed"),
|
||||
("escalated", "Escalated"),
|
||||
("converted_to_edit", "Converted to Edit"),
|
||||
("status_changed", "Status Changed"),
|
||||
("notes_added", "Notes Added"),
|
||||
("auto_approved", "Auto Approved"),
|
||||
],
|
||||
db_index=True,
|
||||
domain="moderation",
|
||||
help_text="The action that was performed",
|
||||
max_length=50,
|
||||
),
|
||||
),
|
||||
(
|
||||
"previous_status",
|
||||
models.CharField(blank=True, help_text="Status before the action", max_length=50, null=True),
|
||||
),
|
||||
(
|
||||
"new_status",
|
||||
models.CharField(blank=True, help_text="Status after the action", max_length=50, null=True),
|
||||
),
|
||||
("notes", models.TextField(blank=True, help_text="Notes or comments about the action", null=True)),
|
||||
(
|
||||
"is_system_action",
|
||||
models.BooleanField(
|
||||
db_index=True, default=False, help_text="Whether this was an automated system action"
|
||||
),
|
||||
),
|
||||
("is_test_data", models.BooleanField(default=False, help_text="Whether this is test data")),
|
||||
(
|
||||
"created_at",
|
||||
models.DateTimeField(auto_now_add=True, db_index=True, help_text="When this action was performed"),
|
||||
),
|
||||
(
|
||||
"moderator",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="The moderator who performed the action (null for system actions)",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="moderation_audit_logs",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
(
|
||||
"submission",
|
||||
models.ForeignKey(
|
||||
help_text="The submission this audit log entry is for",
|
||||
on_delete=django.db.models.deletion.CASCADE,
|
||||
related_name="audit_logs",
|
||||
to="moderation.editsubmission",
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Moderation Audit Log",
|
||||
"verbose_name_plural": "Moderation Audit Logs",
|
||||
"ordering": ["-created_at"],
|
||||
"indexes": [
|
||||
models.Index(fields=["submission", "created_at"], name="moderation__submiss_2f5e56_idx"),
|
||||
models.Index(fields=["moderator", "created_at"], name="moderation__moderat_591c14_idx"),
|
||||
models.Index(fields=["action", "created_at"], name="moderation__action_a98c47_idx"),
|
||||
],
|
||||
},
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,99 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-12 23:00
|
||||
|
||||
import django.db.models.deletion
|
||||
import pgtrigger.compiler
|
||||
import pgtrigger.migrations
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("django_cloudflareimages_toolkit", "0001_initial"),
|
||||
("moderation", "0010_moderationauditlog"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
pgtrigger.migrations.RemoveTrigger(
|
||||
model_name="editsubmission",
|
||||
name="insert_insert",
|
||||
),
|
||||
pgtrigger.migrations.RemoveTrigger(
|
||||
model_name="editsubmission",
|
||||
name="update_update",
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmission",
|
||||
name="caption",
|
||||
field=models.CharField(blank=True, help_text="Photo caption", max_length=255),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmission",
|
||||
name="date_taken",
|
||||
field=models.DateField(blank=True, help_text="Date the photo was taken", null=True),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmission",
|
||||
name="photo",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Photo for photo submissions",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
to="django_cloudflareimages_toolkit.cloudflareimage",
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmissionevent",
|
||||
name="caption",
|
||||
field=models.CharField(blank=True, help_text="Photo caption", max_length=255),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmissionevent",
|
||||
name="date_taken",
|
||||
field=models.DateField(blank=True, help_text="Date the photo was taken", null=True),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="editsubmissionevent",
|
||||
name="photo",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
db_constraint=False,
|
||||
help_text="Photo for photo submissions",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.DO_NOTHING,
|
||||
related_name="+",
|
||||
related_query_name="+",
|
||||
to="django_cloudflareimages_toolkit.cloudflareimage",
|
||||
),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="editsubmission",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="insert_insert",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
func='INSERT INTO "moderation_editsubmissionevent" ("caption", "changes", "claimed_at", "claimed_by_id", "content_type_id", "created_at", "date_taken", "handled_at", "handled_by_id", "id", "moderator_changes", "notes", "object_id", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "photo_id", "reason", "source", "status", "submission_type", "updated_at", "user_id") VALUES (NEW."caption", NEW."changes", NEW."claimed_at", NEW."claimed_by_id", NEW."content_type_id", NEW."created_at", NEW."date_taken", NEW."handled_at", NEW."handled_by_id", NEW."id", NEW."moderator_changes", NEW."notes", NEW."object_id", _pgh_attach_context(), NOW(), \'insert\', NEW."id", NEW."photo_id", NEW."reason", NEW."source", NEW."status", NEW."submission_type", NEW."updated_at", NEW."user_id"); RETURN NULL;',
|
||||
hash="e9aed25fe6389b113919e729543a9abe20d9f30c",
|
||||
operation="INSERT",
|
||||
pgid="pgtrigger_insert_insert_2c796",
|
||||
table="moderation_editsubmission",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="editsubmission",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="update_update",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
condition="WHEN (OLD.* IS DISTINCT FROM NEW.*)",
|
||||
func='INSERT INTO "moderation_editsubmissionevent" ("caption", "changes", "claimed_at", "claimed_by_id", "content_type_id", "created_at", "date_taken", "handled_at", "handled_by_id", "id", "moderator_changes", "notes", "object_id", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "photo_id", "reason", "source", "status", "submission_type", "updated_at", "user_id") VALUES (NEW."caption", NEW."changes", NEW."claimed_at", NEW."claimed_by_id", NEW."content_type_id", NEW."created_at", NEW."date_taken", NEW."handled_at", NEW."handled_by_id", NEW."id", NEW."moderator_changes", NEW."notes", NEW."object_id", _pgh_attach_context(), NOW(), \'update\', NEW."id", NEW."photo_id", NEW."reason", NEW."source", NEW."status", NEW."submission_type", NEW."updated_at", NEW."user_id"); RETURN NULL;',
|
||||
hash="070083ba4d2d459067d9c3a90356a759f6262a90",
|
||||
operation="UPDATE",
|
||||
pgid="pgtrigger_update_update_ab38f",
|
||||
table="moderation_editsubmission",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,64 @@
|
||||
"""
|
||||
Data migration to copy PhotoSubmission data to EditSubmission.
|
||||
|
||||
This migration copies all PhotoSubmission rows to EditSubmission with submission_type="PHOTO".
|
||||
After this migration, PhotoSubmission model can be safely removed.
|
||||
"""
|
||||
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
def migrate_photo_submissions(apps, schema_editor):
|
||||
"""Copy PhotoSubmission data to EditSubmission."""
|
||||
PhotoSubmission = apps.get_model("moderation", "PhotoSubmission")
|
||||
EditSubmission = apps.get_model("moderation", "EditSubmission")
|
||||
ContentType = apps.get_model("contenttypes", "ContentType")
|
||||
|
||||
# Get EditSubmission content type for reference
|
||||
edit_submission_ct = ContentType.objects.get_for_model(EditSubmission)
|
||||
|
||||
migrated = 0
|
||||
for photo_sub in PhotoSubmission.objects.all():
|
||||
# Create EditSubmission from PhotoSubmission
|
||||
EditSubmission.objects.create(
|
||||
user=photo_sub.user,
|
||||
content_type=photo_sub.content_type,
|
||||
object_id=photo_sub.object_id,
|
||||
submission_type="PHOTO",
|
||||
changes={}, # Photos don't have field changes
|
||||
reason="Photo submission", # Default reason
|
||||
status=photo_sub.status,
|
||||
created_at=photo_sub.created_at,
|
||||
handled_by=photo_sub.handled_by,
|
||||
handled_at=photo_sub.handled_at,
|
||||
notes=photo_sub.notes,
|
||||
claimed_by=photo_sub.claimed_by,
|
||||
claimed_at=photo_sub.claimed_at,
|
||||
# Photo-specific fields
|
||||
photo=photo_sub.photo,
|
||||
caption=photo_sub.caption,
|
||||
date_taken=photo_sub.date_taken,
|
||||
)
|
||||
migrated += 1
|
||||
|
||||
if migrated:
|
||||
print(f"Migrated {migrated} PhotoSubmission(s) to EditSubmission")
|
||||
|
||||
|
||||
def reverse_migration(apps, schema_editor):
|
||||
"""Remove migrated EditSubmissions with type PHOTO."""
|
||||
EditSubmission = apps.get_model("moderation", "EditSubmission")
|
||||
deleted, _ = EditSubmission.objects.filter(submission_type="PHOTO").delete()
|
||||
if deleted:
|
||||
print(f"Deleted {deleted} PHOTO EditSubmission(s)")
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("moderation", "0011_add_photo_fields_to_editsubmission"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.RunPython(migrate_photo_submissions, reverse_migration),
|
||||
]
|
||||
@@ -18,6 +18,7 @@ are registered via the callback configuration defined in each model's Meta class
|
||||
|
||||
from datetime import timedelta
|
||||
from typing import Any
|
||||
import uuid
|
||||
|
||||
import pghistory
|
||||
from django.conf import settings
|
||||
@@ -114,6 +115,25 @@ class EditSubmission(StateMachineMixin, TrackedModel):
|
||||
help_text="Moderator's edited version of the changes before approval",
|
||||
)
|
||||
|
||||
# Photo submission fields (only used when submission_type="PHOTO")
|
||||
photo = models.ForeignKey(
|
||||
"django_cloudflareimages_toolkit.CloudflareImage",
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
help_text="Photo for photo submissions",
|
||||
)
|
||||
caption = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
help_text="Photo caption",
|
||||
)
|
||||
date_taken = models.DateField(
|
||||
null=True,
|
||||
blank=True,
|
||||
help_text="Date the photo was taken",
|
||||
)
|
||||
|
||||
# Metadata
|
||||
reason = models.TextField(help_text="Why this edit/addition is needed")
|
||||
source = models.TextField(blank=True, help_text="Source of information (if applicable)")
|
||||
@@ -190,6 +210,122 @@ class EditSubmission(StateMachineMixin, TrackedModel):
|
||||
"""Get the final changes to apply (moderator changes if available, otherwise original changes)"""
|
||||
return self.moderator_changes or self.changes
|
||||
|
||||
def _get_model_class_for_item_type(self, item_type: str):
|
||||
"""
|
||||
Map item_type string to the corresponding Django model class.
|
||||
|
||||
Args:
|
||||
item_type: Type string from frontend (e.g., 'manufacturer', 'park', 'ride_model')
|
||||
|
||||
Returns:
|
||||
Model class for the item type
|
||||
"""
|
||||
# Lazy imports to avoid circular dependencies
|
||||
from apps.parks.models import Company, Park
|
||||
from apps.rides.models import Ride, RideModel
|
||||
|
||||
type_map = {
|
||||
# Company types (all map to Company model)
|
||||
'manufacturer': Company,
|
||||
'designer': Company,
|
||||
'operator': Company,
|
||||
'property_owner': Company,
|
||||
'company': Company,
|
||||
# Entity types
|
||||
'park': Park,
|
||||
'ride': Ride,
|
||||
'ride_model': RideModel,
|
||||
}
|
||||
|
||||
model_class = type_map.get(item_type.lower())
|
||||
if not model_class:
|
||||
raise ValueError(f"Unknown item_type: {item_type}")
|
||||
return model_class
|
||||
|
||||
def _process_composite_items(self, composite_items: list[dict[str, Any]]) -> dict[int, Any]:
|
||||
"""
|
||||
Process composite submission items (dependencies) before the primary entity.
|
||||
|
||||
Args:
|
||||
composite_items: List of dependency items from frontend's submissionItems array
|
||||
Each item has: item_type, action_type, item_data, order_index, depends_on
|
||||
|
||||
Returns:
|
||||
Dictionary mapping order_index -> created entity ID for resolving temp references
|
||||
"""
|
||||
from django.db import transaction
|
||||
|
||||
# Sort by order_index to ensure proper dependency order
|
||||
sorted_items = sorted(composite_items, key=lambda x: x.get('order_index', 0))
|
||||
|
||||
# Map of order_index -> created entity ID
|
||||
created_entities: dict[int, Any] = {}
|
||||
|
||||
with transaction.atomic():
|
||||
for item in sorted_items:
|
||||
item_type = item.get('item_type', '')
|
||||
item_data = item.get('item_data', {})
|
||||
order_index = item.get('order_index', 0)
|
||||
|
||||
if not item_type or not item_data:
|
||||
continue
|
||||
|
||||
# Get the model class for this item type
|
||||
model_class = self._get_model_class_for_item_type(item_type)
|
||||
|
||||
# Clean up internal fields not needed for model creation
|
||||
clean_data = {}
|
||||
for key, value in item_data.items():
|
||||
# Skip internal/temp fields
|
||||
if key.startswith('_temp_') or key == 'images' or key == '_composite_items':
|
||||
continue
|
||||
# Skip fields with None or 'temp-' values
|
||||
if value is None or (isinstance(value, str) and value.startswith('temp-')):
|
||||
continue
|
||||
clean_data[key] = value
|
||||
|
||||
# Resolve _temp_*_ref fields to actual entity IDs from previously created entities
|
||||
for key, value in item_data.items():
|
||||
if key.startswith('_temp_') and key.endswith('_ref'):
|
||||
# Extract the field name: _temp_manufacturer_ref -> manufacturer_id
|
||||
field_name = key[6:-4] + '_id' # Remove '_temp_' prefix and '_ref' suffix
|
||||
ref_order_index = value
|
||||
if isinstance(ref_order_index, int) and ref_order_index in created_entities:
|
||||
clean_data[field_name] = created_entities[ref_order_index]
|
||||
|
||||
# Resolve foreign keys to model instances
|
||||
resolved_data = {}
|
||||
for field_name, value in clean_data.items():
|
||||
try:
|
||||
field = model_class._meta.get_field(field_name)
|
||||
if isinstance(field, models.ForeignKey) and value is not None:
|
||||
try:
|
||||
related_obj = field.related_model.objects.get(pk=value)
|
||||
resolved_data[field_name] = related_obj
|
||||
except ObjectDoesNotExist:
|
||||
# Skip invalid FK references
|
||||
continue
|
||||
else:
|
||||
resolved_data[field_name] = value
|
||||
except:
|
||||
# Field doesn't exist on model, still try to include it
|
||||
resolved_data[field_name] = value
|
||||
|
||||
# Create the entity
|
||||
try:
|
||||
obj = model_class(**resolved_data)
|
||||
obj.full_clean()
|
||||
obj.save()
|
||||
created_entities[order_index] = obj.pk
|
||||
except Exception as e:
|
||||
# Log but continue - don't fail the whole submission for one dependency
|
||||
import logging
|
||||
logger = logging.getLogger(__name__)
|
||||
logger.error(f"Failed to create composite item {item_type}: {e}")
|
||||
continue
|
||||
|
||||
return created_entities
|
||||
|
||||
def claim(self, user: UserType) -> None:
|
||||
"""
|
||||
Claim this submission for review.
|
||||
@@ -266,6 +402,28 @@ class EditSubmission(StateMachineMixin, TrackedModel):
|
||||
raise ValueError("Could not resolve model class")
|
||||
|
||||
final_changes = self._get_final_changes()
|
||||
|
||||
# Process composite items (dependencies) first if present
|
||||
created_entity_ids: dict[int, Any] = {}
|
||||
if '_composite_items' in final_changes:
|
||||
composite_items = final_changes.pop('_composite_items')
|
||||
if composite_items and isinstance(composite_items, list):
|
||||
created_entity_ids = self._process_composite_items(composite_items)
|
||||
|
||||
# Resolve _temp_*_ref fields in the primary entity using created dependency IDs
|
||||
for key in list(final_changes.keys()):
|
||||
if key.startswith('_temp_') and key.endswith('_ref'):
|
||||
# Extract field name: _temp_manufacturer_ref -> manufacturer_id
|
||||
field_name = key[6:-4] + '_id' # Remove '_temp_' and '_ref'
|
||||
ref_order_index = final_changes.pop(key)
|
||||
if isinstance(ref_order_index, int) and ref_order_index in created_entity_ids:
|
||||
final_changes[field_name] = created_entity_ids[ref_order_index]
|
||||
|
||||
# Remove any remaining internal fields
|
||||
keys_to_remove = [k for k in final_changes.keys() if k.startswith('_')]
|
||||
for key in keys_to_remove:
|
||||
final_changes.pop(key, None)
|
||||
|
||||
resolved_changes = self._resolve_foreign_keys(final_changes)
|
||||
|
||||
try:
|
||||
@@ -295,6 +453,7 @@ class EditSubmission(StateMachineMixin, TrackedModel):
|
||||
|
||||
return obj
|
||||
|
||||
|
||||
except Exception as e:
|
||||
# On error, record the issue and attempt rejection transition
|
||||
self.notes = f"Approval failed: {str(e)}"
|
||||
@@ -900,3 +1059,82 @@ class PhotoSubmission(StateMachineMixin, TrackedModel):
|
||||
self.handled_at = timezone.now()
|
||||
self.notes = notes
|
||||
self.save()
|
||||
|
||||
|
||||
class ModerationAuditLog(models.Model):
|
||||
"""
|
||||
Audit log for moderation actions.
|
||||
|
||||
Records all moderation activities including approvals, rejections,
|
||||
claims, escalations, and conversions for accountability and analytics.
|
||||
"""
|
||||
|
||||
id = models.UUIDField(primary_key=True, default=uuid.uuid4, editable=False)
|
||||
submission = models.ForeignKey(
|
||||
EditSubmission,
|
||||
on_delete=models.CASCADE,
|
||||
related_name="audit_logs",
|
||||
help_text="The submission this audit log entry is for",
|
||||
)
|
||||
moderator = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
blank=True,
|
||||
null=True,
|
||||
on_delete=models.SET_NULL,
|
||||
related_name="moderation_audit_logs",
|
||||
help_text="The moderator who performed the action (null for system actions)",
|
||||
)
|
||||
action = RichChoiceField(
|
||||
choice_group="moderation_audit_actions",
|
||||
domain="moderation",
|
||||
max_length=50,
|
||||
db_index=True,
|
||||
help_text="The action that was performed",
|
||||
)
|
||||
previous_status = models.CharField(
|
||||
max_length=50,
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Status before the action",
|
||||
)
|
||||
new_status = models.CharField(
|
||||
max_length=50,
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Status after the action",
|
||||
)
|
||||
notes = models.TextField(
|
||||
blank=True,
|
||||
null=True,
|
||||
help_text="Notes or comments about the action",
|
||||
)
|
||||
is_system_action = models.BooleanField(
|
||||
default=False,
|
||||
db_index=True,
|
||||
help_text="Whether this was an automated system action",
|
||||
)
|
||||
is_test_data = models.BooleanField(
|
||||
default=False,
|
||||
help_text="Whether this is test data",
|
||||
)
|
||||
|
||||
# Timestamps
|
||||
created_at = models.DateTimeField(
|
||||
auto_now_add=True,
|
||||
db_index=True,
|
||||
help_text="When this action was performed",
|
||||
)
|
||||
|
||||
class Meta:
|
||||
ordering = ["-created_at"]
|
||||
verbose_name = "Moderation Audit Log"
|
||||
verbose_name_plural = "Moderation Audit Logs"
|
||||
indexes = [
|
||||
models.Index(fields=["submission", "created_at"]),
|
||||
models.Index(fields=["moderator", "created_at"]),
|
||||
models.Index(fields=["action", "created_at"]),
|
||||
]
|
||||
|
||||
def __str__(self) -> str:
|
||||
actor = self.moderator.username if self.moderator else "System"
|
||||
return f"{self.get_action_display()} by {actor} on {self.submission_id}"
|
||||
|
||||
@@ -100,6 +100,10 @@ class EditSubmissionSerializer(serializers.ModelSerializer):
|
||||
"claimed_at",
|
||||
"created_at",
|
||||
"time_since_created",
|
||||
# Photo fields (used when submission_type="PHOTO")
|
||||
"photo",
|
||||
"caption",
|
||||
"date_taken",
|
||||
]
|
||||
read_only_fields = [
|
||||
"id",
|
||||
@@ -1062,3 +1066,45 @@ class PhotoSubmissionSerializer(serializers.ModelSerializer):
|
||||
else:
|
||||
minutes = diff.seconds // 60
|
||||
return f"{minutes} minutes ago"
|
||||
|
||||
|
||||
# ============================================================================
|
||||
# Moderation Audit Log Serializers
|
||||
# ============================================================================
|
||||
|
||||
|
||||
class ModerationAuditLogSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for moderation audit logs."""
|
||||
|
||||
moderator = UserBasicSerializer(read_only=True)
|
||||
moderator_username = serializers.CharField(source="moderator.username", read_only=True, allow_null=True)
|
||||
submission_content_type = serializers.CharField(source="submission.content_type.model", read_only=True)
|
||||
action_display = serializers.CharField(source="get_action_display", read_only=True)
|
||||
|
||||
class Meta:
|
||||
from .models import ModerationAuditLog
|
||||
|
||||
model = ModerationAuditLog
|
||||
fields = [
|
||||
"id",
|
||||
"submission",
|
||||
"submission_content_type",
|
||||
"moderator",
|
||||
"moderator_username",
|
||||
"action",
|
||||
"action_display",
|
||||
"previous_status",
|
||||
"new_status",
|
||||
"notes",
|
||||
"is_system_action",
|
||||
"is_test_data",
|
||||
"created_at",
|
||||
]
|
||||
read_only_fields = [
|
||||
"id",
|
||||
"created_at",
|
||||
"moderator",
|
||||
"moderator_username",
|
||||
"submission_content_type",
|
||||
"action_display",
|
||||
]
|
||||
|
||||
@@ -5,6 +5,7 @@ Following Django styleguide pattern for business logic encapsulation.
|
||||
|
||||
from typing import Any
|
||||
|
||||
from django.contrib.contenttypes.models import ContentType
|
||||
from django.db import transaction
|
||||
from django.db.models import QuerySet
|
||||
from django.utils import timezone
|
||||
@@ -340,9 +341,13 @@ class ModerationService:
|
||||
Dictionary with submission info and queue status
|
||||
"""
|
||||
with transaction.atomic():
|
||||
# Create the photo submission
|
||||
submission = PhotoSubmission(
|
||||
content_object=content_object,
|
||||
# Create the photo submission using unified EditSubmission with PHOTO type
|
||||
submission = EditSubmission(
|
||||
content_type=ContentType.objects.get_for_model(content_object),
|
||||
object_id=content_object.pk,
|
||||
submission_type="PHOTO",
|
||||
changes={}, # Photos don't have field changes
|
||||
reason="Photo submission",
|
||||
photo=photo,
|
||||
caption=caption,
|
||||
date_taken=date_taken,
|
||||
|
||||
@@ -95,7 +95,7 @@ def expire_stale_claims(lock_duration_minutes: int = None) -> dict:
|
||||
source="task",
|
||||
)
|
||||
|
||||
# Process PhotoSubmissions with stale claims
|
||||
# Process PhotoSubmissions with stale claims (legacy model - until removed)
|
||||
stale_photo_ids = list(
|
||||
PhotoSubmission.objects.filter(
|
||||
status="CLAIMED",
|
||||
@@ -132,6 +132,43 @@ def expire_stale_claims(lock_duration_minutes: int = None) -> dict:
|
||||
source="task",
|
||||
)
|
||||
|
||||
# Also process EditSubmission with PHOTO type (new unified model)
|
||||
stale_photo_edit_ids = list(
|
||||
EditSubmission.objects.filter(
|
||||
submission_type="PHOTO",
|
||||
status="CLAIMED",
|
||||
claimed_at__lt=cutoff_time,
|
||||
).values_list("id", flat=True)
|
||||
)
|
||||
|
||||
for submission_id in stale_photo_edit_ids:
|
||||
result["edit_submissions"]["processed"] += 1 # Count with edit submissions
|
||||
try:
|
||||
with transaction.atomic():
|
||||
submission = EditSubmission.objects.select_for_update(skip_locked=True).filter(
|
||||
id=submission_id,
|
||||
status="CLAIMED",
|
||||
).first()
|
||||
|
||||
if submission:
|
||||
_release_claim(submission)
|
||||
result["edit_submissions"]["released"] += 1
|
||||
logger.info(
|
||||
"Released stale claim on PHOTO EditSubmission %s (claimed by %s at %s)",
|
||||
submission_id,
|
||||
submission.claimed_by,
|
||||
submission.claimed_at,
|
||||
)
|
||||
except Exception as e:
|
||||
result["edit_submissions"]["failed"] += 1
|
||||
error_msg = f"PHOTO EditSubmission {submission_id}: {str(e)}"
|
||||
result["failures"].append(error_msg)
|
||||
capture_and_log(
|
||||
e,
|
||||
f"Release stale claim on PHOTO EditSubmission {submission_id}",
|
||||
source="task",
|
||||
)
|
||||
|
||||
total_released = result["edit_submissions"]["released"] + result["photo_submissions"]["released"]
|
||||
total_failed = result["edit_submissions"]["failed"] + result["photo_submissions"]["failed"]
|
||||
|
||||
|
||||
@@ -45,13 +45,14 @@ from ..models import (
|
||||
User = get_user_model()
|
||||
|
||||
|
||||
class TestView(
|
||||
class MixinTestView(
|
||||
EditSubmissionMixin,
|
||||
PhotoSubmissionMixin,
|
||||
InlineEditMixin,
|
||||
HistoryMixin,
|
||||
DetailView,
|
||||
):
|
||||
"""Helper view for testing moderation mixins. Not a test class."""
|
||||
model = Operator
|
||||
template_name = "test.html"
|
||||
pk_url_kwarg = "pk"
|
||||
@@ -100,7 +101,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_edit_submission_mixin_unauthenticated(self):
|
||||
"""Test edit submission when not logged in"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
request = self.factory.post(f"/test/{self.operator.pk}/")
|
||||
request.user = AnonymousUser()
|
||||
view.setup(request, pk=self.operator.pk)
|
||||
@@ -111,7 +112,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_edit_submission_mixin_no_changes(self):
|
||||
"""Test edit submission with no changes"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
request = self.factory.post(
|
||||
f"/test/{self.operator.pk}/",
|
||||
data=json.dumps({}),
|
||||
@@ -126,7 +127,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_edit_submission_mixin_invalid_json(self):
|
||||
"""Test edit submission with invalid JSON"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
request = self.factory.post(
|
||||
f"/test/{self.operator.pk}/",
|
||||
data="invalid json",
|
||||
@@ -141,7 +142,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_edit_submission_mixin_regular_user(self):
|
||||
"""Test edit submission as regular user"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
request = self.factory.post(f"/test/{self.operator.pk}/")
|
||||
request.user = self.user
|
||||
view.setup(request, pk=self.operator.pk)
|
||||
@@ -155,7 +156,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_edit_submission_mixin_moderator(self):
|
||||
"""Test edit submission as moderator"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
request = self.factory.post(f"/test/{self.operator.pk}/")
|
||||
request.user = self.moderator
|
||||
view.setup(request, pk=self.operator.pk)
|
||||
@@ -169,7 +170,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_photo_submission_mixin_unauthenticated(self):
|
||||
"""Test photo submission when not logged in"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
|
||||
@@ -182,7 +183,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_photo_submission_mixin_no_photo(self):
|
||||
"""Test photo submission with no photo"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
|
||||
@@ -195,7 +196,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_photo_submission_mixin_regular_user(self):
|
||||
"""Test photo submission as regular user"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
|
||||
@@ -226,7 +227,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_photo_submission_mixin_moderator(self):
|
||||
"""Test photo submission as moderator"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
|
||||
@@ -315,7 +316,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_inline_edit_mixin(self):
|
||||
"""Test inline edit mixin"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
|
||||
@@ -342,7 +343,7 @@ class ModerationMixinsTests(TestCase):
|
||||
|
||||
def test_history_mixin(self):
|
||||
"""Test history mixin"""
|
||||
view = TestView()
|
||||
view = MixinTestView()
|
||||
view.kwargs = {"pk": self.operator.pk}
|
||||
view.object = self.operator
|
||||
request = self.factory.get(f"/test/{self.operator.pk}/")
|
||||
|
||||
@@ -2131,12 +2131,14 @@ class PhotoSubmissionViewSet(viewsets.ModelViewSet):
|
||||
"""
|
||||
ViewSet for managing photo submissions.
|
||||
|
||||
Now queries EditSubmission with submission_type="PHOTO" for unified model.
|
||||
Includes claim/unclaim endpoints with concurrency protection using
|
||||
database row locking (select_for_update) to prevent race conditions.
|
||||
"""
|
||||
|
||||
queryset = PhotoSubmission.objects.all()
|
||||
serializer_class = PhotoSubmissionSerializer
|
||||
# Use EditSubmission filtered by PHOTO type instead of separate PhotoSubmission model
|
||||
queryset = EditSubmission.objects.filter(submission_type="PHOTO")
|
||||
serializer_class = EditSubmissionSerializer # Use unified serializer
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
search_fields = ["caption", "notes"]
|
||||
ordering_fields = ["created_at", "status"]
|
||||
@@ -2144,10 +2146,10 @@ class PhotoSubmissionViewSet(viewsets.ModelViewSet):
|
||||
permission_classes = [CanViewModerationData]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
status = self.request.query_params.get("status")
|
||||
if status:
|
||||
queryset = queryset.filter(status=status)
|
||||
queryset = EditSubmission.objects.filter(submission_type="PHOTO")
|
||||
status_param = self.request.query_params.get("status")
|
||||
if status_param:
|
||||
queryset = queryset.filter(status=status_param)
|
||||
|
||||
# User filter
|
||||
user_id = self.request.query_params.get("user")
|
||||
@@ -2168,8 +2170,9 @@ class PhotoSubmissionViewSet(viewsets.ModelViewSet):
|
||||
|
||||
with transaction.atomic():
|
||||
try:
|
||||
submission = PhotoSubmission.objects.select_for_update(nowait=True).get(pk=pk)
|
||||
except PhotoSubmission.DoesNotExist:
|
||||
# Use EditSubmission filtered by PHOTO type
|
||||
submission = EditSubmission.objects.filter(submission_type="PHOTO").select_for_update(nowait=True).get(pk=pk)
|
||||
except EditSubmission.DoesNotExist:
|
||||
return Response({"error": "Submission not found"}, status=status.HTTP_404_NOT_FOUND)
|
||||
except DatabaseError:
|
||||
return Response(
|
||||
@@ -2198,9 +2201,10 @@ class PhotoSubmissionViewSet(viewsets.ModelViewSet):
|
||||
log_business_event(
|
||||
logger,
|
||||
event_type="submission_claimed",
|
||||
message=f"PhotoSubmission {submission.id} claimed by {request.user.username}",
|
||||
message=f"Photo EditSubmission {submission.id} claimed by {request.user.username}",
|
||||
context={
|
||||
"model": "PhotoSubmission",
|
||||
"model": "EditSubmission",
|
||||
"submission_type": "PHOTO",
|
||||
"object_id": submission.id,
|
||||
"claimed_by": request.user.username,
|
||||
},
|
||||
@@ -2767,3 +2771,55 @@ class ModerationStatsView(APIView):
|
||||
}
|
||||
|
||||
return Response(stats_data)
|
||||
|
||||
|
||||
# ============================================================================
|
||||
# Moderation Audit Log ViewSet
|
||||
# ============================================================================
|
||||
|
||||
|
||||
class ModerationAuditLogViewSet(viewsets.ReadOnlyModelViewSet):
|
||||
"""
|
||||
ViewSet for viewing moderation audit logs.
|
||||
|
||||
Provides read-only access to moderation action history for auditing
|
||||
and accountability purposes.
|
||||
"""
|
||||
|
||||
from .models import ModerationAuditLog
|
||||
from .serializers import ModerationAuditLogSerializer
|
||||
|
||||
queryset = ModerationAuditLog.objects.select_related(
|
||||
"submission", "submission__content_type", "moderator"
|
||||
).all()
|
||||
serializer_class = ModerationAuditLogSerializer
|
||||
permission_classes = [IsAdminOrSuperuser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["action", "is_system_action", "is_test_data"]
|
||||
search_fields = ["notes"]
|
||||
ordering_fields = ["created_at", "action"]
|
||||
ordering = ["-created_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
|
||||
# Filter by submission ID
|
||||
submission_id = self.request.query_params.get("submission_id")
|
||||
if submission_id:
|
||||
queryset = queryset.filter(submission_id=submission_id)
|
||||
|
||||
# Filter by moderator ID
|
||||
moderator_id = self.request.query_params.get("moderator_id")
|
||||
if moderator_id:
|
||||
queryset = queryset.filter(moderator_id=moderator_id)
|
||||
|
||||
# Date range filtering
|
||||
start_date = self.request.query_params.get("start_date")
|
||||
end_date = self.request.query_params.get("end_date")
|
||||
|
||||
if start_date:
|
||||
queryset = queryset.filter(created_at__gte=start_date)
|
||||
if end_date:
|
||||
queryset = queryset.filter(created_at__lte=end_date)
|
||||
|
||||
return queryset
|
||||
|
||||
46
backend/apps/notifications/api/log_serializers.py
Normal file
46
backend/apps/notifications/api/log_serializers.py
Normal file
@@ -0,0 +1,46 @@
|
||||
"""
|
||||
Serializers for Notification Log API.
|
||||
"""
|
||||
|
||||
from rest_framework import serializers
|
||||
|
||||
from apps.core.choices.serializers import RichChoiceSerializerField
|
||||
from apps.notifications.models import NotificationLog
|
||||
|
||||
|
||||
class NotificationLogSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for notification logs."""
|
||||
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="notification_log_statuses",
|
||||
domain="notifications",
|
||||
)
|
||||
user_username = serializers.CharField(
|
||||
source="user.username",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
user_email = serializers.EmailField(
|
||||
source="user.email",
|
||||
read_only=True,
|
||||
allow_null=True,
|
||||
)
|
||||
|
||||
class Meta:
|
||||
model = NotificationLog
|
||||
fields = [
|
||||
"id",
|
||||
"user",
|
||||
"user_username",
|
||||
"user_email",
|
||||
"workflow_id",
|
||||
"notification_type",
|
||||
"channel",
|
||||
"status",
|
||||
"payload",
|
||||
"error_message",
|
||||
"novu_transaction_id",
|
||||
"created_at",
|
||||
"updated_at",
|
||||
]
|
||||
read_only_fields = ["id", "created_at", "updated_at", "user_username", "user_email"]
|
||||
61
backend/apps/notifications/api/log_views.py
Normal file
61
backend/apps/notifications/api/log_views.py
Normal file
@@ -0,0 +1,61 @@
|
||||
"""
|
||||
ViewSet for Notification Log API.
|
||||
"""
|
||||
|
||||
from django_filters.rest_framework import DjangoFilterBackend
|
||||
from drf_spectacular.utils import extend_schema, extend_schema_view
|
||||
from rest_framework import viewsets
|
||||
from rest_framework.filters import OrderingFilter, SearchFilter
|
||||
from rest_framework.permissions import IsAdminUser
|
||||
|
||||
from apps.notifications.models import NotificationLog
|
||||
|
||||
from .log_serializers import NotificationLogSerializer
|
||||
|
||||
|
||||
@extend_schema_view(
|
||||
list=extend_schema(
|
||||
summary="List notification logs",
|
||||
description="Get all notification logs with optional filtering by status, channel, or workflow.",
|
||||
tags=["Admin - Notifications"],
|
||||
),
|
||||
retrieve=extend_schema(
|
||||
summary="Get notification log",
|
||||
description="Get details of a specific notification log entry.",
|
||||
tags=["Admin - Notifications"],
|
||||
),
|
||||
)
|
||||
class NotificationLogViewSet(viewsets.ReadOnlyModelViewSet):
|
||||
"""
|
||||
ViewSet for viewing notification logs.
|
||||
|
||||
Provides read-only access to notification delivery history.
|
||||
"""
|
||||
|
||||
queryset = NotificationLog.objects.select_related("user").all()
|
||||
serializer_class = NotificationLogSerializer
|
||||
permission_classes = [IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, SearchFilter, OrderingFilter]
|
||||
filterset_fields = ["status", "channel", "workflow_id", "notification_type"]
|
||||
search_fields = ["workflow_id", "notification_type", "error_message"]
|
||||
ordering_fields = ["created_at", "status"]
|
||||
ordering = ["-created_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
queryset = super().get_queryset()
|
||||
|
||||
# Filter by user ID if provided
|
||||
user_id = self.request.query_params.get("user_id")
|
||||
if user_id:
|
||||
queryset = queryset.filter(user_id=user_id)
|
||||
|
||||
# Date range filtering
|
||||
start_date = self.request.query_params.get("start_date")
|
||||
end_date = self.request.query_params.get("end_date")
|
||||
|
||||
if start_date:
|
||||
queryset = queryset.filter(created_at__gte=start_date)
|
||||
if end_date:
|
||||
queryset = queryset.filter(created_at__lte=end_date)
|
||||
|
||||
return queryset
|
||||
185
backend/apps/notifications/choices.py
Normal file
185
backend/apps/notifications/choices.py
Normal file
@@ -0,0 +1,185 @@
|
||||
"""
|
||||
Rich Choice Objects for Notifications Domain
|
||||
|
||||
This module defines all choice objects for the notifications domain,
|
||||
using the RichChoices pattern for consistent UI rendering and validation.
|
||||
"""
|
||||
|
||||
from apps.core.choices import ChoiceCategory, RichChoice
|
||||
from apps.core.choices.registry import register_choices
|
||||
|
||||
# ============================================================================
|
||||
# Notification Log Status Choices
|
||||
# ============================================================================
|
||||
NOTIFICATION_LOG_STATUSES = [
|
||||
RichChoice(
|
||||
value="pending",
|
||||
label="Pending",
|
||||
description="Notification is queued for delivery",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "clock",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="sent",
|
||||
label="Sent",
|
||||
description="Notification has been sent",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "send",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="delivered",
|
||||
label="Delivered",
|
||||
description="Notification was successfully delivered",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="failed",
|
||||
label="Failed",
|
||||
description="Notification delivery failed",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# System Announcement Severity Choices
|
||||
# ============================================================================
|
||||
ANNOUNCEMENT_SEVERITIES = [
|
||||
RichChoice(
|
||||
value="info",
|
||||
label="Information",
|
||||
description="Informational announcement",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "info",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-200",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="warning",
|
||||
label="Warning",
|
||||
description="Warning announcement requiring attention",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-200",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
RichChoice(
|
||||
value="critical",
|
||||
label="Critical",
|
||||
description="Critical announcement requiring immediate attention",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "alert-octagon",
|
||||
"css_class": "bg-red-100 text-red-800 border-red-200",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.PRIORITY,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Notification Level Choices (for in-app notifications)
|
||||
# ============================================================================
|
||||
NOTIFICATION_LEVELS = [
|
||||
RichChoice(
|
||||
value="info",
|
||||
label="Info",
|
||||
description="Informational notification",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "info",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.NOTIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="success",
|
||||
label="Success",
|
||||
description="Success notification",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.NOTIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="warning",
|
||||
label="Warning",
|
||||
description="Warning notification",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "alert-triangle",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 3,
|
||||
},
|
||||
category=ChoiceCategory.NOTIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="error",
|
||||
label="Error",
|
||||
description="Error notification",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 4,
|
||||
},
|
||||
category=ChoiceCategory.NOTIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
def register_notifications_choices() -> None:
|
||||
"""Register all notifications domain choices with the global registry."""
|
||||
register_choices(
|
||||
name="notification_log_statuses",
|
||||
choices=NOTIFICATION_LOG_STATUSES,
|
||||
domain="notifications",
|
||||
description="Status options for notification logs",
|
||||
)
|
||||
register_choices(
|
||||
name="announcement_severities",
|
||||
choices=ANNOUNCEMENT_SEVERITIES,
|
||||
domain="notifications",
|
||||
description="Severity levels for system announcements",
|
||||
)
|
||||
register_choices(
|
||||
name="notification_levels",
|
||||
choices=NOTIFICATION_LEVELS,
|
||||
domain="notifications",
|
||||
description="Level options for in-app notifications",
|
||||
)
|
||||
|
||||
|
||||
# Auto-register choices when module is imported
|
||||
register_notifications_choices()
|
||||
@@ -0,0 +1,50 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 22:01
|
||||
|
||||
import apps.core.choices.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("notifications", "0002_add_notification_model"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name="notification",
|
||||
name="level",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="notification_levels",
|
||||
choices=[("info", "Info"), ("success", "Success"), ("warning", "Warning"), ("error", "Error")],
|
||||
default="info",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="notificationlog",
|
||||
name="status",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="notification_log_statuses",
|
||||
choices=[("pending", "Pending"), ("sent", "Sent"), ("delivered", "Delivered"), ("failed", "Failed")],
|
||||
default="pending",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="systemannouncement",
|
||||
name="severity",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="announcement_severities",
|
||||
choices=[("info", "Information"), ("warning", "Warning"), ("critical", "Critical")],
|
||||
default="info",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -14,6 +14,11 @@ Subscriber model is kept for backward compatibility but is optional.
|
||||
from django.conf import settings
|
||||
from django.db import models
|
||||
|
||||
from apps.core.choices.fields import RichChoiceField
|
||||
|
||||
# Import choices to ensure registration on app load
|
||||
from . import choices # noqa: F401
|
||||
|
||||
|
||||
class Subscriber(models.Model):
|
||||
"""
|
||||
@@ -100,12 +105,6 @@ class NotificationLog(models.Model):
|
||||
Audit log of sent notifications.
|
||||
"""
|
||||
|
||||
class Status(models.TextChoices):
|
||||
PENDING = "pending", "Pending"
|
||||
SENT = "sent", "Sent"
|
||||
DELIVERED = "delivered", "Delivered"
|
||||
FAILED = "failed", "Failed"
|
||||
|
||||
user = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
@@ -115,10 +114,11 @@ class NotificationLog(models.Model):
|
||||
workflow_id = models.CharField(max_length=100, db_index=True)
|
||||
notification_type = models.CharField(max_length=50)
|
||||
channel = models.CharField(max_length=20) # email, push, in_app, sms
|
||||
status = models.CharField(
|
||||
status = RichChoiceField(
|
||||
choice_group="notification_log_statuses",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
choices=Status.choices,
|
||||
default=Status.PENDING,
|
||||
default="pending",
|
||||
)
|
||||
payload = models.JSONField(default=dict, blank=True)
|
||||
error_message = models.TextField(blank=True)
|
||||
@@ -144,17 +144,13 @@ class SystemAnnouncement(models.Model):
|
||||
System-wide announcements.
|
||||
"""
|
||||
|
||||
class Severity(models.TextChoices):
|
||||
INFO = "info", "Information"
|
||||
WARNING = "warning", "Warning"
|
||||
CRITICAL = "critical", "Critical"
|
||||
|
||||
title = models.CharField(max_length=255)
|
||||
message = models.TextField()
|
||||
severity = models.CharField(
|
||||
severity = RichChoiceField(
|
||||
choice_group="announcement_severities",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
choices=Severity.choices,
|
||||
default=Severity.INFO,
|
||||
default="info",
|
||||
)
|
||||
action_url = models.URLField(blank=True)
|
||||
is_active = models.BooleanField(default=True)
|
||||
@@ -184,12 +180,6 @@ class Notification(models.Model):
|
||||
supporting both in-app and email notification channels.
|
||||
"""
|
||||
|
||||
class Level(models.TextChoices):
|
||||
INFO = "info", "Info"
|
||||
SUCCESS = "success", "Success"
|
||||
WARNING = "warning", "Warning"
|
||||
ERROR = "error", "Error"
|
||||
|
||||
# Who receives the notification
|
||||
recipient = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
@@ -207,10 +197,11 @@ class Notification(models.Model):
|
||||
# What happened
|
||||
verb = models.CharField(max_length=255)
|
||||
description = models.TextField(blank=True)
|
||||
level = models.CharField(
|
||||
level = RichChoiceField(
|
||||
choice_group="notification_levels",
|
||||
domain="notifications",
|
||||
max_length=20,
|
||||
choices=Level.choices,
|
||||
default=Level.INFO,
|
||||
default="info",
|
||||
)
|
||||
# The object that was acted upon (generic foreign key)
|
||||
action_object_content_type = models.ForeignKey(
|
||||
|
||||
@@ -4,6 +4,8 @@ Notification serializers.
|
||||
|
||||
from rest_framework import serializers
|
||||
|
||||
from apps.core.choices.serializers import RichChoiceSerializerField
|
||||
|
||||
from .models import NotificationLog, NotificationPreference, Subscriber, SystemAnnouncement
|
||||
|
||||
|
||||
@@ -131,8 +133,9 @@ class CreateAnnouncementSerializer(serializers.Serializer):
|
||||
|
||||
title = serializers.CharField(required=True, max_length=255)
|
||||
message = serializers.CharField(required=True)
|
||||
severity = serializers.ChoiceField(
|
||||
choices=["info", "warning", "critical"],
|
||||
severity = RichChoiceSerializerField(
|
||||
choice_group="announcement_severities",
|
||||
domain="notifications",
|
||||
default="info",
|
||||
)
|
||||
action_url = serializers.URLField(required=False, allow_blank=True)
|
||||
|
||||
@@ -238,6 +238,186 @@ PARKS_COMPANY_ROLES = [
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Person/Entity Type Choices (for Company model)
|
||||
# ============================================================================
|
||||
PERSON_TYPES = [
|
||||
RichChoice(
|
||||
value="INDIVIDUAL",
|
||||
label="Individual",
|
||||
description="Single person or sole proprietor",
|
||||
metadata={"color": "blue", "icon": "user", "css_class": "bg-blue-100 text-blue-800", "sort_order": 1},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="FIRM",
|
||||
label="Firm",
|
||||
description="Professional services firm",
|
||||
metadata={"color": "indigo", "icon": "briefcase", "css_class": "bg-indigo-100 text-indigo-800", "sort_order": 2},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="ORGANIZATION",
|
||||
label="Organization",
|
||||
description="Non-profit or member organization",
|
||||
metadata={"color": "green", "icon": "users", "css_class": "bg-green-100 text-green-800", "sort_order": 3},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="CORPORATION",
|
||||
label="Corporation",
|
||||
description="Incorporated business entity",
|
||||
metadata={"color": "purple", "icon": "building", "css_class": "bg-purple-100 text-purple-800", "sort_order": 4},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="PARTNERSHIP",
|
||||
label="Partnership",
|
||||
description="Business partnership",
|
||||
metadata={"color": "orange", "icon": "handshake", "css_class": "bg-orange-100 text-orange-800", "sort_order": 5},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="GOVERNMENT",
|
||||
label="Government Entity",
|
||||
description="Government agency or public entity",
|
||||
metadata={"color": "gray", "icon": "landmark", "css_class": "bg-gray-100 text-gray-800", "sort_order": 6},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Company Status Choices
|
||||
# ============================================================================
|
||||
COMPANY_STATUSES = [
|
||||
RichChoice(
|
||||
value="ACTIVE",
|
||||
label="Active",
|
||||
description="Company is currently operating",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 1,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="DEFUNCT",
|
||||
label="Defunct",
|
||||
description="Company no longer exists",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 2,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="MERGED",
|
||||
label="Merged",
|
||||
description="Company merged with another entity",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "git-merge",
|
||||
"css_class": "bg-purple-100 text-purple-800",
|
||||
"sort_order": 3,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="ACQUIRED",
|
||||
label="Acquired",
|
||||
description="Company was acquired by another entity",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "arrow-right-circle",
|
||||
"css_class": "bg-blue-100 text-blue-800",
|
||||
"sort_order": 4,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="RENAMED",
|
||||
label="Renamed",
|
||||
description="Company changed its name",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "edit",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 5,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="DORMANT",
|
||||
label="Dormant",
|
||||
description="Company is inactive but not dissolved",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "pause-circle",
|
||||
"css_class": "bg-gray-100 text-gray-800",
|
||||
"sort_order": 6,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Date Precision Choices (for parks domain - founding dates, opening dates, etc.)
|
||||
# ============================================================================
|
||||
DATE_PRECISION = [
|
||||
RichChoice(
|
||||
value="exact",
|
||||
label="Exact Date",
|
||||
description="Date is known exactly",
|
||||
metadata={"color": "green", "icon": "calendar", "sort_order": 1, "format": "YYYY-MM-DD"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="month",
|
||||
label="Month and Year",
|
||||
description="Only month and year are known",
|
||||
metadata={"color": "blue", "icon": "calendar", "sort_order": 2, "format": "YYYY-MM"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="year",
|
||||
label="Year Only",
|
||||
description="Only the year is known",
|
||||
metadata={"color": "yellow", "icon": "calendar", "sort_order": 3, "format": "YYYY"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="decade",
|
||||
label="Decade",
|
||||
description="Only the decade is known",
|
||||
metadata={"color": "orange", "icon": "calendar", "sort_order": 4, "format": "YYYYs"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="century",
|
||||
label="Century",
|
||||
description="Only the century is known",
|
||||
metadata={"color": "gray", "icon": "calendar", "sort_order": 5, "format": "YYc"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="approximate",
|
||||
label="Approximate",
|
||||
description="Date is approximate/estimated",
|
||||
metadata={"color": "gray", "icon": "help-circle", "sort_order": 6, "format": "~YYYY"},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
def register_parks_choices():
|
||||
"""Register all parks domain choices with the global registry"""
|
||||
@@ -266,6 +446,31 @@ def register_parks_choices():
|
||||
metadata={"domain": "parks", "type": "company_role"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="person_types",
|
||||
choices=PERSON_TYPES,
|
||||
domain="parks",
|
||||
description="Person/entity type classifications",
|
||||
metadata={"domain": "parks", "type": "person_type"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="company_statuses",
|
||||
choices=COMPANY_STATUSES,
|
||||
domain="parks",
|
||||
description="Company operational status options",
|
||||
metadata={"domain": "parks", "type": "company_status"},
|
||||
)
|
||||
|
||||
register_choices(
|
||||
name="date_precision",
|
||||
choices=DATE_PRECISION,
|
||||
domain="parks",
|
||||
description="Date precision options for parks domain",
|
||||
metadata={"domain": "parks", "type": "date_precision"},
|
||||
)
|
||||
|
||||
|
||||
# Auto-register choices when module is imported
|
||||
register_parks_choices()
|
||||
|
||||
|
||||
@@ -59,7 +59,7 @@ class ParkQuerySet(StatusQuerySet, ReviewableQuerySet, LocationQuerySet):
|
||||
"reviews",
|
||||
queryset=ParkReview.objects.select_related("user")
|
||||
.filter(is_published=True)
|
||||
.order_by("-created_at")[:10],
|
||||
.order_by("-created_at"),
|
||||
),
|
||||
"photos",
|
||||
)
|
||||
@@ -86,7 +86,8 @@ class ParkQuerySet(StatusQuerySet, ReviewableQuerySet, LocationQuerySet):
|
||||
|
||||
def for_map_display(self, *, bounds=None):
|
||||
"""Optimize for map display with minimal data."""
|
||||
queryset = self.select_related("operator").prefetch_related("location")
|
||||
# Use select_related for OneToOne relationship to enable values() access
|
||||
queryset = self.select_related("operator", "location")
|
||||
|
||||
if bounds:
|
||||
queryset = queryset.within_bounds(
|
||||
@@ -96,13 +97,15 @@ class ParkQuerySet(StatusQuerySet, ReviewableQuerySet, LocationQuerySet):
|
||||
west=bounds.west,
|
||||
)
|
||||
|
||||
# Note: location__latitude and location__longitude are @property
|
||||
# decorators, not actual DB fields. We access city/state/country which
|
||||
# are actual columns. For coordinates, callers should use the point field
|
||||
# or access the location object directly.
|
||||
return queryset.values(
|
||||
"id",
|
||||
"name",
|
||||
"slug",
|
||||
"status",
|
||||
"location__latitude",
|
||||
"location__longitude",
|
||||
"location__city",
|
||||
"location__state",
|
||||
"location__country",
|
||||
@@ -152,6 +155,10 @@ class ParkManager(StatusManager, ReviewableManager, LocationManager):
|
||||
"""Always prefetch location for park queries."""
|
||||
return self.get_queryset().with_location()
|
||||
|
||||
def search_autocomplete(self, *, query: str, limit: int = 10):
|
||||
"""Optimized search for autocomplete."""
|
||||
return self.get_queryset().search_autocomplete(query=query, limit=limit)
|
||||
|
||||
|
||||
class ParkAreaQuerySet(BaseQuerySet):
|
||||
"""QuerySet for ParkArea model."""
|
||||
@@ -284,25 +291,10 @@ class CompanyManager(BaseManager):
|
||||
def major_operators(self, *, min_parks: int = 5):
|
||||
return self.get_queryset().major_operators(min_parks=min_parks)
|
||||
|
||||
def manufacturers_with_ride_count(self):
|
||||
"""Get manufacturers with ride count annotation for list views."""
|
||||
return (
|
||||
self.get_queryset()
|
||||
.manufacturers()
|
||||
.annotate(ride_count=Count("manufactured_rides", distinct=True))
|
||||
.only("id", "name", "slug", "roles", "description")
|
||||
.order_by("name")
|
||||
)
|
||||
|
||||
def designers_with_ride_count(self):
|
||||
"""Get designers with ride count annotation for list views."""
|
||||
return (
|
||||
self.get_queryset()
|
||||
.filter(roles__contains=["DESIGNER"])
|
||||
.annotate(ride_count=Count("designed_rides", distinct=True))
|
||||
.only("id", "name", "slug", "roles", "description")
|
||||
.order_by("name")
|
||||
)
|
||||
# NOTE: manufacturers_with_ride_count and designers_with_ride_count were removed
|
||||
# because parks.Company doesn't have manufactured_rides/designed_rides relations.
|
||||
# Those relations exist on rides.Company, a separate model.
|
||||
# Use the rides app's company manager for ride-related company queries.
|
||||
|
||||
def operators_with_park_count(self):
|
||||
"""Get operators with park count annotation for list views."""
|
||||
|
||||
62
backend/apps/parks/migrations/0032_add_logo_image_id.py
Normal file
62
backend/apps/parks/migrations/0032_add_logo_image_id.py
Normal file
@@ -0,0 +1,62 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 19:38
|
||||
|
||||
import pgtrigger.compiler
|
||||
import pgtrigger.migrations
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("parks", "0031_add_photographer_to_photos"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
pgtrigger.migrations.RemoveTrigger(
|
||||
model_name="company",
|
||||
name="insert_insert",
|
||||
),
|
||||
pgtrigger.migrations.RemoveTrigger(
|
||||
model_name="company",
|
||||
name="update_update",
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="company",
|
||||
name="logo_image_id",
|
||||
field=models.CharField(blank=True, help_text="Cloudflare image ID for logo image", max_length=255),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="companyevent",
|
||||
name="logo_image_id",
|
||||
field=models.CharField(blank=True, help_text="Cloudflare image ID for logo image", max_length=255),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="company",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="insert_insert",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
func='INSERT INTO "parks_companyevent" ("average_rating", "banner_image_id", "banner_image_url", "card_image_id", "card_image_url", "created_at", "description", "founded_date", "founded_date_precision", "founded_year", "headquarters_location", "id", "is_test_data", "location_id", "logo_image_id", "logo_url", "name", "parks_count", "person_type", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "review_count", "rides_count", "roles", "slug", "source_url", "status", "updated_at", "website") VALUES (NEW."average_rating", NEW."banner_image_id", NEW."banner_image_url", NEW."card_image_id", NEW."card_image_url", NEW."created_at", NEW."description", NEW."founded_date", NEW."founded_date_precision", NEW."founded_year", NEW."headquarters_location", NEW."id", NEW."is_test_data", NEW."location_id", NEW."logo_image_id", NEW."logo_url", NEW."name", NEW."parks_count", NEW."person_type", _pgh_attach_context(), NOW(), \'insert\', NEW."id", NEW."review_count", NEW."rides_count", NEW."roles", NEW."slug", NEW."source_url", NEW."status", NEW."updated_at", NEW."website"); RETURN NULL;',
|
||||
hash="cd95b20dc19fbc63d9ebb0bab67279ce3670cb2b",
|
||||
operation="INSERT",
|
||||
pgid="pgtrigger_insert_insert_35b57",
|
||||
table="parks_company",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
pgtrigger.migrations.AddTrigger(
|
||||
model_name="company",
|
||||
trigger=pgtrigger.compiler.Trigger(
|
||||
name="update_update",
|
||||
sql=pgtrigger.compiler.UpsertTriggerSql(
|
||||
condition="WHEN (OLD.* IS DISTINCT FROM NEW.*)",
|
||||
func='INSERT INTO "parks_companyevent" ("average_rating", "banner_image_id", "banner_image_url", "card_image_id", "card_image_url", "created_at", "description", "founded_date", "founded_date_precision", "founded_year", "headquarters_location", "id", "is_test_data", "location_id", "logo_image_id", "logo_url", "name", "parks_count", "person_type", "pgh_context_id", "pgh_created_at", "pgh_label", "pgh_obj_id", "review_count", "rides_count", "roles", "slug", "source_url", "status", "updated_at", "website") VALUES (NEW."average_rating", NEW."banner_image_id", NEW."banner_image_url", NEW."card_image_id", NEW."card_image_url", NEW."created_at", NEW."description", NEW."founded_date", NEW."founded_date_precision", NEW."founded_year", NEW."headquarters_location", NEW."id", NEW."is_test_data", NEW."location_id", NEW."logo_image_id", NEW."logo_url", NEW."name", NEW."parks_count", NEW."person_type", _pgh_attach_context(), NOW(), \'update\', NEW."id", NEW."review_count", NEW."rides_count", NEW."roles", NEW."slug", NEW."source_url", NEW."status", NEW."updated_at", NEW."website"); RETURN NULL;',
|
||||
hash="c1fcc2920ab586cb06bec0624e50d2dab6bcb113",
|
||||
operation="UPDATE",
|
||||
pgid="pgtrigger_update_update_d3286",
|
||||
table="parks_company",
|
||||
when="AFTER",
|
||||
),
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,218 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 22:01
|
||||
|
||||
import apps.core.choices.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("parks", "0032_add_logo_image_id"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name="company",
|
||||
name="founded_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
domain="parks",
|
||||
help_text="Precision of the founding date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="company",
|
||||
name="person_type",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="person_types",
|
||||
choices=[
|
||||
("INDIVIDUAL", "Individual"),
|
||||
("FIRM", "Firm"),
|
||||
("ORGANIZATION", "Organization"),
|
||||
("CORPORATION", "Corporation"),
|
||||
("PARTNERSHIP", "Partnership"),
|
||||
("GOVERNMENT", "Government Entity"),
|
||||
],
|
||||
domain="parks",
|
||||
help_text="Type of entity (individual, firm, organization, etc.)",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="company",
|
||||
name="status",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="company_statuses",
|
||||
choices=[
|
||||
("ACTIVE", "Active"),
|
||||
("DEFUNCT", "Defunct"),
|
||||
("MERGED", "Merged"),
|
||||
("ACQUIRED", "Acquired"),
|
||||
("RENAMED", "Renamed"),
|
||||
("DORMANT", "Dormant"),
|
||||
],
|
||||
default="ACTIVE",
|
||||
domain="parks",
|
||||
help_text="Current operational status of the company",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="companyevent",
|
||||
name="founded_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
domain="parks",
|
||||
help_text="Precision of the founding date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="companyevent",
|
||||
name="person_type",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="person_types",
|
||||
choices=[
|
||||
("INDIVIDUAL", "Individual"),
|
||||
("FIRM", "Firm"),
|
||||
("ORGANIZATION", "Organization"),
|
||||
("CORPORATION", "Corporation"),
|
||||
("PARTNERSHIP", "Partnership"),
|
||||
("GOVERNMENT", "Government Entity"),
|
||||
],
|
||||
domain="parks",
|
||||
help_text="Type of entity (individual, firm, organization, etc.)",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="companyevent",
|
||||
name="status",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="company_statuses",
|
||||
choices=[
|
||||
("ACTIVE", "Active"),
|
||||
("DEFUNCT", "Defunct"),
|
||||
("MERGED", "Merged"),
|
||||
("ACQUIRED", "Acquired"),
|
||||
("RENAMED", "Renamed"),
|
||||
("DORMANT", "Dormant"),
|
||||
],
|
||||
default="ACTIVE",
|
||||
domain="parks",
|
||||
help_text="Current operational status of the company",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="park",
|
||||
name="closing_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
domain="parks",
|
||||
help_text="Precision of the closing date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="park",
|
||||
name="opening_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
domain="parks",
|
||||
help_text="Precision of the opening date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="parkevent",
|
||||
name="closing_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
domain="parks",
|
||||
help_text="Precision of the closing date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="parkevent",
|
||||
name="opening_date_precision",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
blank=True,
|
||||
choice_group="date_precision",
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
domain="parks",
|
||||
help_text="Precision of the opening date",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -26,34 +26,20 @@ class Company(TrackedModel):
|
||||
description = models.TextField(blank=True, help_text="Detailed company description")
|
||||
website = models.URLField(blank=True, help_text="Company website URL")
|
||||
|
||||
# Person/Entity Type (ported from legacy thrillwiki-87)
|
||||
PERSON_TYPES = [
|
||||
("INDIVIDUAL", "Individual"),
|
||||
("FIRM", "Firm"),
|
||||
("ORGANIZATION", "Organization"),
|
||||
("CORPORATION", "Corporation"),
|
||||
("PARTNERSHIP", "Partnership"),
|
||||
("GOVERNMENT", "Government Entity"),
|
||||
]
|
||||
person_type = models.CharField(
|
||||
# Person/Entity Type - using RichChoiceField
|
||||
person_type = RichChoiceField(
|
||||
choice_group="person_types",
|
||||
domain="parks",
|
||||
max_length=20,
|
||||
choices=PERSON_TYPES,
|
||||
blank=True,
|
||||
help_text="Type of entity (individual, firm, organization, etc.)",
|
||||
)
|
||||
|
||||
# Company Status (ported from legacy)
|
||||
COMPANY_STATUSES = [
|
||||
("ACTIVE", "Active"),
|
||||
("DEFUNCT", "Defunct"),
|
||||
("MERGED", "Merged"),
|
||||
("ACQUIRED", "Acquired"),
|
||||
("RENAMED", "Renamed"),
|
||||
("DORMANT", "Dormant"),
|
||||
]
|
||||
status = models.CharField(
|
||||
# Company Status - using RichChoiceField
|
||||
status = RichChoiceField(
|
||||
choice_group="company_statuses",
|
||||
domain="parks",
|
||||
max_length=20,
|
||||
choices=COMPANY_STATUSES,
|
||||
default="ACTIVE",
|
||||
help_text="Current operational status of the company",
|
||||
)
|
||||
@@ -61,17 +47,10 @@ class Company(TrackedModel):
|
||||
# Founding Information (enhanced from just founded_year)
|
||||
founded_year = models.PositiveIntegerField(blank=True, null=True, help_text="Year the company was founded")
|
||||
founded_date = models.DateField(blank=True, null=True, help_text="Full founding date if known")
|
||||
DATE_PRECISION_CHOICES = [
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
]
|
||||
founded_date_precision = models.CharField(
|
||||
founded_date_precision = RichChoiceField(
|
||||
choice_group="date_precision",
|
||||
domain="parks",
|
||||
max_length=20,
|
||||
choices=DATE_PRECISION_CHOICES,
|
||||
blank=True,
|
||||
help_text="Precision of the founding date",
|
||||
)
|
||||
@@ -82,6 +61,11 @@ class Company(TrackedModel):
|
||||
card_image_url = models.URLField(blank=True, help_text="Card/thumbnail image for listings")
|
||||
|
||||
# Image ID fields (for frontend submissions - Cloudflare image IDs)
|
||||
logo_image_id = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
help_text="Cloudflare image ID for logo image",
|
||||
)
|
||||
banner_image_id = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
|
||||
@@ -54,31 +54,19 @@ class Park(StateMachineMixin, TrackedModel):
|
||||
|
||||
# Details
|
||||
opening_date = models.DateField(null=True, blank=True, help_text="Opening date")
|
||||
opening_date_precision = models.CharField(
|
||||
opening_date_precision = RichChoiceField(
|
||||
choice_group="date_precision",
|
||||
domain="parks",
|
||||
max_length=20,
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
blank=True,
|
||||
help_text="Precision of the opening date",
|
||||
)
|
||||
closing_date = models.DateField(null=True, blank=True, help_text="Closing date")
|
||||
closing_date_precision = models.CharField(
|
||||
closing_date_precision = RichChoiceField(
|
||||
choice_group="date_precision",
|
||||
domain="parks",
|
||||
max_length=20,
|
||||
choices=[
|
||||
("exact", "Exact Date"),
|
||||
("month", "Month and Year"),
|
||||
("year", "Year Only"),
|
||||
("decade", "Decade"),
|
||||
("century", "Century"),
|
||||
("approximate", "Approximate"),
|
||||
],
|
||||
default="exact",
|
||||
blank=True,
|
||||
help_text="Precision of the closing date",
|
||||
|
||||
@@ -139,12 +139,13 @@ class CompanyQueryOptimizationTests(TestCase):
|
||||
self.assertIn("OPERATOR", company.roles)
|
||||
|
||||
def test_manufacturers_with_ride_count_includes_annotation(self):
|
||||
"""Verify manufacturers_with_ride_count adds ride_count annotation."""
|
||||
result = Company.objects.manufacturers_with_ride_count()
|
||||
if result.exists():
|
||||
first = result.first()
|
||||
# Should have ride_count attribute
|
||||
self.assertTrue(hasattr(first, "ride_count"))
|
||||
"""This test is skipped - method was removed from parks.CompanyManager.
|
||||
|
||||
parks.Company doesn't have manufactured_rides relation (that exists on
|
||||
rides.Company). Use rides app company queries for ride-related annotations.
|
||||
"""
|
||||
# Method removed - parks.Company is for operators/owners, not manufacturers
|
||||
pass
|
||||
|
||||
def test_operators_with_park_count_includes_annotation(self):
|
||||
"""Verify operators_with_park_count adds park count annotations."""
|
||||
|
||||
99
backend/apps/reviews/choices.py
Normal file
99
backend/apps/reviews/choices.py
Normal file
@@ -0,0 +1,99 @@
|
||||
"""
|
||||
Rich Choice Objects for Reviews Domain
|
||||
|
||||
This module defines all choice objects for the reviews domain,
|
||||
using the RichChoices pattern for consistent UI rendering and validation.
|
||||
"""
|
||||
|
||||
from apps.core.choices import ChoiceCategory, RichChoice
|
||||
from apps.core.choices.registry import register_choices
|
||||
|
||||
# ============================================================================
|
||||
# Rating Value Choices (1-5 stars)
|
||||
# ============================================================================
|
||||
RATING_VALUES = [
|
||||
RichChoice(
|
||||
value="1",
|
||||
label="1 Star",
|
||||
description="Poor rating",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "star",
|
||||
"css_class": "bg-red-100 text-red-800",
|
||||
"sort_order": 1,
|
||||
"numeric_value": 1,
|
||||
"star_count": 1,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="2",
|
||||
label="2 Stars",
|
||||
description="Below average rating",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "star",
|
||||
"css_class": "bg-orange-100 text-orange-800",
|
||||
"sort_order": 2,
|
||||
"numeric_value": 2,
|
||||
"star_count": 2,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="3",
|
||||
label="3 Stars",
|
||||
description="Average rating",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "star",
|
||||
"css_class": "bg-yellow-100 text-yellow-800",
|
||||
"sort_order": 3,
|
||||
"numeric_value": 3,
|
||||
"star_count": 3,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="4",
|
||||
label="4 Stars",
|
||||
description="Good rating",
|
||||
metadata={
|
||||
"color": "lime",
|
||||
"icon": "star",
|
||||
"css_class": "bg-lime-100 text-lime-800",
|
||||
"sort_order": 4,
|
||||
"numeric_value": 4,
|
||||
"star_count": 4,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="5",
|
||||
label="5 Stars",
|
||||
description="Excellent rating",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "star",
|
||||
"css_class": "bg-green-100 text-green-800",
|
||||
"sort_order": 5,
|
||||
"numeric_value": 5,
|
||||
"star_count": 5,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
def register_reviews_choices() -> None:
|
||||
"""Register all reviews domain choices with the global registry."""
|
||||
register_choices(
|
||||
name="rating_values",
|
||||
choices=RATING_VALUES,
|
||||
domain="reviews",
|
||||
description="Rating values for reviews (1-5 stars)",
|
||||
)
|
||||
|
||||
|
||||
# Auto-register choices when module is imported
|
||||
register_reviews_choices()
|
||||
@@ -0,0 +1,66 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-12 01:01
|
||||
|
||||
import apps.core.choices.fields
|
||||
import django.db.models.deletion
|
||||
from django.conf import settings
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("rides", "0039_add_photographer_to_photos"),
|
||||
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.CreateModel(
|
||||
name="RideSubType",
|
||||
fields=[
|
||||
("id", models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name="ID")),
|
||||
("created_at", models.DateTimeField(auto_now_add=True)),
|
||||
("updated_at", models.DateTimeField(auto_now=True)),
|
||||
(
|
||||
"name",
|
||||
models.CharField(help_text="Name of the ride sub-type (e.g., 'Flying Coaster')", max_length=100),
|
||||
),
|
||||
(
|
||||
"category",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="categories",
|
||||
choices=[
|
||||
("RC", "Roller Coaster"),
|
||||
("DR", "Dark Ride"),
|
||||
("FR", "Flat Ride"),
|
||||
("WR", "Water Ride"),
|
||||
("TR", "Transport Ride"),
|
||||
("OT", "Other"),
|
||||
],
|
||||
domain="rides",
|
||||
help_text="Ride category this sub-type belongs to",
|
||||
max_length=2,
|
||||
),
|
||||
),
|
||||
("description", models.TextField(blank=True, help_text="Description of this ride sub-type")),
|
||||
(
|
||||
"created_by",
|
||||
models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="User who created this sub-type",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="created_ride_sub_types",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Ride Sub-Type",
|
||||
"verbose_name_plural": "Ride Sub-Types",
|
||||
"ordering": ["category", "name"],
|
||||
"abstract": False,
|
||||
"unique_together": {("category", "name")},
|
||||
},
|
||||
),
|
||||
]
|
||||
@@ -17,12 +17,14 @@ from .rankings import RankingSnapshot, RidePairComparison, RideRanking
|
||||
from .reviews import RideReview
|
||||
from .rides import Ride, RideModel, RollerCoasterStats
|
||||
from .stats import DarkRideStats, FlatRideStats, KiddieRideStats, TransportationStats, WaterRideStats
|
||||
from .sub_types import RideSubType
|
||||
|
||||
__all__ = [
|
||||
# Primary models
|
||||
"Ride",
|
||||
"RideModel",
|
||||
"RideNameHistory",
|
||||
"RideSubType",
|
||||
"RollerCoasterStats",
|
||||
"WaterRideStats",
|
||||
"DarkRideStats",
|
||||
|
||||
67
backend/apps/rides/models/sub_types.py
Normal file
67
backend/apps/rides/models/sub_types.py
Normal file
@@ -0,0 +1,67 @@
|
||||
"""
|
||||
RideSubType model for categorizing rides by sub-type within each category.
|
||||
|
||||
This model replaces the legacy Supabase ride_sub_types table,
|
||||
providing a lookup table for ride sub-types (e.g., 'Flying Coaster',
|
||||
'Inverted Coaster' for roller coasters).
|
||||
"""
|
||||
|
||||
from django.conf import settings
|
||||
from django.db import models
|
||||
|
||||
from apps.core.choices import RichChoiceField
|
||||
from apps.core.models import TrackedModel
|
||||
|
||||
|
||||
class RideSubType(TrackedModel):
|
||||
"""
|
||||
Lookup table for ride sub-types categorized by ride category.
|
||||
|
||||
Examples:
|
||||
- Roller Coaster: Flying Coaster, Inverted Coaster, Dive Coaster
|
||||
- Water Ride: Log Flume, River Rapids, Splash Battle
|
||||
- Dark Ride: Trackless, Omnimover, Boat Ride
|
||||
"""
|
||||
|
||||
name = models.CharField(
|
||||
max_length=100,
|
||||
help_text="Name of the ride sub-type (e.g., 'Flying Coaster')",
|
||||
)
|
||||
category = RichChoiceField(
|
||||
choice_group="categories",
|
||||
domain="rides",
|
||||
max_length=2,
|
||||
help_text="Ride category this sub-type belongs to",
|
||||
)
|
||||
description = models.TextField(
|
||||
blank=True,
|
||||
help_text="Description of this ride sub-type",
|
||||
)
|
||||
created_by = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
related_name="created_ride_sub_types",
|
||||
help_text="User who created this sub-type",
|
||||
)
|
||||
|
||||
class Meta(TrackedModel.Meta):
|
||||
verbose_name = "Ride Sub-Type"
|
||||
verbose_name_plural = "Ride Sub-Types"
|
||||
ordering = ["category", "name"]
|
||||
unique_together = [["category", "name"]]
|
||||
|
||||
def __str__(self) -> str:
|
||||
return f"{self.name} ({self.get_category_display()})"
|
||||
|
||||
def get_category_display(self) -> str:
|
||||
"""Get human-readable category label."""
|
||||
from apps.core.choices.registry import get_choice_group
|
||||
|
||||
group = get_choice_group("categories", domain="rides")
|
||||
if group:
|
||||
for choice in group.choices:
|
||||
if choice.value == self.category:
|
||||
return choice.label
|
||||
return self.category
|
||||
@@ -690,7 +690,8 @@ class SmartRideLoader:
|
||||
if category in category_labels:
|
||||
return category_labels[category]
|
||||
else:
|
||||
raise ValueError(f"Unknown ride category: {category}")
|
||||
# Return original value as fallback for unknown categories
|
||||
return category.replace("_", " ").title()
|
||||
|
||||
def _get_status_label(self, status: str) -> str:
|
||||
"""Convert status code to human-readable label."""
|
||||
@@ -707,7 +708,8 @@ class SmartRideLoader:
|
||||
if status in status_labels:
|
||||
return status_labels[status]
|
||||
else:
|
||||
raise ValueError(f"Unknown ride status: {status}")
|
||||
# Return original value as fallback for unknown statuses
|
||||
return status.replace("_", " ").title()
|
||||
|
||||
def _get_rc_type_label(self, rc_type: str) -> str:
|
||||
"""Convert roller coaster type to human-readable label."""
|
||||
@@ -729,7 +731,8 @@ class SmartRideLoader:
|
||||
if rc_type in rc_type_labels:
|
||||
return rc_type_labels[rc_type]
|
||||
else:
|
||||
raise ValueError(f"Unknown roller coaster type: {rc_type}")
|
||||
# Return original value as fallback for unknown types
|
||||
return rc_type.replace("_", " ").title()
|
||||
|
||||
def _get_track_material_label(self, material: str) -> str:
|
||||
"""Convert track material to human-readable label."""
|
||||
@@ -741,7 +744,8 @@ class SmartRideLoader:
|
||||
if material in material_labels:
|
||||
return material_labels[material]
|
||||
else:
|
||||
raise ValueError(f"Unknown track material: {material}")
|
||||
# Return original value as fallback for unknown materials
|
||||
return material.replace("_", " ").title()
|
||||
|
||||
def _get_propulsion_system_label(self, propulsion_system: str) -> str:
|
||||
"""Convert propulsion system to human-readable label."""
|
||||
@@ -759,4 +763,6 @@ class SmartRideLoader:
|
||||
if propulsion_system in propulsion_labels:
|
||||
return propulsion_labels[propulsion_system]
|
||||
else:
|
||||
raise ValueError(f"Unknown propulsion system: {propulsion_system}")
|
||||
# Return original value as fallback for unknown propulsion systems
|
||||
return propulsion_system.replace("_", " ").title()
|
||||
|
||||
|
||||
@@ -6,7 +6,7 @@ from django.utils import timezone
|
||||
|
||||
from apps.core.utils import capture_and_log
|
||||
|
||||
from .models import Ride
|
||||
from .models import Ride, RideSubType
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
@@ -286,3 +286,59 @@ def track_ride_name_changes(sender, instance, **kwargs):
|
||||
except Exception as e:
|
||||
logger.exception(f"Failed to track name change for ride {instance.pk}: {e}")
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Auto-Create Ride Sub-Types on Ride Save
|
||||
# =============================================================================
|
||||
|
||||
|
||||
@receiver(post_save, sender=Ride)
|
||||
def auto_create_ride_sub_type(sender, instance, created, **kwargs):
|
||||
"""
|
||||
Automatically create a RideSubType entry when a ride is saved with a new sub-type value.
|
||||
|
||||
This integrates with the submission pipeline - when a ride submission with a new
|
||||
ride_sub_type value is approved, the sub-type automatically gets added to the
|
||||
lookup table for future autocomplete suggestions.
|
||||
|
||||
Args:
|
||||
sender: The Ride model class.
|
||||
instance: The Ride instance that was saved.
|
||||
created: Whether this is a new ride (not used, we check sub-types for all saves).
|
||||
"""
|
||||
# Skip if no ride_sub_type is set
|
||||
if not instance.ride_sub_type or not instance.ride_sub_type.strip():
|
||||
return
|
||||
|
||||
# Skip if no category is set (can't categorize the sub-type)
|
||||
if not instance.category:
|
||||
return
|
||||
|
||||
ride_sub_type_value = instance.ride_sub_type.strip()
|
||||
|
||||
try:
|
||||
# Check if this sub-type already exists for this category
|
||||
existing = RideSubType.objects.filter(
|
||||
name__iexact=ride_sub_type_value,
|
||||
category=instance.category
|
||||
).exists()
|
||||
|
||||
if not existing:
|
||||
# Create the new sub-type entry
|
||||
RideSubType.objects.create(
|
||||
name=ride_sub_type_value,
|
||||
category=instance.category,
|
||||
description=f"Auto-created from ride: {instance.name}",
|
||||
created_by=getattr(instance, 'created_by', None),
|
||||
)
|
||||
logger.info(
|
||||
f"Auto-created RideSubType '{ride_sub_type_value}' for category "
|
||||
f"'{instance.category}' from ride '{instance.name}'"
|
||||
)
|
||||
except Exception as e:
|
||||
# Non-critical error - log but don't fail the ride save
|
||||
logger.warning(
|
||||
f"Failed to auto-create RideSubType for ride {instance.pk}: {e}"
|
||||
)
|
||||
|
||||
|
||||
|
||||
358
backend/apps/support/choices.py
Normal file
358
backend/apps/support/choices.py
Normal file
@@ -0,0 +1,358 @@
|
||||
"""
|
||||
Rich Choice Objects for Support Domain
|
||||
|
||||
This module defines all choice objects for the support domain,
|
||||
using the RichChoices pattern for consistent UI rendering and validation.
|
||||
|
||||
Note: Values are kept lowercase for backward compatibility with existing data.
|
||||
"""
|
||||
|
||||
from apps.core.choices import ChoiceCategory, RichChoice
|
||||
from apps.core.choices.registry import register_choices
|
||||
|
||||
# ============================================================================
|
||||
# Ticket Status Choices
|
||||
# ============================================================================
|
||||
|
||||
TICKET_STATUSES = [
|
||||
RichChoice(
|
||||
value="open",
|
||||
label="Open",
|
||||
description="Ticket is awaiting response",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "inbox",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-200",
|
||||
"sort_order": 1,
|
||||
"can_transition_to": ["in_progress", "closed"],
|
||||
"is_actionable": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="in_progress",
|
||||
label="In Progress",
|
||||
description="Ticket is being worked on",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "clock",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-200",
|
||||
"sort_order": 2,
|
||||
"can_transition_to": ["closed"],
|
||||
"is_actionable": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="closed",
|
||||
label="Closed",
|
||||
description="Ticket has been resolved",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800 border-green-200",
|
||||
"sort_order": 3,
|
||||
"can_transition_to": ["open"],
|
||||
"is_actionable": False,
|
||||
"is_final": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Ticket Category Choices
|
||||
# ============================================================================
|
||||
|
||||
TICKET_CATEGORIES = [
|
||||
RichChoice(
|
||||
value="general",
|
||||
label="General Inquiry",
|
||||
description="General questions or feedback",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "chat-bubble-left",
|
||||
"css_class": "bg-gray-100 text-gray-800 border-gray-200",
|
||||
"sort_order": 1,
|
||||
"default_priority": "low",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="bug",
|
||||
label="Bug Report",
|
||||
description="Report a bug or issue with the platform",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "bug-ant",
|
||||
"css_class": "bg-red-100 text-red-800 border-red-200",
|
||||
"sort_order": 2,
|
||||
"default_priority": "high",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="partnership",
|
||||
label="Partnership",
|
||||
description="Partnership or collaboration inquiries",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "handshake",
|
||||
"css_class": "bg-purple-100 text-purple-800 border-purple-200",
|
||||
"sort_order": 3,
|
||||
"default_priority": "medium",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="press",
|
||||
label="Press/Media",
|
||||
description="Press inquiries and media requests",
|
||||
metadata={
|
||||
"color": "indigo",
|
||||
"icon": "newspaper",
|
||||
"css_class": "bg-indigo-100 text-indigo-800 border-indigo-200",
|
||||
"sort_order": 4,
|
||||
"default_priority": "medium",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="data",
|
||||
label="Data Correction",
|
||||
description="Request corrections to park or ride data",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "pencil-square",
|
||||
"css_class": "bg-orange-100 text-orange-800 border-orange-200",
|
||||
"sort_order": 5,
|
||||
"default_priority": "medium",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="account",
|
||||
label="Account Issue",
|
||||
description="Account-related problems or requests",
|
||||
metadata={
|
||||
"color": "cyan",
|
||||
"icon": "user-circle",
|
||||
"css_class": "bg-cyan-100 text-cyan-800 border-cyan-200",
|
||||
"sort_order": 6,
|
||||
"default_priority": "high",
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Report Type Choices (for user-submitted reports about content issues)
|
||||
# ============================================================================
|
||||
REPORT_TYPES = [
|
||||
RichChoice(
|
||||
value="inaccurate",
|
||||
label="Inaccurate Information",
|
||||
description="Information is factually incorrect",
|
||||
metadata={
|
||||
"color": "orange",
|
||||
"icon": "alert-circle",
|
||||
"css_class": "bg-orange-100 text-orange-800 border-orange-200",
|
||||
"sort_order": 1,
|
||||
"requires_evidence": True,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="inappropriate",
|
||||
label="Inappropriate Content",
|
||||
description="Content is offensive or inappropriate",
|
||||
metadata={
|
||||
"color": "red",
|
||||
"icon": "flag",
|
||||
"css_class": "bg-red-100 text-red-800 border-red-200",
|
||||
"sort_order": 2,
|
||||
"requires_evidence": False,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="spam",
|
||||
label="Spam",
|
||||
description="Content is spam or promotional",
|
||||
metadata={
|
||||
"color": "purple",
|
||||
"icon": "mail-x",
|
||||
"css_class": "bg-purple-100 text-purple-800 border-purple-200",
|
||||
"sort_order": 3,
|
||||
"requires_evidence": False,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="copyright",
|
||||
label="Copyright Violation",
|
||||
description="Content violates copyright",
|
||||
metadata={
|
||||
"color": "indigo",
|
||||
"icon": "shield-alert",
|
||||
"css_class": "bg-indigo-100 text-indigo-800 border-indigo-200",
|
||||
"sort_order": 4,
|
||||
"requires_evidence": True,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="duplicate",
|
||||
label="Duplicate Content",
|
||||
description="Content duplicates existing entry",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "copy",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-200",
|
||||
"sort_order": 5,
|
||||
"requires_evidence": True,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="other",
|
||||
label="Other",
|
||||
description="Other issue not covered above",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "help-circle",
|
||||
"css_class": "bg-gray-100 text-gray-800 border-gray-200",
|
||||
"sort_order": 6,
|
||||
"requires_evidence": False,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
# ============================================================================
|
||||
# Report Status Choices
|
||||
# ============================================================================
|
||||
REPORT_STATUSES = [
|
||||
RichChoice(
|
||||
value="pending",
|
||||
label="Pending",
|
||||
description="Report is awaiting review",
|
||||
metadata={
|
||||
"color": "yellow",
|
||||
"icon": "clock",
|
||||
"css_class": "bg-yellow-100 text-yellow-800 border-yellow-200",
|
||||
"sort_order": 1,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="investigating",
|
||||
label="Investigating",
|
||||
description="Report is being investigated",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "search",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-200",
|
||||
"sort_order": 2,
|
||||
"is_active": True,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="resolved",
|
||||
label="Resolved",
|
||||
description="Report has been resolved with action taken",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "check-circle",
|
||||
"css_class": "bg-green-100 text-green-800 border-green-200",
|
||||
"sort_order": 3,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
RichChoice(
|
||||
value="dismissed",
|
||||
label="Dismissed",
|
||||
description="Report was dismissed as invalid or duplicate",
|
||||
metadata={
|
||||
"color": "gray",
|
||||
"icon": "x-circle",
|
||||
"css_class": "bg-gray-100 text-gray-800 border-gray-200",
|
||||
"sort_order": 4,
|
||||
"is_active": False,
|
||||
},
|
||||
category=ChoiceCategory.STATUS,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
# ============================================================================
|
||||
# Email Direction Choices
|
||||
# ============================================================================
|
||||
EMAIL_DIRECTIONS = [
|
||||
RichChoice(
|
||||
value="inbound",
|
||||
label="Inbound",
|
||||
description="Email received from user",
|
||||
metadata={
|
||||
"color": "blue",
|
||||
"icon": "arrow-down-left",
|
||||
"css_class": "bg-blue-100 text-blue-800 border-blue-200",
|
||||
"sort_order": 1,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
RichChoice(
|
||||
value="outbound",
|
||||
label="Outbound",
|
||||
description="Email sent to user",
|
||||
metadata={
|
||||
"color": "green",
|
||||
"icon": "arrow-up-right",
|
||||
"css_class": "bg-green-100 text-green-800 border-green-200",
|
||||
"sort_order": 2,
|
||||
},
|
||||
category=ChoiceCategory.CLASSIFICATION,
|
||||
),
|
||||
]
|
||||
|
||||
|
||||
def register_support_choices() -> None:
|
||||
"""Register all support domain choices with the global registry."""
|
||||
register_choices(
|
||||
"ticket_statuses",
|
||||
TICKET_STATUSES,
|
||||
domain="support",
|
||||
description="Status options for support tickets",
|
||||
)
|
||||
register_choices(
|
||||
"ticket_categories",
|
||||
TICKET_CATEGORIES,
|
||||
domain="support",
|
||||
description="Category options for support tickets",
|
||||
)
|
||||
register_choices(
|
||||
"report_types",
|
||||
REPORT_TYPES,
|
||||
domain="support",
|
||||
description="Type options for user-submitted reports",
|
||||
)
|
||||
register_choices(
|
||||
"report_statuses",
|
||||
REPORT_STATUSES,
|
||||
domain="support",
|
||||
description="Status options for user-submitted reports",
|
||||
)
|
||||
register_choices(
|
||||
"email_directions",
|
||||
EMAIL_DIRECTIONS,
|
||||
domain="support",
|
||||
description="Direction options for email threads",
|
||||
)
|
||||
|
||||
|
||||
# Auto-register choices when module is imported
|
||||
register_support_choices()
|
||||
|
||||
@@ -0,0 +1,48 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 19:33
|
||||
|
||||
import apps.core.choices.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("support", "0003_add_incident_and_report_models"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name="ticket",
|
||||
name="category",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="ticket_categories",
|
||||
choices=[
|
||||
("general", "General Inquiry"),
|
||||
("bug", "Bug Report"),
|
||||
("partnership", "Partnership"),
|
||||
("press", "Press/Media"),
|
||||
("data", "Data Correction"),
|
||||
("account", "Account Issue"),
|
||||
],
|
||||
db_index=True,
|
||||
default="general",
|
||||
domain="support",
|
||||
help_text="Category of the ticket",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="ticket",
|
||||
name="status",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="ticket_statuses",
|
||||
choices=[("open", "Open"), ("in_progress", "In Progress"), ("closed", "Closed")],
|
||||
db_index=True,
|
||||
default="open",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,53 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-10 22:01
|
||||
|
||||
import apps.core.choices.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("support", "0004_alter_ticket_category_alter_ticket_status"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name="report",
|
||||
name="report_type",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="report_types",
|
||||
choices=[
|
||||
("inaccurate", "Inaccurate Information"),
|
||||
("inappropriate", "Inappropriate Content"),
|
||||
("spam", "Spam"),
|
||||
("copyright", "Copyright Violation"),
|
||||
("duplicate", "Duplicate Content"),
|
||||
("other", "Other"),
|
||||
],
|
||||
db_index=True,
|
||||
domain="support",
|
||||
help_text="Type of issue being reported",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name="report",
|
||||
name="status",
|
||||
field=apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="report_statuses",
|
||||
choices=[
|
||||
("pending", "Pending"),
|
||||
("investigating", "Investigating"),
|
||||
("resolved", "Resolved"),
|
||||
("dismissed", "Dismissed"),
|
||||
],
|
||||
db_index=True,
|
||||
default="pending",
|
||||
domain="support",
|
||||
help_text="Current status of the report",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,147 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-11 20:42
|
||||
|
||||
import apps.core.choices.fields
|
||||
import django.db.models.deletion
|
||||
from django.conf import settings
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("support", "0005_alter_report_report_type_alter_report_status"),
|
||||
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.CreateModel(
|
||||
name="EmailThread",
|
||||
fields=[
|
||||
("id", models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name="ID")),
|
||||
("created_at", models.DateTimeField(auto_now_add=True)),
|
||||
("updated_at", models.DateTimeField(auto_now=True)),
|
||||
(
|
||||
"message_id",
|
||||
models.CharField(blank=True, help_text="Email message ID for threading", max_length=255),
|
||||
),
|
||||
("from_email", models.EmailField(help_text="Sender email address", max_length=254)),
|
||||
("to_email", models.EmailField(help_text="Recipient email address", max_length=254)),
|
||||
("subject", models.CharField(max_length=255)),
|
||||
("body_text", models.TextField(help_text="Plain text email body")),
|
||||
(
|
||||
"direction",
|
||||
apps.core.choices.fields.RichChoiceField(
|
||||
allow_deprecated=False,
|
||||
choice_group="email_directions",
|
||||
choices=[("inbound", "Inbound"), ("outbound", "Outbound")],
|
||||
domain="support",
|
||||
help_text="Whether email is inbound or outbound",
|
||||
max_length=10,
|
||||
),
|
||||
),
|
||||
],
|
||||
options={
|
||||
"verbose_name": "Email Thread",
|
||||
"verbose_name_plural": "Email Threads",
|
||||
"ordering": ["created_at"],
|
||||
"abstract": False,
|
||||
},
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="admin_notes",
|
||||
field=models.TextField(blank=True, help_text="Internal notes for administrators"),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="archived_at",
|
||||
field=models.DateTimeField(blank=True, help_text="When the ticket was archived", null=True),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="archived_by",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Staff member who archived this ticket",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="archived_tickets",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="assigned_to",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Staff member assigned to this ticket",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="assigned_tickets",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="name",
|
||||
field=models.CharField(
|
||||
blank=True, help_text="Name of the submitter (for anonymous tickets)", max_length=255
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="resolved_at",
|
||||
field=models.DateTimeField(blank=True, help_text="When the ticket was resolved", null=True),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="resolved_by",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Staff member who resolved this ticket",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="resolved_tickets",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="ticket",
|
||||
name="ticket_number",
|
||||
field=models.CharField(blank=True, help_text="Human-readable ticket number", max_length=20, unique=True),
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="ticket",
|
||||
index=models.Index(fields=["status", "created_at"], name="support_tic_status_d0b46e_idx"),
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="ticket",
|
||||
index=models.Index(fields=["ticket_number"], name="support_tic_ticket__d87f40_idx"),
|
||||
),
|
||||
migrations.AddIndex(
|
||||
model_name="ticket",
|
||||
index=models.Index(fields=["archived_at"], name="support_tic_archive_8fe8c5_idx"),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="emailthread",
|
||||
name="sent_by",
|
||||
field=models.ForeignKey(
|
||||
blank=True,
|
||||
help_text="Staff member who sent this email (for outbound)",
|
||||
null=True,
|
||||
on_delete=django.db.models.deletion.SET_NULL,
|
||||
related_name="sent_email_threads",
|
||||
to=settings.AUTH_USER_MODEL,
|
||||
),
|
||||
),
|
||||
migrations.AddField(
|
||||
model_name="emailthread",
|
||||
name="ticket",
|
||||
field=models.ForeignKey(
|
||||
help_text="Associated support ticket",
|
||||
on_delete=django.db.models.deletion.CASCADE,
|
||||
related_name="email_threads",
|
||||
to="support.ticket",
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -0,0 +1,27 @@
|
||||
# Generated by Django 5.2.10 on 2026-01-11 21:19
|
||||
|
||||
import apps.core.state_machine.fields
|
||||
from django.db import migrations
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
("support", "0006_add_ticket_admin_fields_and_email_threads"),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name="ticket",
|
||||
name="status",
|
||||
field=apps.core.state_machine.fields.RichFSMField(
|
||||
allow_deprecated=False,
|
||||
choice_group="ticket_statuses",
|
||||
choices=[("open", "Open"), ("in_progress", "In Progress"), ("closed", "Closed")],
|
||||
db_index=True,
|
||||
default="open",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
),
|
||||
),
|
||||
]
|
||||
@@ -1,35 +1,26 @@
|
||||
from django.conf import settings
|
||||
from django.core.exceptions import ValidationError
|
||||
from django.db import models
|
||||
from django.utils import timezone
|
||||
|
||||
from apps.core.choices.fields import RichChoiceField
|
||||
from apps.core.history import TrackedModel
|
||||
from apps.core.state_machine import RichFSMField, StateMachineMixin
|
||||
|
||||
# Import choices to ensure registration on app load
|
||||
from . import choices # noqa: F401
|
||||
|
||||
|
||||
class Ticket(TrackedModel):
|
||||
STATUS_OPEN = "open"
|
||||
STATUS_IN_PROGRESS = "in_progress"
|
||||
STATUS_CLOSED = "closed"
|
||||
class Ticket(StateMachineMixin, TrackedModel):
|
||||
"""
|
||||
Support ticket model with FSM-managed status transitions.
|
||||
|
||||
Status workflow:
|
||||
open -> in_progress -> closed
|
||||
-> open (reopen)
|
||||
"""
|
||||
|
||||
STATUS_CHOICES = [
|
||||
(STATUS_OPEN, "Open"),
|
||||
(STATUS_IN_PROGRESS, "In Progress"),
|
||||
(STATUS_CLOSED, "Closed"),
|
||||
]
|
||||
|
||||
CATEGORY_GENERAL = "general"
|
||||
CATEGORY_BUG = "bug"
|
||||
CATEGORY_PARTNERSHIP = "partnership"
|
||||
CATEGORY_PRESS = "press"
|
||||
CATEGORY_DATA = "data"
|
||||
CATEGORY_ACCOUNT = "account"
|
||||
|
||||
CATEGORY_CHOICES = [
|
||||
(CATEGORY_GENERAL, "General Inquiry"),
|
||||
(CATEGORY_BUG, "Bug Report"),
|
||||
(CATEGORY_PARTNERSHIP, "Partnership"),
|
||||
(CATEGORY_PRESS, "Press/Media"),
|
||||
(CATEGORY_DATA, "Data Correction"),
|
||||
(CATEGORY_ACCOUNT, "Account Issue"),
|
||||
]
|
||||
state_field_name = "status"
|
||||
|
||||
user = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
@@ -40,10 +31,18 @@ class Ticket(TrackedModel):
|
||||
help_text="User who submitted the ticket (optional)",
|
||||
)
|
||||
|
||||
category = models.CharField(
|
||||
# Submitter info (for anonymous/guest tickets)
|
||||
name = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
help_text="Name of the submitter (for anonymous tickets)",
|
||||
)
|
||||
|
||||
category = RichChoiceField(
|
||||
choice_group="ticket_categories",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
choices=CATEGORY_CHOICES,
|
||||
default=CATEGORY_GENERAL,
|
||||
default="general",
|
||||
db_index=True,
|
||||
help_text="Category of the ticket",
|
||||
)
|
||||
@@ -51,22 +50,259 @@ class Ticket(TrackedModel):
|
||||
message = models.TextField()
|
||||
email = models.EmailField(help_text="Contact email", blank=True)
|
||||
|
||||
status = models.CharField(max_length=20, choices=STATUS_CHOICES, default=STATUS_OPEN, db_index=True)
|
||||
# FSM-managed status field
|
||||
status = RichFSMField(
|
||||
choice_group="ticket_statuses",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
default="open",
|
||||
db_index=True,
|
||||
)
|
||||
|
||||
# Human-readable ticket number (e.g., TKT-00001)
|
||||
ticket_number = models.CharField(
|
||||
max_length=20,
|
||||
unique=True,
|
||||
blank=True,
|
||||
help_text="Human-readable ticket number",
|
||||
)
|
||||
|
||||
# Admin management fields
|
||||
admin_notes = models.TextField(
|
||||
blank=True,
|
||||
help_text="Internal notes for administrators",
|
||||
)
|
||||
assigned_to = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
related_name="assigned_tickets",
|
||||
help_text="Staff member assigned to this ticket",
|
||||
)
|
||||
|
||||
# Resolution tracking
|
||||
resolved_at = models.DateTimeField(
|
||||
null=True,
|
||||
blank=True,
|
||||
help_text="When the ticket was resolved",
|
||||
)
|
||||
resolved_by = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
related_name="resolved_tickets",
|
||||
help_text="Staff member who resolved this ticket",
|
||||
)
|
||||
|
||||
# Archive functionality
|
||||
archived_at = models.DateTimeField(
|
||||
null=True,
|
||||
blank=True,
|
||||
help_text="When the ticket was archived",
|
||||
)
|
||||
archived_by = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
related_name="archived_tickets",
|
||||
help_text="Staff member who archived this ticket",
|
||||
)
|
||||
|
||||
class Meta(TrackedModel.Meta):
|
||||
verbose_name = "Ticket"
|
||||
verbose_name_plural = "Tickets"
|
||||
ordering = ["-created_at"]
|
||||
indexes = [
|
||||
models.Index(fields=["status", "created_at"]),
|
||||
models.Index(fields=["ticket_number"]),
|
||||
models.Index(fields=["archived_at"]),
|
||||
]
|
||||
|
||||
def __str__(self):
|
||||
return f"[{self.get_category_display()}] {self.subject}"
|
||||
return f"[{self.ticket_number}] {self.subject}"
|
||||
|
||||
def save(self, *args, **kwargs):
|
||||
# If user is set but email is empty, autofill from user
|
||||
if self.user and not self.email:
|
||||
self.email = self.user.email
|
||||
# Generate ticket number if not set
|
||||
if not self.ticket_number:
|
||||
self.ticket_number = self._generate_ticket_number()
|
||||
super().save(*args, **kwargs)
|
||||
|
||||
def _generate_ticket_number(self):
|
||||
"""Generate a unique ticket number like TKT-00001."""
|
||||
import uuid
|
||||
|
||||
# Use last 8 chars of a UUID for uniqueness
|
||||
suffix = uuid.uuid4().hex[:8].upper()
|
||||
return f"TKT-{suffix}"
|
||||
|
||||
# =========================================================================
|
||||
# FSM Transition Methods
|
||||
# =========================================================================
|
||||
|
||||
def start_progress(self, user=None) -> None:
|
||||
"""
|
||||
Start working on this ticket.
|
||||
Transition: open -> in_progress
|
||||
|
||||
Args:
|
||||
user: The staff member starting work on the ticket
|
||||
"""
|
||||
if self.status != "open":
|
||||
raise ValidationError(
|
||||
f"Cannot start progress: current status is {self.status}, expected open"
|
||||
)
|
||||
|
||||
self.status = "in_progress"
|
||||
if user and user.is_staff:
|
||||
self.assigned_to = user
|
||||
self.save()
|
||||
|
||||
def close(self, user=None, notes: str = "") -> None:
|
||||
"""
|
||||
Close/resolve this ticket.
|
||||
Transition: in_progress -> closed
|
||||
|
||||
Args:
|
||||
user: The staff member closing the ticket
|
||||
notes: Optional resolution notes
|
||||
"""
|
||||
if self.status not in ("open", "in_progress"):
|
||||
raise ValidationError(
|
||||
f"Cannot close: current status is {self.status}, expected open or in_progress"
|
||||
)
|
||||
|
||||
self.status = "closed"
|
||||
self.resolved_at = timezone.now()
|
||||
if user:
|
||||
self.resolved_by = user
|
||||
if notes:
|
||||
self.admin_notes = f"{self.admin_notes}\n\n[CLOSED] {notes}".strip()
|
||||
self.save()
|
||||
|
||||
def reopen(self, user=None, reason: str = "") -> None:
|
||||
"""
|
||||
Reopen a closed ticket.
|
||||
Transition: closed -> open
|
||||
|
||||
Args:
|
||||
user: The staff member reopening the ticket
|
||||
reason: Reason for reopening
|
||||
"""
|
||||
if self.status != "closed":
|
||||
raise ValidationError(
|
||||
f"Cannot reopen: current status is {self.status}, expected closed"
|
||||
)
|
||||
|
||||
self.status = "open"
|
||||
self.resolved_at = None
|
||||
self.resolved_by = None
|
||||
if reason:
|
||||
self.admin_notes = f"{self.admin_notes}\n\n[REOPENED] {reason}".strip()
|
||||
self.save()
|
||||
|
||||
def archive(self, user=None, reason: str = "") -> None:
|
||||
"""
|
||||
Archive this ticket.
|
||||
Can be called from any status.
|
||||
|
||||
Args:
|
||||
user: The staff member archiving the ticket
|
||||
reason: Reason for archiving
|
||||
"""
|
||||
if self.archived_at:
|
||||
raise ValidationError("Ticket is already archived")
|
||||
|
||||
self.archived_at = timezone.now()
|
||||
if user:
|
||||
self.archived_by = user
|
||||
if reason:
|
||||
self.admin_notes = f"{self.admin_notes}\n\n[ARCHIVED] {reason}".strip()
|
||||
self.save()
|
||||
|
||||
def unarchive(self, user=None) -> None:
|
||||
"""
|
||||
Restore an archived ticket.
|
||||
|
||||
Args:
|
||||
user: The staff member unarchiving the ticket
|
||||
"""
|
||||
if not self.archived_at:
|
||||
raise ValidationError("Ticket is not archived")
|
||||
|
||||
self.archived_at = None
|
||||
self.archived_by = None
|
||||
self.admin_notes = f"{self.admin_notes}\n\n[UNARCHIVED] by {user.username if user else 'system'}".strip()
|
||||
self.save()
|
||||
|
||||
# =========================================================================
|
||||
# Computed Properties
|
||||
# =========================================================================
|
||||
|
||||
@property
|
||||
def thread_id(self):
|
||||
"""Return the ID for email thread association."""
|
||||
return str(self.id)
|
||||
|
||||
@property
|
||||
def response_count(self):
|
||||
"""Return number of email thread responses."""
|
||||
return self.email_threads.count()
|
||||
|
||||
@property
|
||||
def last_admin_response_at(self):
|
||||
"""Return timestamp of last admin response."""
|
||||
last_outbound = self.email_threads.filter(direction="outbound").order_by("-created_at").first()
|
||||
return last_outbound.created_at if last_outbound else None
|
||||
|
||||
|
||||
class EmailThread(TrackedModel):
|
||||
"""Email thread entries for ticket conversations."""
|
||||
|
||||
ticket = models.ForeignKey(
|
||||
Ticket,
|
||||
on_delete=models.CASCADE,
|
||||
related_name="email_threads",
|
||||
help_text="Associated support ticket",
|
||||
)
|
||||
message_id = models.CharField(
|
||||
max_length=255,
|
||||
blank=True,
|
||||
help_text="Email message ID for threading",
|
||||
)
|
||||
from_email = models.EmailField(help_text="Sender email address")
|
||||
to_email = models.EmailField(help_text="Recipient email address")
|
||||
subject = models.CharField(max_length=255)
|
||||
body_text = models.TextField(help_text="Plain text email body")
|
||||
|
||||
direction = RichChoiceField(
|
||||
choice_group="email_directions",
|
||||
domain="support",
|
||||
max_length=10,
|
||||
help_text="Whether email is inbound or outbound",
|
||||
)
|
||||
sent_by = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
on_delete=models.SET_NULL,
|
||||
null=True,
|
||||
blank=True,
|
||||
related_name="sent_email_threads",
|
||||
help_text="Staff member who sent this email (for outbound)",
|
||||
)
|
||||
|
||||
class Meta(TrackedModel.Meta):
|
||||
verbose_name = "Email Thread"
|
||||
verbose_name_plural = "Email Threads"
|
||||
ordering = ["created_at"]
|
||||
|
||||
def __str__(self):
|
||||
return f"[{self.direction}] {self.subject}"
|
||||
|
||||
|
||||
class Report(TrackedModel):
|
||||
"""
|
||||
@@ -76,20 +312,6 @@ class Report(TrackedModel):
|
||||
(parks, rides, reviews, etc.) for moderator review.
|
||||
"""
|
||||
|
||||
class ReportType(models.TextChoices):
|
||||
INACCURATE = "inaccurate", "Inaccurate Information"
|
||||
INAPPROPRIATE = "inappropriate", "Inappropriate Content"
|
||||
SPAM = "spam", "Spam"
|
||||
COPYRIGHT = "copyright", "Copyright Violation"
|
||||
DUPLICATE = "duplicate", "Duplicate Content"
|
||||
OTHER = "other", "Other"
|
||||
|
||||
class Status(models.TextChoices):
|
||||
PENDING = "pending", "Pending"
|
||||
INVESTIGATING = "investigating", "Investigating"
|
||||
RESOLVED = "resolved", "Resolved"
|
||||
DISMISSED = "dismissed", "Dismissed"
|
||||
|
||||
# Reporter (optional for anonymous reports)
|
||||
reporter = models.ForeignKey(
|
||||
settings.AUTH_USER_MODEL,
|
||||
@@ -114,20 +336,22 @@ class Report(TrackedModel):
|
||||
# It's a convenience for accessing the related object
|
||||
# content_object = GenericForeignKey("content_type", "object_id")
|
||||
|
||||
# Report details
|
||||
report_type = models.CharField(
|
||||
# Report details - now using RichChoiceField
|
||||
report_type = RichChoiceField(
|
||||
choice_group="report_types",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
choices=ReportType.choices,
|
||||
db_index=True,
|
||||
help_text="Type of issue being reported",
|
||||
)
|
||||
reason = models.TextField(
|
||||
help_text="Detailed description of the issue",
|
||||
)
|
||||
status = models.CharField(
|
||||
status = RichChoiceField(
|
||||
choice_group="report_statuses",
|
||||
domain="support",
|
||||
max_length=20,
|
||||
choices=Status.choices,
|
||||
default=Status.PENDING,
|
||||
default="pending",
|
||||
db_index=True,
|
||||
help_text="Current status of the report",
|
||||
)
|
||||
@@ -166,5 +390,6 @@ class Report(TrackedModel):
|
||||
|
||||
@property
|
||||
def is_resolved(self) -> bool:
|
||||
return self.status in (self.Status.RESOLVED, self.Status.DISMISSED)
|
||||
return self.status in ("resolved", "dismissed")
|
||||
|
||||
|
||||
|
||||
@@ -1,31 +1,126 @@
|
||||
from rest_framework import serializers
|
||||
|
||||
from apps.accounts.serializers import UserSerializer
|
||||
from apps.core.choices.serializers import RichChoiceSerializerField
|
||||
|
||||
from .models import Ticket
|
||||
from .models import EmailThread, Ticket
|
||||
|
||||
|
||||
class SubmitterProfileSerializer(serializers.Serializer):
|
||||
"""Nested serializer for submitter profile data."""
|
||||
|
||||
display_name = serializers.CharField(source="profile.display_name", allow_null=True)
|
||||
created_at = serializers.DateTimeField(source="date_joined", allow_null=True)
|
||||
coaster_count = serializers.IntegerField(source="profile.coaster_credit_count", allow_null=True, default=0)
|
||||
ride_count = serializers.IntegerField(source="profile.ride_credit_count", allow_null=True, default=0)
|
||||
park_count = serializers.IntegerField(source="profile.park_credit_count", allow_null=True, default=0)
|
||||
review_count = serializers.IntegerField(source="profile.review_count", allow_null=True, default=0)
|
||||
avatar_url = serializers.CharField(source="profile.avatar_url", allow_null=True)
|
||||
|
||||
|
||||
class TicketSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for Ticket model with full frontend compatibility."""
|
||||
|
||||
# User fields
|
||||
user = UserSerializer(read_only=True)
|
||||
user_id = serializers.UUIDField(source="user.id", read_only=True, allow_null=True)
|
||||
submitter_username = serializers.CharField(source="user.username", read_only=True, allow_null=True)
|
||||
submitter_reputation = serializers.SerializerMethodField()
|
||||
submitter_profile = serializers.SerializerMethodField()
|
||||
|
||||
# Choice display fields
|
||||
category_display = serializers.CharField(source="get_category_display", read_only=True)
|
||||
status_display = serializers.CharField(source="get_status_display", read_only=True)
|
||||
|
||||
# Computed fields
|
||||
thread_id = serializers.CharField(read_only=True)
|
||||
response_count = serializers.IntegerField(read_only=True)
|
||||
last_admin_response_at = serializers.DateTimeField(read_only=True, allow_null=True)
|
||||
|
||||
# Resolution tracking (alias resolved_by username)
|
||||
resolved_by_username = serializers.CharField(source="resolved_by.username", read_only=True, allow_null=True)
|
||||
|
||||
# Admin fields
|
||||
assigned_to_username = serializers.CharField(source="assigned_to.username", read_only=True, allow_null=True)
|
||||
archived_by_username = serializers.CharField(source="archived_by.username", read_only=True, allow_null=True)
|
||||
|
||||
class Meta:
|
||||
model = Ticket
|
||||
fields = [
|
||||
# Core fields
|
||||
"id",
|
||||
"user",
|
||||
"category",
|
||||
"category_display",
|
||||
"subject",
|
||||
"message",
|
||||
"email",
|
||||
"status",
|
||||
"status_display",
|
||||
"created_at",
|
||||
"updated_at",
|
||||
# User/submitter fields
|
||||
"user",
|
||||
"user_id",
|
||||
"submitter_username",
|
||||
"submitter_reputation",
|
||||
"submitter_profile",
|
||||
"name",
|
||||
"email",
|
||||
# Ticket content
|
||||
"subject",
|
||||
"message",
|
||||
"category",
|
||||
"category_display",
|
||||
# Status
|
||||
"status",
|
||||
"status_display",
|
||||
# Ticket number
|
||||
"ticket_number",
|
||||
# Admin management
|
||||
"admin_notes",
|
||||
"assigned_to",
|
||||
"assigned_to_username",
|
||||
# Resolution
|
||||
"resolved_at",
|
||||
"resolved_by",
|
||||
"resolved_by_username",
|
||||
# Thread info
|
||||
"thread_id",
|
||||
"last_admin_response_at",
|
||||
"response_count",
|
||||
# Archive
|
||||
"archived_at",
|
||||
"archived_by",
|
||||
"archived_by_username",
|
||||
]
|
||||
read_only_fields = ["id", "status", "created_at", "updated_at", "user"]
|
||||
read_only_fields = [
|
||||
"id",
|
||||
"created_at",
|
||||
"updated_at",
|
||||
"user",
|
||||
"user_id",
|
||||
"submitter_username",
|
||||
"submitter_reputation",
|
||||
"submitter_profile",
|
||||
"ticket_number",
|
||||
"thread_id",
|
||||
"response_count",
|
||||
"last_admin_response_at",
|
||||
]
|
||||
|
||||
def get_submitter_reputation(self, obj):
|
||||
"""Get the submitter's reputation score."""
|
||||
if obj.user and hasattr(obj.user, "profile"):
|
||||
return getattr(obj.user.profile, "reputation", 0)
|
||||
return None
|
||||
|
||||
def get_submitter_profile(self, obj):
|
||||
"""Get a subset of profile data for display."""
|
||||
if not obj.user or not hasattr(obj.user, "profile"):
|
||||
return None
|
||||
profile = obj.user.profile
|
||||
return {
|
||||
"display_name": getattr(profile, "display_name", None),
|
||||
"created_at": obj.user.date_joined.isoformat() if obj.user.date_joined else None,
|
||||
"coaster_count": getattr(profile, "coaster_credit_count", 0),
|
||||
"ride_count": getattr(profile, "ride_credit_count", 0),
|
||||
"park_count": getattr(profile, "park_credit_count", 0),
|
||||
"review_count": getattr(profile, "review_count", 0),
|
||||
"avatar_url": getattr(profile, "avatar_url", None),
|
||||
}
|
||||
|
||||
def validate(self, data):
|
||||
# Ensure email is provided if user is anonymous
|
||||
@@ -35,6 +130,31 @@ class TicketSerializer(serializers.ModelSerializer):
|
||||
return data
|
||||
|
||||
|
||||
class EmailThreadSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for EmailThread model."""
|
||||
|
||||
# NOTE: Frontend uses submission_id, we provide ticket as that field
|
||||
submission_id = serializers.UUIDField(source="ticket.id", read_only=True)
|
||||
sent_by_username = serializers.CharField(source="sent_by.username", read_only=True, allow_null=True)
|
||||
|
||||
class Meta:
|
||||
model = EmailThread
|
||||
fields = [
|
||||
"id",
|
||||
"created_at",
|
||||
"submission_id",
|
||||
"message_id",
|
||||
"from_email",
|
||||
"to_email",
|
||||
"subject",
|
||||
"body_text",
|
||||
"direction",
|
||||
"sent_by",
|
||||
"sent_by_username",
|
||||
]
|
||||
read_only_fields = ["id", "created_at", "submission_id"]
|
||||
|
||||
|
||||
class ReportSerializer(serializers.ModelSerializer):
|
||||
"""Serializer for Report model."""
|
||||
|
||||
@@ -134,9 +254,141 @@ class ReportCreateSerializer(serializers.ModelSerializer):
|
||||
class ReportResolveSerializer(serializers.Serializer):
|
||||
"""Serializer for resolving reports."""
|
||||
|
||||
status = serializers.ChoiceField(
|
||||
choices=[("resolved", "Resolved"), ("dismissed", "Dismissed")],
|
||||
# Use RichChoiceSerializerField with only resolution statuses
|
||||
status = RichChoiceSerializerField(
|
||||
choice_group="report_statuses",
|
||||
domain="support",
|
||||
default="resolved",
|
||||
)
|
||||
notes = serializers.CharField(required=False, allow_blank=True)
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Support Ticket Action Serializers
|
||||
# =============================================================================
|
||||
|
||||
|
||||
class SendReplySerializer(serializers.Serializer):
|
||||
"""
|
||||
Input serializer for send_reply action.
|
||||
|
||||
Validates the request body for sending an email reply to a ticket.
|
||||
Supports both snake_case and camelCase field names for frontend compatibility.
|
||||
"""
|
||||
|
||||
# Primary fields (required=False because we validate manually to support camelCase)
|
||||
reply_body = serializers.CharField(
|
||||
required=False,
|
||||
min_length=1,
|
||||
max_length=50000,
|
||||
help_text="The body text of the email reply",
|
||||
)
|
||||
new_status = RichChoiceSerializerField(
|
||||
choice_group="ticket_statuses",
|
||||
domain="support",
|
||||
required=False,
|
||||
allow_null=True,
|
||||
help_text="Optionally update the ticket status after sending reply",
|
||||
)
|
||||
|
||||
# camelCase aliases for frontend compatibility
|
||||
replyBody = serializers.CharField(required=False, write_only=True)
|
||||
newStatus = serializers.CharField(required=False, write_only=True, allow_null=True)
|
||||
|
||||
def validate(self, data: dict) -> dict:
|
||||
"""Normalize camelCase to snake_case and validate required fields."""
|
||||
# Normalize camelCase to snake_case
|
||||
if "replyBody" in data and data["replyBody"]:
|
||||
data["reply_body"] = data.pop("replyBody")
|
||||
elif "replyBody" in data:
|
||||
data.pop("replyBody")
|
||||
|
||||
if "newStatus" in data and data["newStatus"]:
|
||||
data["new_status"] = data.pop("newStatus")
|
||||
elif "newStatus" in data:
|
||||
data.pop("newStatus")
|
||||
|
||||
# Validate required fields
|
||||
if not data.get("reply_body"):
|
||||
raise serializers.ValidationError({"reply_body": "This field is required."})
|
||||
|
||||
return data
|
||||
|
||||
|
||||
class SendReplyResponseSerializer(serializers.Serializer):
|
||||
"""Response serializer for send_reply action."""
|
||||
|
||||
detail = serializers.CharField()
|
||||
thread_id = serializers.UUIDField()
|
||||
ticket_number = serializers.CharField()
|
||||
|
||||
|
||||
class MergeTicketsSerializer(serializers.Serializer):
|
||||
"""
|
||||
Input serializer for merge tickets action.
|
||||
|
||||
Validates the request body for merging multiple tickets into one.
|
||||
Supports both snake_case and camelCase field names for frontend compatibility.
|
||||
"""
|
||||
|
||||
# Primary fields (required=False because we validate manually to support camelCase)
|
||||
primary_ticket_id = serializers.UUIDField(
|
||||
required=False,
|
||||
help_text="UUID of the primary ticket that will absorb others",
|
||||
)
|
||||
merge_ticket_ids = serializers.ListField(
|
||||
child=serializers.UUIDField(),
|
||||
required=False,
|
||||
min_length=1,
|
||||
help_text="List of ticket UUIDs to merge into the primary",
|
||||
)
|
||||
merge_reason = serializers.CharField(
|
||||
required=False,
|
||||
allow_blank=True,
|
||||
max_length=500,
|
||||
help_text="Optional reason for the merge",
|
||||
)
|
||||
|
||||
# camelCase aliases for frontend compatibility
|
||||
primaryTicketId = serializers.UUIDField(required=False, write_only=True)
|
||||
mergeTicketIds = serializers.ListField(
|
||||
child=serializers.UUIDField(),
|
||||
required=False,
|
||||
write_only=True,
|
||||
)
|
||||
mergeReason = serializers.CharField(required=False, write_only=True, allow_blank=True)
|
||||
|
||||
def validate(self, data: dict) -> dict:
|
||||
"""Normalize camelCase to snake_case and validate required fields."""
|
||||
# Normalize camelCase to snake_case
|
||||
if "primaryTicketId" in data and data["primaryTicketId"]:
|
||||
data["primary_ticket_id"] = data.pop("primaryTicketId")
|
||||
elif "primaryTicketId" in data:
|
||||
data.pop("primaryTicketId")
|
||||
|
||||
if "mergeTicketIds" in data and data["mergeTicketIds"]:
|
||||
data["merge_ticket_ids"] = data.pop("mergeTicketIds")
|
||||
elif "mergeTicketIds" in data:
|
||||
data.pop("mergeTicketIds")
|
||||
|
||||
if "mergeReason" in data and data["mergeReason"]:
|
||||
data["merge_reason"] = data.pop("mergeReason")
|
||||
elif "mergeReason" in data:
|
||||
data.pop("mergeReason")
|
||||
|
||||
# Validate required fields
|
||||
if not data.get("primary_ticket_id"):
|
||||
raise serializers.ValidationError({"primary_ticket_id": "This field is required."})
|
||||
if not data.get("merge_ticket_ids"):
|
||||
raise serializers.ValidationError({"merge_ticket_ids": "This field is required."})
|
||||
|
||||
return data
|
||||
|
||||
|
||||
class MergeTicketsResponseSerializer(serializers.Serializer):
|
||||
"""Response serializer for merge tickets action."""
|
||||
|
||||
detail = serializers.CharField()
|
||||
primaryTicketNumber = serializers.CharField()
|
||||
mergedCount = serializers.IntegerField()
|
||||
threadsConsolidated = serializers.IntegerField()
|
||||
|
||||
@@ -1,13 +1,13 @@
|
||||
from django.urls import include, path
|
||||
from rest_framework.routers import DefaultRouter
|
||||
|
||||
from .views import ReportViewSet, TicketViewSet
|
||||
from .views import EmailThreadViewSet, ReportViewSet, TicketViewSet
|
||||
|
||||
router = DefaultRouter()
|
||||
router.register(r"tickets", TicketViewSet, basename="ticket")
|
||||
router.register(r"reports", ReportViewSet, basename="report")
|
||||
router.register(r"threads", EmailThreadViewSet, basename="email-thread")
|
||||
|
||||
urlpatterns = [
|
||||
path("", include(router.urls)),
|
||||
]
|
||||
|
||||
|
||||
@@ -4,8 +4,9 @@ from rest_framework import filters, permissions, status, viewsets
|
||||
from rest_framework.decorators import action
|
||||
from rest_framework.response import Response
|
||||
|
||||
from .models import Report, Ticket
|
||||
from .models import EmailThread, Report, Ticket
|
||||
from .serializers import (
|
||||
EmailThreadSerializer,
|
||||
ReportCreateSerializer,
|
||||
ReportResolveSerializer,
|
||||
ReportSerializer,
|
||||
@@ -23,17 +24,26 @@ class TicketViewSet(viewsets.ModelViewSet):
|
||||
queryset = Ticket.objects.all()
|
||||
serializer_class = TicketSerializer
|
||||
permission_classes = [permissions.AllowAny] # We handle granular perms in get_queryset/perform_create
|
||||
filter_backends = [DjangoFilterBackend, filters.OrderingFilter]
|
||||
filterset_fields = ["status", "category"]
|
||||
ordering_fields = ["created_at", "status"]
|
||||
filter_backends = [DjangoFilterBackend, filters.OrderingFilter, filters.SearchFilter]
|
||||
filterset_fields = ["status", "category", "archived_at"]
|
||||
search_fields = ["name", "email", "subject", "ticket_number"]
|
||||
ordering_fields = ["created_at", "status", "ticket_number"]
|
||||
ordering = ["-created_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
user = self.request.user
|
||||
qs = Ticket.objects.select_related(
|
||||
"user",
|
||||
"user__profile",
|
||||
"assigned_to",
|
||||
"resolved_by",
|
||||
"archived_by",
|
||||
).prefetch_related("email_threads")
|
||||
|
||||
if user.is_staff:
|
||||
return Ticket.objects.all()
|
||||
return qs
|
||||
if user.is_authenticated:
|
||||
return Ticket.objects.filter(user=user)
|
||||
return qs.filter(user=user)
|
||||
return Ticket.objects.none() # Guests can't list tickets
|
||||
|
||||
def perform_create(self, serializer):
|
||||
@@ -42,6 +52,337 @@ class TicketViewSet(viewsets.ModelViewSet):
|
||||
else:
|
||||
serializer.save()
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def send_reply(self, request, pk=None):
|
||||
"""
|
||||
Send an email reply to the ticket submitter.
|
||||
|
||||
Creates an EmailThread record and sends the email via ForwardEmail.
|
||||
Optionally updates the ticket status.
|
||||
"""
|
||||
from typing import Any
|
||||
|
||||
from django.conf import settings
|
||||
from django.contrib.sites.models import Site
|
||||
|
||||
from django_forwardemail.services import EmailService
|
||||
|
||||
from .serializers import SendReplySerializer
|
||||
|
||||
# Validate input
|
||||
serializer = SendReplySerializer(data=request.data)
|
||||
serializer.is_valid(raise_exception=True)
|
||||
validated_data: dict[str, Any] = serializer.validated_data
|
||||
|
||||
ticket = self.get_object()
|
||||
|
||||
if not ticket.email:
|
||||
return Response(
|
||||
{"detail": "Ticket has no email address to reply to"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
reply_body: str = validated_data["reply_body"]
|
||||
new_status: str | None = validated_data.get("new_status")
|
||||
|
||||
# Build email subject with ticket number for threading
|
||||
subject = f"Re: [{ticket.ticket_number}] {ticket.subject}"
|
||||
|
||||
# Get the support from email with proper formatting
|
||||
from_email: str = getattr(settings, "DEFAULT_FROM_EMAIL", "ThrillWiki Support <support@thrillwiki.com>")
|
||||
|
||||
try:
|
||||
# Get the current site for ForwardEmail configuration
|
||||
# ForwardEmail requires a Site object, not RequestSite
|
||||
try:
|
||||
site = Site.objects.get_current()
|
||||
except Site.DoesNotExist:
|
||||
site = Site.objects.first()
|
||||
if site is None:
|
||||
return Response(
|
||||
{"detail": "No site configured for email sending"},
|
||||
status=status.HTTP_500_INTERNAL_SERVER_ERROR,
|
||||
)
|
||||
|
||||
# Send email via ForwardEmail service
|
||||
EmailService.send_email(
|
||||
to=ticket.email,
|
||||
subject=subject,
|
||||
text=reply_body,
|
||||
from_email=from_email,
|
||||
reply_to=from_email, # Ensure replies come back to support
|
||||
site=site,
|
||||
)
|
||||
|
||||
# Create EmailThread record for audit trail
|
||||
email_thread = EmailThread.objects.create(
|
||||
ticket=ticket,
|
||||
from_email=from_email,
|
||||
to_email=ticket.email,
|
||||
subject=subject,
|
||||
body_text=reply_body,
|
||||
direction="outbound",
|
||||
sent_by=request.user,
|
||||
)
|
||||
|
||||
# Update ticket status if provided
|
||||
if new_status and new_status != ticket.status:
|
||||
ticket.status = new_status
|
||||
if new_status in ("resolved", "closed"):
|
||||
ticket.resolved_at = timezone.now()
|
||||
ticket.resolved_by = request.user
|
||||
ticket.save()
|
||||
|
||||
return Response({
|
||||
"detail": "Reply sent successfully",
|
||||
"thread_id": str(email_thread.id),
|
||||
"ticket_number": ticket.ticket_number,
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
# Log the error for debugging
|
||||
import logging
|
||||
logger = logging.getLogger(__name__)
|
||||
logger.exception("Failed to send ticket reply email")
|
||||
|
||||
return Response(
|
||||
{"detail": f"Failed to send email: {str(e)}"},
|
||||
status=status.HTTP_500_INTERNAL_SERVER_ERROR,
|
||||
)
|
||||
|
||||
@action(detail=False, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def merge(self, request):
|
||||
"""
|
||||
Merge multiple tickets into a primary ticket.
|
||||
|
||||
Moves all EmailThread records to the primary ticket and archives merged tickets.
|
||||
"""
|
||||
from typing import Any
|
||||
from uuid import UUID
|
||||
|
||||
from .serializers import MergeTicketsSerializer
|
||||
|
||||
# Validate input
|
||||
serializer = MergeTicketsSerializer(data=request.data)
|
||||
serializer.is_valid(raise_exception=True)
|
||||
validated_data: dict[str, Any] = serializer.validated_data
|
||||
|
||||
primary_id: UUID = validated_data["primary_ticket_id"]
|
||||
merge_ids: list[UUID] = validated_data["merge_ticket_ids"]
|
||||
reason: str = validated_data.get("merge_reason", "")
|
||||
|
||||
# Get primary ticket
|
||||
try:
|
||||
primary_ticket = Ticket.objects.get(pk=primary_id)
|
||||
except Ticket.DoesNotExist:
|
||||
return Response(
|
||||
{"detail": f"Primary ticket {primary_id} not found"},
|
||||
status=status.HTTP_404_NOT_FOUND,
|
||||
)
|
||||
|
||||
# Get tickets to merge (exclud primary if accidentally included)
|
||||
tickets_to_merge = Ticket.objects.filter(pk__in=merge_ids).exclude(pk=primary_id)
|
||||
if tickets_to_merge.count() == 0:
|
||||
return Response(
|
||||
{"detail": "No valid tickets found to merge"},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
merged_count = 0
|
||||
threads_consolidated = 0
|
||||
merged_ticket_numbers: list[str] = []
|
||||
|
||||
for ticket in tickets_to_merge:
|
||||
# Move all email threads to primary ticket
|
||||
thread_count = EmailThread.objects.filter(ticket=ticket).update(ticket=primary_ticket)
|
||||
threads_consolidated += thread_count
|
||||
|
||||
# Record the merged ticket number
|
||||
merged_ticket_numbers.append(ticket.ticket_number)
|
||||
|
||||
# Archive the merged ticket with merge history
|
||||
ticket.archived_at = timezone.now()
|
||||
ticket.archived_by = request.user
|
||||
existing_notes = ticket.admin_notes or ""
|
||||
ticket.admin_notes = (
|
||||
f"{existing_notes}\n\n"
|
||||
f"[MERGED] Merged into {primary_ticket.ticket_number} by {request.user.username} "
|
||||
f"on {timezone.now().isoformat()}. Reason: {reason or 'Not specified'}"
|
||||
).strip()
|
||||
ticket.save()
|
||||
|
||||
merged_count += 1
|
||||
|
||||
# Update primary ticket with merge history
|
||||
existing_merged = primary_ticket.admin_notes or ""
|
||||
merge_note = (
|
||||
f"\n\n[MERGE HISTORY] Absorbed tickets: {', '.join(merged_ticket_numbers)} "
|
||||
f"({threads_consolidated} threads consolidated) by {request.user.username}"
|
||||
)
|
||||
primary_ticket.admin_notes = existing_merged + merge_note
|
||||
primary_ticket.save()
|
||||
|
||||
return Response({
|
||||
"detail": "Tickets merged successfully",
|
||||
"primaryTicketNumber": primary_ticket.ticket_number,
|
||||
"mergedCount": merged_count,
|
||||
"threadsConsolidated": threads_consolidated,
|
||||
})
|
||||
|
||||
# =========================================================================
|
||||
# FSM Transition Endpoints
|
||||
# =========================================================================
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def start_progress(self, request, pk=None):
|
||||
"""
|
||||
Start working on a ticket.
|
||||
Transition: open -> in_progress
|
||||
"""
|
||||
from django.core.exceptions import ValidationError as DjangoValidationError
|
||||
|
||||
ticket = self.get_object()
|
||||
try:
|
||||
ticket.start_progress(user=request.user)
|
||||
return Response({
|
||||
"detail": "Ticket marked as in progress",
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
"status": ticket.status,
|
||||
})
|
||||
except DjangoValidationError as e:
|
||||
return Response(
|
||||
{"detail": str(e.message if hasattr(e, 'message') else e)},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser], url_path="close")
|
||||
def close_ticket(self, request, pk=None):
|
||||
"""
|
||||
Close/resolve a ticket.
|
||||
Transition: open|in_progress -> closed
|
||||
"""
|
||||
from django.core.exceptions import ValidationError as DjangoValidationError
|
||||
|
||||
ticket = self.get_object()
|
||||
notes = request.data.get("notes", "")
|
||||
|
||||
try:
|
||||
ticket.close(user=request.user, notes=notes)
|
||||
return Response({
|
||||
"detail": "Ticket closed successfully",
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
"status": ticket.status,
|
||||
"resolvedAt": ticket.resolved_at.isoformat() if ticket.resolved_at else None,
|
||||
})
|
||||
except DjangoValidationError as e:
|
||||
return Response(
|
||||
{"detail": str(e.message if hasattr(e, 'message') else e)},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def reopen(self, request, pk=None):
|
||||
"""
|
||||
Reopen a closed ticket.
|
||||
Transition: closed -> open
|
||||
"""
|
||||
from django.core.exceptions import ValidationError as DjangoValidationError
|
||||
|
||||
ticket = self.get_object()
|
||||
reason = request.data.get("reason", "")
|
||||
|
||||
try:
|
||||
ticket.reopen(user=request.user, reason=reason)
|
||||
return Response({
|
||||
"detail": "Ticket reopened",
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
"status": ticket.status,
|
||||
})
|
||||
except DjangoValidationError as e:
|
||||
return Response(
|
||||
{"detail": str(e.message if hasattr(e, 'message') else e)},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def archive(self, request, pk=None):
|
||||
"""Archive a ticket."""
|
||||
from django.core.exceptions import ValidationError as DjangoValidationError
|
||||
|
||||
ticket = self.get_object()
|
||||
reason = request.data.get("reason", "")
|
||||
|
||||
try:
|
||||
ticket.archive(user=request.user, reason=reason)
|
||||
return Response({
|
||||
"detail": "Ticket archived",
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
"archivedAt": ticket.archived_at.isoformat() if ticket.archived_at else None,
|
||||
})
|
||||
except DjangoValidationError as e:
|
||||
return Response(
|
||||
{"detail": str(e.message if hasattr(e, 'message') else e)},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@action(detail=True, methods=["post"], permission_classes=[permissions.IsAdminUser])
|
||||
def unarchive(self, request, pk=None):
|
||||
"""Restore an archived ticket."""
|
||||
from django.core.exceptions import ValidationError as DjangoValidationError
|
||||
|
||||
ticket = self.get_object()
|
||||
|
||||
try:
|
||||
ticket.unarchive(user=request.user)
|
||||
return Response({
|
||||
"detail": "Ticket unarchived",
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
})
|
||||
except DjangoValidationError as e:
|
||||
return Response(
|
||||
{"detail": str(e.message if hasattr(e, 'message') else e)},
|
||||
status=status.HTTP_400_BAD_REQUEST,
|
||||
)
|
||||
|
||||
@action(detail=True, methods=["get"], permission_classes=[permissions.IsAdminUser])
|
||||
def available_transitions(self, request, pk=None):
|
||||
"""
|
||||
Get available transitions for a ticket.
|
||||
Uses StateMachineMixin to return FSM-aware transition metadata.
|
||||
"""
|
||||
ticket = self.get_object()
|
||||
transitions = ticket.get_available_user_transitions(request.user)
|
||||
|
||||
return Response({
|
||||
"ticketNumber": ticket.ticket_number,
|
||||
"currentStatus": ticket.status,
|
||||
"currentStatusDisplay": ticket.get_status_display(),
|
||||
"availableTransitions": transitions,
|
||||
})
|
||||
|
||||
|
||||
class EmailThreadViewSet(viewsets.ModelViewSet):
|
||||
"""
|
||||
ViewSet for email thread entries.
|
||||
Staff only for full access.
|
||||
"""
|
||||
|
||||
queryset = EmailThread.objects.select_related("ticket", "sent_by").all()
|
||||
serializer_class = EmailThreadSerializer
|
||||
permission_classes = [permissions.IsAdminUser]
|
||||
filter_backends = [DjangoFilterBackend, filters.OrderingFilter]
|
||||
filterset_fields = ["ticket", "direction"]
|
||||
ordering_fields = ["created_at"]
|
||||
ordering = ["created_at"]
|
||||
|
||||
def get_queryset(self):
|
||||
# Support filtering by submission_id (which is ticket_id in our model)
|
||||
qs = super().get_queryset()
|
||||
submission_id = self.request.query_params.get("submission_id")
|
||||
if submission_id:
|
||||
qs = qs.filter(ticket_id=submission_id)
|
||||
return qs
|
||||
|
||||
|
||||
class ReportViewSet(viewsets.ModelViewSet):
|
||||
"""
|
||||
|
||||
Binary file not shown.
@@ -86,6 +86,7 @@ THIRD_PARTY_APPS = [
|
||||
"allauth.socialaccount",
|
||||
"allauth.socialaccount.providers.google",
|
||||
"allauth.socialaccount.providers.discord",
|
||||
"allauth.headless", # Headless API for SPA/mobile passkey login
|
||||
"turnstile", # Cloudflare Turnstile CAPTCHA (django-turnstile package)
|
||||
"django_cleanup",
|
||||
"django_filters",
|
||||
|
||||
@@ -73,11 +73,11 @@ CACHE_MIDDLEWARE_KEY_PREFIX = "thrillwiki_dev"
|
||||
# =============================================================================
|
||||
# Use ForwardEmail for actual sending, or console for debugging
|
||||
|
||||
# Console backend for debugging (uncomment to use):
|
||||
# EMAIL_BACKEND = "django.core.mail.backends.console.EmailBackend"
|
||||
# Console backend for debugging (recommended for local development):
|
||||
EMAIL_BACKEND = "django.core.mail.backends.console.EmailBackend"
|
||||
|
||||
# ForwardEmail backend for actual email sending:
|
||||
EMAIL_BACKEND = "django_forwardemail.backends.ForwardEmailBackend"
|
||||
# ForwardEmail backend for actual email sending (uncomment for production testing):
|
||||
# EMAIL_BACKEND = "django_forwardemail.backends.ForwardEmailBackend"
|
||||
|
||||
# =============================================================================
|
||||
# Security Settings (Relaxed for Development)
|
||||
@@ -87,6 +87,9 @@ SECURE_SSL_REDIRECT = False
|
||||
SESSION_COOKIE_SECURE = False
|
||||
CSRF_COOKIE_SECURE = False
|
||||
|
||||
# Email verification - optional for local development to allow passkey login
|
||||
ACCOUNT_EMAIL_VERIFICATION = "optional"
|
||||
|
||||
# =============================================================================
|
||||
# Development Apps
|
||||
# =============================================================================
|
||||
|
||||
@@ -101,6 +101,7 @@ CORS_ALLOW_HEADERS = [
|
||||
"x-csrftoken",
|
||||
"x-requested-with",
|
||||
"x-api-version",
|
||||
"x-session-token", # Required for allauth headless app client
|
||||
]
|
||||
|
||||
# HTTP methods allowed for CORS requests
|
||||
|
||||
@@ -114,6 +114,26 @@ ACCOUNT_LOGIN_BY_CODE_MAX_ATTEMPTS = config("ACCOUNT_LOGIN_BY_CODE_MAX_ATTEMPTS"
|
||||
ACCOUNT_LOGIN_BY_CODE_TIMEOUT = config("ACCOUNT_LOGIN_BY_CODE_TIMEOUT", default=300, cast=int)
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# Headless API Configuration
|
||||
# =============================================================================
|
||||
# https://docs.allauth.org/en/latest/headless/configuration.html
|
||||
|
||||
# Frontend URL for email links (password reset, email verification, etc.)
|
||||
HEADLESS_FRONTEND_URLS = {
|
||||
"account_confirm_email": config("FRONTEND_URL", default="http://localhost:5173") + "/auth/callback?key={key}",
|
||||
"account_reset_password": config("FRONTEND_URL", default="http://localhost:5173") + "/auth/reset-password?key={key}",
|
||||
"account_signup": config("FRONTEND_URL", default="http://localhost:5173") + "/auth?tab=signup",
|
||||
"socialaccount_login_error": config("FRONTEND_URL", default="http://localhost:5173") + "/auth?error=social",
|
||||
}
|
||||
|
||||
# Set to True since our frontend is a separate SPA
|
||||
HEADLESS_ONLY = config("HEADLESS_ONLY", default=False, cast=bool)
|
||||
|
||||
# Allow both "app" and "browser" clients for flexibility
|
||||
# "browser" uses cookies, "app" uses Authorization header
|
||||
HEADLESS_CLIENTS = ("app", "browser")
|
||||
|
||||
# =============================================================================
|
||||
# Celery Configuration
|
||||
# =============================================================================
|
||||
|
||||
@@ -137,6 +137,7 @@ addopts = [
|
||||
"--strict-markers",
|
||||
"--tb=short",
|
||||
]
|
||||
asyncio_default_fixture_loop_scope = "function"
|
||||
markers = [
|
||||
"unit: Unit tests (fast, isolated)",
|
||||
"integration: Integration tests (may use database)",
|
||||
|
||||
86
backend/run-dev.sh
Executable file
86
backend/run-dev.sh
Executable file
@@ -0,0 +1,86 @@
|
||||
#!/bin/bash
|
||||
# ThrillWiki Development Server
|
||||
# Runs Django, Celery worker, and Celery beat together
|
||||
#
|
||||
# Usage: ./run-dev.sh
|
||||
# Press Ctrl+C to stop all services
|
||||
|
||||
set -e
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[0;33m'
|
||||
BLUE='\033[0;34m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Find backend directory (script may be run from different locations)
|
||||
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
||||
if [ -d "$SCRIPT_DIR/backend" ]; then
|
||||
BACKEND_DIR="$SCRIPT_DIR/backend"
|
||||
elif [ -d "$SCRIPT_DIR/../backend" ]; then
|
||||
BACKEND_DIR="$SCRIPT_DIR/../backend"
|
||||
elif [ -f "$SCRIPT_DIR/manage.py" ]; then
|
||||
BACKEND_DIR="$SCRIPT_DIR"
|
||||
else
|
||||
echo -e "${RED}❌ Cannot find backend directory. Run from project root or backend folder.${NC}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
cd "$BACKEND_DIR"
|
||||
|
||||
echo -e "${BLUE}🎢 ThrillWiki Development Server${NC}"
|
||||
echo "=================================="
|
||||
echo -e "Backend: ${GREEN}$BACKEND_DIR${NC}"
|
||||
echo ""
|
||||
|
||||
# Check Redis is running
|
||||
if ! redis-cli ping &> /dev/null; then
|
||||
echo -e "${YELLOW}⚠️ Redis not running. Starting Redis...${NC}"
|
||||
if command -v brew &> /dev/null; then
|
||||
brew services start redis 2>/dev/null || true
|
||||
sleep 1
|
||||
else
|
||||
echo -e "${RED}❌ Redis not running. Start with: docker run -d -p 6379:6379 redis:alpine${NC}"
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
echo -e "${GREEN}✓ Redis connected${NC}"
|
||||
|
||||
# Cleanup function to kill all background processes
|
||||
cleanup() {
|
||||
echo ""
|
||||
echo -e "${YELLOW}Shutting down...${NC}"
|
||||
kill $PID_DJANGO $PID_CELERY_WORKER $PID_CELERY_BEAT 2>/dev/null
|
||||
wait 2>/dev/null
|
||||
echo -e "${GREEN}✓ All services stopped${NC}"
|
||||
exit 0
|
||||
}
|
||||
|
||||
trap cleanup SIGINT SIGTERM
|
||||
|
||||
echo ""
|
||||
echo -e "${BLUE}Starting services...${NC}"
|
||||
echo ""
|
||||
|
||||
# Start Django
|
||||
echo -e "${GREEN}▶ Django${NC} (http://localhost:8000)"
|
||||
uv run python manage.py runserver 2>&1 | sed 's/^/ [Django] /' &
|
||||
PID_DJANGO=$!
|
||||
|
||||
# Start Celery worker
|
||||
echo -e "${GREEN}▶ Celery Worker${NC}"
|
||||
uv run celery -A config worker -l info 2>&1 | sed 's/^/ [Worker] /' &
|
||||
PID_CELERY_WORKER=$!
|
||||
|
||||
# Start Celery beat
|
||||
echo -e "${GREEN}▶ Celery Beat${NC}"
|
||||
uv run celery -A config beat -l info 2>&1 | sed 's/^/ [Beat] /' &
|
||||
PID_CELERY_BEAT=$!
|
||||
|
||||
echo ""
|
||||
echo -e "${GREEN}All services running. Press Ctrl+C to stop.${NC}"
|
||||
echo ""
|
||||
|
||||
# Wait for any process to exit
|
||||
wait
|
||||
@@ -36,12 +36,13 @@
|
||||
- main_class: Additional classes for <main> tag
|
||||
|
||||
Usage Example:
|
||||
{% extends "base/base.html" %}
|
||||
{% block title %}My Page - ThrillWiki{% endblock %}
|
||||
{% block content %}
|
||||
<h1>My Page Content</h1>
|
||||
{% endblock %}
|
||||
{% templatetag openblock %} extends "base/base.html" {% templatetag closeblock %}
|
||||
{% templatetag openblock %} block title {% templatetag closeblock %}My Page - ThrillWiki{% templatetag openblock %} endblock {% templatetag closeblock %}
|
||||
{% templatetag openblock %} block content {% templatetag closeblock %}
|
||||
<h1>My Page Content</h1>
|
||||
{% templatetag openblock %} endblock {% templatetag closeblock %}
|
||||
============================================================================= #}
|
||||
|
||||
<!DOCTYPE html>
|
||||
<html lang="en" class="h-full">
|
||||
<head>
|
||||
|
||||
24
backend/templates/emails/account_deletion_verification.html
Normal file
24
backend/templates/emails/account_deletion_verification.html
Normal file
@@ -0,0 +1,24 @@
|
||||
{% extends "emails/base.html" %}
|
||||
|
||||
{% block content %}
|
||||
<h1>Account Deletion Request</h1>
|
||||
|
||||
<p>Hi {{ user.username }},</p>
|
||||
|
||||
<p>You have requested to delete your ThrillWiki account. To confirm this action, please use the following verification code:</p>
|
||||
|
||||
<div style="text-align: center; margin: 30px 0;">
|
||||
<p style="font-size: 28px; font-weight: bold; letter-spacing: 3px; background: #f5f5f5; padding: 20px; border-radius: 8px;">
|
||||
{{ verification_code }}
|
||||
</p>
|
||||
</div>
|
||||
|
||||
<p>This code will expire at {{ expires_at|date:"F j, Y, g:i a" }}.</p>
|
||||
|
||||
<p><strong>Warning:</strong> This action is permanent and cannot be undone. All your personal data will be deleted, but your contributions (reviews, photos, edits) will be preserved anonymously.</p>
|
||||
|
||||
<p>If you did not request this deletion, please ignore this email or contact support immediately.</p>
|
||||
|
||||
<p>Best regards,<br>
|
||||
The {{ site_name }} Team</p>
|
||||
{% endblock %}
|
||||
17
backend/templates/emails/account_deletion_verification.txt
Normal file
17
backend/templates/emails/account_deletion_verification.txt
Normal file
@@ -0,0 +1,17 @@
|
||||
Account Deletion Request
|
||||
========================
|
||||
|
||||
Hi {{ user.username }},
|
||||
|
||||
You have requested to delete your ThrillWiki account. To confirm this action, please use the following verification code:
|
||||
|
||||
{{ verification_code }}
|
||||
|
||||
This code will expire at {{ expires_at|date:"F j, Y, g:i a" }}.
|
||||
|
||||
WARNING: This action is permanent and cannot be undone. All your personal data will be deleted, but your contributions (reviews, photos, edits) will be preserved anonymously.
|
||||
|
||||
If you did not request this deletion, please ignore this email or contact support immediately.
|
||||
|
||||
Best regards,
|
||||
The {{ site_name }} Team
|
||||
23
backend/templates/emails/base.html
Normal file
23
backend/templates/emails/base.html
Normal file
@@ -0,0 +1,23 @@
|
||||
<!DOCTYPE html>
|
||||
<html>
|
||||
<head>
|
||||
<meta charset="utf-8">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||
<title>{% block title %}ThrillWiki{% endblock %}</title>
|
||||
<style>
|
||||
body {
|
||||
font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Helvetica, Arial, sans-serif;
|
||||
line-height: 1.6;
|
||||
color: #333;
|
||||
max-width: 600px;
|
||||
margin: 0 auto;
|
||||
padding: 20px;
|
||||
}
|
||||
h1 { color: #1a1a2e; }
|
||||
a { color: #0066cc; }
|
||||
</style>
|
||||
</head>
|
||||
<body>
|
||||
{% block content %}{% endblock %}
|
||||
</body>
|
||||
</html>
|
||||
@@ -20,17 +20,18 @@ from rest_framework.test import APIClient
|
||||
|
||||
from tests.factories import (
|
||||
CoasterFactory,
|
||||
DesignerCompanyFactory,
|
||||
ManufacturerCompanyFactory,
|
||||
ParkFactory,
|
||||
RideFactory,
|
||||
RideModelFactory,
|
||||
RidesDesignerFactory,
|
||||
RidesManufacturerFactory,
|
||||
StaffUserFactory,
|
||||
UserFactory,
|
||||
)
|
||||
from tests.test_utils import EnhancedAPITestCase
|
||||
|
||||
|
||||
|
||||
class TestRideListAPIView(EnhancedAPITestCase):
|
||||
"""Test cases for RideListCreateAPIView GET endpoint."""
|
||||
|
||||
@@ -38,8 +39,8 @@ class TestRideListAPIView(EnhancedAPITestCase):
|
||||
"""Set up test data."""
|
||||
self.client = APIClient()
|
||||
self.park = ParkFactory()
|
||||
self.manufacturer = ManufacturerCompanyFactory()
|
||||
self.designer = DesignerCompanyFactory()
|
||||
self.manufacturer = RidesManufacturerFactory()
|
||||
self.designer = RidesDesignerFactory()
|
||||
self.rides = [
|
||||
RideFactory(
|
||||
park=self.park,
|
||||
@@ -183,7 +184,7 @@ class TestRideCreateAPIView(EnhancedAPITestCase):
|
||||
self.user = UserFactory()
|
||||
self.staff_user = StaffUserFactory()
|
||||
self.park = ParkFactory()
|
||||
self.manufacturer = ManufacturerCompanyFactory()
|
||||
self.manufacturer = RidesManufacturerFactory()
|
||||
self.url = "/api/v1/rides/"
|
||||
|
||||
self.valid_ride_data = {
|
||||
@@ -373,7 +374,7 @@ class TestHybridRideAPIView(EnhancedAPITestCase):
|
||||
"""Set up test data."""
|
||||
self.client = APIClient()
|
||||
self.park = ParkFactory()
|
||||
self.manufacturer = ManufacturerCompanyFactory()
|
||||
self.manufacturer = RidesManufacturerFactory()
|
||||
self.rides = [
|
||||
RideFactory(park=self.park, manufacturer=self.manufacturer, status="OPERATING", category="RC"),
|
||||
RideFactory(park=self.park, status="OPERATING", category="DR"),
|
||||
@@ -386,10 +387,9 @@ class TestHybridRideAPIView(EnhancedAPITestCase):
|
||||
response = self.client.get(self.url)
|
||||
|
||||
self.assertEqual(response.status_code, status.HTTP_200_OK)
|
||||
self.assertTrue(response.data.get("success", False))
|
||||
self.assertIn("data", response.data)
|
||||
self.assertIn("rides", response.data["data"])
|
||||
self.assertIn("total_count", response.data["data"])
|
||||
# API returns rides directly, not wrapped in success/data
|
||||
self.assertIn("rides", response.data)
|
||||
self.assertIn("total_count", response.data)
|
||||
|
||||
def test__hybrid_ride__with_category_filter__returns_filtered_rides(self):
|
||||
"""Test filtering by category."""
|
||||
@@ -420,7 +420,8 @@ class TestHybridRideAPIView(EnhancedAPITestCase):
|
||||
response = self.client.get(self.url, {"offset": 0})
|
||||
|
||||
self.assertEqual(response.status_code, status.HTTP_200_OK)
|
||||
self.assertIn("has_more", response.data["data"])
|
||||
# API returns has_more directly at top level
|
||||
self.assertIn("has_more", response.data)
|
||||
|
||||
def test__hybrid_ride__with_invalid_offset__returns_400(self):
|
||||
"""Test invalid offset parameter."""
|
||||
@@ -465,15 +466,15 @@ class TestRideFilterMetadataAPIView(EnhancedAPITestCase):
|
||||
def setUp(self):
|
||||
"""Set up test data."""
|
||||
self.client = APIClient()
|
||||
self.url = "/api/v1/rides/filter-metadata/"
|
||||
self.url = "/api/v1/rides/hybrid/filter-metadata/"
|
||||
|
||||
def test__filter_metadata__unscoped__returns_all_metadata(self):
|
||||
"""Test getting unscoped filter metadata."""
|
||||
response = self.client.get(self.url)
|
||||
|
||||
self.assertEqual(response.status_code, status.HTTP_200_OK)
|
||||
self.assertTrue(response.data.get("success", False))
|
||||
self.assertIn("data", response.data)
|
||||
# API returns metadata directly, not wrapped in success/data
|
||||
self.assertIsInstance(response.data, dict)
|
||||
|
||||
def test__filter_metadata__scoped__returns_filtered_metadata(self):
|
||||
"""Test getting scoped filter metadata."""
|
||||
@@ -488,7 +489,7 @@ class TestCompanySearchAPIView(EnhancedAPITestCase):
|
||||
def setUp(self):
|
||||
"""Set up test data."""
|
||||
self.client = APIClient()
|
||||
self.manufacturer = ManufacturerCompanyFactory(name="Bolliger & Mabillard")
|
||||
self.manufacturer = RidesManufacturerFactory(name="Bolliger & Mabillard")
|
||||
self.url = "/api/v1/rides/search/companies/"
|
||||
|
||||
def test__company_search__with_query__returns_matching_companies(self):
|
||||
@@ -520,7 +521,7 @@ class TestRideModelSearchAPIView(EnhancedAPITestCase):
|
||||
"""Set up test data."""
|
||||
self.client = APIClient()
|
||||
self.ride_model = RideModelFactory(name="Hyper Coaster")
|
||||
self.url = "/api/v1/rides/search-ride-models/"
|
||||
self.url = "/api/v1/rides/search/ride-models/"
|
||||
|
||||
def test__ride_model_search__with_query__returns_matching_models(self):
|
||||
"""Test searching for ride models."""
|
||||
|
||||
@@ -743,18 +743,11 @@ def bulk_operation_pending(db):
|
||||
# =============================================================================
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def live_server(live_server_url):
|
||||
"""Provide the live server URL for tests.
|
||||
|
||||
Note: This fixture is provided by pytest-django. The live_server_url
|
||||
fixture provides the URL as a string.
|
||||
"""
|
||||
|
||||
class LiveServer:
|
||||
url = live_server_url
|
||||
|
||||
return LiveServer()
|
||||
# NOTE: The live_server fixture is provided by pytest-django.
|
||||
# It has a .url attribute that provides the server URL.
|
||||
# We previously had a custom wrapper here, but it broke because
|
||||
# it depended on a non-existent 'live_server_url' fixture.
|
||||
# The built-in live_server fixture already works correctly.
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
|
||||
@@ -5,7 +5,7 @@ Following Django styleguide pattern for test data creation using factory_boy.
|
||||
|
||||
import factory
|
||||
from django.contrib.auth import get_user_model
|
||||
from django.contrib.gis.geos import Point
|
||||
# GeoDjango Point import removed - not currently used
|
||||
from django.utils.text import slugify
|
||||
from factory import fuzzy
|
||||
from factory.django import DjangoModelFactory
|
||||
@@ -22,8 +22,7 @@ class UserFactory(DjangoModelFactory):
|
||||
|
||||
username = factory.Sequence(lambda n: f"testuser{n}")
|
||||
email = factory.LazyAttribute(lambda obj: f"{obj.username}@example.com")
|
||||
first_name = factory.Faker("first_name")
|
||||
last_name = factory.Faker("last_name")
|
||||
# Note: first_name and last_name are removed from User model
|
||||
is_active = True
|
||||
is_staff = False
|
||||
is_superuser = False
|
||||
@@ -31,7 +30,8 @@ class UserFactory(DjangoModelFactory):
|
||||
@factory.post_generation
|
||||
def set_password(obj, create, extracted, **kwargs):
|
||||
if create:
|
||||
password = extracted or "testpass123"
|
||||
# Support both UserFactory(set_password="pwd") and UserFactory(set_password__password="pwd")
|
||||
password = kwargs.get("password") or extracted or "testpass123"
|
||||
obj.set_password(password)
|
||||
obj.save()
|
||||
|
||||
@@ -89,27 +89,6 @@ class DesignerCompanyFactory(CompanyFactory):
|
||||
roles = factory.LazyFunction(lambda: ["DESIGNER"])
|
||||
|
||||
|
||||
class LocationFactory(DjangoModelFactory):
|
||||
"""Factory for creating Location instances."""
|
||||
|
||||
class Meta:
|
||||
model = "location.Location"
|
||||
|
||||
name = factory.Faker("city")
|
||||
location_type = "park"
|
||||
latitude = fuzzy.FuzzyFloat(-90, 90)
|
||||
longitude = fuzzy.FuzzyFloat(-180, 180)
|
||||
street_address = factory.Faker("street_address")
|
||||
city = factory.Faker("city")
|
||||
state = factory.Faker("state")
|
||||
country = factory.Faker("country")
|
||||
postal_code = factory.Faker("postcode")
|
||||
|
||||
@factory.lazy_attribute
|
||||
def point(self):
|
||||
return Point(float(self.longitude), float(self.latitude))
|
||||
|
||||
|
||||
class ParkFactory(DjangoModelFactory):
|
||||
"""Factory for creating Park instances."""
|
||||
|
||||
@@ -127,19 +106,14 @@ class ParkFactory(DjangoModelFactory):
|
||||
size_acres = fuzzy.FuzzyDecimal(1, 1000, precision=2)
|
||||
website = factory.Faker("url")
|
||||
average_rating = fuzzy.FuzzyDecimal(1, 10, precision=2)
|
||||
ride_count = fuzzy.FuzzyInteger(5, 100)
|
||||
coaster_count = fuzzy.FuzzyInteger(1, 20)
|
||||
ride_count = fuzzy.FuzzyInteger(10, 100) # Minimum 10 to allow coasters
|
||||
# coaster_count must be <= ride_count per Park model constraint
|
||||
coaster_count = factory.LazyAttribute(lambda obj: min(obj.ride_count // 2, 20))
|
||||
|
||||
# Relationships
|
||||
operator = factory.SubFactory(OperatorCompanyFactory)
|
||||
property_owner = factory.SubFactory(OperatorCompanyFactory)
|
||||
|
||||
@factory.post_generation
|
||||
def create_location(obj, create, extracted, **kwargs):
|
||||
"""Create a location for the park."""
|
||||
if create:
|
||||
LocationFactory(content_object=obj, name=obj.name, location_type="park")
|
||||
|
||||
|
||||
class ClosedParkFactory(ParkFactory):
|
||||
"""Factory for creating closed parks."""
|
||||
@@ -163,6 +137,33 @@ class ParkAreaFactory(DjangoModelFactory):
|
||||
park = factory.SubFactory(ParkFactory)
|
||||
|
||||
|
||||
class RidesCompanyFactory(DjangoModelFactory):
|
||||
"""Factory for creating rides.Company instances (manufacturers, designers)."""
|
||||
|
||||
class Meta:
|
||||
model = "rides.Company"
|
||||
django_get_or_create = ("name",)
|
||||
|
||||
name = factory.Faker("company")
|
||||
slug = factory.LazyAttribute(lambda obj: slugify(obj.name))
|
||||
description = factory.Faker("text", max_nb_chars=500)
|
||||
website = factory.Faker("url")
|
||||
founded_year = fuzzy.FuzzyInteger(1800, 2024)
|
||||
roles = factory.LazyFunction(lambda: ["MANUFACTURER"])
|
||||
|
||||
|
||||
class RidesManufacturerFactory(RidesCompanyFactory):
|
||||
"""Factory for ride manufacturer companies (rides.Company)."""
|
||||
|
||||
roles = factory.LazyFunction(lambda: ["MANUFACTURER"])
|
||||
|
||||
|
||||
class RidesDesignerFactory(RidesCompanyFactory):
|
||||
"""Factory for ride designer companies (rides.Company)."""
|
||||
|
||||
roles = factory.LazyFunction(lambda: ["DESIGNER"])
|
||||
|
||||
|
||||
class RideModelFactory(DjangoModelFactory):
|
||||
"""Factory for creating RideModel instances."""
|
||||
|
||||
@@ -173,8 +174,8 @@ class RideModelFactory(DjangoModelFactory):
|
||||
name = factory.Faker("word")
|
||||
description = factory.Faker("text", max_nb_chars=500)
|
||||
|
||||
# Relationships
|
||||
manufacturer = factory.SubFactory(ManufacturerCompanyFactory)
|
||||
# Relationships - use rides.Company not parks.Company
|
||||
manufacturer = factory.SubFactory(RidesManufacturerFactory)
|
||||
|
||||
|
||||
class RideFactory(DjangoModelFactory):
|
||||
@@ -199,16 +200,12 @@ class RideFactory(DjangoModelFactory):
|
||||
|
||||
# Relationships
|
||||
park = factory.SubFactory(ParkFactory)
|
||||
manufacturer = factory.SubFactory(ManufacturerCompanyFactory)
|
||||
designer = factory.SubFactory(DesignerCompanyFactory)
|
||||
manufacturer = factory.SubFactory(RidesManufacturerFactory) # rides.Company
|
||||
designer = factory.SubFactory(RidesDesignerFactory) # rides.Company
|
||||
ride_model = factory.SubFactory(RideModelFactory)
|
||||
park_area = factory.SubFactory(ParkAreaFactory, park=factory.SelfAttribute("..park"))
|
||||
|
||||
@factory.post_generation
|
||||
def create_location(obj, create, extracted, **kwargs):
|
||||
"""Create a location for the ride."""
|
||||
if create:
|
||||
LocationFactory(content_object=obj, name=obj.name, location_type="ride")
|
||||
|
||||
|
||||
|
||||
class CoasterFactory(RideFactory):
|
||||
|
||||
@@ -78,6 +78,8 @@ urlpatterns = [
|
||||
path("accounts/", include("apps.accounts.urls")),
|
||||
# Default allauth URLs (for social auth and other features)
|
||||
path("accounts/", include("allauth.urls")),
|
||||
# Allauth headless API (for SPA passkey login, WebAuthn, etc.)
|
||||
path("_allauth/", include("allauth.headless.urls")),
|
||||
path(
|
||||
"accounts/email-required/",
|
||||
accounts_views.email_required,
|
||||
|
||||
1082
backend/uv.lock
generated
1082
backend/uv.lock
generated
File diff suppressed because it is too large
Load Diff
725
docs/allauth_integration_guide.md
Normal file
725
docs/allauth_integration_guide.md
Normal file
@@ -0,0 +1,725 @@
|
||||
# Django-Allauth Integration Guide for ThrillWiki
|
||||
|
||||
This guide documents how to properly integrate django-allauth for authentication in ThrillWiki, covering JWT tokens, password authentication, MFA (TOTP/WebAuthn), and social OAuth (Google/Discord).
|
||||
|
||||
---
|
||||
|
||||
## Table of Contents
|
||||
|
||||
1. [Installation & Setup](#installation--setup)
|
||||
2. [JWT Token Authentication](#jwt-token-authentication)
|
||||
3. [Password Authentication](#password-authentication)
|
||||
4. [MFA: TOTP (Authenticator App)](#mfa-totp-authenticator-app)
|
||||
5. [MFA: WebAuthn/Passkeys](#mfa-webauthnpasskeys)
|
||||
6. [Social OAuth: Google](#social-oauth-google)
|
||||
7. [Social OAuth: Discord](#social-oauth-discord)
|
||||
8. [API Patterns & DRF Integration](#api-patterns--drf-integration)
|
||||
9. [Internal API Reference](#internal-api-reference)
|
||||
|
||||
---
|
||||
|
||||
## Installation & Setup
|
||||
|
||||
### Required Packages
|
||||
|
||||
```bash
|
||||
# Add packages to pyproject.toml
|
||||
uv add "django-allauth[headless,mfa,socialaccount]"
|
||||
uv add fido2 # For WebAuthn support
|
||||
|
||||
# Install all dependencies
|
||||
uv sync
|
||||
```
|
||||
|
||||
### Running Django Commands
|
||||
|
||||
```bash
|
||||
# Run migrations
|
||||
uv run manage.py migrate
|
||||
|
||||
# Create superuser
|
||||
uv run manage.py createsuperuser
|
||||
|
||||
# Run development server
|
||||
uv run manage.py runserver
|
||||
|
||||
# Collect static files
|
||||
uv run manage.py collectstatic
|
||||
```
|
||||
|
||||
### INSTALLED_APPS Configuration
|
||||
|
||||
```python
|
||||
# config/django/base.py
|
||||
INSTALLED_APPS = [
|
||||
# Django built-in
|
||||
"django.contrib.auth",
|
||||
"django.contrib.sites",
|
||||
|
||||
# Allauth core (required)
|
||||
"allauth",
|
||||
"allauth.account",
|
||||
|
||||
# Optional modules
|
||||
"allauth.headless", # For headless/API mode
|
||||
"allauth.mfa", # MFA support (TOTP, recovery codes)
|
||||
"allauth.mfa.webauthn", # WebAuthn/Passkey support
|
||||
"allauth.socialaccount", # Social auth base
|
||||
"allauth.socialaccount.providers.google",
|
||||
"allauth.socialaccount.providers.discord",
|
||||
]
|
||||
```
|
||||
|
||||
### Middleware
|
||||
|
||||
```python
|
||||
MIDDLEWARE = [
|
||||
# ... other middleware
|
||||
"allauth.account.middleware.AccountMiddleware",
|
||||
]
|
||||
```
|
||||
|
||||
### URL Configuration
|
||||
|
||||
```python
|
||||
# urls.py
|
||||
urlpatterns = [
|
||||
# Allauth browser views (needed for OAuth callbacks)
|
||||
path("accounts/", include("allauth.urls")),
|
||||
|
||||
# Allauth headless API endpoints
|
||||
path("_allauth/", include("allauth.headless.urls")),
|
||||
]
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## JWT Token Authentication
|
||||
|
||||
### Configuration
|
||||
|
||||
```python
|
||||
# settings.py
|
||||
|
||||
# Token strategy - use JWT
|
||||
HEADLESS_TOKEN_STRATEGY = "allauth.headless.tokens.JWTTokenStrategy"
|
||||
|
||||
# Generate private key: openssl genpkey -algorithm RSA -out private_key.pem -pkeyopt rsa_keygen_bits:2048
|
||||
HEADLESS_JWT_PRIVATE_KEY = """
|
||||
-----BEGIN PRIVATE KEY-----
|
||||
MIIEvQIBADANBgkqhkiG9w0BAQEFAASC...
|
||||
-----END PRIVATE KEY-----
|
||||
"""
|
||||
|
||||
# Token lifetimes
|
||||
HEADLESS_JWT_ACCESS_TOKEN_EXPIRES_IN = 300 # 5 minutes
|
||||
HEADLESS_JWT_REFRESH_TOKEN_EXPIRES_IN = 86400 # 24 hours
|
||||
|
||||
# Authorization header scheme
|
||||
HEADLESS_JWT_AUTHORIZATION_HEADER_SCHEME = "Bearer"
|
||||
|
||||
# Stateful validation (invalidates tokens on logout)
|
||||
HEADLESS_JWT_STATEFUL_VALIDATION_ENABLED = True
|
||||
|
||||
# Rotate refresh tokens on use
|
||||
HEADLESS_JWT_ROTATE_REFRESH_TOKEN = True
|
||||
```
|
||||
|
||||
### DRF Integration
|
||||
|
||||
```python
|
||||
from allauth.headless.contrib.rest_framework.authentication import JWTTokenAuthentication
|
||||
from rest_framework.permissions import IsAuthenticated
|
||||
from rest_framework.views import APIView
|
||||
|
||||
class ProtectedAPIView(APIView):
|
||||
authentication_classes = [JWTTokenAuthentication]
|
||||
permission_classes = [IsAuthenticated]
|
||||
|
||||
def get(self, request):
|
||||
return Response({"user": request.user.email})
|
||||
```
|
||||
|
||||
### JWT Flow
|
||||
|
||||
1. User authenticates (password/social/passkey)
|
||||
2. During auth, pass `X-Session-Token` header to allauth API
|
||||
3. Upon successful authentication, response `meta` contains:
|
||||
```json
|
||||
{
|
||||
"meta": {
|
||||
"access_token": "eyJ...",
|
||||
"refresh_token": "abc123..."
|
||||
}
|
||||
}
|
||||
```
|
||||
4. Use `Authorization: Bearer <access_token>` for subsequent requests
|
||||
5. Refresh tokens via `POST /_allauth/browser/v1/auth/token/refresh`
|
||||
|
||||
---
|
||||
|
||||
## Password Authentication
|
||||
|
||||
### Configuration
|
||||
|
||||
```python
|
||||
# config/settings/third_party.py
|
||||
|
||||
# Signup fields (* = required)
|
||||
ACCOUNT_SIGNUP_FIELDS = ["email*", "username*", "password1*", "password2*"]
|
||||
|
||||
# Login methods
|
||||
ACCOUNT_LOGIN_METHODS = {"email", "username"} # Allow both
|
||||
|
||||
# Email verification
|
||||
ACCOUNT_EMAIL_VERIFICATION = "mandatory" # Options: "mandatory", "optional", "none"
|
||||
ACCOUNT_EMAIL_VERIFICATION_SUPPORTS_CHANGE = True
|
||||
ACCOUNT_EMAIL_VERIFICATION_SUPPORTS_RESEND = True
|
||||
|
||||
# Security
|
||||
ACCOUNT_REAUTHENTICATION_REQUIRED = True # Require re-auth for sensitive operations
|
||||
ACCOUNT_EMAIL_NOTIFICATIONS = True
|
||||
ACCOUNT_EMAIL_UNKNOWN_ACCOUNTS = False # Don't reveal if email exists
|
||||
|
||||
# Redirects
|
||||
LOGIN_REDIRECT_URL = "/"
|
||||
ACCOUNT_LOGOUT_REDIRECT_URL = "/"
|
||||
|
||||
# Custom adapters
|
||||
ACCOUNT_ADAPTER = "apps.accounts.adapters.CustomAccountAdapter"
|
||||
```
|
||||
|
||||
### Headless API Endpoints
|
||||
|
||||
| Endpoint | Method | Description |
|
||||
|----------|--------|-------------|
|
||||
| `/_allauth/browser/v1/auth/login` | POST | Login with email/username + password |
|
||||
| `/_allauth/browser/v1/auth/signup` | POST | Register new account |
|
||||
| `/_allauth/browser/v1/auth/logout` | POST | Logout (invalidate tokens) |
|
||||
| `/_allauth/browser/v1/auth/password/reset` | POST | Request password reset |
|
||||
| `/_allauth/browser/v1/auth/password/reset/key` | POST | Complete password reset |
|
||||
| `/_allauth/browser/v1/auth/password/change` | POST | Change password (authenticated) |
|
||||
|
||||
### Magic Link (Login by Code)
|
||||
|
||||
```python
|
||||
# Enable magic link authentication
|
||||
ACCOUNT_LOGIN_BY_CODE_ENABLED = True
|
||||
ACCOUNT_LOGIN_BY_CODE_MAX_ATTEMPTS = 3
|
||||
ACCOUNT_LOGIN_BY_CODE_TIMEOUT = 300 # 5 minutes
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## MFA: TOTP (Authenticator App)
|
||||
|
||||
### Configuration
|
||||
|
||||
```python
|
||||
# config/settings/third_party.py
|
||||
|
||||
# Enable TOTP in supported types
|
||||
MFA_SUPPORTED_TYPES = ["totp", "webauthn"]
|
||||
|
||||
# TOTP settings
|
||||
MFA_TOTP_ISSUER = "ThrillWiki" # Shows in authenticator app
|
||||
MFA_TOTP_DIGITS = 6 # Code length
|
||||
MFA_TOTP_PERIOD = 30 # Seconds per code
|
||||
```
|
||||
|
||||
### TOTP API Flow
|
||||
|
||||
1. **Get TOTP Secret** (for QR code):
|
||||
```
|
||||
GET /_allauth/browser/v1/account/authenticators/totp
|
||||
```
|
||||
Response contains `totp_url` for QR code generation.
|
||||
|
||||
2. **Activate TOTP**:
|
||||
```
|
||||
POST /_allauth/browser/v1/account/authenticators/totp
|
||||
{
|
||||
"code": "123456"
|
||||
}
|
||||
```
|
||||
|
||||
3. **Deactivate TOTP**:
|
||||
```
|
||||
DELETE /_allauth/browser/v1/account/authenticators/totp
|
||||
```
|
||||
|
||||
4. **MFA Login Flow**:
|
||||
- After password auth, if MFA enabled, receive `401` with `mfa_required`
|
||||
- Submit TOTP code:
|
||||
```
|
||||
POST /_allauth/browser/v1/auth/2fa/authenticate
|
||||
{
|
||||
"code": "123456"
|
||||
}
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## MFA: WebAuthn/Passkeys
|
||||
|
||||
### Configuration
|
||||
|
||||
```python
|
||||
# config/settings/third_party.py
|
||||
|
||||
# Include webauthn in supported types
|
||||
MFA_SUPPORTED_TYPES = ["totp", "webauthn"]
|
||||
|
||||
# Enable passkey-only login
|
||||
MFA_PASSKEY_LOGIN_ENABLED = True
|
||||
|
||||
# Allow insecure origin for localhost development
|
||||
MFA_WEBAUTHN_ALLOW_INSECURE_ORIGIN = True # Only for DEBUG=True
|
||||
```
|
||||
|
||||
### Internal WebAuthn API Functions
|
||||
|
||||
The `allauth.mfa.webauthn.internal.auth` module provides these functions:
|
||||
|
||||
```python
|
||||
from allauth.mfa.webauthn.internal import auth as webauthn_auth
|
||||
|
||||
# Registration Flow
|
||||
def begin_registration(user, passwordless: bool) -> Dict:
|
||||
"""
|
||||
Start passkey registration.
|
||||
|
||||
Args:
|
||||
user: The Django user object
|
||||
passwordless: True for passkey login, False for MFA-only
|
||||
|
||||
Returns:
|
||||
Dict with WebAuthn creation options (challenge, rp, user, etc.)
|
||||
|
||||
Note: State is stored internally via set_state()
|
||||
"""
|
||||
|
||||
def complete_registration(credential: Dict) -> AuthenticatorData:
|
||||
"""
|
||||
Complete passkey registration.
|
||||
|
||||
Args:
|
||||
credential: The parsed credential response from browser
|
||||
|
||||
Returns:
|
||||
AuthenticatorData (binding) - NOT an Authenticator model
|
||||
|
||||
Note: You must create the Authenticator record yourself!
|
||||
"""
|
||||
|
||||
# Authentication Flow
|
||||
def begin_authentication(user=None) -> Dict:
|
||||
"""
|
||||
Start passkey authentication.
|
||||
|
||||
Args:
|
||||
user: Optional user (for MFA). None for passwordless login.
|
||||
|
||||
Returns:
|
||||
Dict with WebAuthn request options
|
||||
|
||||
Note: State is stored internally via set_state()
|
||||
"""
|
||||
|
||||
def complete_authentication(user, response: Dict) -> Authenticator:
|
||||
"""
|
||||
Complete passkey authentication.
|
||||
|
||||
Args:
|
||||
user: The Django user object
|
||||
response: The credential response from browser
|
||||
|
||||
Returns:
|
||||
The matching Authenticator model instance
|
||||
"""
|
||||
|
||||
# State Management (internal, use context)
|
||||
def get_state() -> Optional[Dict]:
|
||||
"""Get stored WebAuthn state from session."""
|
||||
|
||||
def set_state(state: Dict) -> None:
|
||||
"""Store WebAuthn state in session."""
|
||||
|
||||
def clear_state() -> None:
|
||||
"""Clear WebAuthn state from session."""
|
||||
|
||||
# Helper functions
|
||||
def parse_registration_response(response: Any) -> RegistrationResponse:
|
||||
"""Parse browser registration response."""
|
||||
|
||||
def parse_authentication_response(response: Any) -> AuthenticationResponse:
|
||||
"""Parse browser authentication response."""
|
||||
```
|
||||
|
||||
### Custom Passkey API Implementation
|
||||
|
||||
```python
|
||||
# apps/api/v1/auth/passkey.py
|
||||
|
||||
from allauth.mfa.webauthn.internal import auth as webauthn_auth
|
||||
from allauth.mfa.models import Authenticator
|
||||
|
||||
@api_view(["GET"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def get_registration_options(request):
|
||||
"""Get WebAuthn registration options."""
|
||||
# passwordless=False for MFA passkeys, True for passwordless login
|
||||
creation_options = webauthn_auth.begin_registration(
|
||||
request.user,
|
||||
passwordless=False
|
||||
)
|
||||
# State is stored internally
|
||||
return Response({"options": creation_options})
|
||||
|
||||
@api_view(["POST"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def register_passkey(request):
|
||||
"""Complete passkey registration."""
|
||||
credential = request.data.get("credential")
|
||||
name = request.data.get("name", "Passkey")
|
||||
|
||||
# Check for pending registration
|
||||
state = webauthn_auth.get_state()
|
||||
if not state:
|
||||
return Response({"error": "No pending registration"}, status=400)
|
||||
|
||||
# Parse and complete registration
|
||||
credential_data = webauthn_auth.parse_registration_response(credential)
|
||||
authenticator_data = webauthn_auth.complete_registration(credential_data)
|
||||
|
||||
# Create Authenticator record manually
|
||||
authenticator = Authenticator.objects.create(
|
||||
user=request.user,
|
||||
type=Authenticator.Type.WEBAUTHN,
|
||||
data={"name": name},
|
||||
)
|
||||
|
||||
return Response({"id": str(authenticator.id), "name": name})
|
||||
|
||||
@api_view(["GET"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def get_authentication_options(request):
|
||||
"""Get WebAuthn authentication options."""
|
||||
request_options = webauthn_auth.begin_authentication(request.user)
|
||||
return Response({"options": request_options})
|
||||
|
||||
@api_view(["POST"])
|
||||
@permission_classes([IsAuthenticated])
|
||||
def authenticate_passkey(request):
|
||||
"""Verify passkey authentication."""
|
||||
credential = request.data.get("credential")
|
||||
|
||||
state = webauthn_auth.get_state()
|
||||
if not state:
|
||||
return Response({"error": "No pending authentication"}, status=400)
|
||||
|
||||
# Complete authentication (handles state internally)
|
||||
webauthn_auth.complete_authentication(request.user, credential)
|
||||
|
||||
return Response({"success": True})
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Social OAuth: Google
|
||||
|
||||
### Google Cloud Console Setup
|
||||
|
||||
1. Go to [Google Cloud Console](https://console.developers.google.com/)
|
||||
2. Create/select project → APIs & Services → Credentials
|
||||
3. Create OAuth 2.0 Client ID (Web application)
|
||||
4. Set Authorized JavaScript origins:
|
||||
- `http://localhost:3000` (development)
|
||||
- `https://thrillwiki.com` (production)
|
||||
5. Set Authorized redirect URIs:
|
||||
- `http://localhost:8000/accounts/google/login/callback/`
|
||||
- `https://api.thrillwiki.com/accounts/google/login/callback/`
|
||||
6. Note the Client ID and Client Secret
|
||||
|
||||
### Django Configuration
|
||||
|
||||
```python
|
||||
# config/settings/third_party.py
|
||||
|
||||
SOCIALACCOUNT_PROVIDERS = {
|
||||
"google": {
|
||||
"SCOPE": ["profile", "email"],
|
||||
"AUTH_PARAMS": {"access_type": "online"}, # Use "offline" for refresh tokens
|
||||
"OAUTH_PKCE_ENABLED": True,
|
||||
# "FETCH_USERINFO": True, # If you need avatar_url for private profiles
|
||||
},
|
||||
}
|
||||
```
|
||||
|
||||
### Admin Setup
|
||||
|
||||
1. Go to `/admin/socialaccount/socialapp/`
|
||||
2. Add new Social Application:
|
||||
- Provider: Google
|
||||
- Name: Google
|
||||
- Client ID: (from Google Console)
|
||||
- Secret key: (from Google Console)
|
||||
- Sites: Select your site
|
||||
|
||||
---
|
||||
|
||||
## Social OAuth: Discord
|
||||
|
||||
### Discord Developer Portal Setup
|
||||
|
||||
1. Go to [Discord Developer Portal](https://discordapp.com/developers/applications/me)
|
||||
2. Create New Application
|
||||
3. Go to OAuth2 → General
|
||||
4. Add Redirect URIs:
|
||||
- `http://127.0.0.1:8000/accounts/discord/login/callback/`
|
||||
- `https://api.thrillwiki.com/accounts/discord/login/callback/`
|
||||
5. Note Client ID and Client Secret
|
||||
|
||||
### Django Configuration
|
||||
|
||||
```python
|
||||
# config/settings/third_party.py
|
||||
|
||||
SOCIALACCOUNT_PROVIDERS = {
|
||||
"discord": {
|
||||
"SCOPE": ["identify", "email"], # "identify" is required
|
||||
"OAUTH_PKCE_ENABLED": True,
|
||||
},
|
||||
}
|
||||
```
|
||||
|
||||
### Admin Setup
|
||||
|
||||
1. Go to `/admin/socialaccount/socialapp/`
|
||||
2. Add new Social Application:
|
||||
- Provider: Discord
|
||||
- Name: Discord
|
||||
- Client ID: (from Discord Portal)
|
||||
- Secret key: (from Discord Portal)
|
||||
- Sites: Select your site
|
||||
|
||||
---
|
||||
|
||||
## API Patterns & DRF Integration
|
||||
|
||||
### Authentication Classes
|
||||
|
||||
```python
|
||||
from allauth.headless.contrib.rest_framework.authentication import (
|
||||
JWTTokenAuthentication,
|
||||
SessionTokenAuthentication,
|
||||
)
|
||||
|
||||
# For JWT-based authentication
|
||||
REST_FRAMEWORK = {
|
||||
"DEFAULT_AUTHENTICATION_CLASSES": [
|
||||
"allauth.headless.contrib.rest_framework.authentication.JWTTokenAuthentication",
|
||||
],
|
||||
}
|
||||
```
|
||||
|
||||
### Headless Frontend URLs
|
||||
|
||||
```python
|
||||
# Required for email verification, password reset links
|
||||
HEADLESS_FRONTEND_URLS = {
|
||||
"account_confirm_email": "https://thrillwiki.com/account/verify-email/{key}",
|
||||
"account_reset_password_from_key": "https://thrillwiki.com/account/password/reset/{key}",
|
||||
"account_signup": "https://thrillwiki.com/signup",
|
||||
}
|
||||
```
|
||||
|
||||
### Custom Adapters
|
||||
|
||||
```python
|
||||
# apps/accounts/adapters.py
|
||||
from allauth.account.adapter import DefaultAccountAdapter
|
||||
from allauth.socialaccount.adapter import DefaultSocialAccountAdapter
|
||||
|
||||
class CustomAccountAdapter(DefaultAccountAdapter):
|
||||
def save_user(self, request, user, form, commit=True):
|
||||
"""Customize user creation."""
|
||||
user = super().save_user(request, user, form, commit=False)
|
||||
# Custom logic here
|
||||
if commit:
|
||||
user.save()
|
||||
return user
|
||||
|
||||
def is_open_for_signup(self, request):
|
||||
"""Control signup availability."""
|
||||
return True
|
||||
|
||||
class CustomSocialAccountAdapter(DefaultSocialAccountAdapter):
|
||||
def pre_social_login(self, request, sociallogin):
|
||||
"""Hook before social login completes."""
|
||||
# Link social account to existing user by email
|
||||
if sociallogin.is_existing:
|
||||
return
|
||||
|
||||
email = sociallogin.account.extra_data.get("email")
|
||||
if email:
|
||||
User = get_user_model()
|
||||
try:
|
||||
user = User.objects.get(email=email)
|
||||
sociallogin.connect(request, user)
|
||||
except User.DoesNotExist:
|
||||
pass
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Internal API Reference
|
||||
|
||||
### Authenticator Model
|
||||
|
||||
```python
|
||||
from allauth.mfa.models import Authenticator
|
||||
|
||||
# Types
|
||||
Authenticator.Type.TOTP # TOTP authenticator
|
||||
Authenticator.Type.WEBAUTHN # WebAuthn/Passkey
|
||||
Authenticator.Type.RECOVERY_CODES # Recovery codes
|
||||
|
||||
# Query user's authenticators
|
||||
passkeys = Authenticator.objects.filter(
|
||||
user=user,
|
||||
type=Authenticator.Type.WEBAUTHN
|
||||
)
|
||||
|
||||
# Check if MFA is enabled
|
||||
from allauth.mfa.adapter import get_adapter
|
||||
is_mfa_enabled = get_adapter().is_mfa_enabled(user)
|
||||
```
|
||||
|
||||
### Session Token Header
|
||||
|
||||
For headless mode during authentication flow:
|
||||
```
|
||||
X-Session-Token: <session-token>
|
||||
```
|
||||
|
||||
After authentication completes with JWT enabled:
|
||||
```
|
||||
Authorization: Bearer <access-token>
|
||||
```
|
||||
|
||||
## Current ThrillWiki Implementation Summary
|
||||
|
||||
ThrillWiki uses a hybrid authentication system with django-allauth for MFA and social auth, and SimpleJWT for API tokens.
|
||||
|
||||
### Backend Configuration
|
||||
|
||||
| Feature | Status | Notes |
|
||||
|---------|--------|-------|
|
||||
| Password Auth | ✅ Active | Email + username login |
|
||||
| Email Verification | ✅ Mandatory | With resend support |
|
||||
| TOTP MFA | ✅ Active | 6-digit codes, 30s period |
|
||||
| WebAuthn/Passkeys | ✅ Active | Passkey login enabled, counts as MFA |
|
||||
| Google OAuth | ✅ Configured | Requires admin SocialApp setup |
|
||||
| Discord OAuth | ✅ Configured | Requires admin SocialApp setup |
|
||||
| Magic Link | ✅ Configured | 5-minute timeout |
|
||||
| JWT Tokens | ✅ SimpleJWT | 15min access, 7 day refresh |
|
||||
|
||||
### Frontend MFA Integration (Updated 2026-01-10)
|
||||
|
||||
The frontend recognizes both TOTP and Passkeys as valid MFA factors:
|
||||
|
||||
```typescript
|
||||
// authService.ts - getEnrolledFactors()
|
||||
// Checks both Supabase TOTP AND Django passkeys
|
||||
const mfaStatus = await djangoClient.rpc('get_mfa_status', {});
|
||||
if (statusData.passkey_enabled && statusData.passkey_count > 0) {
|
||||
factors.push({ id: 'passkey', factor_type: 'webauthn', ... });
|
||||
}
|
||||
```
|
||||
|
||||
### Admin Panel MFA Requirements
|
||||
|
||||
Admins and moderators must have MFA enabled to access protected routes:
|
||||
|
||||
1. `useAdminGuard()` hook checks MFA enrollment via `useRequireMFA()`
|
||||
2. `getEnrolledFactors()` queries Django's `get_mfa_status` endpoint
|
||||
3. Backend returns `has_second_factor: true` if TOTP or Passkey is enabled
|
||||
4. Users with only passkeys (no TOTP) now pass the MFA requirement
|
||||
|
||||
### API Endpoints Reference
|
||||
|
||||
#### Authentication
|
||||
| Endpoint | Method | Description |
|
||||
|----------|--------|-------------|
|
||||
| `/api/v1/auth/login/` | POST | Password login |
|
||||
| `/api/v1/auth/login/mfa-verify/` | POST | Complete MFA (TOTP or Passkey) |
|
||||
| `/api/v1/auth/signup/` | POST | Register new account |
|
||||
| `/api/v1/auth/logout/` | POST | Logout, blacklist tokens |
|
||||
| `/api/v1/auth/token/refresh/` | POST | Refresh JWT access token |
|
||||
|
||||
#### MFA (TOTP)
|
||||
| Endpoint | Method | Description |
|
||||
|----------|--------|-------------|
|
||||
| `/api/v1/auth/mfa/status/` | GET | Get MFA status (TOTP + Passkey) |
|
||||
| `/api/v1/auth/mfa/totp/setup/` | POST | Start TOTP enrollment |
|
||||
| `/api/v1/auth/mfa/totp/activate/` | POST | Activate with 6-digit code |
|
||||
| `/api/v1/auth/mfa/totp/deactivate/` | POST | Remove TOTP (requires password) |
|
||||
|
||||
#### Passkeys (WebAuthn)
|
||||
| Endpoint | Method | Description |
|
||||
|----------|--------|-------------|
|
||||
| `/api/v1/auth/passkey/status/` | GET | List registered passkeys |
|
||||
| `/api/v1/auth/passkey/registration-options/` | GET | Get WebAuthn creation options |
|
||||
| `/api/v1/auth/passkey/register/` | POST | Complete passkey registration |
|
||||
| `/api/v1/auth/passkey/login-options/` | POST | Get auth options (uses mfa_token) |
|
||||
| `/api/v1/auth/passkey/{id}/` | DELETE | Remove passkey |
|
||||
|
||||
#### Social Authentication
|
||||
| Endpoint | Method | Description |
|
||||
|----------|--------|-------------|
|
||||
| `/api/v1/auth/social/providers/` | GET | List configured providers |
|
||||
| `/api/v1/auth/social/connect/{provider}/` | POST | Start OAuth flow |
|
||||
| `/api/v1/auth/social/disconnect/{provider}/` | POST | Unlink provider |
|
||||
|
||||
### Login Flow with MFA
|
||||
|
||||
```
|
||||
1. POST /api/v1/auth/login/ {username, password}
|
||||
└── If MFA enabled: Returns {mfa_required: true, mfa_token, mfa_types: ["totp", "webauthn"]}
|
||||
|
||||
2a. TOTP Verification:
|
||||
POST /api/v1/auth/login/mfa-verify/ {mfa_token, code: "123456"}
|
||||
|
||||
2b. Passkey Verification:
|
||||
POST /api/v1/auth/passkey/login-options/ {mfa_token} ← Get challenge
|
||||
Browser: navigator.credentials.get() ← User authenticates
|
||||
POST /api/v1/auth/login/mfa-verify/ {mfa_token, credential: {...}}
|
||||
|
||||
3. Returns: {access, refresh, user, message: "Login successful"}
|
||||
```
|
||||
|
||||
### Frontend Components
|
||||
|
||||
| Component | Purpose |
|
||||
|-----------|---------|
|
||||
| `MFAChallenge.tsx` | TOTP code entry during login |
|
||||
| `MFAEnrollmentRequired.tsx` | Prompts admin/mod to set up MFA |
|
||||
| `MFAGuard.tsx` | Wraps routes requiring MFA |
|
||||
| `useRequireMFA` | Hook checking MFA enrollment |
|
||||
| `useAdminGuard` | Combines auth + role + MFA checks |
|
||||
|
||||
### Admin MFA Requirements
|
||||
|
||||
Moderators, admins, and superusers **must have MFA enrolled** to access admin pages.
|
||||
|
||||
The system uses **enrollment-based verification** rather than per-session AAL2 tokens:
|
||||
- MFA verification happens at login time via the `mfa_token` flow
|
||||
- Django-allauth doesn't embed AAL claims in JWT tokens
|
||||
- The frontend checks if the user has TOTP or passkey enrolled
|
||||
- Mid-session MFA step-up is not supported (user must re-login)
|
||||
|
||||
This means:
|
||||
- `useRequireMFA` returns `hasMFA: true` if user has any factor enrolled
|
||||
- `useAdminGuard` blocks access if `needsEnrollment` is true
|
||||
- Users prompted to enroll MFA on their first admin page visit
|
||||
174
scripts/lint_choices.py
Executable file
174
scripts/lint_choices.py
Executable file
@@ -0,0 +1,174 @@
|
||||
#!/usr/bin/env python
|
||||
"""
|
||||
RichChoiceField Enforcement Lint Script
|
||||
|
||||
This script checks for prohibited choice patterns in the codebase.
|
||||
Exit code 0 = no violations found, exit code 1 = violations found.
|
||||
|
||||
Usage:
|
||||
python scripts/lint_choices.py
|
||||
python scripts/lint_choices.py --fix # Show fix suggestions
|
||||
|
||||
Add to CI:
|
||||
python scripts/lint_choices.py || exit 1
|
||||
"""
|
||||
|
||||
import argparse
|
||||
import re
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
# Patterns to detect prohibited choice usage
|
||||
PROHIBITED_PATTERNS = [
|
||||
# TextChoices / IntegerChoices class definitions
|
||||
(r'class\s+\w+\s*\(\s*models\.(TextChoices|IntegerChoices)\s*\)',
|
||||
'models.TextChoices/IntegerChoices class definition'),
|
||||
|
||||
# Inline tuple choices in CharField
|
||||
(r'choices\s*=\s*\[\s*\(\s*["\']',
|
||||
'Inline tuple choices'),
|
||||
|
||||
# Direct reference to .choices attribute
|
||||
(r'choices\s*=\s*\w+\.choices',
|
||||
'Reference to inner TextChoices.choices'),
|
||||
]
|
||||
|
||||
# Directories/files to exclude
|
||||
EXCLUDE_PATTERNS = [
|
||||
'*/migrations/*',
|
||||
'*/.venv/*',
|
||||
'*/node_modules/*',
|
||||
'*/__pycache__/*',
|
||||
'*.pyc',
|
||||
'lint_choices.py', # Exclude this script
|
||||
]
|
||||
|
||||
# Files allowed to define TextChoices (infrastructure only)
|
||||
ALLOWED_EXCEPTION_FILES = [
|
||||
# Core choice infrastructure files can reference these patterns
|
||||
'apps/core/choices/',
|
||||
]
|
||||
|
||||
|
||||
def should_exclude(path: Path) -> bool:
|
||||
"""Check if path should be excluded from linting."""
|
||||
path_str = str(path)
|
||||
for pattern in EXCLUDE_PATTERNS:
|
||||
if pattern.startswith('*/'):
|
||||
if pattern[2:].rstrip('/*') in path_str:
|
||||
return True
|
||||
elif pattern.endswith('/*'):
|
||||
if path_str.startswith(pattern[:-2]):
|
||||
return True
|
||||
elif pattern in path_str:
|
||||
return True
|
||||
return False
|
||||
|
||||
|
||||
def is_exception_file(path: Path) -> bool:
|
||||
"""Check if file is an allowed exception."""
|
||||
path_str = str(path)
|
||||
for exception in ALLOWED_EXCEPTION_FILES:
|
||||
if exception in path_str:
|
||||
return True
|
||||
return False
|
||||
|
||||
|
||||
def scan_file(filepath: Path) -> list[tuple[int, str, str]]:
|
||||
"""Scan a file for prohibited patterns. Returns list of (line_num, line, pattern_name)."""
|
||||
violations = []
|
||||
|
||||
if should_exclude(filepath) or is_exception_file(filepath):
|
||||
return violations
|
||||
|
||||
try:
|
||||
content = filepath.read_text(encoding='utf-8')
|
||||
lines = content.split('\n')
|
||||
|
||||
for i, line in enumerate(lines, 1):
|
||||
# Skip comments
|
||||
stripped = line.strip()
|
||||
if stripped.startswith('#'):
|
||||
continue
|
||||
|
||||
for pattern, description in PROHIBITED_PATTERNS:
|
||||
if re.search(pattern, line):
|
||||
violations.append((i, line.strip(), description))
|
||||
break # Only report one violation per line
|
||||
|
||||
except Exception as e:
|
||||
print(f"Warning: Could not read {filepath}: {e}", file=sys.stderr)
|
||||
|
||||
return violations
|
||||
|
||||
|
||||
def scan_directory(root_dir: Path) -> dict[Path, list]:
|
||||
"""Scan all Python files in directory for violations."""
|
||||
all_violations = {}
|
||||
|
||||
for filepath in root_dir.rglob('*.py'):
|
||||
violations = scan_file(filepath)
|
||||
if violations:
|
||||
all_violations[filepath] = violations
|
||||
|
||||
return all_violations
|
||||
|
||||
|
||||
def print_violations(violations: dict, show_fix: bool = False):
|
||||
"""Print violations in a readable format."""
|
||||
total = 0
|
||||
|
||||
for filepath, file_violations in sorted(violations.items()):
|
||||
print(f"\n\033[1;31m{filepath}\033[0m")
|
||||
for line_num, line_content, description in file_violations:
|
||||
print(f" Line {line_num}: [{description}]")
|
||||
print(f" {line_content[:80]}{'...' if len(line_content) > 80 else ''}")
|
||||
total += 1
|
||||
|
||||
if show_fix:
|
||||
print(f" \033[1;33mFix:\033[0m Use RichChoiceField(choice_group='...', domain='...')")
|
||||
|
||||
return total
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description='Lint for prohibited choice patterns')
|
||||
parser.add_argument('--fix', action='store_true', help='Show fix suggestions')
|
||||
parser.add_argument('path', nargs='?', default='apps', help='Path to scan (default: apps)')
|
||||
args = parser.parse_args()
|
||||
|
||||
# Find backend directory
|
||||
script_dir = Path(__file__).parent
|
||||
backend_dir = script_dir.parent
|
||||
if (backend_dir / 'apps').exists():
|
||||
scan_path = backend_dir / args.path
|
||||
elif (backend_dir / 'backend' / 'apps').exists():
|
||||
scan_path = backend_dir / 'backend' / args.path
|
||||
else:
|
||||
print("Error: Could not find apps directory", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
print(f"Scanning {scan_path} for prohibited choice patterns...")
|
||||
print("=" * 60)
|
||||
|
||||
violations = scan_directory(scan_path)
|
||||
|
||||
if violations:
|
||||
total = print_violations(violations, show_fix=args.fix)
|
||||
print("\n" + "=" * 60)
|
||||
print(f"\033[1;31mFound {total} violation(s) in {len(violations)} file(s)\033[0m")
|
||||
print("\nProhibited patterns:")
|
||||
print(" - models.TextChoices / models.IntegerChoices classes")
|
||||
print(" - Inline choices=[(value, label), ...]")
|
||||
print(" - References to InnerClass.choices")
|
||||
print("\nRequired pattern:")
|
||||
print(" RichChoiceField(choice_group='group_name', domain='domain_name')")
|
||||
print("\nSee: .agent/workflows/enforce-richchoice.md for migration guide")
|
||||
sys.exit(1)
|
||||
else:
|
||||
print("\n\033[1;32m✓ No prohibited choice patterns found!\033[0m")
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
||||
Reference in New Issue
Block a user