Files
breakpilot-compliance/backend-compliance/compliance/services/dse_parser.py
T
Benjamin Admin 275bdf9848
Build + Deploy / build-admin-compliance (push) Successful in 1m49s
Build + Deploy / build-backend-compliance (push) Successful in 2m57s
Build + Deploy / build-ai-sdk (push) Successful in 50s
Build + Deploy / build-developer-portal (push) Successful in 1m2s
Build + Deploy / build-tts (push) Successful in 1m23s
Build + Deploy / build-document-crawler (push) Successful in 39s
Build + Deploy / build-dsms-gateway (push) Successful in 23s
Build + Deploy / build-dsms-node (push) Successful in 10s
CI / branch-name (push) Has been skipped
CI / guardrail-integrity (push) Has been skipped
CI / loc-budget (push) Failing after 21s
CI / secret-scan (push) Has been skipped
CI / go-lint (push) Has been skipped
CI / python-lint (push) Has been skipped
CI / nodejs-lint (push) Has been skipped
CI / nodejs-build (push) Successful in 2m31s
CI / dep-audit (push) Has been skipped
CI / sbom-scan (push) Has been skipped
CI / test-go (push) Failing after 41s
CI / test-python-backend (push) Successful in 40s
CI / test-python-document-crawler (push) Successful in 25s
CI / test-python-dsms-gateway (push) Successful in 20s
CI / validate-canonical-controls (push) Successful in 13s
Build + Deploy / trigger-orca (push) Successful in 2m46s
fix: Add missing service modules required by agent_scan_routes
These files existed on the feature branch but were never cherry-picked
to main, causing ModuleNotFoundError on import:
- dse_parser.py — parses DSE HTML into structured sections
- dse_matcher.py — matches detected services against DSE sections
- mandatory_content_checker.py — checks Art. 13 DSGVO mandatory fields
- legal_basis_validator.py — validates legal basis (lit. a-f)

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-05-04 23:23:02 +02:00

225 lines
7.4 KiB
Python

"""
DSE Parser — parses privacy policy HTML into structured sections.
Extracts headings, section numbers, content blocks and builds a
hierarchical structure that enables precise text references.
"""
import logging
import re
from dataclasses import dataclass, field
from html.parser import HTMLParser
logger = logging.getLogger(__name__)
@dataclass
class DSESection:
"""A section in a privacy policy."""
heading: str
heading_level: int # 1-4
section_number: str # "2.5" or "" if no number
content: str # Plain text content
html: str # Original HTML content
parent_heading: str = ""
url: str = ""
element_id: str = ""
paragraph_count: int = 0
class _HeadingExtractor(HTMLParser):
"""Extract headings and their content from HTML."""
def __init__(self):
super().__init__()
self.sections: list[dict] = []
self._current_tag = ""
self._in_heading = False
self._heading_level = 0
self._heading_text = ""
self._heading_id = ""
self._content_parts: list[str] = []
self._html_parts: list[str] = []
self._skip_tags = {"script", "style", "nav", "footer", "header"}
self._skip_depth = 0
self._p_count = 0
def handle_starttag(self, tag, attrs):
attrs_dict = dict(attrs)
if tag in self._skip_tags:
self._skip_depth += 1
return
if self._skip_depth > 0:
return
if tag in ("h1", "h2", "h3", "h4"):
# Save previous section
if self._heading_text:
self._save_section()
self._in_heading = True
self._heading_level = int(tag[1])
self._heading_text = ""
self._heading_id = attrs_dict.get("id", "")
self._content_parts = []
self._html_parts = []
self._p_count = 0
if tag == "p":
self._p_count += 1
# Reconstruct HTML
attr_str = " ".join(f'{k}="{v}"' for k, v in attrs)
self._html_parts.append(f"<{tag}{' ' + attr_str if attr_str else ''}>")
def handle_endtag(self, tag):
if tag in self._skip_tags and self._skip_depth > 0:
self._skip_depth -= 1
return
if self._skip_depth > 0:
return
if tag in ("h1", "h2", "h3", "h4"):
self._in_heading = False
self._html_parts.append(f"</{tag}>")
def handle_data(self, data):
if self._skip_depth > 0:
return
if self._in_heading:
self._heading_text += data.strip()
else:
self._content_parts.append(data)
self._html_parts.append(data)
def _save_section(self):
if not self._heading_text:
return
content = " ".join(self._content_parts)
content = re.sub(r"\s+", " ", content).strip()
self.sections.append({
"heading": self._heading_text.strip(),
"heading_level": self._heading_level,
"element_id": self._heading_id,
"content": content,
"html": "".join(self._html_parts),
"paragraph_count": self._p_count,
})
def finalize(self):
"""Call after feeding all data to save the last section."""
if self._heading_text:
self._save_section()
def parse_dse(html: str, url: str = "") -> list[DSESection]:
"""Parse privacy policy HTML into structured sections."""
extractor = _HeadingExtractor()
try:
extractor.feed(html)
extractor.finalize()
except Exception as e:
logger.warning("HTML parsing failed, falling back to regex: %s", e)
return _regex_fallback(html, url)
if not extractor.sections:
return _regex_fallback(html, url)
# Build parent hierarchy
sections: list[DSESection] = []
parent_stack: list[str] = [""] # Stack of parent headings by level
for raw in extractor.sections:
heading = raw["heading"]
level = raw["heading_level"]
# Extract section number (e.g., "2.5" from "2.5 Webanalyse")
num_match = re.match(r"^(\d+(?:\.\d+)*)\s*[.:]?\s*", heading)
section_number = num_match.group(1) if num_match else ""
# Track parent headings
while len(parent_stack) > level:
parent_stack.pop()
parent = parent_stack[-1] if parent_stack else ""
parent_stack.append(heading)
sections.append(DSESection(
heading=heading,
heading_level=level,
section_number=section_number,
content=raw["content"][:2000], # Cap content length
html=raw["html"][:3000],
parent_heading=parent,
url=url,
element_id=raw["element_id"],
paragraph_count=raw["paragraph_count"],
))
logger.info("Parsed DSE: %d sections from %s", len(sections), url)
return sections
def _regex_fallback(html: str, url: str) -> list[DSESection]:
"""Fallback parser using regex when HTML parsing fails."""
# Strip scripts and styles
clean = re.sub(r"<(script|style)[^>]*>.*?</\1>", "", html, flags=re.DOTALL | re.IGNORECASE)
sections = []
# Find all headings
for match in re.finditer(r"<h([1-4])[^>]*(?:id=[\"']([^\"']*)[\"'])?[^>]*>(.*?)</h\1>", clean, re.DOTALL | re.IGNORECASE):
level = int(match.group(1))
elem_id = match.group(2) or ""
heading = re.sub(r"<[^>]+>", "", match.group(3)).strip()
# Get content until next heading
start = match.end()
next_heading = re.search(r"<h[1-4]", clean[start:], re.IGNORECASE)
end = start + next_heading.start() if next_heading else start + 2000
content = clean[start:end]
content = re.sub(r"<[^>]+>", " ", content)
content = re.sub(r"\s+", " ", content).strip()
num_match = re.match(r"^(\d+(?:\.\d+)*)", heading)
sections.append(DSESection(
heading=heading,
heading_level=level,
section_number=num_match.group(1) if num_match else "",
content=content[:2000],
html="",
url=url,
element_id=elem_id,
))
return sections
def find_section_by_content(sections: list[DSESection], search_text: str) -> DSESection | None:
"""Find the section that contains specific text."""
search_lower = search_text.lower()
for section in sections:
if search_lower in section.content.lower():
return section
return None
def find_section_by_category(sections: list[DSESection], category: str) -> DSESection | None:
"""Find the section most likely to contain a service category."""
category_keywords = {
"tracking": ["cookie", "tracking", "webanalyse", "analytics", "statistik"],
"marketing": ["marketing", "werbung", "newsletter", "remarketing"],
"payment": ["zahlung", "payment", "bezahlung", "zahlungsabwicklung"],
"chatbot": ["chat", "kommunikation", "kundenservice", "kontakt"],
"cdn": ["hosting", "bereitstellung", "technisch", "infrastruktur", "cdn"],
"other": ["sonstig", "weitere", "dritte", "extern"],
}
keywords = category_keywords.get(category, category_keywords["other"])
for section in sections:
heading_lower = section.heading.lower()
content_lower = section.content.lower()[:500]
for kw in keywords:
if kw in heading_lower or kw in content_lower:
return section
return None