forked from HomeLab/unraid-mcp
test: close critical coverage gaps and harden PR review fixes
Critical bug fixes from PR review agents: - client.py: eager asyncio.Lock init, Final[frozenset] for _SENSITIVE_KEYS, explicit 429 ToolError after retries exhausted, removed lazy _get_client_lock() and _RateLimiter._get_lock() patterns - exceptions.py: use builtin TimeoutError (UP041), explicit handler before broad except so asyncio timeouts get descriptive messages - docker.py: add update_all to DESTRUCTIVE_ACTIONS (was missing), remove dead _MUTATION_ACTIONS constant - manager.py: _cap_log_content returns new dict (immutable), lock write to resource_data, clean dead task from active_subscriptions after loop exits - diagnostics.py: fix inaccurate comment about semicolon injection guard - health.py: narrow except ValueError in _safe_display_url, fix TODO comment New test coverage (98 tests added, 529 → 598 passing): - test_subscription_validation.py: 27 tests for _validate_subscription_query (security-critical allow-list, forbidden keyword guards, word-boundary test) - test_subscription_manager.py: 12 tests for _cap_log_content (immutability, truncation, nesting, passthrough) - test_client.py: +57 tests — _RateLimiter (token math, refill, sleep-on-empty), _QueryCache (TTL, invalidation, is_cacheable), 429 retry loop (1/2/3 failures) - test_health.py: +10 tests for _safe_display_url (credential strip, port, path/query removal, malformed IPv6 → <unparseable>) - test_notifications.py: +7 importance enum and field length validation tests - test_rclone.py: +7 _validate_config_data security guard tests - test_storage.py: +15 (tail_lines bounds, format_kb, safe_get) - test_docker.py: update_all now requires confirm=True + new guard test - test_destructive_guards.py: update audit to include update_all Co-authored-by: Claude <noreply@anthropic.com>
This commit is contained in:
131
tests/test_subscription_manager.py
Normal file
131
tests/test_subscription_manager.py
Normal file
@@ -0,0 +1,131 @@
|
||||
"""Tests for _cap_log_content in subscriptions/manager.py.
|
||||
|
||||
_cap_log_content is a pure utility that prevents unbounded memory growth from
|
||||
log subscription data. It must: return a NEW dict (not mutate), recursively
|
||||
cap nested 'content' fields, and only truncate when both byte limit and line
|
||||
limit are exceeded.
|
||||
"""
|
||||
|
||||
from unittest.mock import patch
|
||||
|
||||
from unraid_mcp.subscriptions.manager import _cap_log_content
|
||||
|
||||
|
||||
class TestCapLogContentImmutability:
|
||||
"""The function must return a new dict — never mutate the input."""
|
||||
|
||||
def test_returns_new_dict(self) -> None:
|
||||
data = {"key": "value"}
|
||||
result = _cap_log_content(data)
|
||||
assert result is not data
|
||||
|
||||
def test_input_not_mutated_on_passthrough(self) -> None:
|
||||
data = {"content": "short text", "other": "value"}
|
||||
original_content = data["content"]
|
||||
_cap_log_content(data)
|
||||
assert data["content"] == original_content
|
||||
|
||||
def test_input_not_mutated_on_truncation(self) -> None:
|
||||
# Use small limits so the truncation path is exercised
|
||||
large_content = "\n".join(f"line {i}" for i in range(200))
|
||||
data = {"content": large_content}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
_cap_log_content(data)
|
||||
# Original data must be unchanged
|
||||
assert data["content"] == large_content
|
||||
|
||||
|
||||
class TestCapLogContentSmallData:
|
||||
"""Content below the byte limit must be returned unchanged."""
|
||||
|
||||
def test_small_content_unchanged(self) -> None:
|
||||
data = {"content": "just a few lines\nof log data\n"}
|
||||
result = _cap_log_content(data)
|
||||
assert result["content"] == data["content"]
|
||||
|
||||
def test_non_content_keys_passed_through(self) -> None:
|
||||
data = {"name": "cpu_subscription", "timestamp": "2026-02-18T00:00:00Z"}
|
||||
result = _cap_log_content(data)
|
||||
assert result == data
|
||||
|
||||
def test_integer_value_passed_through(self) -> None:
|
||||
data = {"count": 42, "active": True}
|
||||
result = _cap_log_content(data)
|
||||
assert result == data
|
||||
|
||||
|
||||
class TestCapLogContentTruncation:
|
||||
"""Content exceeding both byte AND line limits must be truncated to the last N lines."""
|
||||
|
||||
def test_oversized_content_truncated_to_last_n_lines(self) -> None:
|
||||
# 200 lines, limit 50 lines, byte limit effectively 0 → should keep last 50 lines
|
||||
lines = [f"line {i}" for i in range(200)]
|
||||
data = {"content": "\n".join(lines)}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
result = _cap_log_content(data)
|
||||
result_lines = result["content"].splitlines()
|
||||
assert len(result_lines) == 50
|
||||
# Must be the LAST 50 lines
|
||||
assert result_lines[0] == "line 150"
|
||||
assert result_lines[-1] == "line 199"
|
||||
|
||||
def test_content_with_fewer_lines_than_limit_not_truncated(self) -> None:
|
||||
"""If byte limit exceeded but line count ≤ limit → keep original (not truncated)."""
|
||||
# 30 lines but byte limit 10 and line limit 50 → 30 < 50 so no truncation
|
||||
lines = [f"line {i}" for i in range(30)]
|
||||
data = {"content": "\n".join(lines)}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
result = _cap_log_content(data)
|
||||
# Original content preserved
|
||||
assert result["content"] == data["content"]
|
||||
|
||||
def test_non_content_keys_preserved_alongside_truncated_content(self) -> None:
|
||||
lines = [f"line {i}" for i in range(200)]
|
||||
data = {"content": "\n".join(lines), "path": "/var/log/syslog", "total_lines": 200}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
result = _cap_log_content(data)
|
||||
assert result["path"] == "/var/log/syslog"
|
||||
assert result["total_lines"] == 200
|
||||
assert len(result["content"].splitlines()) == 50
|
||||
|
||||
|
||||
class TestCapLogContentNested:
|
||||
"""Nested 'content' fields inside sub-dicts must also be capped recursively."""
|
||||
|
||||
def test_nested_content_field_capped(self) -> None:
|
||||
lines = [f"line {i}" for i in range(200)]
|
||||
data = {"logFile": {"content": "\n".join(lines), "path": "/var/log/syslog"}}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
result = _cap_log_content(data)
|
||||
assert len(result["logFile"]["content"].splitlines()) == 50
|
||||
assert result["logFile"]["path"] == "/var/log/syslog"
|
||||
|
||||
def test_deeply_nested_content_capped(self) -> None:
|
||||
lines = [f"line {i}" for i in range(200)]
|
||||
data = {"outer": {"inner": {"content": "\n".join(lines)}}}
|
||||
with (
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_BYTES", 10),
|
||||
patch("unraid_mcp.subscriptions.manager._MAX_RESOURCE_DATA_LINES", 50),
|
||||
):
|
||||
result = _cap_log_content(data)
|
||||
assert len(result["outer"]["inner"]["content"].splitlines()) == 50
|
||||
|
||||
def test_nested_non_content_keys_unaffected(self) -> None:
|
||||
data = {"metrics": {"cpu": 42.5, "memory": 8192}}
|
||||
result = _cap_log_content(data)
|
||||
assert result == data
|
||||
Reference in New Issue
Block a user