forked from HomeLab/unraid-mcp
Addresses all critical, high, medium, and low issues from full codebase review. 494 tests pass, ruff clean, ty type-check clean. Security: - Add tool_error_handler context manager (exceptions.py) — standardised error handling, eliminates 11 bare except-reraise patterns - Remove unused exception subclasses (ConfigurationError, UnraidAPIError, SubscriptionError, ValidationError, IdempotentOperationError) - Harden GraphQL subscription query validator with allow-list and forbidden-keyword regex (diagnostics.py) - Add input validation for rclone create_remote config_data: injection, path-traversal, and key-count limits (rclone.py) - Validate notifications importance enum before GraphQL request (notifications.py) - Sanitise HTTP/network/JSON error messages — no raw exception strings leaked to clients (client.py) - Strip path/creds from displayed API URL via _safe_display_url (health.py) - Enable Ruff S (bandit) rule category in pyproject.toml - Harden container mutations to strict-only matching — no fuzzy/substring for destructive operations (docker.py) Performance: - Token-bucket rate limiter (90 tokens, 9 req/s) with 429 retry backoff (client.py) - Lazy asyncio.Lock init via _get_client_lock() — fixes event-loop module-load crash (client.py) - Double-checked locking in get_http_client() for fast-path (client.py) - Short hex container ID fast-path skips list fetch (docker.py) - Cap resource_data log content to 1 MB / 5,000 lines (manager.py) - Reset reconnect counter after 30 s stable connection (manager.py) - Move tail_lines validation to module level; enforce 10,000 line cap (storage.py, docker.py) - force_terminal=True removed from logging RichHandler (logging.py) Architecture: - Register diagnostic tools in server startup (server.py) - Move ALL_ACTIONS computation to module level in all tools - Consolidate format_kb / format_bytes into shared core/utils.py - Add _safe_get() helper in core/utils.py for nested dict traversal - Extract _analyze_subscription_status() from health.py diagnose handler - Validate required config at startup — fail fast with CRITICAL log (server.py) Code quality: - Remove ~90 lines of dead Rich formatting helpers from logging.py - Remove dead self.websocket attribute from SubscriptionManager - Remove dead setup_uvicorn_logging() wrapper - Move _VALID_IMPORTANCE to module level (N806 fix) - Add slots=True to all three dataclasses (SubscriptionData, SystemHealth, APIResponse) - Fix None rendering as literal "None" string in info.py summaries - Change fuzzy-match log messages from INFO to DEBUG (docker.py) - UTC-aware datetimes throughout (manager.py, diagnostics.py) Infrastructure: - Upgrade base image python:3.11-slim → python:3.12-slim (Dockerfile) - Add non-root appuser (UID/GID 1000) with HEALTHCHECK (Dockerfile) - Add read_only, cap_drop: ALL, tmpfs /tmp to docker-compose.yml - Single-source version via importlib.metadata (pyproject.toml → __init__.py) - Add open_timeout to all websockets.connect() calls Tests: - Update error message matchers to match sanitised messages (test_client.py) - Fix patch targets for UNRAID_API_URL → utils module (test_subscriptions.py) - Fix importance="info" → importance="normal" (test_notifications.py, http_layer) - Fix naive datetime fixtures → UTC-aware (test_subscriptions.py) Co-authored-by: Claude <claude@anthropic.com>
205 lines
7.9 KiB
Python
205 lines
7.9 KiB
Python
"""Tests for unraid_storage tool."""
|
|
|
|
from collections.abc import Generator
|
|
from unittest.mock import AsyncMock, patch
|
|
|
|
import pytest
|
|
from conftest import make_tool_fn
|
|
|
|
from unraid_mcp.core.exceptions import ToolError
|
|
from unraid_mcp.core.utils import format_bytes
|
|
|
|
|
|
# --- Unit tests for helpers ---
|
|
|
|
|
|
class TestFormatBytes:
|
|
def test_none(self) -> None:
|
|
assert format_bytes(None) == "N/A"
|
|
|
|
def test_bytes(self) -> None:
|
|
assert format_bytes(512) == "512.00 B"
|
|
|
|
def test_kilobytes(self) -> None:
|
|
assert format_bytes(2048) == "2.00 KB"
|
|
|
|
def test_megabytes(self) -> None:
|
|
assert format_bytes(1048576) == "1.00 MB"
|
|
|
|
def test_gigabytes(self) -> None:
|
|
assert format_bytes(1073741824) == "1.00 GB"
|
|
|
|
def test_terabytes(self) -> None:
|
|
assert format_bytes(1099511627776) == "1.00 TB"
|
|
|
|
|
|
# --- Integration tests ---
|
|
|
|
|
|
@pytest.fixture
|
|
def _mock_graphql() -> Generator[AsyncMock, None, None]:
|
|
with patch("unraid_mcp.tools.storage.make_graphql_request", new_callable=AsyncMock) as mock:
|
|
yield mock
|
|
|
|
|
|
def _make_tool():
|
|
return make_tool_fn("unraid_mcp.tools.storage", "register_storage_tool", "unraid_storage")
|
|
|
|
|
|
class TestStorageValidation:
|
|
async def test_disk_details_requires_disk_id(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="disk_id"):
|
|
await tool_fn(action="disk_details")
|
|
|
|
async def test_logs_requires_log_path(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="log_path"):
|
|
await tool_fn(action="logs")
|
|
|
|
async def test_logs_rejects_invalid_path(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="log_path must start with"):
|
|
await tool_fn(action="logs", log_path="/etc/shadow")
|
|
|
|
async def test_logs_rejects_path_traversal(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
# Traversal that escapes /var/log/ to reach /etc/shadow
|
|
with pytest.raises(ToolError, match="log_path must start with"):
|
|
await tool_fn(action="logs", log_path="/var/log/../../etc/shadow")
|
|
# Traversal that escapes /mnt/ to reach /etc/passwd
|
|
with pytest.raises(ToolError, match="log_path must start with"):
|
|
await tool_fn(action="logs", log_path="/mnt/../etc/passwd")
|
|
|
|
async def test_logs_allows_valid_paths(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"logFile": {"path": "/var/log/syslog", "content": "ok"}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="logs", log_path="/var/log/syslog")
|
|
assert result["content"] == "ok"
|
|
|
|
|
|
class TestStorageActions:
|
|
async def test_shares(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"shares": [{"id": "s:1", "name": "media"}, {"id": "s:2", "name": "backups"}]
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="shares")
|
|
assert len(result["shares"]) == 2
|
|
|
|
async def test_disks(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"disks": [{"id": "d:1", "device": "sda"}]}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="disks")
|
|
assert len(result["disks"]) == 1
|
|
|
|
async def test_disk_details(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"disk": {
|
|
"id": "d:1",
|
|
"device": "sda",
|
|
"name": "WD",
|
|
"serialNum": "SN1",
|
|
"size": 1073741824,
|
|
"temperature": 35,
|
|
}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="disk_details", disk_id="d:1")
|
|
assert result["summary"]["temperature"] == "35\u00b0C"
|
|
assert "1.00 GB" in result["summary"]["size_formatted"]
|
|
|
|
async def test_disk_details_temperature_zero(self, _mock_graphql: AsyncMock) -> None:
|
|
"""Temperature of 0 should display as '0\u00b0C', not 'N/A'."""
|
|
_mock_graphql.return_value = {
|
|
"disk": {
|
|
"id": "d:1",
|
|
"device": "sda",
|
|
"name": "WD",
|
|
"serialNum": "SN1",
|
|
"size": 1073741824,
|
|
"temperature": 0,
|
|
}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="disk_details", disk_id="d:1")
|
|
assert result["summary"]["temperature"] == "0\u00b0C"
|
|
|
|
async def test_disk_details_temperature_null(self, _mock_graphql: AsyncMock) -> None:
|
|
"""Null temperature should display as 'N/A'."""
|
|
_mock_graphql.return_value = {
|
|
"disk": {
|
|
"id": "d:1",
|
|
"device": "sda",
|
|
"name": "WD",
|
|
"serialNum": "SN1",
|
|
"size": 1073741824,
|
|
"temperature": None,
|
|
}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="disk_details", disk_id="d:1")
|
|
assert result["summary"]["temperature"] == "N/A"
|
|
|
|
async def test_logs_null_log_file(self, _mock_graphql: AsyncMock) -> None:
|
|
"""logFile being null should return an empty dict."""
|
|
_mock_graphql.return_value = {"logFile": None}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="logs", log_path="/var/log/syslog")
|
|
assert result == {}
|
|
|
|
async def test_disk_details_not_found(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"disk": None}
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="not found"):
|
|
await tool_fn(action="disk_details", disk_id="d:missing")
|
|
|
|
async def test_unassigned(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"unassignedDevices": []}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="unassigned")
|
|
assert result["devices"] == []
|
|
|
|
async def test_log_files(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"logFiles": [{"name": "syslog", "path": "/var/log/syslog"}]}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="log_files")
|
|
assert len(result["log_files"]) == 1
|
|
|
|
async def test_logs(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"logFile": {"path": "/var/log/syslog", "content": "log line", "totalLines": 1}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="logs", log_path="/var/log/syslog")
|
|
assert result["content"] == "log line"
|
|
|
|
|
|
class TestStorageNetworkErrors:
|
|
"""Tests for network-level failures in storage operations."""
|
|
|
|
async def test_logs_json_decode_error(self, _mock_graphql: AsyncMock) -> None:
|
|
"""Invalid JSON response when reading logs should propagate as ToolError."""
|
|
_mock_graphql.side_effect = ToolError(
|
|
"Invalid JSON response from Unraid API: Expecting value: line 1 column 1"
|
|
)
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="Invalid JSON"):
|
|
await tool_fn(action="logs", log_path="/var/log/syslog")
|
|
|
|
async def test_shares_connection_refused(self, _mock_graphql: AsyncMock) -> None:
|
|
"""Connection refused when listing shares should propagate as ToolError."""
|
|
_mock_graphql.side_effect = ToolError(
|
|
"Network connection error: [Errno 111] Connection refused"
|
|
)
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="Connection refused"):
|
|
await tool_fn(action="shares")
|
|
|
|
async def test_disks_http_500(self, _mock_graphql: AsyncMock) -> None:
|
|
"""HTTP 500 when listing disks should propagate as ToolError."""
|
|
_mock_graphql.side_effect = ToolError("HTTP error 500: Internal Server Error")
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="HTTP error 500"):
|
|
await tool_fn(action="disks")
|