mirror of
https://github.com/jmagar/unraid-mcp.git
synced 2026-03-02 00:04:45 -08:00
**Project Configuration:** - Enhance pyproject.toml with comprehensive metadata, keywords, and classifiers - Add LICENSE file (MIT) for proper open-source distribution - Add PUBLISHING.md with comprehensive publishing guidelines - Update .gitignore to exclude tool artifacts (.cache, .pytest_cache, .ruff_cache, .ty_cache) - Ignore documentation working directories (.docs, .full-review, docs/plans, docs/sessions) **Documentation:** - Add extensive Unraid API research documentation - API source code analysis and resolver mapping - Competitive analysis and feature gap assessment - Release notes analysis (7.0.0, 7.1.0, 7.2.0) - Connect platform overview and remote access documentation - Document known API patterns, limitations, and edge cases **Testing & Code Quality:** - Expand test coverage across all tool modules - Add destructive action confirmation tests - Improve test assertions and error case validation - Refine type annotations for better static analysis **Tool Improvements:** - Enhance error handling consistency across all tools - Improve type safety with explicit type annotations - Refine GraphQL query construction patterns - Better handling of optional parameters and edge cases This commit prepares the project for v0.2.0 release with improved metadata, comprehensive documentation, and enhanced code quality. Co-authored-by: Claude <noreply@anthropic.com>
131 lines
5.3 KiB
Python
131 lines
5.3 KiB
Python
"""Tests for unraid_vm tool."""
|
|
|
|
from unittest.mock import AsyncMock, patch
|
|
|
|
import pytest
|
|
from conftest import make_tool_fn
|
|
|
|
from unraid_mcp.core.exceptions import ToolError
|
|
|
|
|
|
@pytest.fixture
|
|
def _mock_graphql() -> AsyncMock:
|
|
with patch("unraid_mcp.tools.virtualization.make_graphql_request", new_callable=AsyncMock) as mock:
|
|
yield mock
|
|
|
|
|
|
def _make_tool():
|
|
return make_tool_fn("unraid_mcp.tools.virtualization", "register_vm_tool", "unraid_vm")
|
|
|
|
|
|
class TestVmValidation:
|
|
async def test_actions_except_list_require_vm_id(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
for action in ("details", "start", "stop", "pause", "resume", "reboot"):
|
|
with pytest.raises(ToolError, match="vm_id"):
|
|
await tool_fn(action=action)
|
|
|
|
async def test_destructive_actions_require_confirm(self, _mock_graphql: AsyncMock) -> None:
|
|
tool_fn = _make_tool()
|
|
for action in ("force_stop", "reset"):
|
|
with pytest.raises(ToolError, match="destructive"):
|
|
await tool_fn(action=action, vm_id="uuid-1")
|
|
|
|
async def test_destructive_vm_id_check_before_confirm(self, _mock_graphql: AsyncMock) -> None:
|
|
"""Destructive actions without vm_id should fail on vm_id first (validated before confirm)."""
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="vm_id"):
|
|
await tool_fn(action="force_stop")
|
|
|
|
|
|
class TestVmActions:
|
|
async def test_list(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"vms": {
|
|
"domains": [
|
|
{"id": "vm:1", "name": "Windows 11", "state": "RUNNING", "uuid": "uuid-1"},
|
|
]
|
|
}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="list")
|
|
assert len(result["vms"]) == 1
|
|
assert result["vms"][0]["name"] == "Windows 11"
|
|
|
|
async def test_list_empty(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vms": {"domains": []}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="list")
|
|
assert result["vms"] == []
|
|
|
|
async def test_list_no_vms_key(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="list")
|
|
assert result["vms"] == []
|
|
|
|
async def test_details_by_uuid(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"vms": {"domains": [{"id": "vm:1", "name": "Win11", "state": "RUNNING", "uuid": "uuid-1"}]}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="details", vm_id="uuid-1")
|
|
assert result["name"] == "Win11"
|
|
|
|
async def test_details_by_name(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"vms": {"domains": [{"id": "vm:1", "name": "Win11", "state": "RUNNING", "uuid": "uuid-1"}]}
|
|
}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="details", vm_id="Win11")
|
|
assert result["uuid"] == "uuid-1"
|
|
|
|
async def test_details_not_found(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {
|
|
"vms": {"domains": [{"id": "vm:1", "name": "Win11", "state": "RUNNING", "uuid": "uuid-1"}]}
|
|
}
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="not found"):
|
|
await tool_fn(action="details", vm_id="nonexistent")
|
|
|
|
async def test_start_vm(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {"start": True}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="start", vm_id="uuid-1")
|
|
assert result["success"] is True
|
|
assert result["action"] == "start"
|
|
|
|
async def test_force_stop(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {"forceStop": True}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="force_stop", vm_id="uuid-1", confirm=True)
|
|
assert result["success"] is True
|
|
assert result["action"] == "force_stop"
|
|
|
|
async def test_stop_vm(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {"stop": True}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="stop", vm_id="uuid-1")
|
|
assert result["success"] is True
|
|
assert result["action"] == "stop"
|
|
|
|
async def test_pause_vm(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {"pause": True}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="pause", vm_id="uuid-1")
|
|
assert result["success"] is True
|
|
assert result["action"] == "pause"
|
|
|
|
async def test_resume_vm(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {"resume": True}}
|
|
tool_fn = _make_tool()
|
|
result = await tool_fn(action="resume", vm_id="uuid-1")
|
|
assert result["success"] is True
|
|
assert result["action"] == "resume"
|
|
|
|
async def test_mutation_unexpected_response(self, _mock_graphql: AsyncMock) -> None:
|
|
_mock_graphql.return_value = {"vm": {}}
|
|
tool_fn = _make_tool()
|
|
with pytest.raises(ToolError, match="Failed to start"):
|
|
await tool_fn(action="start", vm_id="uuid-1")
|