An MCP server that provides AI-powered visual analysis for UI/UX assessment using Google's Gemini Vision API.
- Screenshot capture from URLs
- AI-powered UI element detection and analysis
- WCAG contrast ratio analysis
- Color palette and typography extraction
- Visual accessibility auditing
- Comprehensive JSON reporting
git clone <repository-url>
cd mcp-ux-vision
npm install
npm run buildAdd to your MCP client configuration:
{
"servers": {
"ux-vision": {
"command": "node",
"args": [
"path/to/mcp-ux-vision/build/index.js"
]
}
}
}Capture a screenshot of any webpage.
url(required): URL to screenshotfullPage(optional): Capture full page vs viewportwaitTime(optional): Delay before capture (ms)
Analyze the most recent screenshot with AI, extracting:
- UI elements with geometry and styling
- Color palette and typography
- Accessibility metrics
- Visual hierarchy assessment
Create a comprehensive JSON report from the last analysis.
testUrl(required): URL that was analyzedappName(optional): Application nameoutput_path(optional): Report output directory
One-step workflow combines all above tools with the same parameters: screenshot → analyze → report.
Once the server is running and configured in your MCP client (like Claude), you can use natural language prompts to call the tools:
screenshot_url
- "Take a screenshot of https://example.com"
- "Capture a full page screenshot of localhost:3000"
- "Screenshot https://myapp.com and wait for the .loading element to disappear"
analyze_screen
- "Analyze the current screenshot for UI elements"
- "Examine the accessibility of this interface"
- "What UI components do you see in the screenshot?"
generate_report
- "Create a UX report for the analysis of https://example.com"
- "Generate a comprehensive UI audit report"
analyze_url_full_report
- "Take a screenshot of https://example.com and create a full UX analysis report"
- "Analyze https://myapp.com and generate a complete accessibility audit"
MIT