Curated Claude Code catalog
Updated 07.05.2026 · 19:39 CET
01 / Skill
ai-dashboad

flutter-skill

Quality
9.0

This skill provides AI agents with "eyes and hands" to perform end-to-end testing and UI automation across 10 different platforms. It excels at eliminating the need for manual test code, allowing agents to interact with applications through natural language prompts, making it ideal for comprehensive, AI-driven quality assurance.

USP

Unlike traditional testing tools, flutter-skill enables AI agents to autonomously test across 10 platforms with zero test code, using natural language and direct MCP integration. Its "zero config" setup simplifies adoption significantly.

Use cases

  • 01AI-driven end-to-end testing
  • 02Cross-platform UI automation
  • 03Debugging applications with AI agents
  • 04Automated regression testing
  • 05Exploratory testing by AI

Detected files (3)

  • skills-submission/flutter-skill/SKILL.mdskill
    Show content (3675 bytes)
    ---
    name: flutter-skill
    description: Control and automate Flutter applications - inspect UI, perform gestures, validate state, take screenshots, and debug. Connects AI agents to running Flutter apps via Dart VM Service Protocol.
    ---
    
    # Flutter Skill
    
    Give your AI Agent eyes and hands inside your Flutter app. This skill enables comprehensive control of Flutter applications for testing, debugging, and automation.
    
    ## Installation
    
    ### Option 1: npx (Recommended)
    ```json
    {
      "flutter-skill": {
        "command": "npx",
        "args": ["flutter-skill"]
      }
    }
    ```
    
    ### Option 2: Global Install
    ```bash
    dart pub global activate flutter_skill
    ```
    
    Then configure:
    ```json
    {
      "flutter-skill": {
        "command": "flutter_skill",
        "args": ["server"]
      }
    }
    ```
    
    ## Available Tools
    
    ### Connection
    - `connect_app` - Connect to a running Flutter app via WebSocket URI
    - `launch_app` - Launch a Flutter app with auto-setup (adds dependencies, patches main.dart)
    
    ### UI Inspection
    - `inspect` - Get interactive elements (buttons, text fields, etc.)
    - `get_widget_tree` - Full widget tree structure with configurable depth
    - `get_widget_properties` - Widget details (size, position, visibility)
    - `get_text_content` - Extract all visible text from screen
    - `find_by_type` - Find all widgets of a specific type
    
    ### Interactions
    - `tap` - Tap element by key or text
    - `double_tap` - Double tap gesture
    - `long_press` - Long press gesture
    - `swipe` - Swipe up/down/left/right
    - `drag` - Drag from one element to another
    - `scroll_to` - Scroll element into view
    - `enter_text` - Input text into text field
    
    ### State Validation
    - `get_text_value` - Get text field value
    - `get_checkbox_state` - Get checkbox checked state
    - `get_slider_value` - Get slider current value
    - `wait_for_element` - Wait for element to appear (with timeout)
    - `wait_for_gone` - Wait for element to disappear
    
    ### Screenshots
    - `screenshot` - Capture full app screenshot (base64 PNG)
    - `screenshot_element` - Capture specific element screenshot
    
    ### Navigation
    - `get_current_route` - Get current route name
    - `go_back` - Navigate back
    - `get_navigation_stack` - Get navigation history
    
    ### Debug & Logs
    - `get_logs` - Application logs
    - `get_errors` - Error messages
    - `get_performance` - Performance metrics
    - `clear_logs` - Clear log buffer
    - `hot_reload` - Trigger hot reload
    
    ## Usage Examples
    
    ### Test a Counter App
    ```
    1. Launch the app: launch_app with project_path="/path/to/app"
    2. Inspect UI: inspect
    3. Tap increment: tap with key="increment_button"
    4. Verify: get_text_content to see updated counter
    ```
    
    ### Test a Login Flow
    ```
    1. Enter email: enter_text with key="email_field", text="user@example.com"
    2. Enter password: enter_text with key="password_field", text="password123"
    3. Tap login: tap with key="login_button"
    4. Wait for home: wait_for_element with key="home_screen", timeout=5000
    ```
    
    ### Debug an Issue
    ```
    1. Connect: connect_app with uri="ws://127.0.0.1:xxxxx/ws"
    2. Check errors: get_errors
    3. View logs: get_logs
    4. Take screenshot: screenshot
    ```
    
    ## Best Practices
    
    ### Use Widget Keys
    For reliable element identification, target apps should use `ValueKey`:
    ```dart
    ElevatedButton(
      key: const ValueKey('submit_button'),
      onPressed: _submit,
      child: const Text('Submit'),
    )
    ```
    
    ### Element Finding Priority
    1. **Key** (most reliable): `tap with key="submit_button"`
    2. **Text content**: `tap with text="Submit"`
    3. **Widget type**: `find_by_type with type="ElevatedButton"`
    
    ## Links
    
    - [GitHub Repository](https://github.com/ai-dashboad/flutter-skill)
    - [pub.dev Package](https://pub.dev/packages/flutter_skill)
    - [npm Package](https://www.npmjs.com/package/flutter-skill)
    
  • skills/e2e-testing/SKILL.mdskill
    Show content (6391 bytes)
    ---
    name: e2e-testing
    description: AI-powered E2E testing for any app — Flutter, React Native, iOS, Android, Electron, Tauri, KMP, .NET MAUI. Test 8 platforms with natural language through MCP. No test code needed. Just describe what to test and the agent sees screenshots, taps elements, enters text, scrolls, and verifies UI state automatically.
    version: 0.9.36
    ---
    
    # AI E2E Testing — 8 Platforms, Zero Test Code
    
    > Give your AI agent eyes and hands inside any running app.
    
    flutter-skill is an MCP server that connects AI agents to running apps. The agent can see screenshots, tap elements, enter text, scroll, navigate, inspect UI trees, and verify state — all through natural language.
    
    ## Supported Platforms
    
    | Platform | Setup |
    |----------|-------|
    | Flutter (iOS/Android/Web) | `flutter pub add flutter_skill` |
    | React Native | `npm install flutter-skill-react-native` |
    | Electron | `npm install flutter-skill-electron` |
    | iOS (Swift/UIKit) | SPM: `FlutterSkillSDK` |
    | Android (Kotlin) | Gradle: `flutter-skill-android` |
    | Tauri (Rust) | `cargo add flutter-skill-tauri` |
    | KMP Desktop | Gradle dependency |
    | .NET MAUI | NuGet package |
    
    **Test scorecard: 562/567 (99.1%) across all 8 platforms.**
    
    ## Install
    
    ```bash
    # npm (recommended)
    npm install -g flutter-skill
    
    # Homebrew
    brew install ai-dashboad/flutter-skill/flutter-skill
    
    # Or download binary from GitHub Releases
    ```
    
    ## MCP Configuration
    
    Add to your AI agent's MCP config (Claude Desktop, Cursor, Windsurf, OpenClaw, etc.):
    
    ```json
    {
      "mcpServers": {
        "flutter-skill": {
          "command": "flutter-skill",
          "args": ["server"]
        }
      }
    }
    ```
    
    ### OpenClaw
    
    If using OpenClaw, add to your gateway config under `mcp.servers`:
    
    ```yaml
    mcp:
      servers:
        flutter-skill:
          command: flutter-skill
          args: ["server"]
    ```
    
    ## Quick Start
    
    ### 1. Initialize your app (one-time)
    
    ```bash
    cd /path/to/your/app
    flutter-skill init
    ```
    
    Auto-detects project type and patches your app with the testing bridge.
    
    ### 2. Launch and connect
    
    ```bash
    flutter-skill launch .
    ```
    
    ### 3. Test with natural language
    
    Tell the agent what to test:
    
    > "Test the login flow — enter admin@test.com and password123, tap Login, verify Dashboard appears"
    
    The agent will automatically:
    1. `screenshot()` → see the current screen
    2. `inspect_interactive()` → discover all tappable/typeable elements with semantic refs
    3. `tap(ref: "button:Login")` → tap using stable semantic reference
    4. `enter_text(ref: "input:Email", text: "admin@test.com")` → type into field
    5. `wait_for_element(key: "Dashboard")` → verify navigation
    6. `screenshot()` → confirm final state
    
    ## Available MCP Tools
    
    ### Core Actions
    | Tool | Description |
    |------|-------------|
    | `screenshot` | Capture current screen as image |
    | `tap` | Tap element by key, text, ref, or coordinates |
    | `enter_text` | Type text into a field |
    | `scroll` | Scroll up/down/left/right |
    | `swipe` | Swipe gesture between points |
    | `long_press` | Long press an element |
    | `drag` | Drag from point A to B |
    | `go_back` | Navigate back |
    | `press_key` | Send keyboard key events |
    
    ### Inspection (v0.8.0+)
    | Tool | Description |
    |------|-------------|
    | `inspect_interactive` | **NEW** — Get all interactive elements with semantic ref IDs |
    | `get_elements` | List all elements on screen |
    | `find_element` | Find element by key or text |
    | `wait_for_element` | Wait for element to appear (with timeout) |
    | `get_element_properties` | Get detailed properties of an element |
    
    ### Text Manipulation
    | Tool | Description |
    |------|-------------|
    | `set_text` | Replace text in a field |
    | `clear_text` | Clear a text field |
    | `get_text` | Read text content |
    
    ### App Control
    | Tool | Description |
    |------|-------------|
    | `get_logs` | Read app logs |
    | `clear_logs` | Clear log buffer |
    
    ## Semantic Refs (v0.8.0)
    
    `inspect_interactive` returns elements with stable semantic reference IDs:
    
    ```
    button:Login          → Login button
    input:Email           → Email text field
    toggle:Dark Mode      → Dark mode switch
    button:Submit[1]      → Second Submit button (disambiguated)
    ```
    
    Format: `{role}:{content}[{index}]`
    
    7 roles: `button`, `input`, `toggle`, `slider`, `select`, `link`, `item`
    
    Use refs for reliable element targeting that survives UI changes:
    ```
    tap(ref: "button:Login")
    enter_text(ref: "input:Email", text: "test@example.com")
    ```
    
    ## Testing Workflow
    
    ### Basic Flow
    ```
    screenshot() → inspect_interactive() → tap/enter_text → screenshot() → verify
    ```
    
    ### Comprehensive Testing
    > "Explore every screen of this app. Test all buttons, forms, navigation, and edge cases. Report any bugs you find."
    
    The agent will systematically:
    - Navigate every screen via tab bars, menus, links
    - Interact with every interactive element
    - Test form validation (empty, invalid, valid inputs)
    - Test edge cases (long text, special characters, emoji)
    - Verify navigation flows (forward, back, deep links)
    - Take screenshots at each step for verification
    
    ### Example Prompts
    
    **Quick smoke test:**
    > "Tap every tab and screenshot each page"
    
    **Form testing:**
    > "Fill the registration form with edge case data — emoji name, very long email, short password — and verify error messages"
    
    **Navigation:**
    > "Test the complete user journey: sign up → create post → like → comment → delete → sign out"
    
    **Accessibility:**
    > "Check every screen for missing labels, small tap targets, and contrast issues"
    
    ## Tips
    
    1. **Always start with `screenshot()`** — see before you act
    2. **Use `inspect_interactive()` to discover elements** — don't guess at selectors
    3. **Prefer `ref:` selectors** — more stable than text or coordinates
    4. **`wait_for_element()` after navigation** — apps need time to transition
    5. **Screenshot after every action** — verify the expected effect
    6. **Use `press_key` for keyboard shortcuts** — test keyboard navigation
    
    ## Links
    
    - [GitHub](https://github.com/ai-dashboad/flutter-skill)
    - [npm](https://www.npmjs.com/package/flutter-skill)
    - [Documentation](https://github.com/ai-dashboad/flutter-skill/blob/main/docs/USAGE_GUIDE.md)
    - [Demo Video](https://github.com/user-attachments/assets/d4617c73-043f-424c-9a9a-1a61d4c2d3c6)
    - [pub.dev](https://pub.dev/packages/flutter_skill)
    - [VSCode Extension](https://marketplace.visualstudio.com/items?itemName=AIDashboard.flutter-skill)
    
  • server.jsonmcp_server
    Show content (585 bytes)
    {
      "$schema": "https://static.modelcontextprotocol.io/schemas/2025-12-11/server.schema.json",
      "name": "io.github.ai-dashboad/flutter-skill",
      "description": "AI-powered E2E testing for 10 platforms. 253 MCP tools. Zero test code needed.",
      "repository": {
        "url": "https://github.com/ai-dashboad/flutter-skill",
        "source": "github"
      },
      "version": "0.9.36",
      "packages": [
        {
          "registryType": "npm",
          "identifier": "flutter-skill",
          "version": "0.9.36",
          "transport": {
            "type": "stdio"
          },
          "environmentVariables": []
        }
      ]
    }
    

README

flutter-skill CLI demo — nav, tap, type, screenshot across websites

flutter-skill

Give any AI agent eyes and hands inside any running app.
10 platforms. Zero test code. One MCP server.

Stars pub.dev npm CI License

DemoQuick StartAI PlatformsPlatformsvs OthersDocs

🚀 Zero config. Zero test code. Just talk to your AI.

If this saves you time, please consider starring the repo ⭐ — it helps others find it!


30-Second Demo

https://github.com/user-attachments/assets/d4617c73-043f-424c-9a9a-1a61d4c2d3c6

One prompt. 28 AI-driven actions. Zero test code. The AI explores a TikTok clone, navigates tabs, scrolls feeds, tests search, fills forms — all autonomously.


Why This Exists

Writing E2E tests is painful. Maintaining them is worse. flutter-skill takes a different approach:

  • 🔌 Connects any AI agent (Claude, Cursor, Windsurf, Copilot, OpenClaw) directly to your running app via MCP
  • 👀 The agent sees your screen — taps buttons, types text, scrolls, navigates — like a human tester who never sleeps
  • Zero test code — no Page Objects, no XPath, no brittle selectors. Just plain English
  • Zero config — 2 lines of code, works on all 10 platforms
You: "Test the checkout flow with an empty cart, then add 3 items and complete purchase"

Your AI agent handles the rest — screenshots, taps, text entry, assertions, navigation.
No Page Objects. No XPath. No brittle selectors. Just plain English.

Quick Start

1. Install (30 seconds)

npm install -g flutter-skill

2. Add to your AI (copy-paste into MCP config)

{
  "mcpServers": {
    "flutter-skill": {
      "command": "flutter-skill",
      "args": ["server"]
    }
  }
}

Works with Claude Desktop, Cursor, Windsurf, Copilot, Cline, OpenClaw — any MCP-compatible agent.

3. Add to your app (2 lines for Flutter)

import 'package:flutter_skill/flutter_skill.dart';

void main() {
  if (kDebugMode) FlutterSkillBinding.ensureInitialized();
  runApp(MyApp());
}

4. Test — just talk to your AI:

"Launch my app, explore every screen, and report any bugs"

That's it. Zero configuration. Zero test code. Works in under 60 seconds.

📦 More install methods (Homebrew, Scoop, Docker, IDE, Agent Skill)
MethodCommand
npmnpm install -g flutter-skill
Homebrewbrew install ai-dashboad/flutter-skill/flutter-skill
Scoopscoop install flutter-skill
Dockerdocker pull ghcr.io/ai-dashboad/flutter-skill
pub.devdart pub global activate flutter_skill
VSCodeExtensions → "Flutter Skill"
JetBrainsPlugins → "Flutter Skill"
Agent Skillnpx skills add ai-dashboad/flutter-skill
Zero-configflutter-skill init (auto-detects & patches your app)

Use with AI Platforms

MCP Server Mode (IDE Integration)

Works with any MCP-compatible AI tool. One config line:

{
  "mcpServers": {
    "flutter-skill": {
      "command": "flutter-skill",
      "args": ["server"]
    }
  }
}
PlatformConfig FileStatus
Cursor.cursor/mcp.json
Claude Desktopclaude_desktop_config.json
Windsurf~/.codeium/windsurf/mcp_config.json
VSCode Copilot.vscode/mcp.json
ClineVSCode Settings → Cline → MCP
OpenClawSkill or MCP config
Continue.dev.continue/config.json

HTTP Serve Mode (CLI & Automation)

For standalone browser automation, CI/CD pipelines, or remote access:

# Start server
flutter-skill serve https://your-app.com

# Use CLI client commands
flutter-skill nav https://google.com
flutter-skill snap                    # Accessibility tree (99% fewer tokens)
flutter-skill screenshot /tmp/ss.jpg
flutter-skill tap "Login"
flutter-skill type "hello@example.com"
flutter-skill eval "document.title"
flutter-skill tools                   # List all available tools
CommandDescription
nav <url>Navigate to URL
snapAccessibility tree snapshot
screenshot [path]Take screenshot
tap <text|ref|x y>Tap element
type <text>Type via keyboard
key <key> [mod]Press key
eval <js>Execute JavaScript
titleGet page title
textGet visible text
hover <text>Hover element
upload <sel> <file>Upload file
toolsList tools
call <tool> [json]Call any tool

Supports --port=N, --host=H flags and FS_PORT/FS_HOST env vars.

Two Modes Compared

server (MCP stdio)serve (HTTP)
Use caseIDE / AI agent integrationCLI / automation / CI/CD
ProtocolMCP (JSON-RPC over stdio)HTTP REST
Tools253 (dynamic per page)246 (generic)
BrowserAuto-launches ChromeConnects to existing Chrome
Best forCursor, Claude, VSCodeOpenClaw, scripts, pipelines

Full CLI client reference: docs/CLI_CLIENT.md


10 Platforms, One Tool

Most testing tools work on 1-2 platforms. flutter-skill works on 10.

PlatformSDKTest Score
Flutter (iOS/Android/Web)flutter_skill✅ 188/195
React Nativesdks/react-native✅ 75/75
Electronsdks/electron✅ 75/75
Tauri (Rust)sdks/tauri✅ 75/75
Android (Kotlin)sdks/android✅ 74/75
KMP Desktopsdks/kmp✅ 75/75
.NET MAUIsdks/dotnet-maui✅ 75/75
iOS (Swift/UIKit)sdks/ios✅ 19/19
Web (any website)sdks/web
Web CDP (zero-config)No SDK needed✅ 141/156

Total: 656/664 tests passing (98.8%) — each platform tested against a complex social media app with 50+ elements.


⚡ Performance

Real benchmarks from automated test runs against a complex social media app:

OperationWeb (CDP)ElectronAndroid
connect93 ms55 ms103 ms
tap1 ms1 ms2 ms
enter_text1 ms1 ms2 ms
inspect3 ms12 ms10 ms
snapshot2 ms8 ms29 ms
screenshot31 ms80 ms88 ms
eval1 ms

Token efficiency: snapshot() returns a structured element tree instead of an image — 87–99% fewer tokens than sending screenshots to your AI agent.

How fast is that? A tap takes 1–2 ms end-to-end. Browser automation tools like Playwright and Selenium typically take 50–100 ms for the same operation. That's 50–100× faster, because flutter-skill talks directly to the app runtime instead of going through WebDriver or CDP indirection.

Heavy DOM Sites (Real-World)

Tested 15 MCP tools against production websites — 75/75 passed, zero timeouts:

SiteToolsTotal Timesnapshotscreenshotcount_elements
YouTube15/15 ✅6.9s43 ms30 ms4 ms
Amazon15/15 ✅14.2s1 ms5 ms2 ms
Reddit15/15 ✅17.9s6 ms32 ms51 ms
Hacker News15/15 ✅4.8s53 ms188 ms1 ms
Wikipedia15/15 ✅7.8s15 ms336 ms1 ms

Total time includes page load. Tool execution is consistently sub-100ms even on heavy DOM sites.


Why Not Playwright / Appium / Detox?

flutter-skillPlaywright MCPAppiumDetox
MCP tools253~33
Platforms101 (web)MobileReact Native
Setup time30 secMinutesHoursHours
Test code needed❌ None✅ Yes✅ Yes✅ Yes
AI-native (MCP)
Self-healing tests
Monkey/fuzz testing
Visual regression
Network mock/replay
API + UI testing
Multi-device syncPartial
Accessibility audit
i18n testing
Performance monitoring
Natural language
Flutter support✅ NativePartialPartial
Desktop apps

| AI page understanding | ✅ AX Tree | ❌ Screenshots | ❌ | ❌ | | Boundary/security test | ✅ 13 payloads | ❌ | ❌ | ❌ | | Batch actions | ✅ 5+/call | 1/call | 1/call | 1/call |

flutter-skill is the only AI-native E2E testing tool that works across mobile, web, and desktop — with 7× more tools than the nearest competitor.


CLI Commands

# 🤖 AI autonomous exploration — finds bugs automatically
flutter-skill explore https://my-app.com --depth=3

# 🐒 Monkey/fuzz testing — random actions, crash detection
flutter-skill monkey https://my-app.com --actions=100 --seed=42

# 🚀 Parallel multi-platform testing
flutter-skill test --url https://my-app.com --platforms web,electron,android

# 🌐 Zero-config WebMCP server — any website becomes testable
flutter-skill serve https://my-app.com

🧠 AI-Native: 95% Fewer Tokens

Most AI testing tools send screenshots to the LLM — each one costs ~4,000 tokens.

flutter-skill uses Chrome's Accessibility Tree to give your AI a compact semantic summary of any page:

// page_summary → ~200 tokens (vs ~4,000 for a screenshot)
{
  "title": "Shopping Cart",
  "nav": ["Home", "Products", "Cart", "Account"],
  "forms": [{"input:Coupon Code": "text"}],
  "buttons": ["Apply", "Checkout", "Continue Shopping"],
  "features": {"search": true, "pagination": true},
  "links": 47, "inputs": 3
}

Then batch multiple actions in one call:

// explore_actions → 5 actions per call (vs 5 separate tool calls)
{"actions": [
  {"type": "fill", "target": "input:Coupon Code", "value": "SAVE20"},
  {"type": "tap", "target": "button:Apply"},
  {"type": "tap", "target": "button:Checkout"},
  {"type": "fill", "target": "input:Email", "value": "test@example.com"},
  {"type": "tap", "target": "button:Continue"}
]}

Result: Your AI agent tests faster, costs less, and understands pages better than screenshot-based tools.

flutter-skillScreenshot-based tools
Tokens per page~200~4,000
Actions per call5+1
Understands semantics✅ roles, names, state❌ pixels only
Works with Shadow DOM

What It Can Do

👀 See

  • screenshot — capture the screen
  • inspect_interactive — all tappable/typeable elements with semantic refs
  • find_element / wait_for_element
  • get_elements — full element tree

👆 Interact

  • tap / long_press / swipe / drag
  • enter_text / set_text / clear_text
  • scroll — all directions
  • go_back / press_key

🔍 Inspect (v0.8.0)

  • Semantic refs: button:Login, input:Email
  • Stable across UI changes
  • tap(ref: "button:Submit")
  • 7 roles: button, input, toggle, slider, select, link, item

🚀 Control

  • launch_app — launch with flavors
  • hot_reload / hot_restart
  • get_logs / get_errors
  • scan_and_connect — auto-find apps
253 tools — full reference

AI Explore: page_summary, explore_actions, boundary_test, explore_report

Launch & Connect: launch_app, scan_and_connect, connect_cdp, hot_reload, hot_restart, list_sessions, switch_session, close_session, disconnect, stop_app

Screen: screenshot, screenshot_region, screenshot_element, native_screenshot, inspect, inspect_interactive, snapshot, get_widget_tree, find_by_type, get_text_content, get_visible_text

Interaction: tap, double_tap, long_press, enter_text, set_text, clear_text, swipe, scroll_to, drag, go_back, press_key, type_text, hover, fill, select_option, set_checkbox, focus, blur, native_tap, native_input_text, native_swipe

Smart Testing: smart_tap, smart_enter_text, smart_assert (self-healing with fuzzy match)

Assertions: assert_text, assert_visible, assert_not_visible, assert_element_count, assert_batch, wait_for_element, wait_for_gone, wait_for_idle, wait_for_stable, wait_for_url, wait_for_text, wait_for_element_count

Visual Regression: visual_baseline_save, visual_baseline_compare, visual_baseline_update, visual_regression_report, visual_verify, visual_diff, compare_screenshot

Network Mock: mock_api, mock_clear, record_network, replay_network, intercept_requests, clear_interceptions, block_urls, http_request

API Testing: api_request, api_assert

Coverage & Reliability: coverage_start, coverage_stop, coverage_report, coverage_gaps, retry_on_fail, stability_check

Data-Driven: test_with_data, generate_test_data

Multi-Device: multi_connect, multi_action, multi_compare, multi_disconnect, parallel_snapshot, parallel_tap

Accessibility: accessibility_audit, a11y_full_audit, a11y_tab_order, a11y_color_contrast, a11y_screen_reader

i18n: set_locale, verify_translations, i18n_snapshot

Performance: perf_start, perf_stop, perf_report, get_performance, get_frame_stats, get_memory_stats

Session: save_session, restore_session, session_diff

Recording & Export: record_start, record_stop, record_export (Playwright, Cypress, XCUITest, Espresso, Detox, Maestro, +5 more), video_start, video_stop

Auth: auth_inject_session, auth_biometric, auth_otp, auth_deeplink

CDP Browser: navigate, reload, go_forward, get_title, get_page_source, eval, get_tabs, new_tab, switch_tab, close_tab, get_cookies, set_cookie, clear_cookies, get_local_storage, set_local_storage, clear_local_storage, generate_pdf, set_viewport, emulate_device, throttle_network, go_offline, set_geolocation, set_timezone, set_color_scheme

Debug: get_logs, get_errors, get_console_messages, get_network_requests, diagnose, diagnose_project, reset_app


Platform Setup

Flutter (iOS / Android / Web)
dependencies:
  flutter_skill: ^0.9.36
import 'package:flutter_skill/flutter_skill.dart';

void main() {
  if (kDebugMode) FlutterSkillBinding.ensureInitialized();
  runApp(MyApp());
}
React Native
npm install flutter-skill-react-native
import FlutterSkill from 'flutter-skill-react-native';
FlutterSkill.start();
Electron
npm install flutter-skill-electron
const { FlutterSkillBridge } = require('flutter-skill-electron');
FlutterSkillBridge.start(mainWindow);
iOS (Swift)
// Swift Package Manager: FlutterSkillSDK
import FlutterSkill
FlutterSkillBridge.shared.start()

Text("Hello").flutterSkillId("greeting")
Android (Kotlin)
implementation("com.flutterskill:flutter-skill:0.8.0")

FlutterSkillBridge.start(this)
Tauri (Rust)
[dependencies]
flutter-skill-tauri = "0.8.0"
KMP Desktop

Add Gradle dependency — see sdks/kmp for details.

.NET MAUI

Add NuGet package — see sdks/dotnet-maui for details.


Example Prompts

Just tell your AI what to test:

PromptWhat happens
"Test login with wrong password"Screenshots → enters creds → taps login → verifies error
"Explore every screen and report bugs"Systematically navigates all screens, tests all elements
"Fill registration with edge cases"Tests emoji 🌍, long strings, empty fields, special chars
"Compare checkout flow on iOS and Android"Runs same test on both platforms, compares screenshots
"Take screenshots of all 5 tabs"Taps each tab, captures state

Contributing

See CONTRIBUTING.md for guidelines.

git clone https://github.com/ai-dashboad/flutter-skill
cd flutter-skill
dart pub get
dart run bin/flutter_skill.dart server  # Start MCP server

Links

📦 pub.dev🧩 VSCode
📦 npm🧩 JetBrains
🍺 Homebrew📖 Docs
🤖 Agent Skill📋 Changelog

⭐ If flutter-skill saves you time, star it so others can find it too!

MIT License © 2025