Pepper gives AI agents eyes and hands inside iOS Simulator apps.
It injects a shared library into any running simulator app โ no source changes, no SDK, no build step. Your agent sees the screen as structured data, taps buttons, inspects live objects, intercepts network calls, reads the heap, and debugs layout issues. Dylib injection requires the simulator; device support uses a different mechanism.
PepperIceCubesDemo_2x.mov
Claude navigating and inspecting Ice Cubes (Mastodon client) with zero source access.
Works with Claude Code ยท Cursor ยท Claude Desktop ยท any MCP client
Requires macOS 14+, Python 3.10+, and an iOS Simulator runtime.
pip install pepper-ios
pepper-ctl deploy # inject into the frontmost simulator app
pepper-ctl look # see what's on screenThen ask your agent:
"Tap through the onboarding flow and make sure every screen looks right"
That's it. The agent uses Pepper's MCP tools to see, tap, and inspect โ no extra config needed.
Other MCP clients (Cursor, Claude Desktop, etc.)
{
"mcpServers": {
"pepper": {
"command": "pepper-mcp"
}
}
}$ pepper-ctl look
Screen: navigation_stack (9 interactive, 4 text)
seg "Trending" โ tap text:"Trending"
btn "Sheryl Weikal, Right wing tech bros: we love St..."
โ tap text:"Sheryl Weikal, Right wing tech bros: we ..."
seg "Settings" โ tap text:"Settings"
$ pepper-ctl tap --text "Settings"
Action: Tapped Settings
Screen: navigation_stack Title: "Settings" (15 interactive, 3 text)
btn "App Icon" โ tap text:"App Icon"
btn "Display Settings" โ tap text:"Display Settings"
...
Every element comes with its tap command. The agent sees the screen, acts, and gets the new state back in one round trip.
These work against any app โ no setup beyond deploy.
"Why is this list empty? Check the network request and see what the API returned"
"What objects are holding a reference to DeviceActuationService? I think it's leaking"
"Switch to dark mode and look for any text that's invisible against the background"
"The login button isn't responding โ check the responder chain and see what's intercepting taps"
"Profile the scroll performance on the feed. Is anything hitching?"
"Read what's in the keychain after logout โ nothing should be left"
"Turn off the feature flag for new-nav and make sure the old flow still works"
View hierarchy, network interception, heap inspection, console capture, crash logs, performance profiling, accessibility audits, constraint debugging, keychain/UserDefaults/CoreData access, locale and orientation control, feature flags, push notifications โ 60+ tools in total. Parameter docs are built into every tool; your MCP client surfaces them automatically.
Pepper uses macOS's dynamic linker (DYLD_INSERT_LIBRARIES) to load a dylib into the simulator process at launch. The dylib starts a WebSocket server on a local port. pepper-mcp connects to that WebSocket and translates MCP tool calls into commands executed inside the app.
Because it runs in-process, Pepper has access to the full view hierarchy, the ObjC runtime, live object graphs, network delegates, and the HID event system. All touch input goes through native input events (IOHIDEvent) โ the same path real fingers take. No private APIs, no entitlements โ just dyld.
dylib/DYLIB.md โ architecture and adding commands ยท tools/TOOLS.md โ MCP tool layer ยท docs/TROUBLESHOOTING.md โ common issues
