When to use
Use LLM mode when you need:- Zero setup: No selector configuration or step definition required
- Self-healing: Automatically adapts when pages change
- Complex interactions: AI handles multi-step sequences intelligently
- Dynamic pages: Works with pages that change frequently
LLM mode uses the vx14 driver and incurs additional token consumption costs based on prompt and response length. This is more expensive than deterministic flows but requires zero maintenance.
How it works
- Send a prompt: Describe what you want to accomplish in natural language
- AI takes control: The LLM agent analyzes the page and determines the best action sequence
- Real-time execution: Actions are performed live in the browser, not pre-generated
- Adaptive behavior: AI adjusts to page variations and handles unexpected elements
The LLM agent executes actions in real-time based on the current page state. It’s not translating your prompt into predefined steps—it’s actually controlling the browser dynamically.
Supported parameters
Available in - Extract.| Parameter | Type | Description | Default |
|---|---|---|---|
render | Boolean | Enable or disable JS rendering (required to be true) | false |
browser_actions_prompt | String | Described the required page interactions with natural language prompt | - |
Usage
Basic example
Let the AI handle all interactions:Infinite scroll
Handle dynamic content loading:Complex interaction
Multi-step workflows:Example response
When the LLM agent completes your prompt, you receive the final page state and execution details. The response includes:- data: All related extacted data
- data.html: Final DOM state after AI execution
- llm_execution: Details about actions the AI performed
- actions_performed: Step-by-step log of what the agent did
- metadata: Execution details including task id, driver used, execution time and more
- metadata.browser_actions: The browser actions results per step
Best practices
Writing effective prompts
Be specific but flexible:Pricing
LLM mode costs include:- vx14 driver usage: Higher tier driver for AI capabilities
- Token consumption: Based on prompt length and AI response
- API call: Standard request fee

