OpenAI-Compatible API
BLAST provides an OpenAI-compatible API for web browsing AI.
Start the Server
You can then use any OpenAI API client:
Chat Completions API
1. Basic Usage
2. Streaming
Enable streaming to receive real-time updates:
The streaming response includes:
- Initial role message (
role: "assistant"
) - Each chunk’s
delta.content
contains either:- Thought (if the string contains
" "
) - Screenshot (no spaces in content)
- Final result
- Thought (if the string contains
- Final chunk with
finish_reason: "stop"
3. Conversation
BLAST lets you run multi-turn conversations. The engine’s prefix caching ensures that already-computed browser actions are not repeated unless needed.
4. Caching
BLAST will by default cache both results and the LLM-generated steps to create those results (in the case of queries that have results that change but the steps to access to new result value doesn’t).
Control caching behavior with cache_control
options:
Available cache control options:
no-cache
- Skip results cache lookupno-store
- Don’t store in results cacheno-cache-plan
- Skip plan cache lookupno-store-plan
- Don’t store plan in cache
Responses API
1. Basic Usage
2. Streaming
Enable streaming to receive detailed event updates:
BLAST emits a sequence of events during streaming:
response.created
- Initial event when the response is createdresponse.in_progress
- Task processing has startedresponse.output_text.delta
- Each streaming event’s delta is either:- Thought (if the content contains
" "
) - Screenshot (no spaces in content)
- Thought (if the content contains
response.output_text.done
- An event is completeresponse.completed
- Indicates all events are sent.
3. Conversation
Support for stateful conversations using previous response IDs:
4. Caching
Next Steps
- Learn about the Engine API for direct access
- Understand Concurrency and Parallelism
- Configure Settings and Constraints