Skip to content

Latest commit

 

History

History
208 lines (151 loc) · 6.16 KB

README.md

File metadata and controls

208 lines (151 loc) · 6.16 KB

browsy

What is browsy?

browsy is a service that lets you run browser automation tasks without managing browser instances yourself. It provides:

  • Simple Job Definition: Write Playwright-powered automation tasks in Python
  • HTTP API: Queue jobs and retrieve results through HTTP endpoints
  • Docker Ready: Run everything in containers without worrying about browser dependencies
  • Queue System: Jobs are processed in order, with automatic retries and status tracking
  • Extensible: Create any browser automation task - from screenshots and PDFs to complex scraping operations

Think of it as a way to turn your Playwright scripts into HTTP services that can be called from anywhere.

Quick Start

Download files

To get started, download the necessary files using the following script:

curl -LsSf https://raw.githubusercontent.com/mbroton/browsy/main/scripts/get.sh | sh

This script will download the docker-compose file and example jobs. Once downloaded, navigate to the browsy/ directory.

Alternatively, you can clone the repository directly and navigate to the quickstart directory.

Start browsy

To start the service, run:

docker compose up --build --scale worker=3

You can adjust the number of workers by modifying the --scale worker parameter.

Visit http://localhost:8000/docs to access the interactive API documentation provided by FastAPI.

Defining custom jobs

A job is defined as any class that inherits from browsy.BaseJob. Browsy will automatically search for these classes within the jobs/ directory.

Here's an example implementation:

from browsy import BaseJob, Page

class ScreenshotJob(BaseJob):
    NAME = "screenshot"

    url: str | None = None
    html: str | None = None
    full_page: bool = False

    async def execute(self, page: Page) -> bytes:
        if self.url:
            await page.goto(self.url)
        elif self.html:
            await page.set_content(self.html)
        return await page.screenshot(full_page=self.full_page)

    async def validate_logic(self) -> bool:
        return bool(self.url) != bool(self.html)
  • Class Definition: The ScreenshotJob class inherits from BaseJob, which itself is based on Pydantic's BaseModel. This provides automatic data validation and serialization.
  • Job Name: The NAME attribute uniquely identifies the job type when making API calls.
  • Parameters: Defined as class attributes, these are automatically validated by Pydantic during API calls. This ensures that input data meets the expected types and constraints before processing.
  • Validation Logic: The validate_logic method runs during API calls to verify that the job's input parameters satisfy specific conditions. This validation occurs before the job is submitted for execution, allowing for early detection of configuration errors.
  • Execution Method: The execute method carries out the browser automation using a Playwright Page object. Workers use this method to execute jobs.

Refer to Playwright's documentation for more details on what you can do with page.

Client

To interact with the service using Python, you can use browsy client:

pip install browsy

Here's how you can use it:

from browsy import BrowsyClient
# Also available:
# from browsy import AsyncBrowsyClient

client = BrowsyClient(base_url="http://localhost:8000")

job = client.create_job(name="screenshot", parameters={
    "url": "https://example.com",
    "full_page": True
})

result = client.get_job_output(job_id=job.id)
if result:
  with open("screenshot.png", "wb") as f:
      f.write(result)

This example demonstrates how to submit a screenshot job, retrieve the result, and save it locally.

API

You can explore and interact with the API using the Swagger UI documentation provided by FastAPI. Visit http://localhost:8000/docs to access it.

If you prefer to interact with the service directly via HTTP, here are example requests:

Submit a job

POST /api/v1/jobs

Example request:

{
  "name": "screenshot",
  "parameters": {
    "url": "https://example.com",
    "html": null,
    "full_page": true
  }
}

Example response:

{
  "id": 1,
  "name": "screenshot",
  "input": {
    "url": "https://example.com",
    "html": null,
    "full_page": true
  },
  "processing_time": null,
  "status": "pending",
  "created_at": "2024-12-30T15:02:04.720000",
  "updated_at": null,
  "worker": null
}

Check job status

GET /api/v1/jobs/{job_id}

Example response:

{
  "id": 1,
  "name": "screenshot",
  "input": {
    "url": "https://example.com",
    "html": null,
    "full_page": true
  },
  "processing_time": 1298,
  "status": "done",
  "created_at": "2024-12-30T15:06:39.204000",
  "updated_at": "2024-12-30T15:06:44.743000",
  "worker": "worker_lze4nFMy"
}

Retrieve job result

To retrieve the result of a job, use the following endpoint:

GET /api/v1/jobs/{job_id}/result

Status Codes:

  • 200: The job is complete, and the output is available. The response type is application/octet-stream.
  • 202: The job is pending or currently in progress.
  • 204: The job is complete or has failed, and there is no output available.
  • 404: No job exists with the provided ID.

Response Headers:

  • X-Job-Status: Indicates the current status of the job.
  • X-Job-Last-Updated: Shows the last time the job's status was updated.

Internal Dashboard

Browsy provides a built-in monitoring dashboard accessible at /internal. This interface gives you real-time visibility:

  • View all active workers and their current status
  • Monitor job queues and execution status
  • Track execution times
  • Identify potential bottlenecks or issues

The dashboard updates automatically to show you the latest state of your system:

Internal Dashboard

How it works

flow

  1. You define jobs using Playwright's API
  2. Send job requests through HTTP
  3. Workers execute jobs in Docker containers
  4. Get results when ready

License

MIT License - see LICENSE for details.