Parseable is a lightweight, cloud native log observability and analytics engine. It is written in Rust and uses Apache Arrow and Parquet.
Parseable uses a simple, index-free mechanism to organize and query data allowing low latency, and high throughput ingestion and query. It can use either a local mount point or object storage (S3/compatible stores) for data storage.
For comparison, Parseable consumes up to ~80% lower memory and ~50% lower CPU than Elastic for similar ingestion throughput. Read more in the benchmarks directory.
- Choose your own storage backend - local drive or S3 (or compatible) object store.
- Ingestion API compatible with HTTP + JSON output of log agents.
- Query log data with PostgreSQL compatible SQL.
- Grafana ↗︎ for visualization.
- Send alerts ↗︎ to webhook targets including Slack.
- Stats API ↗︎ to track ingestion and compressed data.
- Single binary includes all components - ingestion, store and query. Built-in UI.
Run the below command to deploy Parseable in local storage mode with Docker.
mkdir -p $HOME/parseable/data
mkdir -p $HOME/parseable/staging
docker run -p 8000:8000 \
-v $HOME/parseable/data:/parseable/data \
-v $HOME/parseable/staging:/parseable/staging \
-e P_FS_DIR=/parseable/data \
-e P_STAGING_DIR=/parseable/staging \
parseable/parseable:latest \
parseable local-store
Once this runs successfully, you'll see dashboard at http://localhost:8000. You can login to the dashboard default credentials admin
, admin
.
curl --location --request POST 'http://localhost:8000/api/v1/ingest' \
--header 'X-P-META-meta1: value1' \
--header 'X-P-TAG-tag1: value1' \
--header 'X-P-Stream: demo' \
--header 'Authorization: Basic YWRtaW46YWRtaW4=' \
--header 'Content-Type: application/json' \
--data-raw '[
{
"id": "434a5f5e-2f5f-11ed-a261-0242ac120002",
"datetime": "24/Jun/2022:14:12:15 +0000",
"host": "153.10.110.81",
"user-identifier": "Mozilla/5.0 Gecko/20100101 Firefox/64.0",
"method": "PUT",
"status": 500,
"referrer": "http://www.google.com/"
}
]'
Note: The X-P-Stream
header is used to specify the log stream where data is sent. Parseable will create the stream automatically if it doesn't exist already.
You can see the events in Parseable UI, or use the below curl command to see the query response on CLI.
NOTE: Please change the startTime
and endTime
to the time range corresponding to the event you sent in the previous step.
curl --location --request POST 'http://localhost:8000/api/v1/query' \
--header 'Authorization: Basic YWRtaW46YWRtaW4=' \
--header 'Content-Type: application/json' \
--data-raw '{
"query":"select * from demo",
"startTime":"2023-06-09T00:00:00+00:00",
"endTime":"2023-06-09T23:59:00+00:00"
}'
Parseable is benchmarked with K6. Please find the results and details on how to run the benchmark in your environment in the benchmarks directory.
Traditionally, logging has been seen as a text search problem. Log volumes were not high, and data ingestion or storage were not really issues. This led us to today, where all the logging platforms are primarily text search engines.
But with log data growing exponentially, today's log data challenges involve whole lot more – Data ingestion, storage, and observation, all at scale. We are building Parseable to address these challenges.
- For bugs, please create issue on GitHub ↗︎.
- For commercial support and consultation, please reach out to us at
[email protected]
↗︎. - Please consider supporting us on GitHub Sponsors ↗︎.
Refer to the contributing guide here ↗︎.