Skip to content

loadmill/loadmill-kafka-relay

Repository files navigation

Loadmill Kafka Relay Client

The Loadmill Kafka Relay is a powerful Kafka client that allows you to produce and consume messages from Kafka topics easily. It provides a set of APIs to interact with Kafka brokers, manage message production, consumption, and schema registry operations. It was built to help you test your Kafka infrastructure and applications, and to make it easier to debug and troubleshoot Kafka-related issues.

Use Cases

The Loadmill Kafka Relay aims to simplify the testing of services that use Kafka, a common challenge due to the complex nature of message systems. Two common use cases are:

  1. Message Consumption Testing: This involves testing the functionality of applications that produce Kafka messages. For instance, after you simulate an action like creating an order through the application's API, the relay is used to consume messages from the corresponding Kafka topic. This process allows you to verify whether the application is correctly sending the expected messages to the right topic, ensuring its proper functionality.

  2. Message Production Testing: The relay also facilitates testing the response of applications that consume Kafka messages. By producing and sending messages to a specific Kafka topic through the relay, you can observe and analyze how consumer applications behave. This helps in ensuring that these applications correctly process and react to the messages as they are intended in various scenarios.

Table of Contents

Getting Started

To use the Loadmill Kafka Relay Client, you have 2 options:

Option 1:

Docker

The Loadmill Kafka Relay Client is available as a Docker image, which you either pull from Docker Hub or build locally.

Docker hub

To use the ready-made docker from docker-hub run the following commands:

  • Download the image
docker pull loadmill/kafka-relay:latest
  • Run the image
docker run -p 3000:3000 loadmill/kafka-relay 

OR you can build the image locally, in the following way:

Local Docker Build

  • Build the image:
docker build -t loadmill/kafka-relay .
  • Run the image:
docker run -p 3000:3000 loadmill/kafka-relay

Option 2:

Node.js

You can fork/clone this repository and run the Kafka Relay Client locally using Node.js.

git clone [email protected]:loadmill/loadmill-kafka-relay.git
cd loadmill-kafka-relay
npm install # or yarn install
npm run build # or yarn build
npm run start # or yarn start

API Reference

The Loadmill Kafka Relay Client provides the following API endpoints for interacting with Kafka. You can make HTTP requests to these endpoints to perform various actions.

Basic Usage

Subscribe to a Topic

Endpoint: POST /subscribe

To subscribe to a Kafka topic and start receiving messages, send a POST request to the /subscribe endpoint with the following parameters in the request body:

  • Required parameters:
    • brokers (array of strings): The list of Kafka brokers.
    • topic (string): The Kafka topic to subscribe to.
  • Optional parameters:
    • ssl (boolean, optional): Whether to use SSL for the Kafka brokers.
    • sasl (object, optional): SASL authentication credentials for the Kafka brokers.
      • mechanism (string): The SASL mechanism to use for authentication.
      • username (string): The username for authentication.
      • password (string): The password for authentication.
    • timestamp (number, optional): The timestamp from which to start consuming messages. Should be a number in the format of milliseconds since the Unix epoch (Example: 1623346800000). If not provided, the relay will start consuming messages from 1 minute ago as the default.

Example Request:

POST /subscribe
Content-Type: application/json
{
  "brokers": ["kafka1.example.com:9092", "kafka2.example.com:9092"],
  "topic": "my-topic"
}

Example Response:

{
  "id": "c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55"
}

Use the id field in the response to consume messages from the subscribed Kafka topic, with the /consume/:id endpoint.

Consume Messages

Endpoint: GET /consume/:id

To consume messages from a subscribed Kafka topic, send a GET request to the /consume/:id endpoint, where :id represents the unique identifier of the subscription. The default result will be the latest message on the topic.

If no message is available, the relay waits for 1 second and tries to consume again. If no message is available after 25 seconds (or a provided timeout value), an error will be returned.

Optional query parameters:

  • filter (string, optional): A regular expression to filter the topic's messages by.
  • multiple (number, optional): The number of messages to consume. Will return the latest multiple messages. Minimum value is 1, maximum value is 100. Defaults to 1.
  • timeout (number, optional): The maximum time (in seconds) to wait for a message to be available. If no message is available after the timeout, an error will be returned. Minimum value is 5, maximum value is 25. Defaults to 25.
  • text (boolean, optional): Whether to return the message value as a string (and not parse it as JSON). Defaults to false.

Example Request 1:

GET /consume/c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55

Example Response 1:

{
  "messages": [
    {
      "key": "my-key",
      "value": "Hello, Kafka!",
      "timestamp": 1623346800000,
      ...
    }
  ]
}

Example Request 2 (using a regex filter):

GET /consume/c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55?filter=myregex

Example Request 3 (using a timeout):

GET /consume/c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55?filter=myregex&timeout=7

Example Request 4 (using multiple messages option):

GET /consume/c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55?multiple=2

Example Response 4 (using multiple messages option):

{
  "messages": [
    {
      "key": "my-key-1",
      "value": "Hello, Kafka 1!",
      "timestamp": 1623346800000,
      ...
    },
    {
      "key": "my-key-2",
      "value": "Hello, Kafka 2!",
      "timestamp": 1623346800001,
      ...
    }
  ]
}

Compression codecs

To turn on LZ4 compression codec decoding on message consuming add the following environemt variable:

LOADMILL_KAFKA_LZ4_COMPRESSION_CODEC='true'

Unsubscribe from a Topic

Endpoint: DELETE /subscriptions/:id

To unsubscribe from a Kafka topic, send a DELETE request to the /subscriptions/:id endpoint, where :id represents the unique identifier of the subscriptions.

Example Request:

DELETE /subscriptions/c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55
{
  "id": "c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55"
}

Example Response:

{
  "id": "c8bf8b9b-5bb6-4f17-babd-3d027eb7ad55"
}

Advanced Usage

Produce a Message

Endpoint: POST /produce

To produce a message to a Kafka topic, send a POST request to the /produce endpoint with the following parameters in the request body:

  • Required Parameters:
    • brokers (array of strings): The list of Kafka brokers.
    • topic (string): The Kafka topic to which the message will be produced.
    • message (object): The message to be produced.
      • value (string): The value of the message.
      • key (string, optional): The key of the message.
      • headers (object, optional): The headers of the message.
  • Optional Parameters:
    • conversions (array, optional): Options for certain keys inside the message value and headers. Currently supports only decimal and bytes conversions. (Note: For bytes conversion, the value must be an encoded base64 string. See example below.) Each item in the array should be an object with the following keys:
      • key (string): The key to convert.
      • type (string): The type to convert the key to. Can be one of the following: decimal, bytes.
    • encode (object, optional): Options for schema encoding.
      • value (object, optional): Options for encoding the message value.
        • subject (string): The subject of the schema.
        • version (number, optional): The version of the schema.
      • headers (object, optional): Options for encoding the message headers.
        • key (string): The key of the header to encode.
          • subject (string): The subject of the schema.
          • version (number, optional): The version of the schema.
    • ssl (boolean, optional): Whether to use SSL for the Kafka brokers.
    • sasl (object, optional): SASL authentication credentials for the Kafka brokers.
      • mechanism (string): The SASL mechanism to use for authentication.
      • username (string): The username for authentication.
      • password (string): The password for authentication.

Note: Using encode will only apply your encoding to the current message. It will not override the global encode (this is set via the /registry/encode endpoint).

Example Request:

POST /produce
Content-Type: application/json
{
  // required parameters
  "brokers": ["kafka1.example.com:9092", "kafka2.example.com:9092"],
  "topic": "my-topic",
  "message": {
    "value": "Hello, Kafka!",
    "key": "my-key",
    "headers": {
      "my-header": "${__encode_base64(my-header-value)}" // __encode_base64 is a Loadmill function
    }
  }
  // optional parameters
  "conversions": [
    {
      "key": "my-key",
      "type": "decimal"
    },
    {
      "key": "my-header",
      "type": "bytes"
    }
  ],
  "encode": {
    "value": {
      "subject": "my-schema",
      "version": 1
    }
  },
  "ssl": true,
  "sasl": {
    "mechanism": "plain",
    "username": "user",
    "password": "password"
  }
}

Schema Registry

Setting a Schema Registry

You may use the following env vars to connect to a schema registry on the app startup:

# Minimal configuration for decoding messages:
LOADMILL_KAFKA_SCHEMA_REGISTRY_URL= # The URL of the Schema Registry
# For encoding messages add this var:
LOADMILL_KAFKA_SCHEMA_SUBJECT= # The subject of the schema
# Optional configuration
LOADMILL_KAFKA_SCHEMA_VERSION= # The version of the schema
LOADMILL_KAFKA_SCHEMA_REGISTRY_USERNAME= # The username for authentication
LOADMILL_KAFKA_SCHEMA_REGISTRY_PASSWORD= # The password for authentication

Another option is to use the following endpoint to set (connect to) a Schema Registry:

Endpoint: POST /registry

To initialize the Schema Registry for your Kafka environment, send a POST request to the /registry endpoint with the following parameters in the request body:

  • Required parameters:
    • url (string): The URL of the Schema Registry.
  • Optional parameters:
    • auth (object, optional): Authentication credentials for the Schema Registry.
      • username (string): The username for authentication.
      • password (string): The password for authentication.
    • encode (object, optional): Options for schema encoding.
      • subject (string): The subject of the schema.
      • version (number, optional): The version of the schema.

Note: that the optional encode option is used to set the encode schema for the Schema Registry. If you don't provide this option, messages will not be encoded.

Example Request:

POST /registry
Content-Type: application/json
{
  "url": "https://schema-registry.example.com",
  "auth": {
    "username": "user",
    "password": "password"
  },
  "encode": {
    "subject": "my-schema",
    "version": 1
  }
}

Set Encode Schema

Endpoint: PUT /registry/encode

This endpoint is useful when you want to encode your produced message according to a specific schema, or want to update the current encode schema that is set for the Schema Registry.

Note 1: Before you call this endpoint, you must initialize the Schema Registry using the /registry endpoint. Note 2: Changing the encode schema will affect all future produced messages.

To set the encode schema for the Schema Registry, send a PUT request to the /registry/encode endpoint with the following parameters in the request body:

  • Required parameters:
    • subject (string): The subject of the schema.
  • Optional parameters:
    • version (number, optional): The version of the schema. (If not provided, the latest version will be used.)

Example Request:

PUT /registry/encode
Content-Type: application/json
{
  "subject": "my-schema",
  "version": 1
}

Environment Variables

Why Environment Variables?

The main advantage of using environment variables is security. You can store sensitive information like passwords and private urls in environment variables, and then use them in your requests without exposing them in your code.

How to Use Env Vars Here?

If you pass a value in the request body that looks like this: <VAR_NAME> the Loadmill Kafka Relay Client will try to replace it with the value of the corresponding environment variable.

Note: You can use any name you want, as long as it matches the name of an environment variable.

Example: Assuming the following environment variables are defined:

BROKERS=kafka1.example.com:9092,kafka2.example.com:9092
TOPIC=my-topic
SSL=true
MECHANISM=plain
USERNAME=user
PASSWORD=password

Then you can send the following request to the /subscribe endpoint:

POST /subscribe
{
  "brokers": "<BROKERS>",
  "topic": "<TOPIC>",
  "ssl": "<SSL>",
  "sasl": {
    "mechanism": "<MECHANISM>",
    "username": "<USERNAME>",
    "password": "<PASSWORD>"
  }
}

And the server will receive the following request body:

POST /subscribe
{
  "brokers": ["kafka1.example.com:9092", "kafka2.example.com:9092"],
  "topic": "my-topic",
  "ssl": true,
  "sasl": {
    "mechanism": "plain",
    "username": "user",
    "password": "password"
  }
}

This example uses the /subscribe call, but the env var replacement will work for all endpoints.

NOTE: "brokers" key is converted to an array, so the string value of the env var will be split by commas (,).

To recap: <FOO> -> process.env.FOO

Meaning, any value in the request body using the < BAR > syntax will be replaced with the corresponding environment variable, specifically process.env.BAR.

Env Var Names Used in This App

You may use the following environment variables in this app:

LOADMILL_KAFKA_SERVER_PORT= # The port on which the server will listen, defaults to 3000

Debugging

Debug API

The Loadmill Kafka Relay Client provides a debug API that can be used to get information about the current state of the relay, such as the list of active subscriptions and the current schema registry configuration.

Endpoint: GET /debug

Example Request:

GET /debug

Example Response:

{
  "schemaRegistry": {
    "url": "https://schema-registry.example.com"
  },
  "subscriptions": {
    "11920dc8-f571-43e7-a0e2-af5f66b422f1": {
      "topic": "my_topic",
      "timeOfSubscription": 1721803796114,
      "messages": [
        {
          "timestamp": "1721719271555",
          "offset": "16",
          "key": null,
          "headers": {},
          "value": "{\"a\":\"b\",\"..."
        }
      ]
    },
    "f1b1b1b1-1b1b-1b1b-1b1b-1b1b1b1b1b1b": {
      "topic": "his_topic",
      "timeOfSubscription": 1721803796222,
      "messages": []
    }
  }
}

Logging

The Loadmill Kafka Relay Client writes its logs to loadmill-kafka-relay.log in the root directory of the app, as well as to the console. By default, logs are written the info level. You can customize the logging behavior by setting the following environment variables:

NODE_ENV=development # Setting this will enable debug logs to console only

LOG_LEVEL= # The level of logging. Can be one of the following: "fatal" | "error" | "warn" | "info" | "debug" | "trace". Defaults to info.

Multi Instance

The Loadmill Kafka Relay supports a multi-instance mode, which allows the relay to manage its data on Redis. This feature is enabled by passing the REDIS_URL environment variable. When running in multi-instance mode, the relay will store its state in Redis and can take over other instances if they are shut down.

Enabling Multi-Instance Mode

To enable multi-instance mode, set the REDIS_URL environment variable to the URL of your Redis instance. This can be done by adding the following line to your environment configuration:

REDIS_URL=rediss://your-redis-host:6379
# Optional:
REDIS_TLS_REJECT_UNAUTHORIZED=false # If you're Redis instance is using a self-signed certificate

Behavior in Multi-Instance Mode

  • Memory Management: The relay will store its Subscribers, Kafka Messages and Kafka connections in Redis, ensuring that the state is preserved across different instances.
  • Instance Takeover: If an instance is shut down, other running instances will take over its responsibilities, ensuring continuous operation.

API Behavior in Multi-Instance Mode

  • /subscribe API: This API will create a subscriber maintained by the instance that handles the request.
  • /consume, /unsubscribe, and /debug APIs: These APIs are agnostic of the instance and will return the desired result regardless of whether the subscriber ID belongs to that instance or another.

Best Practices for Multi-Instance Mode

Multi-instance mode is best suited for scenarios where you have multiple instances running in a virtual environment (in the background) and only expose a single URL for all the instances. This can be achieved using a load balancer (to ensure no single relay becomes overwhelmed) or a reverse proxy.

About

Loadmill's Kafka relay

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published