Skip to content

deliveryhero/kinesis-logback-appender

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

57 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Maven Central

LOGBack Appender for Amazon Kinesis

This is an implementation of the AWS - Labs log4j appender for LOGBack.

Supports both Kinesis and Kinesis Firehose streams.

Sample Configuration

<configuration>
  <appender name="KINESIS" class="com.gu.logback.appender.kinesis.KinesisAppender">
    <bufferSize>1000</bufferSize>
    <threadCount>20</threadCount>
    <endpoint>kinesis.us-east-1.amazonaws.com</endpoint><!-- Specify endpoint OR region -->
    <region>us-east-1</region>
    <roleToAssumeArn>foo</roleToAssumeArn><!-- Optional: ARN of role for cross account access -->
    <maxRetries>3</maxRetries>
    <shutdownTimeout>30</shutdownTimeout>
    <streamName>testStream</streamName>
    <encoding>UTF-8</encoding>
    <layout class="ch.qos.logback.classic.PatternLayout">
      <pattern>%m</pattern>
    </layout>
  </appender>
  <appender name="stdout" class="ch.qos.logback.core.ConsoleAppender">
    <encoder>
      <pattern>%5p [%t] (%F:%L) - %m%n</pattern>
    </encoder>
  </appender>
  <logger name="KinesisLogger" additivity="false" level="INFO">
    <appender-ref ref="KINESIS"/>
  </logger>
  <root level="INFO">
    <appender-ref ref="stdout"/>
  </root>
</configuration>

Use com.gu.logback.appender.kinesis.KinesisAppender for Kinesis or com.gu.logback.appender.kinesis.FirehoseAppender for Kinesis Firehose.

Performance and reliability notes

This appender is performant but will block if the Kinesis stream throughput is exceeded. In order to guard against this you might want to consider:

  • ensure you have calculated how many shards you need based on your expected throughput
  • alerting on write throughput exceeded on the Kinesis stream(s)
  • setting up an autoscaling approach that will automatically scale your shards up and down appropriately AWS docs
  • configuring the AWS client to not retry on failure so that log lines are discarded when stream throughput is exceeded rather than backing up and causing a cascading failure
  • wrapping the appender in AsyncAppender, which can be configured to automatically drop overflowing messages on blocking

Testing locally

In order to test this you can simply use mvn install (to deploy to your local machine).

Releasing

Some notes for Guardian employees shipping updates to this.

First of all confirm that your pom.xml has a SNAPSHOT version in it (e.g. https://github.com/guardian/kinesis-logback-appender/blob/08de9295a41ef99f72fb0d75d7717d61b7c5f4f2/pom.xml#L22).

In order to release this to maven you'll need to have a settings file at ~/.m2/settings.xml containing your sonatype credentials (you can probably find these in .sbt/0.13/sonatype.sbt if you've shipped Scala libraries):

<settings>
  <servers>
    <server>
      <id>ossrh</id>
      <username>username</username>
      <password>password</password>
    </server>
  </servers>
</settings>

Once you've got that you can use mvn clean deploy to deploy your snapshot to sonatype. This will only release to the snapshot repo (you can add resolvers += "Sonatype snapshots" at "https://oss.sonatype.org/content/repositories/snapshots" to test resolution of this).

Finally when ready run mvn release:clean release:prepare and follow the prompts. The tagging convention we've been using is to use the same as the version we are shipping (no project name). Once this has completed you need to do one more step to actually release it on maven central: mvn release:perform.

About

LOGBack Appender for writing data into a Kinesis Stream

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Java 100.0%