Displaying 1 to 14 from 14 results

librdkafka - The Apache Kafka C/C++ library

  •    C

Copyright (c) 2012-2016, Magnus Edenhill.librdkafka is a C library implementation of the Apache Kafka protocol, containing both Producer and Consumer support. It was designed with message delivery reliability and high performance in mind, current figures exceed 1 million msgs/second for the producer and 3 million msgs/second for the consumer.

Debezium - Stream changes from your databases.

  •    Java

Debezium is a distributed platform that turns your existing databases into event streams, so applications can see and respond immediately to each row-level change in the databases. Debezium is built on top of Apache Kafka and provides Kafka Connect compatible connectors that monitor specific database management systems. Debezium records the history of data changes in Kafka logs, from where your application consumes them. This makes it possible for your application to easily consume all of the events correctly and completely.

kq - Kafka-based Job Queue for Python

  •    Python

KQ (Kafka Queue) is a lightweight Python library which lets you queue and execute jobs asynchronously using Apache Kafka. It uses kafka-python under the hood. You may need to use sudo depending on your environment.




delivery_boy - A simple way to publish messages to Kafka from Ruby applications

  •    Ruby

In addition to improving response time, delivering messages asynchronously also protects your application against Kafka availability issues -- if messages cannot be delivered, they'll be buffered for later and retried automatically. Regarding partition and partition_key: if none are specified, DeliveryBoy will pick a partition at random. If you want to ensure that e.g. all messages related to a user always get written to the same partition, you can pass the user id to partition_key. Don't use partition directly unless you know what you're doing, since it requires you to know exactly how many partitions each topic has, which can change and cause you pain and misery. Just use partition_key or let DeliveryBoy choose at random.

rafka - Interact with Kafka in a simple way

  •    Go

Rafka is a gateway service that exposes Kafka using simple semantics. It implements the Redis protocol.

neo4j-kafka - Neo4j Kafka Integrations

  •    Kotlin

Ideas based on a now removed repository from workday. You can set the following configuration values in your neo4j.conf, here are the defaults.


reactive-interaction-gateway - Your UI deserves an API, too!

  •    Elixir

In short: handling asynchronous events. Instead, let RIG handle those connections for you. RIG is designed for scalability and allows you to concentrate on the actual business logic. Back-end (micro)services no longer have to care about connection state, which means they can be stateless, making it very easy to roll out updates to them. Sending notifications to all online devices of a certain user becomes as easy as POSTing a message to an HTTP endpoint.

kafka-0.11-examples - Code snippets that demonstrate how to leverage the new Kafka 0.11 APIs

  •    Scala

This repository contains code snippets in Scala, using the Java clients, that demonstrate how to use the Kafka 0.11 new APIs. The code is incomplete, and that's intended: the goal is to bring the focus on the new APIs only.

neo4j-streams - Neo4j Kafka Integrations

  •    Kotlin

You can set the following configuration values in your neo4j.conf, here are the defaults. See the Apache Kafka documentation for details on these settings.

kafkabridge - The Apache Kafka Client SDK

  •    C++

Navigate to the cxx/c installation directory,and run build.sh -release, you will get a new file named libqbus.so in the "./lib/release" directory. Navigate to the Go installation directory,and run build.sh, you will get the new files qbus.go and libQBus_go.so in the directory "gopath/src/qbus".

prometheus-kafka-adapter - Use Kafka as a remote storage database for Prometheus (remote write only)

  •    Go

Prometheus-kafka-adapter is a service which receives Prometheus metrics through remote_write, marshal into JSON and sends them into Kafka. We use prometheus-kafka-adapter internally at Telefonica for dumping Prometheus metrics into an object storage in diferent clouds, through Kafka and Kafka-Connect.