Interface KafkaReadStream<K,​V>

  • All Superinterfaces:
    ReadStream<org.apache.kafka.clients.consumer.ConsumerRecord<K,​V>>, StreamBase

    public interface KafkaReadStream<K,​V>
    extends ReadStream<org.apache.kafka.clients.consumer.ConsumerRecord<K,​V>>
    A ReadStream for consuming Kafka ConsumerRecord.

    The pause() and resume() provides global control over reading the records from the consumer.

    The pause(Set) and resume(Set) provides finer grained control over reading records for specific Topic/Partition, these are Kafka's specific operations.

    • Method Summary

      All Methods Static Methods Instance Methods Abstract Methods 
      Modifier and Type Method Description
      Future<Void> assign​(Set<org.apache.kafka.common.TopicPartition> partitions)
      Manually assign a set of partitions to this consumer.
      Future<Set<org.apache.kafka.common.TopicPartition>> assignment()
      Get the set of partitions currently assigned to this consumer.
      KafkaReadStream<K,​V> batchHandler​(Handler<org.apache.kafka.clients.consumer.ConsumerRecords<K,​V>> handler)
      Set the handler that will be called when a new batch of records is returned from Kafka.
      Future<Map<org.apache.kafka.common.TopicPartition,​Long>> beginningOffsets​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Get the first offset for the given partitions.
      Future<Long> beginningOffsets​(org.apache.kafka.common.TopicPartition topicPartition)
      Get the first offset for the given partition.
      Future<Void> close()
      Close the stream
      Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata>> commit()
      Commit current offsets for all the subscribed list of topics and partition.
      Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata>> commit​(Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata> offsets)
      Commit the specified offsets for the specified list of topics and partitions to Kafka.
      Future<org.apache.kafka.clients.consumer.OffsetAndMetadata> committed​(org.apache.kafka.common.TopicPartition topicPartition)
      Get the last committed offset for the given partition (whether the commit happened by this process or another).
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, KafkaClientOptions options)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, KafkaClientOptions options, Class<K> keyType, Class<V> valueType)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, KafkaClientOptions options, org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer, org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Map<String,​Object> config)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Map<String,​Object> config, Class<K> keyType, Class<V> valueType)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Map<String,​Object> config, org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer, org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Properties config)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Properties config, Class<K> keyType, Class<V> valueType)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, Properties config, org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer, org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, org.apache.kafka.clients.consumer.Consumer<K,​V> consumer)
      Create a new KafkaReadStream instance
      static <K,​V>
      KafkaReadStream<K,​V>
      create​(Vertx vertx, org.apache.kafka.clients.consumer.Consumer<K,​V> consumer, KafkaClientOptions options)
      Create a new KafkaReadStream instance
      long demand()
      Returns the current demand.
      KafkaReadStream<K,​V> endHandler​(Handler<Void> endHandler)
      Set an end handler.
      Future<Map<org.apache.kafka.common.TopicPartition,​Long>> endOffsets​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Get the last offset for the given partitions.
      Future<Long> endOffsets​(org.apache.kafka.common.TopicPartition topicPartition)
      Get the last offset for the given partition.
      KafkaReadStream<K,​V> exceptionHandler​(Handler<Throwable> handler)
      Set an exception handler on the read stream.
      KafkaReadStream<K,​V> fetch​(long amount)
      Fetch the specified amount of elements.
      KafkaReadStream<K,​V> handler​(Handler<org.apache.kafka.clients.consumer.ConsumerRecord<K,​V>> handler)
      Set a data handler.
      Future<Map<String,​List<org.apache.kafka.common.PartitionInfo>>> listTopics()
      Get metadata about partitions for all topics that the user is authorized to view.
      Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndTimestamp>> offsetsForTimes​(Map<org.apache.kafka.common.TopicPartition,​Long> topicPartitionTimestamps)
      Look up the offsets for the given partitions by timestamp.
      Future<org.apache.kafka.clients.consumer.OffsetAndTimestamp> offsetsForTimes​(org.apache.kafka.common.TopicPartition topicPartition, long timestamp)
      * Look up the offset for the given partition by timestamp.
      KafkaReadStream<K,​V> partitionsAssignedHandler​(Handler<Set<org.apache.kafka.common.TopicPartition>> handler)
      Set the handler called when topic partitions are assigned to the consumer
      Future<List<org.apache.kafka.common.PartitionInfo>> partitionsFor​(String topic)
      Get metadata about the partitions for a given topic.
      KafkaReadStream<K,​V> partitionsRevokedHandler​(Handler<Set<org.apache.kafka.common.TopicPartition>> handler)
      Set the handler called when topic partitions are revoked to the consumer
      KafkaReadStream<K,​V> pause()
      Pause the ReadStream, it sets the buffer in fetch mode and clears the actual demand.
      Future<Void> pause​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Suspend fetching from the requested partitions.
      Future<Set<org.apache.kafka.common.TopicPartition>> paused()
      Get the set of partitions that were previously paused by a call to pause(Set).
      Future<org.apache.kafka.clients.consumer.ConsumerRecords<K,​V>> poll​(java.time.Duration timeout)
      Executes a poll for getting messages from Kafka.
      KafkaReadStream<K,​V> pollTimeout​(java.time.Duration timeout)
      Sets the poll timeout for the underlying native Kafka Consumer.
      Future<Long> position​(org.apache.kafka.common.TopicPartition partition)
      Get the offset of the next record that will be fetched (if a record with that offset exists).
      KafkaReadStream<K,​V> resume()
      Resume reading, and sets the buffer in flowing mode.
      Future<Void> resume​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Resume specified partitions which have been paused with pause.
      Future<Void> seek​(org.apache.kafka.common.TopicPartition topicPartition, long offset)
      Overrides the fetch offsets that the consumer will use on the next poll.
      Future<Void> seek​(org.apache.kafka.common.TopicPartition topicPartition, org.apache.kafka.clients.consumer.OffsetAndMetadata offsetAndMetadata)
      Overrides the fetch offsets that the consumer will use on the next poll.
      Future<Void> seekToBeginning​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Seek to the first offset for each of the given partitions.
      Future<Void> seekToEnd​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
      Seek to the last offset for each of the given partitions.
      Future<Void> subscribe​(Pattern pattern)
      Subscribe to all topics matching specified pattern to get dynamically assigned partitions.
      Future<Void> subscribe​(Set<String> topics)
      Subscribe to the given list of topics to get dynamically assigned partitions.
      Future<Set<String>> subscription()
      Get the current subscription.
      Future<Void> unsubscribe()
      Unsubscribe from topics currently subscribed with subscribe.
      org.apache.kafka.clients.consumer.Consumer<K,​V> unwrap()  
    • Method Detail

      • handler

        KafkaReadStream<K,​V> handler​(Handler<org.apache.kafka.clients.consumer.ConsumerRecord<K,​V>> handler)
        Description copied from interface: ReadStream
        Set a data handler. As data is read, the handler will be called with the data.
        Specified by:
        handler in interface ReadStream<K>
        Returns:
        a reference to this, so the API can be used fluently
      • pause

        KafkaReadStream<K,​V> pause()
        Description copied from interface: ReadStream
        Pause the ReadStream, it sets the buffer in fetch mode and clears the actual demand.

        While it's paused, no data will be sent to the data handler.

        Specified by:
        pause in interface ReadStream<K>
        Returns:
        a reference to this, so the API can be used fluently
      • resume

        KafkaReadStream<K,​V> resume()
        Description copied from interface: ReadStream
        Resume reading, and sets the buffer in flowing mode.

        If the ReadStream has been paused, reading will recommence on it.

        Specified by:
        resume in interface ReadStream<K>
        Returns:
        a reference to this, so the API can be used fluently
      • fetch

        KafkaReadStream<K,​V> fetch​(long amount)
        Description copied from interface: ReadStream
        Fetch the specified amount of elements. If the ReadStream has been paused, reading will recommence with the specified amount of items, otherwise the specified amount will be added to the current stream demand.
        Specified by:
        fetch in interface ReadStream<K>
        Returns:
        a reference to this, so the API can be used fluently
      • endHandler

        KafkaReadStream<K,​V> endHandler​(Handler<Void> endHandler)
        Description copied from interface: ReadStream
        Set an end handler. Once the stream has ended, and there is no more data to be read, this handler will be called.
        Specified by:
        endHandler in interface ReadStream<K>
        Returns:
        a reference to this, so the API can be used fluently
      • demand

        long demand()
        Returns the current demand.
          If the stream is in flowing mode will return Long.MAX_VALUE.
        • If the stream is in fetch mode, will return the current number of elements still to be delivered or 0 if paused.
        Returns:
        current demand
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Properties config)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Properties config,
                                                             Class<K> keyType,
                                                             Class<V> valueType)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        keyType - class type for the key deserialization
        valueType - class type for the value deserialization
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Properties config,
                                                             org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer,
                                                             org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        keyDeserializer - key deserializer
        valueDeserializer - value deserializer
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Map<String,​Object> config)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Map<String,​Object> config,
                                                             Class<K> keyType,
                                                             Class<V> valueType)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        keyType - class type for the key deserialization
        valueType - class type for the value deserialization
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             Map<String,​Object> config,
                                                             org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer,
                                                             org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        config - Kafka consumer configuration
        keyDeserializer - key deserializer
        valueDeserializer - value deserializer
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             KafkaClientOptions options)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        options - Kafka consumer options
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             KafkaClientOptions options,
                                                             Class<K> keyType,
                                                             Class<V> valueType)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        options - Kafka consumer options
        keyType - class type for the key deserialization
        valueType - class type for the value deserialization
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             KafkaClientOptions options,
                                                             org.apache.kafka.common.serialization.Deserializer<K> keyDeserializer,
                                                             org.apache.kafka.common.serialization.Deserializer<V> valueDeserializer)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        options - Kafka consumer options
        keyDeserializer - key deserializer
        valueDeserializer - value deserializer
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             org.apache.kafka.clients.consumer.Consumer<K,​V> consumer)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        consumer - native Kafka consumer instance
        Returns:
        an instance of the KafkaReadStream
      • create

        static <K,​V> KafkaReadStream<K,​V> create​(Vertx vertx,
                                                             org.apache.kafka.clients.consumer.Consumer<K,​V> consumer,
                                                             KafkaClientOptions options)
        Create a new KafkaReadStream instance
        Parameters:
        vertx - Vert.x instance to use
        consumer - native Kafka consumer instance
        options - options used only for tracing settings
        Returns:
        an instance of the KafkaReadStream
      • committed

        Future<org.apache.kafka.clients.consumer.OffsetAndMetadata> committed​(org.apache.kafka.common.TopicPartition topicPartition)
        Get the last committed offset for the given partition (whether the commit happened by this process or another).
        Parameters:
        topicPartition - topic partition for getting last committed offset
        Returns:
        a future notified on operation completed
      • pause

        Future<Void> pause​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Suspend fetching from the requested partitions.

        Due to internal buffering of messages, the record handler will continue to observe messages from the given topicPartitions until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will not see messages from the given topicPartitions.

        Parameters:
        topicPartitions - topic partition from which suspend fetching
        Returns:
        a Future completed with the operation result
      • paused

        Future<Set<org.apache.kafka.common.TopicPartition>> paused()
        Get the set of partitions that were previously paused by a call to pause(Set).
        Returns:
        a Future completed with the operation result
      • resume

        Future<Void> resume​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Resume specified partitions which have been paused with pause.
        Parameters:
        topicPartitions - topic partition from which resume fetching
        Returns:
        a Future completed with the operation result
      • seekToEnd

        Future<Void> seekToEnd​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Seek to the last offset for each of the given partitions.

        Due to internal buffering of messages, the record handler will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new offset.

        Parameters:
        topicPartitions - topic partition for which seek
        Returns:
        a Future completed with the operation result
      • seekToBeginning

        Future<Void> seekToBeginning​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Seek to the first offset for each of the given partitions.

        Due to internal buffering of messages, the record handler will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new offset.

        Parameters:
        topicPartitions - topic partition for which seek
        Returns:
        a Future completed with the operation result
      • seek

        Future<Void> seek​(org.apache.kafka.common.TopicPartition topicPartition,
                          long offset)
        Overrides the fetch offsets that the consumer will use on the next poll.

        Due to internal buffering of messages, the record handler will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new offset.

        Parameters:
        topicPartition - topic partition for which seek
        offset - offset to seek inside the topic partition
        Returns:
        a Future completed with the operation result
      • seek

        Future<Void> seek​(org.apache.kafka.common.TopicPartition topicPartition,
                          org.apache.kafka.clients.consumer.OffsetAndMetadata offsetAndMetadata)
        Overrides the fetch offsets that the consumer will use on the next poll.

        Due to internal buffering of messages, the record handler will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new offset.

        Parameters:
        topicPartition - topic partition for which seek
        offsetAndMetadata - offset to seek inside the topic partition
        Returns:
        a Future completed with the operation result
      • partitionsRevokedHandler

        KafkaReadStream<K,​V> partitionsRevokedHandler​(Handler<Set<org.apache.kafka.common.TopicPartition>> handler)
        Set the handler called when topic partitions are revoked to the consumer
        Parameters:
        handler - handler called on revoked topic partitions
        Returns:
        current KafkaReadStream instance
      • partitionsAssignedHandler

        KafkaReadStream<K,​V> partitionsAssignedHandler​(Handler<Set<org.apache.kafka.common.TopicPartition>> handler)
        Set the handler called when topic partitions are assigned to the consumer
        Parameters:
        handler - handler called on assigned topic partitions
        Returns:
        current KafkaReadStream instance
      • subscribe

        Future<Void> subscribe​(Set<String> topics)
        Subscribe to the given list of topics to get dynamically assigned partitions.

        Due to internal buffering of messages, when changing the subscribed topics the old set of topics may remain in effect (as observed by the handler(Handler) record handler}) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new set of topics.

        Parameters:
        topics - topics to subscribe to
        Returns:
        a Future completed with the operation result
      • subscribe

        Future<Void> subscribe​(Pattern pattern)
        Subscribe to all topics matching specified pattern to get dynamically assigned partitions.

        Due to internal buffering of messages, when changing the subscribed topics the old set of topics may remain in effect (as observed by the handler(Handler) record handler}) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new set of topics.

        Parameters:
        pattern - Pattern to subscribe to
        Returns:
        a Future completed with the operation result
      • unsubscribe

        Future<Void> unsubscribe()
        Unsubscribe from topics currently subscribed with subscribe.
        Returns:
        current KafkaReadStream instance
      • subscription

        Future<Set<String>> subscription()
        Get the current subscription.
      • assign

        Future<Void> assign​(Set<org.apache.kafka.common.TopicPartition> partitions)
        Manually assign a set of partitions to this consumer.

        Due to internal buffering of messages, when reassigning the old set of partitions may remain in effect (as observed by the handler(Handler) record handler)} until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the batchHandler(Handler) will only see messages consistent with the new set of partitions.

        Parameters:
        partitions - partitions which want assigned
        Returns:
        a future notified with the result
      • assignment

        Future<Set<org.apache.kafka.common.TopicPartition>> assignment()
        Get the set of partitions currently assigned to this consumer.
      • listTopics

        Future<Map<String,​List<org.apache.kafka.common.PartitionInfo>>> listTopics()
        Get metadata about partitions for all topics that the user is authorized to view.
      • commit

        Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata>> commit()
        Commit current offsets for all the subscribed list of topics and partition.
      • commit

        Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata>> commit​(Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndMetadata> offsets)
        Commit the specified offsets for the specified list of topics and partitions to Kafka.
        Parameters:
        offsets - offsets list to commit
        Returns:
        a Future completed with the operation result
      • partitionsFor

        Future<List<org.apache.kafka.common.PartitionInfo>> partitionsFor​(String topic)
        Get metadata about the partitions for a given topic.
        Parameters:
        topic - topic partition for which getting partitions info
        Returns:
        a future notified on operation completed
      • close

        Future<Void> close()
        Close the stream
        Returns:
        a Future completed with the operation result
      • position

        Future<Long> position​(org.apache.kafka.common.TopicPartition partition)
        Get the offset of the next record that will be fetched (if a record with that offset exists).
        Parameters:
        partition - The partition to get the position for
        Returns:
        a future notified on operation completed
      • offsetsForTimes

        Future<Map<org.apache.kafka.common.TopicPartition,​org.apache.kafka.clients.consumer.OffsetAndTimestamp>> offsetsForTimes​(Map<org.apache.kafka.common.TopicPartition,​Long> topicPartitionTimestamps)
        Look up the offsets for the given partitions by timestamp.
        Parameters:
        topicPartitionTimestamps - A map with pairs of (TopicPartition, Timestamp).
        Returns:
        a future notified on operation completed
      • offsetsForTimes

        Future<org.apache.kafka.clients.consumer.OffsetAndTimestamp> offsetsForTimes​(org.apache.kafka.common.TopicPartition topicPartition,
                                                                                     long timestamp)
        * Look up the offset for the given partition by timestamp.
        Parameters:
        topicPartition - Partition to query.
        timestamp - Timestamp used to determine the offset.
        Returns:
        a future notified on operation completed
      • beginningOffsets

        Future<Map<org.apache.kafka.common.TopicPartition,​Long>> beginningOffsets​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Get the first offset for the given partitions.
        Parameters:
        topicPartitions - the partitions to get the earliest offsets.
        Returns:
        a future notified on operation completed
      • beginningOffsets

        Future<Long> beginningOffsets​(org.apache.kafka.common.TopicPartition topicPartition)
        Get the first offset for the given partition.
        Parameters:
        topicPartition - the partition to get the earliest offset.
        Returns:
        a future notified on operation completed
      • endOffsets

        Future<Map<org.apache.kafka.common.TopicPartition,​Long>> endOffsets​(Set<org.apache.kafka.common.TopicPartition> topicPartitions)
        Get the last offset for the given partitions. The last offset of a partition is the offset of the upcoming message, i.e. the offset of the last available message + 1.
        Parameters:
        topicPartitions - the partitions to get the end offsets.
        Returns:
        a future notified on operation completed
      • endOffsets

        Future<Long> endOffsets​(org.apache.kafka.common.TopicPartition topicPartition)
        Get the last offset for the given partition. The last offset of a partition is the offset of the upcoming message, i.e. the offset of the last available message + 1.
        Parameters:
        topicPartition - the partition to get the end offset.
        Returns:
        a future notified on operation completed
      • unwrap

        org.apache.kafka.clients.consumer.Consumer<K,​V> unwrap()
        Returns:
        the underlying consumer
      • batchHandler

        KafkaReadStream<K,​V> batchHandler​(Handler<org.apache.kafka.clients.consumer.ConsumerRecords<K,​V>> handler)
        Set the handler that will be called when a new batch of records is returned from Kafka. Batch handlers need to take care not to block the event loop when dealing with large batches. It is better to process records individually using the record handler.
        Parameters:
        handler - handler called each time Kafka returns a batch of records.
        Returns:
        current KafkaReadStream instance.
      • pollTimeout

        KafkaReadStream<K,​V> pollTimeout​(java.time.Duration timeout)
        Sets the poll timeout for the underlying native Kafka Consumer. Defaults to 1000 ms. Setting timeout to a lower value results in a more 'responsive' client, because it will block for a shorter period if no data is available in the assigned partition and therefore allows subsequent actions to be executed with a shorter delay. At the same time, the client will poll more frequently and thus will potentially create a higher load on the Kafka Broker.
        Parameters:
        timeout - The time, spent waiting in poll if data is not available in the buffer. If 0, returns immediately with any records that are available currently in the native Kafka consumer's buffer, else returns empty. Must not be negative.
      • poll

        Future<org.apache.kafka.clients.consumer.ConsumerRecords<K,​V>> poll​(java.time.Duration timeout)
        Executes a poll for getting messages from Kafka.
        Parameters:
        timeout - The maximum time to block (must not be greater than Long.MAX_VALUE milliseconds)
        Returns:
        a future notified on operation completed