Packages

class KafkaConsumer[K, V] extends ReadStream[KafkaConsumerRecord[K, V]]

Vert.x Kafka consumer.

You receive Kafka records by providing a io.vertx.scala.kafka.client.consumer.KafkaConsumer#handler. As messages arrive the handler will be called with the records.

The io.vertx.scala.kafka.client.consumer.KafkaConsumer#pause and io.vertx.scala.kafka.client.consumer.KafkaConsumer#resume provides global control over reading the records from the consumer.

The io.vertx.scala.kafka.client.consumer.KafkaConsumer#pause and io.vertx.scala.kafka.client.consumer.KafkaConsumer#resume provides finer grained control over reading records for specific Topic/Partition, these are Kafka's specific operations.

Linear Supertypes
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. KafkaConsumer
  2. ReadStream
  3. StreamBase
  4. AnyRef
  5. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new KafkaConsumer(_asJava: AnyRef)(implicit arg0: scala.reflect.api.JavaUniverse.TypeTag[K], arg1: scala.reflect.api.JavaUniverse.TypeTag[V])

Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def asJava: AnyRef
    Definition Classes
    KafkaConsumerReadStreamStreamBase
  6. def assign(topicPartitions: Set[TopicPartition], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Manually assign a list of partition to this consumer.

    Manually assign a list of partition to this consumer.

    Due to internal buffering of messages, when reassigning the old set of partitions may remain in effect (as observed by the record handler) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new set of partitions.

    topicPartitions

    partitions which want assigned

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  7. def assign(topicPartition: TopicPartition, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Manually assign a partition to this consumer.

    Manually assign a partition to this consumer.

    Due to internal buffering of messages, when reassigning the old partition may remain in effect (as observed by the record handler) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new partition.

    topicPartition

    partition which want assignedsee TopicPartition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  8. def assign(topicPartitions: Set[TopicPartition]): KafkaConsumer[K, V]

    Manually assign a list of partition to this consumer.

    Manually assign a list of partition to this consumer.

    topicPartitions

    partitions which want assigned

    returns

    current KafkaConsumer instance

  9. def assign(topicPartition: TopicPartition): KafkaConsumer[K, V]

    Manually assign a partition to this consumer.

    Manually assign a partition to this consumer.

    topicPartition

    partition which want assignedsee TopicPartition

    returns

    current KafkaConsumer instance

  10. def assignFuture(topicPartitions: Set[TopicPartition]): Future[Unit]

    Like assign but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  11. def assignFuture(topicPartition: TopicPartition): Future[Unit]

    Like assign but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  12. def assignment(handler: Handler[AsyncResult[Set[TopicPartition]]]): KafkaConsumer[K, V]

    Get the set of partitions currently assigned to this consumer.

    Get the set of partitions currently assigned to this consumer.

    handler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  13. def assignmentFuture(): Future[Set[TopicPartition]]

    Like assignment but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  14. def batchHandler(handler: Handler[KafkaConsumerRecords[K, V]]): KafkaConsumer[K, V]

    Set the handler to be used when batches of messages are fetched from the Kafka server.

    Set the handler to be used when batches of messages are fetched from the Kafka server. Batch handlers need to take care not to block the event loop when dealing with large batches. It is better to process records individually using the io.vertx.scala.kafka.client.consumer.KafkaConsumer.

    handler

    handler called when batches of messages are fetched

    returns

    current KafkaConsumer instance

  15. def beginningOffsets(topicPartition: TopicPartition, handler: Handler[AsyncResult[Long]]): Unit

    Get the first offset for the given partitions.

    Get the first offset for the given partitions.

    topicPartition

    the partition to get the earliest offset.see TopicPartition

    handler

    handler called on operation completed. Returns the earliest available offset for the given partition

  16. def beginningOffsetsFuture(topicPartition: TopicPartition): Future[Long]

    Like beginningOffsets but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  17. def clone(): AnyRef
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )
  18. def close(completionHandler: Handler[AsyncResult[Unit]]): Unit

    Close the consumer

    Close the consumer

    completionHandler

    handler called on operation completed

  19. def close(): Unit

    Close the consumer

  20. def closeFuture(): Future[Unit]

    Like close but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  21. def commit(completionHandler: Handler[AsyncResult[Unit]]): Unit

    Commit current offsets for all the subscribed list of topics and partition.

    Commit current offsets for all the subscribed list of topics and partition.

    completionHandler

    handler called on operation completed

  22. def commit(): Unit

    Commit current offsets for all the subscribed list of topics and partition.

  23. def commitFuture(): Future[Unit]

    Like commit but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  24. def committed(topicPartition: TopicPartition, handler: Handler[AsyncResult[OffsetAndMetadata]]): Unit

    Get the last committed offset for the given partition (whether the commit happened by this process or another).

    Get the last committed offset for the given partition (whether the commit happened by this process or another).

    topicPartition

    topic partition for getting last committed offsetsee TopicPartition

    handler

    handler called on operation completed

  25. def committedFuture(topicPartition: TopicPartition): Future[OffsetAndMetadata]

    Like committed but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  26. def endHandler(endHandler: Handler[Unit]): KafkaConsumer[K, V]

    Set an end handler.

    Set an end handler. Once the stream has ended, and there is no more data to be read, this handler will be called.

    returns

    a reference to this, so the API can be used fluently

    Definition Classes
    KafkaConsumerReadStream
  27. def endOffsets(topicPartition: TopicPartition, handler: Handler[AsyncResult[Long]]): Unit

    Get the last offset for the given partition.

    Get the last offset for the given partition. The last offset of a partition is the offset of the upcoming message, i.e. the offset of the last available message + 1.

    topicPartition

    the partition to get the end offset.see TopicPartition

    handler

    handler called on operation completed. The end offset for the given partition.

  28. def endOffsetsFuture(topicPartition: TopicPartition): Future[Long]

    Like endOffsets but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  29. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  30. def equals(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  31. def exceptionHandler(handler: Handler[Throwable]): KafkaConsumer[K, V]

    Set an exception handler on the read stream.

    Set an exception handler on the read stream.

    handler

    the exception handler

    returns

    a reference to this, so the API can be used fluently

    Definition Classes
    KafkaConsumerReadStreamStreamBase
  32. def finalize(): Unit
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  33. final def getClass(): Class[_]
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  34. def handler(handler: Handler[KafkaConsumerRecord[K, V]]): KafkaConsumer[K, V]

    Set a data handler.

    Set a data handler. As data is read, the handler will be called with the data.

    returns

    a reference to this, so the API can be used fluently

    Definition Classes
    KafkaConsumerReadStream
  35. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @native()
  36. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  37. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  38. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  39. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @native()
  40. def offsetsForTimes(topicPartition: TopicPartition, timestamp: Long, handler: Handler[AsyncResult[OffsetAndTimestamp]]): Unit

    Look up the offset for the given partition by timestamp.

    Look up the offset for the given partition by timestamp. Note: the result might be null in case for the given timestamp no offset can be found -- e.g., when the timestamp refers to the future

    topicPartition

    TopicPartition to query.see TopicPartition

    timestamp

    Timestamp to be used in the query.

    handler

    handler called on operation completed

  41. def offsetsForTimesFuture(topicPartition: TopicPartition, timestamp: Long): Future[OffsetAndTimestamp]

    Like offsetsForTimes but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  42. def partitionsAssignedHandler(handler: Handler[Set[TopicPartition]]): KafkaConsumer[K, V]

    Set the handler called when topic partitions are assigned to the consumer

    Set the handler called when topic partitions are assigned to the consumer

    handler

    handler called on assigned topic partitions

    returns

    current KafkaConsumer instance

  43. def partitionsFor(topic: String, handler: Handler[AsyncResult[Buffer[PartitionInfo]]]): KafkaConsumer[K, V]

    Get metadata about the partitions for a given topic.

    Get metadata about the partitions for a given topic.

    topic

    topic partition for which getting partitions info

    handler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  44. def partitionsForFuture(topic: String): Future[Buffer[PartitionInfo]]

    Like partitionsFor but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  45. def partitionsRevokedHandler(handler: Handler[Set[TopicPartition]]): KafkaConsumer[K, V]

    Set the handler called when topic partitions are revoked to the consumer

    Set the handler called when topic partitions are revoked to the consumer

    handler

    handler called on revoked topic partitions

    returns

    current KafkaConsumer instance

  46. def pause(topicPartitions: Set[TopicPartition], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Suspend fetching from the requested partitions.

    Suspend fetching from the requested partitions.

    Due to internal buffering of messages, the will continue to observe messages from the given topicParations until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will not see messages from the given topicParations.

    topicPartitions

    topic partition from which suspend fetching

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  47. def pause(topicPartition: TopicPartition, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Suspend fetching from the requested partition.

    Suspend fetching from the requested partition.

    Due to internal buffering of messages, the will continue to observe messages from the given topicParation until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will not see messages from the given topicParation.

    topicPartition

    topic partition from which suspend fetchingsee TopicPartition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  48. def pause(topicPartitions: Set[TopicPartition]): KafkaConsumer[K, V]

    Suspend fetching from the requested partitions.

    Suspend fetching from the requested partitions.

    topicPartitions

    topic partition from which suspend fetching

    returns

    current KafkaConsumer instance

  49. def pause(topicPartition: TopicPartition): KafkaConsumer[K, V]

    Suspend fetching from the requested partition.

    Suspend fetching from the requested partition.

    topicPartition

    topic partition from which suspend fetchingsee TopicPartition

    returns

    current KafkaConsumer instance

  50. def pause(): KafkaConsumer[K, V]

    Pause the ReadSupport.

    Pause the ReadSupport. While it's paused, no data will be sent to the dataHandler

    returns

    a reference to this, so the API can be used fluently

    Definition Classes
    KafkaConsumerReadStream
  51. def pauseFuture(topicPartitions: Set[TopicPartition]): Future[Unit]

    Like pause but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  52. def pauseFuture(topicPartition: TopicPartition): Future[Unit]

    Like pause but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  53. def paused(handler: Handler[AsyncResult[Set[TopicPartition]]]): Unit

    Get the set of partitions that were previously paused by a call to pause(Set).

    Get the set of partitions that were previously paused by a call to pause(Set).

    handler

    handler called on operation completed

  54. def pausedFuture(): Future[Set[TopicPartition]]

    Like paused but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  55. def pollTimeout(timeout: Long): KafkaConsumer[K, V]

    Sets the poll timeout (in ms) for the underlying native Kafka Consumer.

    Sets the poll timeout (in ms) for the underlying native Kafka Consumer. Defaults to 1000. Setting timeout to a lower value results in a more 'responsive' client, because it will block for a shorter period if no data is available in the assigned partition and therefore allows subsequent actions to be executed with a shorter delay. At the same time, the client will poll more frequently and thus will potentially create a higher load on the Kafka Broker.

    timeout

    The time, in milliseconds, spent waiting in poll if data is not available in the buffer. If 0, returns immediately with any records that are available currently in the native Kafka consumer's buffer, else returns empty. Must not be negative.

  56. def position(partition: TopicPartition, handler: Handler[AsyncResult[Long]]): Unit

    Get the offset of the next record that will be fetched (if a record with that offset exists).

    Get the offset of the next record that will be fetched (if a record with that offset exists).

    partition

    The partition to get the position forsee TopicPartition

    handler

    handler called on operation completed

  57. def positionFuture(partition: TopicPartition): Future[Long]

    Like position but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  58. def resume(topicPartitions: Set[TopicPartition], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Resume specified partitions which have been paused with pause.

    Resume specified partitions which have been paused with pause.

    topicPartitions

    topic partition from which resume fetching

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  59. def resume(topicPartition: TopicPartition, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Resume specified partition which have been paused with pause.

    Resume specified partition which have been paused with pause.

    topicPartition

    topic partition from which resume fetchingsee TopicPartition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  60. def resume(topicPartitions: Set[TopicPartition]): KafkaConsumer[K, V]

    Resume specified partitions which have been paused with pause.

    Resume specified partitions which have been paused with pause.

    topicPartitions

    topic partition from which resume fetching

    returns

    current KafkaConsumer instance

  61. def resume(topicPartition: TopicPartition): KafkaConsumer[K, V]

    Resume specified partition which have been paused with pause.

    Resume specified partition which have been paused with pause.

    topicPartition

    topic partition from which resume fetchingsee TopicPartition

    returns

    current KafkaConsumer instance

  62. def resume(): KafkaConsumer[K, V]

    Resume reading.

    Resume reading. If the ReadSupport has been paused, reading will recommence on it.

    returns

    a reference to this, so the API can be used fluently

    Definition Classes
    KafkaConsumerReadStream
  63. def resumeFuture(topicPartitions: Set[TopicPartition]): Future[Unit]

    Like resume but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  64. def resumeFuture(topicPartition: TopicPartition): Future[Unit]

    Like resume but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  65. def seek(topicPartition: TopicPartition, offset: Long, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Overrides the fetch offsets that the consumer will use on the next poll.

    Overrides the fetch offsets that the consumer will use on the next poll.

    Due to internal buffering of messages, the will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new offset.

    topicPartition

    topic partition for which seeksee TopicPartition

    offset

    offset to seek inside the topic partition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  66. def seek(topicPartition: TopicPartition, offset: Long): KafkaConsumer[K, V]

    Overrides the fetch offsets that the consumer will use on the next poll.

    Overrides the fetch offsets that the consumer will use on the next poll.

    topicPartition

    topic partition for which seeksee TopicPartition

    offset

    offset to seek inside the topic partition

    returns

    current KafkaConsumer instance

  67. def seekFuture(topicPartition: TopicPartition, offset: Long): Future[Unit]

    Like seek but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  68. def seekToBeginning(topicPartitions: Set[TopicPartition], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Seek to the first offset for each of the given partitions.

    Seek to the first offset for each of the given partitions.

    Due to internal buffering of messages, the will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new offset.

    topicPartitions

    topic partition for which seek

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  69. def seekToBeginning(topicPartition: TopicPartition, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Seek to the first offset for each of the given partition.

    Seek to the first offset for each of the given partition.

    Due to internal buffering of messages, the will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new offset.

    topicPartition

    topic partition for which seeksee TopicPartition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  70. def seekToBeginning(topicPartitions: Set[TopicPartition]): KafkaConsumer[K, V]

    Seek to the first offset for each of the given partitions.

    Seek to the first offset for each of the given partitions.

    topicPartitions

    topic partition for which seek

    returns

    current KafkaConsumer instance

  71. def seekToBeginning(topicPartition: TopicPartition): KafkaConsumer[K, V]

    Seek to the first offset for each of the given partition.

    Seek to the first offset for each of the given partition.

    topicPartition

    topic partition for which seeksee TopicPartition

    returns

    current KafkaConsumer instance

  72. def seekToBeginningFuture(topicPartitions: Set[TopicPartition]): Future[Unit]

    Like seekToBeginning but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  73. def seekToBeginningFuture(topicPartition: TopicPartition): Future[Unit]

    Like seekToBeginning but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  74. def seekToEnd(topicPartitions: Set[TopicPartition], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Seek to the last offset for each of the given partitions.

    Seek to the last offset for each of the given partitions.

    Due to internal buffering of messages, the will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new offset.

    topicPartitions

    topic partition for which seek

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  75. def seekToEnd(topicPartition: TopicPartition, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Seek to the last offset for each of the given partition.

    Seek to the last offset for each of the given partition.

    Due to internal buffering of messages, the will continue to observe messages fetched with respect to the old offset until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new offset.

    topicPartition

    topic partition for which seeksee TopicPartition

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  76. def seekToEnd(topicPartitions: Set[TopicPartition]): KafkaConsumer[K, V]

    Seek to the last offset for each of the given partitions.

    Seek to the last offset for each of the given partitions.

    topicPartitions

    topic partition for which seek

    returns

    current KafkaConsumer instance

  77. def seekToEnd(topicPartition: TopicPartition): KafkaConsumer[K, V]

    Seek to the last offset for each of the given partition.

    Seek to the last offset for each of the given partition.

    topicPartition

    topic partition for which seeksee TopicPartition

    returns

    current KafkaConsumer instance

  78. def seekToEndFuture(topicPartitions: Set[TopicPartition]): Future[Unit]

    Like seekToEnd but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  79. def seekToEndFuture(topicPartition: TopicPartition): Future[Unit]

    Like seekToEnd but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  80. def subscribe(topics: Set[String], completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Subscribe to the given list of topics to get dynamically assigned partitions.

    Subscribe to the given list of topics to get dynamically assigned partitions.

    Due to internal buffering of messages, when changing the subscribed topics the old set of topics may remain in effect (as observed by the record handler) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new set of topics.

    topics

    topics to subscribe to

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  81. def subscribe(topic: String, completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Subscribe to the given topic to get dynamically assigned partitions.

    Subscribe to the given topic to get dynamically assigned partitions.

    Due to internal buffering of messages, when changing the subscribed topic the old topic may remain in effect (as observed by the record handler) until some time after the given completionHandler is called. In contrast, the once the given completionHandler is called the io.vertx.scala.kafka.client.consumer.KafkaConsumer#batchHandler will only see messages consistent with the new topic.

    topic

    topic to subscribe to

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  82. def subscribe(topics: Set[String]): KafkaConsumer[K, V]

    Subscribe to the given list of topics to get dynamically assigned partitions.

    Subscribe to the given list of topics to get dynamically assigned partitions.

    topics

    topics to subscribe to

    returns

    current KafkaConsumer instance

  83. def subscribe(topic: String): KafkaConsumer[K, V]

    Subscribe to the given topic to get dynamically assigned partitions.

    Subscribe to the given topic to get dynamically assigned partitions.

    topic

    topic to subscribe to

    returns

    current KafkaConsumer instance

  84. def subscribeFuture(topics: Set[String]): Future[Unit]

    Like subscribe but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  85. def subscribeFuture(topic: String): Future[Unit]

    Like subscribe but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  86. def subscription(handler: Handler[AsyncResult[Set[String]]]): KafkaConsumer[K, V]

    Get the current subscription.

    Get the current subscription.

    handler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  87. def subscriptionFuture(): Future[Set[String]]

    Like subscription but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  88. final def synchronized[T0](arg0: ⇒ T0): T0
    Definition Classes
    AnyRef
  89. def toString(): String
    Definition Classes
    AnyRef → Any
  90. def unsubscribe(completionHandler: Handler[AsyncResult[Unit]]): KafkaConsumer[K, V]

    Unsubscribe from topics currently subscribed with subscribe.

    Unsubscribe from topics currently subscribed with subscribe.

    completionHandler

    handler called on operation completed

    returns

    current KafkaConsumer instance

  91. def unsubscribe(): KafkaConsumer[K, V]

    Unsubscribe from topics currently subscribed with subscribe.

    Unsubscribe from topics currently subscribed with subscribe.

    returns

    current KafkaConsumer instance

  92. def unsubscribeFuture(): Future[Unit]

    Like unsubscribe but returns a scala.concurrent.Future instead of taking an AsyncResultHandler.

  93. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  94. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  95. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @native() @throws( ... )

Inherited from ReadStream[KafkaConsumerRecord[K, V]]

Inherited from StreamBase

Inherited from AnyRef

Inherited from Any

Ungrouped