This repository contains a pure java-based implementation for a Couchbase
DCP (Database Change Protocol) client. It is currently a work in progress
and intended to be the primary DCP client going forward, eventually
deprecating the experimental supported inside core-io
.
It supports:
- Low Overhead Streaming
- Async from top to bottom
- Stream specific vbuckets
- Manual Start/Stop of streams
- Noop Acknowledgements
- Flow Control
- Session Management for restartability
- Rebalance Support
- Export current session state for durability (ships with JSON)
- Start and Restart from specific session state (import durable state)
- Pausing and Restarts
- Stream up to a specific point in time for a vbucket and then stop
- Proper shutdown/disconnect and cleanup
We publish the releases (including pre-releases to maven central):
<dependency>
<groupId>com.couchbase.client</groupId>
<artifactId>dcp-client</artifactId>
<version>0.7.0</version>
</dependency>
If you want the bleeding edge, you can check out the project from github and build it on your own. It is a maven-based project so simply run
$ git clone https://github.com/couchbaselabs/java-dcp-client.git
$ cd java-dcp-client
$ mvn install
Right now it will install the com.couchbase.client:dcp-client
artifact
with the 0.8.0-SNAPSHOT
version. You can then depend on it in your
project.
The simplest way is to initiate a stream against localhost
and open
all streams available. You always need to attach a callback for both the
config and the data events - in the simplest case all the messages are
just discarded. It's important to release the buffers!
Please check out the examples!
The following example connects to the travel-sample
bucket and prints
out all subsequent mutations and deletions that occur.
// Connect to localhost and use the travel-sample bucket
final Client client = Client.configure()
.hostnames("localhost")
.bucket("travel-sample")
.build();
// Don't do anything with control events in this example
client.controlEventHandler(new ControlEventHandler() {
@Override
public void onEvent(ByteBuf event) {
event.release();
}
});
// Print out Mutations and Deletions
client.dataEventHandler(new DataEventHandler() {
@Override
public void onEvent(ByteBuf event) {
if (DcpMutationMessage.is(event)) {
System.out.println("Mutation: " + DcpMutationMessage.toString(event));
// You can print the content via DcpMutationMessage.content(event).toString(CharsetUtil.UTF_8);
} else if (DcpDeletionMessage.is(event)) {
System.out.println("Deletion: " + DcpDeletionMessage.toString(event));
}
event.release();
}
});
// Connect the sockets
client.connect().await();
// Initialize the state (start now, never stop)
client.initializeState(StreamFrom.NOW, StreamTo.INFINITY).await();
// Start streaming on all partitions
client.startStreaming().await();
// Sleep for some time to print the mutations
// The printing happens on the IO threads!
Thread.sleep(TimeUnit.MINUTES.toMillis(10));
// Once the time is over, shutdown.
client.disconnect().await();
To save allocations the actual data you are interacting with are raw
netty ByteBuf
s that may be pooled, depending on the configuration. So
it is always important to release()
them when not needed anymore.
Since working with the raw buffers is not fun, the client provides flyweights that allow you to extract the information out of the buffers easily. Consult the docs for information on which message types to expect when, but as an example if you want to print the key and content of an incoming mutation in the data handler you can do it like this:
if (DcpMutationMessage.is(event)) {
String key = DcpMutationMessage.key(event).toString(CharsetUtil.UTF_8);
String content = DcpMutationMessage.content(event).toString(CharsetUtil.UTF_8);
System.out.println("Found Key " + key + " with Content " + content);
}
To handle slow clients better and to make it possible that the client signals backpressure to the server (that it should stop sending new data when the client is busy processing the previous ones) flow control tuneables are available.
Handling flow control consist of two stages: first, you need to enable it during bootstrap and then acknowledge specific message types as soon as you are done processing them.
To activate flow control, the DcpControl.Names.CONNECTION_BUFFER_SIZE
control param needs to be set to a value greater than zero. A reasonable
start value to test would be "10240" (10K).
Next, you also need to set the bufferAckWatermark
to a value which is
equal or smaller than the connection buffer size. Every time a message
is acknowledged the client accumulates up to the watermark and only if
the watermark is exceeded the acknowledgement is sent. This helps with
cutting down on network traffic and to reduce the workload on the server
side for accounting.
If you do not acknowledge the bytes read for specific messages, the server
will stop streaming new messages when the CONNECTION_BUFFER_SIZE
is
reached.
The following messages need to be acknowledged by the user:
DcpSnapshotMarkerRequest
(on theControlEventHandler
)DcpMutationMessage
(on theDataEventHandler
)DcpDeletionMessage
(on theDataEventHandler
)DcpExpirationMessage
(on theDataEventHandler
)
Acknowledging works by passing the number of bytes from the event to the
Client#acknowledgeBytes
method. Note that the vbucket id also needs to
be passed since the client needs to know against which connection the
acknowledge message should be performed.
A simple way to do this is the following:
client.acknowledgeBuffer(event);
This method extracts the vbucket ID and gets the number of readable bytes out of it. When you already did consume the bytes and the reader index of the buffer is not the number of bytes orginally, you can fall back to the lower level API:
client.acknowledgeBuffer(vbid, numBytes);
Read in details about SSL in Couchbase on our documentation. Here we will just post quick start steps:
-
Download and store in file cluster certificate from "Security" -> "Root Certificate" section on Admin Console.
-
Import this certificate using keytool:
keytool -importcert -keystore /tmp/keystore \ -storepass secret \ -file /tmp/cluster.cert
-
And update configuration of the DCP client:
final Client client = Client.configure() .hostnames("localhost") .bucket("travel-sample") .sslEnabled(true) .sslKeystoreFile("/tmp/keystore") .sslKeystorePassword("secret") .build();
Since 0.7.0 release, the client implements notification service, which allows to react on events, which are not tied directly to protocol and data transmission, such as connection errors, or notification about stream completion when the end sequence number wasn't set to infinity. The following example subscribes handler to system events to find out when partition 42 has done with data transmission:
client.systemEventHandler(new SystemEventHandler() {
@Override
public void onEvent(CouchbaseEvent event) {
if (event instanceof StreamEndEvent) {
StreamEndEvent streamEnd = (StreamEndEvent) event;
if (streamEnd.partition() == 42) {
System.out.println("Stream for partition 42 has ended (reason: " + streamEnd.reason() + ")");
}
}
}
});