Skip to content
This repository has been archived by the owner on Sep 13, 2024. It is now read-only.

Commit

Permalink
[FIX] checkstyle errors
Browse files Browse the repository at this point in the history
  • Loading branch information
gaurav7261 committed May 6, 2024
1 parent 7acc1f0 commit 9f869be
Show file tree
Hide file tree
Showing 4 changed files with 20 additions and 7 deletions.
5 changes: 5 additions & 0 deletions src/main/java/io/debezium/connector/jdbc/Buffer.java
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
/*
* Copyright Debezium Authors.
*
* Licensed under the Apache Software License version 2.0, available at http://www.apache.org/licenses/LICENSE-2.0
*/
package io.debezium.connector.jdbc;

import java.util.List;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -130,7 +130,8 @@ public void execute(Collection<SinkRecord> records) {
Buffer tableIdBuffer;
if (config.isUseReductionBuffer()) {
tableIdBuffer = deleteBufferByTable.computeIfAbsent(tableId, k -> new ReducedRecordBuffer(config));
} else {
}
else {
tableIdBuffer = deleteBufferByTable.computeIfAbsent(tableId, k -> new RecordBuffer(config));
}

Expand All @@ -153,7 +154,8 @@ public void execute(Collection<SinkRecord> records) {
Buffer tableIdBuffer;
if (config.isUseReductionBuffer()) {
tableIdBuffer = updateBufferByTable.computeIfAbsent(tableId, k -> new ReducedRecordBuffer(config));
} else {
}
else {
tableIdBuffer = updateBufferByTable.computeIfAbsent(tableId, k -> new RecordBuffer(config));
}

Expand Down
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
/*
* Copyright Debezium Authors.
*
* Licensed under the Apache Software License version 2.0, available at http://www.apache.org/licenses/LICENSE-2.0
*/
package io.debezium.connector.jdbc;

import org.apache.kafka.connect.data.Schema;
Expand Down Expand Up @@ -46,7 +51,8 @@ public List<SinkRecordDescriptor> add(SinkRecordDescriptor recordDescriptor) {
Struct keyStruct = recordDescriptor.getKeyStruct(connectorConfig.getPrimaryKeyMode());
if (keyStruct != null) {
records.put(keyStruct, recordDescriptor);
} else {
}
else {
throw new ConnectException("No struct-based primary key defined for record key/value, reduction buffer require struct based primary key");
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ void setUp() {
@DisplayName("When 10 sink records arrives and buffer size is 5 then the buffer will be flushed 2 times")
void correctlyBuffer(SinkRecordFactory factory) {

JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key","primary.key.fields", "id"));
JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key", "primary.key.fields", "id"));

ReducedRecordBuffer reducedRecordBuffer = new ReducedRecordBuffer(config);

Expand All @@ -90,7 +90,7 @@ void correctlyBuffer(SinkRecordFactory factory) {
@DisplayName("When key schema changes then the buffer will be flushed")
void keySchemaChange(SinkRecordFactory factory) {

JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key","primary.key.fields", "id"));
JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key", "primary.key.fields", "id"));

ReducedRecordBuffer reducedRecordBuffer = new ReducedRecordBuffer(config);

Expand Down Expand Up @@ -135,7 +135,7 @@ void keySchemaChange(SinkRecordFactory factory) {
@DisplayName("When value schema changes then the buffer will be flushed")
void valueSchemaChange(SinkRecordFactory factory) {

JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key","primary.key.fields", "id"));
JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key", "primary.key.fields", "id"));

ReducedRecordBuffer reducedRecordBuffer = new ReducedRecordBuffer(config);

Expand Down Expand Up @@ -180,7 +180,7 @@ void valueSchemaChange(SinkRecordFactory factory) {
@DisplayName("When 10 sink records arrives and buffer size is 5 with every alternate duplicate sink record then the buffer will be flushed 1 time")
void correctlyBufferWithDuplicate(SinkRecordFactory factory) {

JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key","primary.key.fields", "id"));
JdbcSinkConnectorConfig config = new JdbcSinkConnectorConfig(Map.of("batch.size", "5", "primary.key.mode", "record_key", "primary.key.fields", "id"));

ReducedRecordBuffer reducedRecordBuffer = new ReducedRecordBuffer(config);

Expand Down

0 comments on commit 9f869be

Please sign in to comment.