31
31
import com .linkedin .davinci .client .DaVinciRecordTransformerConfig ;
32
32
import com .linkedin .davinci .client .factory .CachingDaVinciClientFactory ;
33
33
import com .linkedin .venice .D2 .D2ClientUtils ;
34
+ import com .linkedin .venice .client .store .ClientConfig ;
34
35
import com .linkedin .venice .compression .CompressionStrategy ;
35
36
import com .linkedin .venice .controllerapi .ControllerClient ;
36
37
import com .linkedin .venice .controllerapi .SchemaResponse ;
37
38
import com .linkedin .venice .controllerapi .UpdateStoreQueryParams ;
38
39
import com .linkedin .venice .duckdb .DuckDBDaVinciRecordTransformer ;
39
40
import com .linkedin .venice .integration .utils .ServiceFactory ;
41
+ import com .linkedin .venice .integration .utils .VeniceClusterCreateOptions ;
40
42
import com .linkedin .venice .integration .utils .VeniceClusterWrapper ;
41
43
import com .linkedin .venice .integration .utils .VeniceRouterWrapper ;
44
+ import com .linkedin .venice .producer .online .OnlineProducerFactory ;
45
+ import com .linkedin .venice .producer .online .OnlineVeniceProducer ;
42
46
import com .linkedin .venice .utils .PropertyBuilder ;
43
47
import com .linkedin .venice .utils .PushInputSchemaBuilder ;
44
48
import com .linkedin .venice .utils .TestUtils ;
@@ -91,20 +95,24 @@ public void setUp() {
91
95
clusterConfig .put (SERVER_PROMOTION_TO_LEADER_REPLICA_DELAY_SECONDS , 1L );
92
96
clusterConfig .put (PUSH_STATUS_STORE_ENABLED , true );
93
97
clusterConfig .put (DAVINCI_PUSH_STATUS_SCAN_INTERVAL_IN_SECONDS , 3 );
94
- cluster = ServiceFactory .getVeniceCluster (1 , 2 , 1 , 2 , 100 , false , false , clusterConfig );
95
- d2Client = new D2ClientBuilder ().setZkHosts (cluster .getZk ().getAddress ())
98
+ this .cluster = ServiceFactory .getVeniceCluster (
99
+ new VeniceClusterCreateOptions .Builder ().numberOfControllers (1 )
100
+ .numberOfServers (2 )
101
+ .numberOfRouters (1 )
102
+ .replicationFactor (2 )
103
+ .extraProperties (clusterConfig )
104
+ .build ());
105
+ this .d2Client = new D2ClientBuilder ().setZkHosts (this .cluster .getZk ().getAddress ())
96
106
.setZkSessionTimeout (3 , TimeUnit .SECONDS )
97
107
.setZkStartupTimeout (3 , TimeUnit .SECONDS )
98
108
.build ();
99
- D2ClientUtils .startClient (d2Client );
109
+ D2ClientUtils .startClient (this . d2Client );
100
110
}
101
111
102
112
@ AfterClass
103
113
public void cleanUp () {
104
- if (d2Client != null ) {
105
- D2ClientUtils .shutdownClient (d2Client );
106
- }
107
- Utils .closeQuietlyWithErrorLogged (cluster );
114
+ D2ClientUtils .shutdownClient (this .d2Client );
115
+ Utils .closeQuietlyWithErrorLogged (this .cluster );
108
116
}
109
117
110
118
@ BeforeMethod
@@ -158,24 +166,40 @@ public void testRecordTransformer() throws Exception {
158
166
159
167
clientWithRecordTransformer .subscribeAll ().get ();
160
168
161
- assertRowCount (duckDBUrl , storeName , "subscribeAll() finishes!" );
169
+ assertRowCount (duckDBUrl , storeName , DEFAULT_USER_DATA_RECORD_COUNT , "subscribeAll() finishes!" );
170
+
171
+ try (OnlineVeniceProducer producer = OnlineProducerFactory .createProducer (
172
+ ClientConfig .defaultGenericClientConfig (storeName )
173
+ .setD2Client (d2Client )
174
+ .setD2ServiceName (VeniceRouterWrapper .CLUSTER_DISCOVERY_D2_SERVICE_NAME ),
175
+ VeniceProperties .empty (),
176
+ null )) {
177
+ producer .asyncDelete (getKey (1 )).get ();
178
+ }
179
+
180
+ TestUtils .waitForNonDeterministicAssertion (
181
+ 10 ,
182
+ TimeUnit .SECONDS ,
183
+ true ,
184
+ () -> assertRowCount (duckDBUrl , storeName , DEFAULT_USER_DATA_RECORD_COUNT - 1 , "deleting 1 row!" ));
162
185
163
186
clientWithRecordTransformer .unsubscribeAll ();
164
187
}
165
188
166
- assertRowCount (duckDBUrl , storeName , "DVC gets closed!" );
189
+ assertRowCount (duckDBUrl , storeName , DEFAULT_USER_DATA_RECORD_COUNT - 1 , "DVC gets closed!" );
167
190
}
168
191
169
- private void assertRowCount (String duckDBUrl , String storeName , String assertionErrorMsg ) throws SQLException {
192
+ private void assertRowCount (String duckDBUrl , String storeName , int expectedCount , String assertionErrorMsg )
193
+ throws SQLException {
170
194
try (Connection connection = DriverManager .getConnection (duckDBUrl );
171
195
Statement statement = connection .createStatement ();
172
196
ResultSet rs = statement .executeQuery ("SELECT count(*) FROM " + storeName )) {
173
197
assertTrue (rs .next ());
174
198
int rowCount = rs .getInt (1 );
175
199
assertEquals (
176
200
rowCount ,
177
- DEFAULT_USER_DATA_RECORD_COUNT ,
178
- "The DB should contain " + DEFAULT_USER_DATA_RECORD_COUNT + " right after " + assertionErrorMsg );
201
+ expectedCount ,
202
+ "The DB should contain " + expectedCount + " rows right after " + assertionErrorMsg );
179
203
}
180
204
}
181
205
@@ -195,8 +219,7 @@ protected void setUpStore(
195
219
String lastName = "last_name_" ;
196
220
Schema valueSchema = writeSimpleAvroFile (inputDir , pushRecordSchema , i -> {
197
221
GenericRecord keyValueRecord = new GenericData .Record (pushRecordSchema );
198
- GenericRecord key = new GenericData .Record (SINGLE_FIELD_RECORD_SCHEMA );
199
- key .put ("key" , i .toString ());
222
+ GenericRecord key = getKey (i );
200
223
keyValueRecord .put (DEFAULT_KEY_FIELD_PROP , key );
201
224
GenericRecord valueRecord = new GenericData .Record (NAME_RECORD_V1_SCHEMA );
202
225
valueRecord .put ("firstName" , firstName + i );
@@ -214,7 +237,9 @@ protected void setUpStore(
214
237
final int numPartitions = 3 ;
215
238
UpdateStoreQueryParams params = new UpdateStoreQueryParams ().setPartitionCount (numPartitions )
216
239
.setChunkingEnabled (chunkingEnabled )
217
- .setCompressionStrategy (compressionStrategy );
240
+ .setCompressionStrategy (compressionStrategy )
241
+ .setHybridOffsetLagThreshold (10 )
242
+ .setHybridRewindSeconds (1 );
218
243
219
244
paramsConsumer .accept (params );
220
245
@@ -231,6 +256,12 @@ protected void setUpStore(
231
256
}
232
257
}
233
258
259
+ private GenericRecord getKey (Integer i ) {
260
+ GenericRecord key = new GenericData .Record (SINGLE_FIELD_RECORD_SCHEMA );
261
+ key .put ("key" , i .toString ());
262
+ return key ;
263
+ }
264
+
234
265
private static void runVPJ (Properties vpjProperties , int expectedVersionNumber , VeniceClusterWrapper cluster ) {
235
266
long vpjStart = System .currentTimeMillis ();
236
267
String jobName = Utils .getUniqueString ("batch-job-" + expectedVersionNumber );
0 commit comments