[server] Validate block-cache config against system memory #4325
VeniceCI-StaticAnalysisAndUnitTests.yml
on: pull_request
Matrix: Clients / UT & CodeCov
Matrix: Controller / UT & CodeCov
Matrix: Integrations / UT & CodeCov
Matrix: Internal / UT & CodeCov
Matrix: Router / UT & CodeCov
Matrix: Server / UT & CodeCov
Matrix: StaticAnalysis
ValidateGradleWrapper
8s
StaticAnalysisAndUnitTestsCompletionCheck
0s
Annotations
41 errors
Internal / UT & CodeCov (8)
Process completed with exit code 1.
|
Internal / UT & CodeCov (17)
Process completed with exit code 1.
|
ClassSizeEstimatorTest.testClassOverhead[1](EMPIRICAL_MEASUREMENT):
internal/venice-common/src/test/java/com/linkedin/venice/memory/ClassSizeEstimatorTest.java#L36
java.lang.AssertionError: Class ClassWithOnePublicPrimitiveIntField has a memoryAllocatedPerInstance (13.37644) which is too far from the predictedUsage (16); delta: 0.164; maxAllowedDelta: 0.075. No more attempts left.
|
ClassSizeEstimatorTest.testClassOverhead[1](EMPIRICAL_MEASUREMENT):
internal/venice-common/src/test/java/com/linkedin/venice/memory/ClassSizeEstimatorTest.java#L36
java.lang.AssertionError: Class ClassWithOnePublicPrimitiveIntField has a memoryAllocatedPerInstance (13.37644) which is too far from the predictedUsage (16); delta: 0.164; maxAllowedDelta: 0.075. No more attempts left.
|
Internal / UT & CodeCov (11)
Process completed with exit code 1.
|
Server / UT & CodeCov (8)
Process completed with exit code 1.
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testMetadataColumnFamily:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[0](true, false, false, true):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[1](true, false, false, false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[2](true, true, true, false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[3](true, true, false, false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
SITWithTWiseWithoutBufferAfterLeaderTest.testOffsetSyncBeforeGracefulShutDown[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
java.lang.AssertionError: pcs.getLatestProcessedLocalVersionTopicOffset() for PARTITION_FOO is expected to be zero! expected [0] but found [2]
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testMetadataColumnFamily:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[0](true, false, false, true):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[1](true, false, false, false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
ReplicationMeadataRocksDBStoragePartitionCFTest.testReplicationMetadataIngestion[2](true, true, true, false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/rocksdb/ReplicationMeadataRocksDBStoragePartitionCFTest.java#L1
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
Server / UT & CodeCov (17)
Process completed with exit code 1.
|
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_6b84cafb8d_976d602d_v1",
1,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$35(StoreIngestionTaskTest.java:1619)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_6b84cafb8d_976d602d_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_6b84cafb8d_976d602d_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_6b84cafb8d_976d602d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@18bcb4ba
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.computeStoreVersionState(
"TestTopic_6b84cafb8d_976d602d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@34265fb6
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.put(
"TestTopic_6b84cafb8d_976d602d_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_6b84cafb8d_976d602d_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_6b84cafb8d_976d602d_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_6b84cafb8d_976d602d_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_6b84cafb8d_976d602d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@24ae8c69
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.computeStoreVersionState(
"TestTopic_6b84cafb8d_976d602d_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@5964fb4f
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
|
SITWithPWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
Wanted 2 times:
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499)
But was 1 time:
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629)
|
SITWithPWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499)
However, there were exactly 27 interactions with this mock:
hostLevelIngestionStats.recordProcessConsumerActionLatency(
4.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.01068d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.00529d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.127086d,
1738973161451L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.055062d,
1738973161453L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordTotalRecordsConsumed();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2621)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.004719d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.003937d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.038733000000000004d,
1738973161463L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordTotalRecordsConsumed();
-> at com.linkedin.davinci.kafka.con
|
SITWithPWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[1](true):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
Wanted 3 times:
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499)
But was 2 times:
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629)
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629)
|
SITWithTWiseAndBufferAfterLeaderTest.testOffsetSyncBeforeGracefulShutDown[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
java.lang.AssertionError: pcs.getLatestProcessedLocalVersionTopicOffset() for PARTITION_FOO is expected to be zero! expected [0] but found [2]
|
SITWithTWiseWithoutBufferAfterLeaderTest.testNotifier[1](AA_OFF):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted:
storageMetadataService.put(
"TestTopic_676455648d_bdd8b2d9_v1",
2,
OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$35(StoreIngestionTaskTest.java:1621)
Actual invocations have different arguments:
storageMetadataService.getLastOffset(
"TestTopic_676455648d_bdd8b2d9_v1",
1
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.getLastOffset(
"TestTopic_676455648d_bdd8b2d9_v1",
2
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49)
storageMetadataService.computeStoreVersionState(
"TestTopic_676455648d_bdd8b2d9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@697e5b42
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.computeStoreVersionState(
"TestTopic_676455648d_bdd8b2d9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6d5caaec
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.put(
"TestTopic_676455648d_bdd8b2d9_v1",
1,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150676, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_676455648d_bdd8b2d9_v1",
1,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150677, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_676455648d_bdd8b2d9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6db16be2
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataService.put(
"TestTopic_676455648d_bdd8b2d9_v1",
2,
OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150676, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.put(
"TestTopic_676455648d_bdd8b2d9_v1",
2,
OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150677, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}
);
-> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38)
storageMetadataService.computeStoreVersionState(
"TestTopic_676455648d_bdd8b2d9_v1",
com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6b7b4d04
);
-> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
storageMetadataS
|
SITWithTWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked:
leaderFollowerStoreIngestionTask.reportError(
<any string>,
1,
<Capturing argument>
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4353)
However, there were exactly 50 interactions with this mock:
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_682c8d928e_f9d0a3aa_v1-1
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:887)
leaderFollowerStoreIngestionTask.subscribePartition(
TestTopic_682c8d928e_f9d0a3aa_v1-1,
true
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:644)
leaderFollowerStoreIngestionTask.throwIfNotRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:658)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:605)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:4186)
leaderFollowerStoreIngestionTask.nextSeqNum();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:667)
leaderFollowerStoreIngestionTask.run();
-> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
leaderFollowerStoreIngestionTask.isRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1657)
leaderFollowerStoreIngestionTask.getIsRunning();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:4186)
leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged(
Mock for Store, hashCode: 312791945
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1659)
leaderFollowerStoreIngestionTask.isHybridMode();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1575)
leaderFollowerStoreIngestionTask.processConsumerActions(
Mock for Store, hashCode: 312791945
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1660)
leaderFollowerStoreIngestionTask.processConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_682c8d928e_f9d0a3aa_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1738973154070},
Mock for Store, hashCode: 312791945
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1949)
leaderFollowerStoreIngestionTask.processCommonConsumerAction(
KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_682c8d928e_f9d0a3aa_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1738973154070}
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:572)
leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset(
PCS{replicaId=TestTopic_682c8d928e_f9d0a3aa_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY}
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2199)
leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower(
PCS{replicaId=TestTopic_682c8d928e_f9d0a3aa_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-
|
SITWithTWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false):
clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked:
hostLevelIngestionStats.recordTotalBytesConsumed(
<any long>
);
-> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499)
However, there were exactly 27 interactions with this mock:
hostLevelIngestionStats.recordProcessConsumerActionLatency(
13.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.01098d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.004919d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.11496500000000001d,
1738973160969L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.069609d,
1738973160971L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.004588d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordTotalRecordsConsumed();
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2621)
hostLevelIngestionStats.recordProcessConsumerActionLatency(
0.0d
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995)
hostLevelIngestionStats.recordCheckLongRunningTasksLatency(
0.004428d
);
-> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635)
hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency(
0.047809000000000004d,
1738973160987L
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302)
hostLevelIngestionStats.recordStorageQuotaUsed(
NaNd
);
-> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307)
hostLevelIngestionStats.recordTotalRecordsConsumed();
-> at com.linkedin.davi
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
InMemoryStorageEngineTest.setUp:
clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
|
Server / UT & CodeCov (11)
Process completed with exit code 1.
|
DispatchingAvroGenericStoreClientTest.testBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/DispatchingAvroGenericStoreClientTest.java#L664
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.DispatchingAvroGenericStoreClientTest.testBatchGet() didn't finish within the time-out 10000
|
DispatchingAvroGenericStoreClientTest.testBatchGet:
clients/venice-client/src/test/java/com/linkedin/venice/fastclient/DispatchingAvroGenericStoreClientTest.java#L664
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.DispatchingAvroGenericStoreClientTest.testBatchGet() didn't finish within the time-out 10000
|
StaticAnalysisAndUnitTestsCompletionCheck
Process completed with exit code 1.
|
Artifacts
Produced during runtime
Name | Size | |
---|---|---|
StaticAnalysis
|
666 KB |
|
clients-jdk11
|
2.57 MB |
|
clients-jdk17
|
2.59 MB |
|
clients-jdk8
|
2.52 MB |
|
controller-jdk11
|
1.62 MB |
|
controller-jdk17
|
1.61 MB |
|
controller-jdk8
|
1.61 MB |
|
integrations-jdk11
|
521 KB |
|
integrations-jdk17
|
528 KB |
|
integrations-jdk8
|
507 KB |
|
internal-jdk11
|
3.49 MB |
|
internal-jdk17
|
3.49 MB |
|
internal-jdk8
|
3.47 MB |
|
router-jdk11
|
1.04 MB |
|
router-jdk17
|
1.03 MB |
|
router-jdk8
|
1.02 MB |
|
server-jdk11
|
7.68 MB |
|
server-jdk17
|
7.47 MB |
|
server-jdk8
|
7.44 MB |
|