Skip to content

[server] Validate block-cache config against system memory #4325

[server] Validate block-cache config against system memory

[server] Validate block-cache config against system memory #4325

Triggered via pull request February 8, 2025 00:00
Status Failure
Total duration 13m 10s
Artifacts 19
Matrix: Clients / UT & CodeCov
Matrix: Controller / UT & CodeCov
Matrix: Integrations / UT & CodeCov
Matrix: Internal / UT & CodeCov
Matrix: Router / UT & CodeCov
Matrix: Server / UT & CodeCov
Matrix: StaticAnalysis
ValidateGradleWrapper
8s
ValidateGradleWrapper
StaticAnalysisAndUnitTestsCompletionCheck
0s
StaticAnalysisAndUnitTestsCompletionCheck
Fit to window
Zoom out
Zoom in

Annotations

41 errors
Internal / UT & CodeCov (8)
Process completed with exit code 1.
Internal / UT & CodeCov (17)
Process completed with exit code 1.
ClassSizeEstimatorTest.testClassOverhead[1](EMPIRICAL_MEASUREMENT): internal/venice-common/src/test/java/com/linkedin/venice/memory/ClassSizeEstimatorTest.java#L36
java.lang.AssertionError: Class ClassWithOnePublicPrimitiveIntField has a memoryAllocatedPerInstance (13.37644) which is too far from the predictedUsage (16); delta: 0.164; maxAllowedDelta: 0.075. No more attempts left.
ClassSizeEstimatorTest.testClassOverhead[1](EMPIRICAL_MEASUREMENT): internal/venice-common/src/test/java/com/linkedin/venice/memory/ClassSizeEstimatorTest.java#L36
java.lang.AssertionError: Class ClassWithOnePublicPrimitiveIntField has a memoryAllocatedPerInstance (13.37644) which is too far from the predictedUsage (16); delta: 0.164; maxAllowedDelta: 0.075. No more attempts left.
Internal / UT & CodeCov (11)
Process completed with exit code 1.
Server / UT & CodeCov (8)
Process completed with exit code 1.
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
SITWithTWiseWithoutBufferAfterLeaderTest.testOffsetSyncBeforeGracefulShutDown[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
java.lang.AssertionError: pcs.getLatestProcessedLocalVersionTopicOffset() for PARTITION_FOO is expected to be zero! expected [0] but found [2]
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
Server / UT & CodeCov (17)
Process completed with exit code 1.
SITWithPWiseAndBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_6b84cafb8d_976d602d_v1", 1, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$35(StoreIngestionTaskTest.java:1619) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_6b84cafb8d_976d602d_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_6b84cafb8d_976d602d_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_6b84cafb8d_976d602d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@18bcb4ba ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.computeStoreVersionState( "TestTopic_6b84cafb8d_976d602d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@34265fb6 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.put( "TestTopic_6b84cafb8d_976d602d_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_6b84cafb8d_976d602d_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_6b84cafb8d_976d602d_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_6b84cafb8d_976d602d_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973168400, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_6b84cafb8d_976d602d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@24ae8c69 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.computeStoreVersionState( "TestTopic_6b84cafb8d_976d602d_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$959/0x0000000800811c40@5964fb4f ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34)
SITWithPWiseAndBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseAndBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); Wanted 2 times: -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499) But was 1 time: -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629)
SITWithPWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499) However, there were exactly 27 interactions with this mock: hostLevelIngestionStats.recordProcessConsumerActionLatency( 4.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.01068d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.00529d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.127086d, 1738973161451L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.055062d, 1738973161453L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordTotalRecordsConsumed(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2621) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.004719d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.003937d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.038733000000000004d, 1738973161463L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordTotalRecordsConsumed(); -> at com.linkedin.davinci.kafka.con
SITWithPWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[1](true): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithPWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.TooFewActualInvocations: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); Wanted 3 times: -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499) But was 2 times: -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629) -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2629)
SITWithTWiseAndBufferAfterLeaderTest.testOffsetSyncBeforeGracefulShutDown[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseAndBufferAfterLeaderTest.java#L1
java.lang.AssertionError: pcs.getLatestProcessedLocalVersionTopicOffset() for PARTITION_FOO is expected to be zero! expected [0] but found [2]
SITWithTWiseWithoutBufferAfterLeaderTest.testNotifier[1](AA_OFF): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
org.mockito.exceptions.verification.ArgumentsAreDifferent: Argument(s) are different! Wanted: storageMetadataService.put( "TestTopic_676455648d_bdd8b2d9_v1", 2, OffsetRecord{localVersionTopicOffset=3, upstreamOffset=-1, leaderTopic=null, offsetLag=0, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=true, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.lambda$testNotifier$35(StoreIngestionTaskTest.java:1621) Actual invocations have different arguments: storageMetadataService.getLastOffset( "TestTopic_676455648d_bdd8b2d9_v1", 1 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.getLastOffset( "TestTopic_676455648d_bdd8b2d9_v1", 2 ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.getLastOffset(DeepCopyOffsetManager.java:49) storageMetadataService.computeStoreVersionState( "TestTopic_676455648d_bdd8b2d9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@697e5b42 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.computeStoreVersionState( "TestTopic_676455648d_bdd8b2d9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6d5caaec ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.put( "TestTopic_676455648d_bdd8b2d9_v1", 1, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150676, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_676455648d_bdd8b2d9_v1", 1, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150677, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_676455648d_bdd8b2d9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6db16be2 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataService.put( "TestTopic_676455648d_bdd8b2d9_v1", 2, OffsetRecord{localVersionTopicOffset=1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150676, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.put( "TestTopic_676455648d_bdd8b2d9_v1", 2, OffsetRecord{localVersionTopicOffset=2, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=1738973150677, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null} ); -> at com.linkedin.venice.offsets.DeepCopyOffsetManager.put(DeepCopyOffsetManager.java:38) storageMetadataService.computeStoreVersionState( "TestTopic_676455648d_bdd8b2d9_v1", com.linkedin.venice.offsets.DeepCopyStorageMetadataService$$Lambda$1008/0x0000000800839440@6b7b4d04 ); -> at com.linkedin.venice.offsets.DeepCopyStorageMetadataService.computeStoreVersionState(DeepCopyStorageMetadataService.java:34) storageMetadataS
SITWithTWiseWithoutBufferAfterLeaderTest.testProcessConsumerActionsError: clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked: leaderFollowerStoreIngestionTask.reportError( <any string>, 1, <Capturing argument> ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.reportError(StoreIngestionTask.java:4353) However, there were exactly 50 interactions with this mock: leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_682c8d928e_f9d0a3aa_v1-1 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTaskTest.runTest(StoreIngestionTaskTest.java:887) leaderFollowerStoreIngestionTask.subscribePartition( TestTopic_682c8d928e_f9d0a3aa_v1-1, true ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:644) leaderFollowerStoreIngestionTask.throwIfNotRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:658) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.throwIfNotRunning(StoreIngestionTask.java:605) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:4186) leaderFollowerStoreIngestionTask.nextSeqNum(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.subscribePartition(StoreIngestionTask.java:667) leaderFollowerStoreIngestionTask.run(); -> at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) leaderFollowerStoreIngestionTask.isRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1657) leaderFollowerStoreIngestionTask.getIsRunning(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.isRunning(StoreIngestionTask.java:4186) leaderFollowerStoreIngestionTask.updateIngestionRoleIfStoreChanged( Mock for Store, hashCode: 312791945 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1659) leaderFollowerStoreIngestionTask.isHybridMode(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.updateIngestionRoleIfStoreChanged(StoreIngestionTask.java:1575) leaderFollowerStoreIngestionTask.processConsumerActions( Mock for Store, hashCode: 312791945 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.run(StoreIngestionTask.java:1660) leaderFollowerStoreIngestionTask.processConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_682c8d928e_f9d0a3aa_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1738973154070}, Mock for Store, hashCode: 312791945 ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1949) leaderFollowerStoreIngestionTask.processCommonConsumerAction( KafkaTaskMessage{type=SUBSCRIBE, topicPartition=TestTopic_682c8d928e_f9d0a3aa_v1-1, attempts=1, sequenceNumber=1, createdTimestampInMs=1738973154070} ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.processConsumerAction(LeaderFollowerStoreIngestionTask.java:572) leaderFollowerStoreIngestionTask.reportIfCatchUpVersionTopicOffset( PCS{replicaId=TestTopic_682c8d928e_f9d0a3aa_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-1, latestProcessedUpstreamVersionTopicOffset=-1, latestProcessedUpstreamRTOffsetMap={}, latestIgnoredUpstreamRTOffsetMap={}, latestRTOffsetTriedToProduceToVTMap{}, offsetRecord=OffsetRecord{localVersionTopicOffset=-1, upstreamOffset=-1, leaderTopic=null, offsetLag=9223372036854775807, eventTimeEpochMs=-1, latestProducerProcessingTimeInMs=0, isEndOfPushReceived=false, databaseInfo={}, realTimeProducerState={}, recordTransformerClassHash=null}, errorReported=false, started=false, lagCaughtUp=false, processedRecordSizeSinceLastSync=0, leaderFollowerState=STANDBY} ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processCommonConsumerAction(StoreIngestionTask.java:2199) leaderFollowerStoreIngestionTask.updateLeaderTopicOnFollower( PCS{replicaId=TestTopic_682c8d928e_f9d0a3aa_v1-1, hybrid=false, latestProcessedLocalVersionTopicOffset=-
SITWithTWiseWithoutBufferAfterLeaderTest.testRecordLevelMetricForCurrentVersion[0](false): clients/da-vinci-client/src/test/java/com/linkedin/davinci/kafka/consumer/SITWithTWiseWithoutBufferAfterLeaderTest.java#L1
Wanted but not invoked: hostLevelIngestionStats.recordTotalBytesConsumed( <any long> ); -> at com.linkedin.davinci.stats.HostLevelIngestionStats.recordTotalBytesConsumed(HostLevelIngestionStats.java:499) However, there were exactly 27 interactions with this mock: hostLevelIngestionStats.recordProcessConsumerActionLatency( 13.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.01098d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.004919d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.11496500000000001d, 1738973160969L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.069609d, 1738973160971L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.004588d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordTotalRecordsConsumed(); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerRecord(StoreIngestionTask.java:2621) hostLevelIngestionStats.recordProcessConsumerActionLatency( 0.0d ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.processConsumerActions(StoreIngestionTask.java:1995) hostLevelIngestionStats.recordCheckLongRunningTasksLatency( 0.004428d ); -> at com.linkedin.davinci.kafka.consumer.LeaderFollowerStoreIngestionTask.checkLongRunningTaskState(LeaderFollowerStoreIngestionTask.java:756) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.recordQuotaMetrics(StoreIngestionTask.java:1635) hostLevelIngestionStats.recordConsumerRecordsQueuePutLatency( 0.047809000000000004d, 1738973160987L ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1302) hostLevelIngestionStats.recordStorageQuotaUsed( NaNd ); -> at com.linkedin.davinci.kafka.consumer.StoreIngestionTask.produceToStoreBufferServiceOrKafka(StoreIngestionTask.java:1307) hostLevelIngestionStats.recordTotalRecordsConsumed(); -> at com.linkedin.davi
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
InMemoryStorageEngineTest.setUp: clients/da-vinci-client/src/test/java/com/linkedin/davinci/store/memory/InMemoryStorageEngineTest.java#L30
java.lang.RuntimeException: Cannot setup rocksdb instance with block-cache size 16.0 GiB. System memory : 15.6 GiB
Server / UT & CodeCov (11)
Process completed with exit code 1.
DispatchingAvroGenericStoreClientTest.testBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/DispatchingAvroGenericStoreClientTest.java#L664
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.DispatchingAvroGenericStoreClientTest.testBatchGet() didn't finish within the time-out 10000
DispatchingAvroGenericStoreClientTest.testBatchGet: clients/venice-client/src/test/java/com/linkedin/venice/fastclient/DispatchingAvroGenericStoreClientTest.java#L664
org.testng.internal.thread.ThreadTimeoutException: Method com.linkedin.venice.fastclient.DispatchingAvroGenericStoreClientTest.testBatchGet() didn't finish within the time-out 10000
StaticAnalysisAndUnitTestsCompletionCheck
Process completed with exit code 1.

Artifacts

Produced during runtime
Name Size
StaticAnalysis
666 KB
clients-jdk11
2.57 MB
clients-jdk17
2.59 MB
clients-jdk8
2.52 MB
controller-jdk11
1.62 MB
controller-jdk17
1.61 MB
controller-jdk8
1.61 MB
integrations-jdk11
521 KB
integrations-jdk17
528 KB
integrations-jdk8
507 KB
internal-jdk11
3.49 MB
internal-jdk17
3.49 MB
internal-jdk8
3.47 MB
router-jdk11
1.04 MB
router-jdk17
1.03 MB
router-jdk8
1.02 MB
server-jdk11
7.68 MB
server-jdk17
7.47 MB
server-jdk8
7.44 MB