INFO [main] 2025-08-11 05:03:06,384 YamlConfigurationLoader.java:89 - Configuration location: file:/etc/cassandra/cassandra.yaml INFO [main] 2025-08-11 05:03:06,624 Config.java:495 - Node configuration:[allocate_tokens_for_keyspace=null; authenticator=AllowAllAuthenticator; authorizer=AllowAllAuthorizer; auto_bootstrap=true; auto_snapshot=true; back_pressure_enabled=false; back_pressure_strategy=org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}; batch_size_fail_threshold_in_kb=50; batch_size_warn_threshold_in_kb=5; batchlog_replay_throttle_in_kb=1024; broadcast_address=10.0.0.38; broadcast_rpc_address=10.0.0.38; buffer_pool_use_heap_if_exhausted=true; cas_contention_timeout_in_ms=1000; cdc_enabled=false; cdc_free_space_check_interval_ms=250; cdc_raw_directory=null; cdc_total_space_in_mb=0; client_encryption_options=; cluster_name=contrail_database; column_index_cache_size_in_kb=2; column_index_size_in_kb=64; commit_failure_policy=stop; commitlog_compression=null; commitlog_directory=/var/lib/cassandra/commitlog; commitlog_max_compression_buffers_in_pool=3; commitlog_periodic_queue_size=-1; commitlog_segment_size_in_mb=32; commitlog_sync=periodic; commitlog_sync_batch_window_in_ms=NaN; commitlog_sync_period_in_ms=10000; commitlog_total_space_in_mb=null; compaction_large_partition_warning_threshold_mb=100; compaction_throughput_mb_per_sec=256; concurrent_compactors=4; concurrent_counter_writes=32; concurrent_materialized_view_writes=32; concurrent_reads=64; concurrent_replicates=null; concurrent_writes=64; counter_cache_keys_to_save=2147483647; counter_cache_save_period=7200; counter_cache_size_in_mb=null; counter_write_request_timeout_in_ms=5000; credentials_cache_max_entries=1000; credentials_update_interval_in_ms=-1; credentials_validity_in_ms=2000; cross_node_timeout=false; data_file_directories=[Ljava.lang.String;@6b19b79; disk_access_mode=auto; disk_failure_policy=stop; disk_optimization_estimate_percentile=0.95; disk_optimization_page_cross_chance=0.1; disk_optimization_strategy=ssd; dynamic_snitch=true; dynamic_snitch_badness_threshold=0.1; dynamic_snitch_reset_interval_in_ms=600000; dynamic_snitch_update_interval_in_ms=100; enable_materialized_views=true; enable_scripted_user_defined_functions=false; enable_user_defined_functions=false; enable_user_defined_functions_threads=true; encryption_options=null; endpoint_snitch=SimpleSnitch; file_cache_round_up=null; file_cache_size_in_mb=null; gc_log_threshold_in_ms=200; gc_warn_threshold_in_ms=1000; hinted_handoff_disabled_datacenters=[]; hinted_handoff_enabled=true; hinted_handoff_throttle_in_kb=1024; hints_compression=null; hints_directory=null; hints_flush_period_in_ms=10000; incremental_backups=false; index_interval=null; index_summary_capacity_in_mb=null; index_summary_resize_interval_in_minutes=60; initial_token=null; inter_dc_stream_throughput_outbound_megabits_per_sec=200; inter_dc_tcp_nodelay=false; internode_authenticator=null; internode_compression=dc; internode_recv_buff_size_in_bytes=0; internode_send_buff_size_in_bytes=0; key_cache_keys_to_save=2147483647; key_cache_save_period=14400; key_cache_size_in_mb=null; listen_address=10.0.0.38; listen_interface=null; listen_interface_prefer_ipv6=false; listen_on_broadcast_address=false; max_hint_window_in_ms=10800000; max_hints_delivery_threads=2; max_hints_file_size_in_mb=128; max_mutation_size_in_kb=null; max_streaming_retries=3; max_value_size_in_mb=256; memtable_allocation_type=offheap_objects; memtable_cleanup_threshold=null; memtable_flush_writers=4; memtable_heap_space_in_mb=null; memtable_offheap_space_in_mb=null; min_free_space_per_drive_in_mb=50; native_transport_max_concurrent_connections=-1; native_transport_max_concurrent_connections_per_ip=-1; native_transport_max_frame_size_in_mb=256; native_transport_max_threads=128; native_transport_port=9042; native_transport_port_ssl=null; num_tokens=256; otc_backlog_expiration_interval_ms=200; otc_coalescing_enough_coalesced_messages=8; otc_coalescing_strategy=DISABLED; otc_coalescing_window_us=200; partitioner=org.apache.cassandra.dht.Murmur3Partitioner; permissions_cache_max_entries=1000; permissions_update_interval_in_ms=-1; permissions_validity_in_ms=2000; phi_convict_threshold=8.0; prepared_statements_cache_size_mb=null; range_request_timeout_in_ms=10000; read_request_timeout_in_ms=5000; request_scheduler=org.apache.cassandra.scheduler.NoScheduler; request_scheduler_id=null; request_scheduler_options=null; request_timeout_in_ms=10000; role_manager=CassandraRoleManager; roles_cache_max_entries=1000; roles_update_interval_in_ms=-1; roles_validity_in_ms=2000; row_cache_class_name=org.apache.cassandra.cache.OHCProvider; row_cache_keys_to_save=2147483647; row_cache_save_period=0; row_cache_size_in_mb=0; rpc_address=10.0.0.38; rpc_interface=null; rpc_interface_prefer_ipv6=false; rpc_keepalive=true; rpc_listen_backlog=50; rpc_max_threads=2147483647; rpc_min_threads=16; rpc_port=9160; rpc_recv_buff_size_in_bytes=null; rpc_send_buff_size_in_bytes=null; rpc_server_type=sync; saved_caches_directory=/var/lib/cassandra/saved_caches; seed_provider=org.apache.cassandra.locator.SimpleSeedProvider{seeds=10.0.0.254,10.0.0.38}; server_encryption_options=; slow_query_log_timeout_in_ms=500; snapshot_before_compaction=false; ssl_storage_port=7001; sstable_preemptive_open_interval_in_mb=50; start_native_transport=true; start_rpc=true; storage_port=7000; stream_throughput_outbound_megabits_per_sec=200; streaming_keep_alive_period_in_secs=300; streaming_socket_timeout_in_ms=86400000; thrift_framed_transport_size_in_mb=15; thrift_max_message_length_in_mb=16; thrift_prepared_statements_cache_size_mb=null; tombstone_failure_threshold=100000; tombstone_warn_threshold=1000; tracetype_query_ttl=86400; tracetype_repair_ttl=604800; transparent_data_encryption_options=org.apache.cassandra.config.TransparentDataEncryptionOptions@2a32de6c; trickle_fsync=false; trickle_fsync_interval_in_kb=10240; truncate_request_timeout_in_ms=60000; unlogged_batch_across_partitions_warn_threshold=10; user_defined_function_fail_timeout=1500; user_defined_function_warn_timeout=500; user_function_timeout_policy=die; windows_timer_interval=1; write_request_timeout_in_ms=2000] INFO [main] 2025-08-11 05:03:06,625 DatabaseDescriptor.java:367 - DiskAccessMode 'auto' determined to be mmap, indexAccessMode is mmap INFO [main] 2025-08-11 05:03:06,625 DatabaseDescriptor.java:425 - Global memtable on-heap threshold is enabled at 502MB INFO [main] 2025-08-11 05:03:06,625 DatabaseDescriptor.java:429 - Global memtable off-heap threshold is enabled at 502MB INFO [main] 2025-08-11 05:03:06,652 RateBasedBackPressure.java:123 - Initialized back-pressure with high ratio: 0.9, factor: 5, flow: FAST, window size: 2000. INFO [main] 2025-08-11 05:03:06,652 DatabaseDescriptor.java:729 - Back-pressure is disabled with strategy org.apache.cassandra.net.RateBasedBackPressure{high_ratio=0.9, factor=5, flow=FAST}. INFO [main] 2025-08-11 05:03:06,891 JMXServerUtils.java:246 - Configured JMX server at: service:jmx:rmi://0.0.0.0/jndi/rmi://0.0.0.0:7201/jmxrmi INFO [main] 2025-08-11 05:03:06,908 CassandraDaemon.java:473 - Hostname: cn-jenkins-deploy-platform-ansible-os-3797-2. INFO [main] 2025-08-11 05:03:06,910 CassandraDaemon.java:480 - JVM vendor/version: OpenJDK 64-Bit Server VM/1.8.0_322 INFO [main] 2025-08-11 05:03:06,913 CassandraDaemon.java:481 - Heap size: 984.000MiB/1.961GiB INFO [main] 2025-08-11 05:03:06,917 CassandraDaemon.java:486 - Code Cache Non-heap memory: init = 2555904(2496K) used = 4295616(4194K) committed = 4325376(4224K) max = 251658240(245760K) INFO [main] 2025-08-11 05:03:06,918 CassandraDaemon.java:486 - Metaspace Non-heap memory: init = 0(0K) used = 19170104(18720K) committed = 19660800(19200K) max = -1(-1K) INFO [main] 2025-08-11 05:03:06,918 CassandraDaemon.java:486 - Compressed Class Space Non-heap memory: init = 0(0K) used = 2238512(2186K) committed = 2490368(2432K) max = 1073741824(1048576K) INFO [main] 2025-08-11 05:03:06,918 CassandraDaemon.java:486 - Par Eden Space Heap memory: init = 335544320(327680K) used = 93992952(91789K) committed = 335544320(327680K) max = 335544320(327680K) INFO [main] 2025-08-11 05:03:06,918 CassandraDaemon.java:486 - Par Survivor Space Heap memory: init = 41943040(40960K) used = 0(0K) committed = 41943040(40960K) max = 41943040(40960K) INFO [main] 2025-08-11 05:03:06,918 CassandraDaemon.java:486 - CMS Old Gen Heap memory: init = 654311424(638976K) used = 0(0K) committed = 654311424(638976K) max = 1728053248(1687552K) INFO [main] 2025-08-11 05:03:06,919 CassandraDaemon.java:488 - Classpath: /opt/cassandra/conf:/opt/cassandra/build/classes/main:/opt/cassandra/build/classes/thrift:/opt/cassandra/lib/airline-0.6.jar:/opt/cassandra/lib/antlr-runtime-3.5.2.jar:/opt/cassandra/lib/apache-cassandra-3.11.3.jar:/opt/cassandra/lib/apache-cassandra-thrift-3.11.3.jar:/opt/cassandra/lib/asm-5.0.4.jar:/opt/cassandra/lib/caffeine-2.2.6.jar:/opt/cassandra/lib/cassandra-driver-core-3.0.1-shaded.jar:/opt/cassandra/lib/commons-cli-1.1.jar:/opt/cassandra/lib/commons-codec-1.9.jar:/opt/cassandra/lib/commons-lang3-3.1.jar:/opt/cassandra/lib/commons-math3-3.2.jar:/opt/cassandra/lib/compress-lzf-0.8.4.jar:/opt/cassandra/lib/concurrentlinkedhashmap-lru-1.4.jar:/opt/cassandra/lib/concurrent-trees-2.4.0.jar:/opt/cassandra/lib/disruptor-3.0.1.jar:/opt/cassandra/lib/ecj-4.4.2.jar:/opt/cassandra/lib/guava-18.0.jar:/opt/cassandra/lib/HdrHistogram-2.1.9.jar:/opt/cassandra/lib/high-scale-lib-1.0.6.jar:/opt/cassandra/lib/hppc-0.5.4.jar:/opt/cassandra/lib/jackson-core-asl-1.9.13.jar:/opt/cassandra/lib/jackson-mapper-asl-1.9.13.jar:/opt/cassandra/lib/jamm-0.3.0.jar:/opt/cassandra/lib/javax.inject.jar:/opt/cassandra/lib/jbcrypt-0.3m.jar:/opt/cassandra/lib/jcl-over-slf4j-1.7.7.jar:/opt/cassandra/lib/jctools-core-1.2.1.jar:/opt/cassandra/lib/jflex-1.6.0.jar:/opt/cassandra/lib/jna-4.2.2.jar:/opt/cassandra/lib/joda-time-2.4.jar:/opt/cassandra/lib/json-simple-1.1.jar:/opt/cassandra/lib/jstackjunit-0.0.1.jar:/opt/cassandra/lib/libthrift-0.13.0.jar:/opt/cassandra/lib/log4j-over-slf4j-1.7.7.jar:/opt/cassandra/lib/logback-classic-1.2.9.jar:/opt/cassandra/lib/logback-core-1.2.9.jar:/opt/cassandra/lib/lz4-1.3.0.jar:/opt/cassandra/lib/metrics-core-3.1.5.jar:/opt/cassandra/lib/metrics-jvm-3.1.5.jar:/opt/cassandra/lib/metrics-logback-3.1.5.jar:/opt/cassandra/lib/netty-all-4.1.39.Final.jar:/opt/cassandra/lib/ohc-core-0.4.4.jar:/opt/cassandra/lib/ohc-core-j8-0.4.4.jar:/opt/cassandra/lib/reporter-config3-3.0.3.jar:/opt/cassandra/lib/reporter-config-base-3.0.3.jar:/opt/cassandra/lib/sigar-1.6.4.jar:/opt/cassandra/lib/slf4j-api-1.7.7.jar:/opt/cassandra/lib/snakeyaml-1.11.jar:/opt/cassandra/lib/snappy-java-1.1.1.7.jar:/opt/cassandra/lib/snowball-stemmer-1.3.0.581.1.jar:/opt/cassandra/lib/ST4-4.0.8.jar:/opt/cassandra/lib/stream-2.5.2.jar:/opt/cassandra/lib/thrift-server-0.3.7.jar:/opt/cassandra/lib/jsr223/*/*.jar:/opt/cassandra/lib/jamm-0.3.0.jar INFO [main] 2025-08-11 05:03:06,919 CassandraDaemon.java:490 - JVM Arguments: [-Xloggc:/opt/cassandra/logs/gc.log, -ea, -XX:+UseThreadPriorities, -XX:ThreadPriorityPolicy=42, -XX:+HeapDumpOnOutOfMemoryError, -Xss256k, -XX:StringTableSize=1000003, -XX:+AlwaysPreTouch, -XX:-UseBiasedLocking, -XX:+UseTLAB, -XX:+ResizeTLAB, -XX:+UseNUMA, -XX:+PerfDisableSharedMem, -Djava.net.preferIPv4Stack=true, -Xms1g, -Xmx2g, -XX:+UseParNewGC, -XX:+UseConcMarkSweepGC, -XX:+CMSParallelRemarkEnabled, -XX:SurvivorRatio=8, -XX:MaxTenuringThreshold=1, -XX:CMSInitiatingOccupancyFraction=75, -XX:+UseCMSInitiatingOccupancyOnly, -XX:CMSWaitDuration=10000, -XX:+CMSParallelInitialMarkEnabled, -XX:+CMSEdenChunksRecordAlways, -XX:+CMSClassUnloadingEnabled, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -XX:+PrintHeapAtGC, -XX:+PrintTenuringDistribution, -XX:+PrintGCApplicationStoppedTime, -XX:+PrintPromotionFailure, -XX:+UseGCLogFileRotation, -XX:NumberOfGCLogFiles=10, -XX:GCLogFileSize=10M, -Xmn400M, -XX:+UseCondCardMark, -XX:CompileCommandFile=/opt/cassandra/conf/hotspot_compiler, -javaagent:/opt/cassandra/lib/jamm-0.3.0.jar, -Dcassandra.jmx.remote.port=7199, -Dcom.sun.management.jmxremote.rmi.port=7199, -Dcom.sun.management.jmxremote.authenticate=true, -Dcom.sun.management.jmxremote.password.file=/etc/cassandra/jmxremote.password, -Djava.library.path=/opt/cassandra/lib/sigar-bin, -Dcassandra.rpc_port=9161, -Dcassandra.native_transport_port=9041, -Dcassandra.ssl_storage_port=7013, -Dcassandra.storage_port=7012, -Dcassandra.jmx.local.port=7201, -Dcom.sun.management.jmxremote.access.file=/etc/cassandra/jmxremote.access, -Dcassandra.jmx.remote.port=7201, -Dcom.sun.management.jmxremote.rmi.port=7201, -Dcassandra.libjemalloc=/usr/lib64/libjemalloc.so.1, -XX:OnOutOfMemoryError=kill -9 %p, -Dlogback.configurationFile=logback.xml, -Dcassandra.logdir=/opt/cassandra/logs, -Dcassandra.storagedir=/opt/cassandra/data, -Dcassandra-foreground=yes] WARN [main] 2025-08-11 05:03:06,962 NativeLibrary.java:187 - Unable to lock JVM memory (ENOMEM). This can result in part of the JVM being swapped out, especially with mmapped I/O enabled. Increase RLIMIT_MEMLOCK or run Cassandra as root. INFO [main] 2025-08-11 05:03:06,963 StartupChecks.java:140 - jemalloc seems to be preloaded from /usr/lib64/libjemalloc.so.1 INFO [main] 2025-08-11 05:03:06,963 StartupChecks.java:176 - JMX is enabled to receive remote connections on port: 7201 INFO [main] 2025-08-11 05:03:06,964 SigarLibrary.java:44 - Initializing SIGAR library INFO [main] 2025-08-11 05:03:06,979 SigarLibrary.java:180 - Checked OS settings and found them configured for optimal performance. WARN [main] 2025-08-11 05:03:06,982 StartupChecks.java:311 - Maximum number of memory map areas per process (vm.max_map_count) 128960 is too low, recommended value: 1048575, you can change it with sysctl. WARN [main] 2025-08-11 05:03:07,006 StartupChecks.java:332 - Directory /var/lib/cassandra/commitlog doesn't exist WARN [main] 2025-08-11 05:03:07,008 StartupChecks.java:332 - Directory /var/lib/cassandra/saved_caches doesn't exist WARN [main] 2025-08-11 05:03:07,009 StartupChecks.java:332 - Directory /opt/cassandra/data/hints doesn't exist INFO [main] 2025-08-11 05:03:07,102 QueryProcessor.java:116 - Initialized prepared statement caches with 10 MB (native) and 10 MB (Thrift) INFO [main] 2025-08-11 05:03:07,626 ColumnFamilyStore.java:411 - Initializing system.IndexInfo INFO [main] 2025-08-11 05:03:08,535 ColumnFamilyStore.java:411 - Initializing system.batches INFO [main] 2025-08-11 05:03:08,539 ColumnFamilyStore.java:411 - Initializing system.paxos INFO [main] 2025-08-11 05:03:08,591 ColumnFamilyStore.java:411 - Initializing system.local INFO [main] 2025-08-11 05:03:08,597 ColumnFamilyStore.java:411 - Initializing system.peers INFO [main] 2025-08-11 05:03:08,618 ColumnFamilyStore.java:411 - Initializing system.peer_events INFO [main] 2025-08-11 05:03:08,622 ColumnFamilyStore.java:411 - Initializing system.range_xfers INFO [main] 2025-08-11 05:03:08,640 ColumnFamilyStore.java:411 - Initializing system.compaction_history INFO [main] 2025-08-11 05:03:08,653 ColumnFamilyStore.java:411 - Initializing system.sstable_activity INFO [main] 2025-08-11 05:03:08,668 ColumnFamilyStore.java:411 - Initializing system.size_estimates INFO [main] 2025-08-11 05:03:08,690 ColumnFamilyStore.java:411 - Initializing system.available_ranges INFO [main] 2025-08-11 05:03:08,704 ColumnFamilyStore.java:411 - Initializing system.transferred_ranges INFO [main] 2025-08-11 05:03:08,717 ColumnFamilyStore.java:411 - Initializing system.views_builds_in_progress INFO [main] 2025-08-11 05:03:08,724 ColumnFamilyStore.java:411 - Initializing system.built_views INFO [main] 2025-08-11 05:03:08,741 ColumnFamilyStore.java:411 - Initializing system.hints INFO [main] 2025-08-11 05:03:08,759 ColumnFamilyStore.java:411 - Initializing system.batchlog INFO [main] 2025-08-11 05:03:08,765 ColumnFamilyStore.java:411 - Initializing system.prepared_statements INFO [main] 2025-08-11 05:03:08,773 ColumnFamilyStore.java:411 - Initializing system.schema_keyspaces INFO [main] 2025-08-11 05:03:08,795 ColumnFamilyStore.java:411 - Initializing system.schema_columnfamilies INFO [main] 2025-08-11 05:03:08,810 ColumnFamilyStore.java:411 - Initializing system.schema_columns INFO [main] 2025-08-11 05:03:08,821 ColumnFamilyStore.java:411 - Initializing system.schema_triggers INFO [main] 2025-08-11 05:03:08,856 ColumnFamilyStore.java:411 - Initializing system.schema_usertypes INFO [main] 2025-08-11 05:03:08,880 ColumnFamilyStore.java:411 - Initializing system.schema_functions INFO [main] 2025-08-11 05:03:08,889 ColumnFamilyStore.java:411 - Initializing system.schema_aggregates INFO [main] 2025-08-11 05:03:08,901 ViewManager.java:137 - Not submitting build tasks for views in keyspace system as storage service is not initialized INFO [main] 2025-08-11 05:03:09,049 ApproximateTime.java:44 - Scheduling approximate time-check task with a precision of 10 milliseconds INFO [main] 2025-08-11 05:03:09,128 ColumnFamilyStore.java:411 - Initializing system_schema.keyspaces INFO [main] 2025-08-11 05:03:09,188 ColumnFamilyStore.java:411 - Initializing system_schema.tables INFO [main] 2025-08-11 05:03:09,228 ColumnFamilyStore.java:411 - Initializing system_schema.columns INFO [main] 2025-08-11 05:03:09,232 ColumnFamilyStore.java:411 - Initializing system_schema.triggers INFO [main] 2025-08-11 05:03:09,256 ColumnFamilyStore.java:411 - Initializing system_schema.dropped_columns INFO [main] 2025-08-11 05:03:09,259 ColumnFamilyStore.java:411 - Initializing system_schema.views INFO [main] 2025-08-11 05:03:09,262 ColumnFamilyStore.java:411 - Initializing system_schema.types INFO [main] 2025-08-11 05:03:09,275 ColumnFamilyStore.java:411 - Initializing system_schema.functions INFO [main] 2025-08-11 05:03:09,278 ColumnFamilyStore.java:411 - Initializing system_schema.aggregates INFO [main] 2025-08-11 05:03:09,292 ColumnFamilyStore.java:411 - Initializing system_schema.indexes INFO [main] 2025-08-11 05:03:09,303 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_schema as storage service is not initialized INFO [MemtableFlushWriter:1] 2025-08-11 05:03:10,356 CacheService.java:112 - Initializing key cache with capacity of 49 MBs. INFO [MemtableFlushWriter:1] 2025-08-11 05:03:10,366 CacheService.java:134 - Initializing row cache with capacity of 0 MBs INFO [MemtableFlushWriter:1] 2025-08-11 05:03:10,370 CacheService.java:163 - Initializing counter cache with capacity of 24 MBs INFO [MemtableFlushWriter:1] 2025-08-11 05:03:10,372 CacheService.java:174 - Scheduling counter cache save to every 7200 seconds (going to save all keys). INFO [CompactionExecutor:4] 2025-08-11 05:03:10,724 BufferPool.java:230 - Global buffer pool is enabled, when pool is exhausted (max is 502.000MiB) it will allocate on heap INFO [main] 2025-08-11 05:03:10,772 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-08-11 05:03:10,922 StorageService.java:607 - Token metadata: INFO [pool-4-thread-1] 2025-08-11 05:03:11,003 AutoSavingCache.java:174 - Completed loading (15 ms; 8 keys) KeyCache cache INFO [main] 2025-08-11 05:03:11,038 CommitLog.java:152 - No commitlog files found; skipping replay INFO [main] 2025-08-11 05:03:11,041 StorageService.java:600 - Populating token metadata from system tables INFO [main] 2025-08-11 05:03:11,113 StorageService.java:607 - Token metadata: INFO [main] 2025-08-11 05:03:11,221 QueryProcessor.java:163 - Preloaded 0 prepared statements INFO [main] 2025-08-11 05:03:11,222 StorageService.java:618 - Cassandra version: 3.11.3 INFO [main] 2025-08-11 05:03:11,222 StorageService.java:619 - Thrift API version: 20.1.0 INFO [main] 2025-08-11 05:03:11,222 StorageService.java:620 - CQL supported versions: 3.4.4 (default: 3.4.4) INFO [main] 2025-08-11 05:03:11,222 StorageService.java:622 - Native protocol supported versions: 3/v3, 4/v4, 5/v5-beta (default: 4/v4) INFO [main] 2025-08-11 05:03:11,252 IndexSummaryManager.java:85 - Initializing index summary manager with a memory pool size of 49 MB and a resize interval of 60 minutes INFO [main] 2025-08-11 05:03:11,273 MessagingService.java:761 - Starting Messaging Service on /10.0.0.38:7012 (ens3) WARN [main] 2025-08-11 05:03:11,286 SystemKeyspace.java:1087 - No host ID found, created 0810adf8-6338-48cc-ace0-ca7ec1cd98b5 (Note: This should happen exactly once per node). INFO [main] 2025-08-11 05:03:11,312 OutboundTcpConnection.java:108 - OutboundTcpConnection using coalescing strategy DISABLED INFO [HANDSHAKE-/10.0.0.48] 2025-08-11 05:03:13,517 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.48 INFO [HANDSHAKE-/10.0.0.254] 2025-08-11 05:03:13,866 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.254 INFO [main] 2025-08-11 05:03:14,330 StorageService.java:550 - Unable to gossip with any peers but continuing anyway since node is in its own seed list INFO [main] 2025-08-11 05:03:14,368 StorageService.java:704 - Loading persisted ring state INFO [main] 2025-08-11 05:03:14,369 StorageService.java:822 - Starting up server gossip INFO [main] 2025-08-11 05:03:14,510 StorageService.java:883 - This node will not auto bootstrap because it is configured to be a seed node. INFO [main] 2025-08-11 05:03:14,532 BootStrapper.java:228 - Generated random tokens. tokens are [-922243857338436063, -1254710546114676975, 7573046190170620098, -8568999916604067684, 2512267426623338901, 4926241062438929646, -1268870939699010660, -7555667343842968954, -1171326297555531395, 7504791548999054178, 4816713694942964866, -6848127382100436032, -6427479982938010142, -8967715448292354507, 1257980168599514316, -5001146597323969596, -546828275492705581, -8485744976274770563, 3481671551440980367, -5929607739379917705, -4643543610983418808, -4182873870133288487, 3963891088909437959, 7885584630150876720, -1846291705564092085, 6907362380601563723, -7914785638039302323, -1542397939566254547, -8882128040132651219, 4784777247834383665, 6617256402486531008, 4769798069225834233, -3582810340464399377, -5563615872537676051, -1608583484966736162, 8845413344052430257, -1806295944319099269, -1446920288003778265, -4866549717400992653, 558524953172463336, -5016630611073257815, -7028232176096410422, 1477512856889126157, 8569455317317121017, 1116708922940365470, -2141813573857186585, -3931563397989115875, 4950013527172079026, -6555905943234577524, 749380500001527714, 2770797050572678954, 4085183372716964672, -1429548174989199446, -1060553579488441280, -3264027398293592314, -2086827993395524129, 952549208811754875, -8023054686500304400, -7212455001860833646, 2437032372739120748, -5008899575247804144, 1650338058903525779, -8130643348871856049, -2201939323178462528, -691896742073347053, 2989137655164105203, 6559031572143340484, 8932775237577642643, 8524360820666279619, 4389108279028247236, 3266403023397922599, -4363726928328598895, 6148499716310660677, 2698943763081315318, 5916613128602870567, 4737517462060563802, 4670532865859697623, -6404373445655993516, 3652750788935728889, -1087882488792597341, 740166614000479488, 3043763225958671312, 6197773603955753619, 9222616138366552740, 7987682581017123067, 7790517425208209537, -2789659765304744841, -1894325621602526999, 3063351872026905652, 5883374631255046118, -240258265225009428, -6293629845788826990, 6666539558938086822, 6647493893754023204, -2880920461020533636, 6104001047399267878, -7978452654010448796, 3171943397372914021, -6118400494076722659, 3983533382731384201, -3476334309187650276, -1716798735616185176, -8403199008751002376, -8172994629395579541, 4034142872765197003, 1672424803056763769, -6709569642982662238, -6311841966587873645, 179959044105206601, -6599345731267618469, -3678317532107149807, 1778265078521073986, -651304408707871817, -4043125838027732808, -1124505467890544325, 6803964913107125401, 9045785024214577620, 3663013921285439391, 588711067040114424, 8716660416407963685, -927211589115407638, -3546733851887505184, 2469397696259290099, 60906045551544760, -993121776818419688, -5180059753629049466, -8646284163048957675, 1239581197410188149, -8939609920862827889, -7850980005527901521, 5250069707066718904, 7002293997081287601, 8762257729049103408, -4837557454784790074, 9028673770639881056, 6380828976955673700, -2247311330917296828, -765086873987583988, -45804441924628233, -8425439749007716208, -7901130291680335726, -1226925813006651588, 2855983433490814646, -972757864284397317, -2114712713921344347, -6679517581065528386, 863404494307707854, 3414669472537978397, -4520693506647507743, -2945950615679574208, -9144056520884262949, 755612848760212701, 8768164781100851720, -2386117526298026533, 2793786079596281076, -5635316107034208137, 1218050885637301166, -4543089804315977402, -3639949218377858694, -3464636789101786544, 2664574135574508809, -3283380489110430266, 7060308475822545323, -525450892870755483, 5107693848146233293, -6378143067224168510, -7476938030703834267, -6852712117039131407, -1975039030747348661, -6794738856000594612, -4901126140408021329, 8432896802672793172, -258306875197330808, -3233960077794428066, 9170180804550629302, -5043940682038362835, -773586314555113291, -5519430711398302813, -6460304998359728668, -4489343154216130560, -1788380262251033539, -5665478351294757381, 641118275614900680, -6116954164253983161, -5001839338556504256, -9152077769878549738, 7368891829797258887, -363031662845818849, 3841301817784867674, 5004655754497029869, -277563842710521302, 1429911359206002507, 1750370019182943094, -2980485276585820281, 798695152453518939, -90410882511504646, 9116516713723865808, -2986487423200768669, -5137190648384952586, -398799177535181863, 7044575058566789793, -4117263105595949479, -890872958428966121, 6818639074613373173, -2799565270652251143, -6007442849984542788, 486124473178059997, -7238299673173297589, -956676727930219428, -1098098127341220938, -8562951480920334679, 3690223523304299633, 7999356140737312456, -49636482160936453, 2751814167882168676, -3972994772518353237, -4666989509806246705, -3071555821375031922, 7596262759960300957, 8162697046427510330, 4010897839308287514, -1036939135596558827, -5027042771387939205, -4095308468303096980, -4205929198910729807, 2533115391372435516, -551282799413825407, 6558098939162783836, -5020661699595964036, 7050531043973890392, 1622293110476668470, -9212282612430673986, -2046851833247632848, 860793568901980748, -987909108278351673, 6150419517318485522, 9025220959025073183, 6274703961646649948, 4258792644963256496, 577151709998422143, -2247082755960374008, -4823955968790116520, -7894294849441303057, 6707896277171033097, 7933280872367474140, -7366715826314521204, 5679480193311957823, 4348256590348645893, 3405114400480136987, -4056859189394733037, 8909277375643874415, 6991008435266982917, 1224061759620822013, 424170382482724016, -7703350894792556420, 3326810319329634097] INFO [main] 2025-08-11 05:03:14,538 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_traces, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=2}}, tables=[org.apache.cassandra.config.CFMetaData@49a01709[cfId=c5e99f16-8677-3914-b17e-960613512345,ksName=system_traces,cfName=sessions,flags=[COMPOUND],params=TableParams{comment=tracing sessions, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [client command coordinator duration request started_at parameters]],partitionKeyColumns=[session_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[client, command, session_id, coordinator, request, started_at, duration, parameters],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@106d5d3e[cfId=8826e8e9-e16a-3728-8753-3bc1fc713c25,ksName=system_traces,cfName=events,flags=[COMPOUND],params=TableParams{comment=tracing events, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=0, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [activity source source_elapsed thread]],partitionKeyColumns=[session_id],clusteringColumns=[event_id],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[activity, event_id, session_id, source, thread, source_elapsed],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-08-11 05:03:15,100 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_traces as storage service is not initialized INFO [MigrationStage:1] 2025-08-11 05:03:15,113 ColumnFamilyStore.java:411 - Initializing system_traces.events INFO [MigrationStage:1] 2025-08-11 05:03:15,127 ColumnFamilyStore.java:411 - Initializing system_traces.sessions INFO [main] 2025-08-11 05:03:15,165 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_distributed, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[org.apache.cassandra.config.CFMetaData@78351296[cfId=759fffad-624b-3181-80ee-fa9a52d1f627,ksName=system_distributed,cfName=repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | [coordinator exception_message exception_stacktrace finished_at parent_id range_begin range_end started_at status participants]],partitionKeyColumns=[keyspace_name, columnfamily_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[status, id, coordinator, finished_at, participants, exception_stacktrace, parent_id, range_end, range_begin, exception_message, keyspace_name, started_at, columnfamily_name],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@56003f32[cfId=deabd734-b99d-3b9c-92e5-fd92eb5abf14,ksName=system_distributed,cfName=parent_repair_history,flags=[COMPOUND],params=TableParams{comment=Repair history, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [exception_message exception_stacktrace finished_at keyspace_name started_at columnfamily_names options requested_ranges successful_ranges]],partitionKeyColumns=[parent_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[requested_ranges, exception_message, keyspace_name, successful_ranges, started_at, finished_at, options, exception_stacktrace, parent_id, columnfamily_names],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@5e03518c[cfId=5582b59f-8e4e-35e1-b913-3acada51eb04,ksName=system_distributed,cfName=view_build_status,flags=[COMPOUND],params=TableParams{comment=Materialized View build status, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UUIDType),partitionColumns=[[] | [status]],partitionKeyColumns=[keyspace_name, view_name],clusteringColumns=[host_id],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[view_name, status, keyspace_name, host_id],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-08-11 05:03:15,299 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_distributed as storage service is not initialized INFO [MigrationStage:1] 2025-08-11 05:03:15,311 ColumnFamilyStore.java:411 - Initializing system_distributed.parent_repair_history INFO [MigrationStage:1] 2025-08-11 05:03:15,316 ColumnFamilyStore.java:411 - Initializing system_distributed.repair_history INFO [MigrationStage:1] 2025-08-11 05:03:15,324 ColumnFamilyStore.java:411 - Initializing system_distributed.view_build_status INFO [main] 2025-08-11 05:03:15,361 StorageService.java:1446 - JOINING: Finish joining ring INFO [HANDSHAKE-/10.0.0.254] 2025-08-11 05:03:15,454 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.254 INFO [main] 2025-08-11 05:03:15,535 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=system_auth, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=1}}, tables=[org.apache.cassandra.config.CFMetaData@5c0ec2ce[cfId=5bc52802-de25-35ed-aeab-188eecebb090,ksName=system_auth,cfName=roles,flags=[COMPOUND],params=TableParams{comment=role definitions, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [can_login is_superuser salted_hash member_of]],partitionKeyColumns=[role],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[salted_hash, member_of, role, can_login, is_superuser],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@4e0ddd6b[cfId=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d,ksName=system_auth,cfName=role_members,flags=[COMPOUND],params=TableParams{comment=role memberships lookup table, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | []],partitionKeyColumns=[role],clusteringColumns=[member],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[role, member],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@48ac4921[cfId=3afbe79f-2194-31a7-add7-f5ab90d8ec9c,ksName=system_auth,cfName=role_permissions,flags=[COMPOUND],params=TableParams{comment=permissions granted to db roles, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [permissions]],partitionKeyColumns=[role],clusteringColumns=[resource],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[role, resource, permissions],droppedColumns={},triggers=[],indexes=[]], org.apache.cassandra.config.CFMetaData@662879fc[cfId=5f2fbdad-91f1-3946-bd25-d5da3a5c35ec,ksName=system_auth,cfName=resource_role_permissons_index,flags=[COMPOUND],params=TableParams{comment=index of db roles with permissions granted on a resource, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=7776000, default_time_to_live=0, memtable_flush_period_in_ms=3600000, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | []],partitionKeyColumns=[resource],clusteringColumns=[role],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[resource, role],droppedColumns={},triggers=[],indexes=[]]], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-08-11 05:03:15,715 ViewManager.java:137 - Not submitting build tasks for views in keyspace system_auth as storage service is not initialized INFO [MigrationStage:1] 2025-08-11 05:03:15,717 ColumnFamilyStore.java:411 - Initializing system_auth.resource_role_permissons_index INFO [MigrationStage:1] 2025-08-11 05:03:15,725 ColumnFamilyStore.java:411 - Initializing system_auth.role_members INFO [MigrationStage:1] 2025-08-11 05:03:15,731 ColumnFamilyStore.java:411 - Initializing system_auth.role_permissions INFO [MigrationStage:1] 2025-08-11 05:03:15,736 ColumnFamilyStore.java:411 - Initializing system_auth.roles INFO [main] 2025-08-11 05:03:15,749 Gossiper.java:1692 - Waiting for gossip to settle... INFO [GossipStage:1] 2025-08-11 05:03:16,227 Gossiper.java:1055 - Node /10.0.0.48 is now part of the cluster INFO [GossipStage:1] 2025-08-11 05:03:16,229 Gossiper.java:1055 - Node /10.0.0.254 is now part of the cluster INFO [RequestResponseStage-1] 2025-08-11 05:03:16,233 Gossiper.java:1019 - InetAddress /10.0.0.254 is now UP INFO [RequestResponseStage-1] 2025-08-11 05:03:16,234 Gossiper.java:1019 - InetAddress /10.0.0.48 is now UP WARN [GossipTasks:1] 2025-08-11 05:03:16,433 FailureDetector.java:288 - Not marking nodes down due to local pause of 5463051652 > 5000000000 INFO [HANDSHAKE-/10.0.0.48] 2025-08-11 05:03:17,100 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.48 INFO [main] 2025-08-11 05:03:23,750 Gossiper.java:1723 - No gossip backlog; proceeding INFO [main] 2025-08-11 05:03:23,982 NativeTransportService.java:70 - Netty using native Epoll event loop INFO [main] 2025-08-11 05:03:24,045 Server.java:155 - Using Netty Version: [netty-buffer=netty-buffer-4.1.39.Final.88c2a4c (repository: dirty), netty-codec=netty-codec-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-dns=netty-codec-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-haproxy=netty-codec-haproxy-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http=netty-codec-http-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-http2=netty-codec-http2-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-memcache=netty-codec-memcache-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-mqtt=netty-codec-mqtt-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-redis=netty-codec-redis-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-smtp=netty-codec-smtp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-socks=netty-codec-socks-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-stomp=netty-codec-stomp-4.1.39.Final.88c2a4c (repository: dirty), netty-codec-xml=netty-codec-xml-4.1.39.Final.88c2a4c (repository: dirty), netty-common=netty-common-4.1.39.Final.88c2a4c (repository: dirty), netty-handler=netty-handler-4.1.39.Final.88c2a4c (repository: dirty), netty-handler-proxy=netty-handler-proxy-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver=netty-resolver-4.1.39.Final.88c2a4c (repository: dirty), netty-resolver-dns=netty-resolver-dns-4.1.39.Final.88c2a4c (repository: dirty), netty-tcnative=netty-tcnative-2.0.25.Final.c46c351, netty-transport=netty-transport-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-epoll=netty-transport-native-epoll-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-native-kqueue=netty-transport-native-kqueue-4.1.39.Final.88c2a4cab5 (repository: dirty), netty-transport-native-unix-common=netty-transport-native-unix-common-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-rxtx=netty-transport-rxtx-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-sctp=netty-transport-sctp-4.1.39.Final.88c2a4c (repository: dirty), netty-transport-udt=netty-transport-udt-4.1.39.Final.88c2a4c (repository: dirty)] INFO [main] 2025-08-11 05:03:24,046 Server.java:156 - Starting listening for CQL clients on /10.0.0.38:9041 (unencrypted)... INFO [main] 2025-08-11 05:03:24,106 ThriftServer.java:116 - Binding thrift service to /10.0.0.38:9161 INFO [Thread-2] 2025-08-11 05:03:24,137 ThriftServer.java:133 - Listening for thrift clients... INFO [OptionalTasks:1] 2025-08-11 05:03:25,929 CassandraRoleManager.java:356 - Created default superuser role 'cassandra' INFO [Native-Transport-Requests-2] 2025-08-11 05:03:39,870 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@ca6b316[cfId=8b9507e0-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:39,971 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration INFO [MigrationStage:1] 2025-08-11 05:03:40,303 ColumnFamilyStore.java:411 - Initializing reaper_db.schema_migration_leader INFO [HANDSHAKE-/10.0.0.38] 2025-08-11 05:03:40,953 OutboundTcpConnection.java:561 - Handshaking version with /10.0.0.38 WARN [Native-Transport-Requests-1] 2025-08-11 05:03:40,960 TimeFcts.java:99 - The function 'dateof' is deprecated. Use the function 'toTimestamp' instead. INFO [MigrationStage:1] 2025-08-11 05:03:41,134 ColumnFamilyStore.java:411 - Initializing reaper_db.running_reapers INFO [Native-Transport-Requests-4] 2025-08-11 05:03:41,667 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@5a4d58bf[cfId=8ca73b30-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:41,762 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_unit_v1 INFO [MigrationStage:1] 2025-08-11 05:03:43,321 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_by_cluster_and_keyspace INFO [Native-Transport-Requests-2] 2025-08-11 05:03:43,587 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@439afb2[cfId=8dcc3330-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_run_by_cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:43,680 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster INFO [MigrationStage:1] 2025-08-11 05:03:44,666 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_schedule_v1 INFO [Native-Transport-Requests-3] 2025-08-11 05:03:45,113 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@2ed8ec7c[cfId=8eb50c90-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[partitioner, seed_hosts, name],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:45,224 ColumnFamilyStore.java:411 - Initializing reaper_db.cluster INFO [MigrationStage:1] 2025-08-11 05:03:45,554 ColumnFamilyStore.java:411 - Initializing reaper_db.snapshot INFO [Native-Transport-Requests-6] 2025-08-11 05:03:46,310 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@6bc67784[cfId=8f6bb260-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_metrics_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=120, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [active_anticompactions cluster datacenter has_repair_running pending_compactions requested]],partitionKeyColumns=[run_id, time_partition],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.LongType),columnMetadata=[cluster, node, has_repair_running, pending_compactions, active_anticompactions, time_partition, datacenter, requested, run_id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:46,430 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v1 INFO [MigrationStage:1] 2025-08-11 05:03:47,403 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run INFO [Native-Transport-Requests-2] 2025-08-11 05:03:47,623 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@73308776[cfId=90340b70-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:47,722 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_unit INFO [MigrationStage:1] 2025-08-11 05:03:48,282 ColumnFamilyStore.java:411 - Initializing reaper_db.leader INFO [MigrationStage:1] 2025-08-11 05:03:50,514 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v2 INFO [STREAM-INIT-/10.0.0.48:57972] 2025-08-11 05:03:50,566 StreamResultFuture.java:116 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e ID#0] Creating new streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.48:57972] 2025-08-11 05:03:50,574 StreamResultFuture.java:123 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e, ID#0] Received streaming plan for Bootstrap INFO [STREAM-INIT-/10.0.0.48:57974] 2025-08-11 05:03:50,574 StreamResultFuture.java:123 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e, ID#0] Received streaming plan for Bootstrap INFO [STREAM-IN-/10.0.0.48:57974] 2025-08-11 05:03:50,686 StreamResultFuture.java:173 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e ID#0] Prepare completed. Receiving 0 files(0.000KiB), sending 1 files(0.079KiB) INFO [STREAM-IN-/10.0.0.48:57974] 2025-08-11 05:03:50,993 StreamResultFuture.java:187 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e] Session with /10.0.0.48 is complete INFO [STREAM-IN-/10.0.0.48:57974] 2025-08-11 05:03:50,994 StreamResultFuture.java:219 - [Stream #91db79e0-7670-11f0-993e-6d60ca3e2f4e] All sessions completed INFO [Native-Transport-Requests-1] 2025-08-11 05:03:51,209 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@64f666c8[cfId=92571280-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:51,306 ColumnFamilyStore.java:411 - Initializing reaper_db.node_operations INFO [Native-Transport-Requests-2] 2025-08-11 05:03:52,725 MigrationManager.java:454 - Update table 'reaper_db/cluster' From org.apache.cassandra.config.CFMetaData@48bbf0d0[cfId=8eb50c90-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@7242e2a6[cfId=8eb50c90-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=cluster,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [partitioner properties state seed_hosts]],partitionKeyColumns=[name],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[properties, state, name, partitioner, seed_hosts],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:54,374 ColumnFamilyStore.java:411 - Initializing reaper_db.diagnostic_event_subscription INFO [Native-Transport-Requests-1] 2025-08-11 05:03:56,247 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@21f7eef6[cfId=9557cf60-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:56,343 ColumnFamilyStore.java:411 - Initializing reaper_db.node_metrics_v3 INFO [Native-Transport-Requests-2] 2025-08-11 05:03:56,638 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@5b5af494[cfId=95939fe0-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_run_by_cluster_v2,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimeUUIDType)),partitionColumns=[[] | [repair_run_state]],partitionKeyColumns=[cluster_name],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.UTF8Type,columnMetadata=[cluster_name, repair_run_state, id],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:56,728 ColumnFamilyStore.java:411 - Initializing reaper_db.repair_run_by_cluster_v2 INFO [Native-Transport-Requests-1] 2025-08-11 05:03:57,248 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@61273d56[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, last_event, id, segment_end_time, state, cluster_name, end_time, end_token, start_token, segment_start_time, segment_state, cause, creation_time, start_time, coordinator_host, token_ranges, owner, repair_parallelism, tables, segment_id, pause_time, repair_unit_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@56717850[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, segment_count, last_event, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-08-11 05:03:57,677 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@612afcb6[cfId=963229d0-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=running_repairs,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [reaper_instance_host reaper_instance_id segment_id]],partitionKeyColumns=[repair_id],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, repair_id, node, segment_id, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:58,120 ColumnFamilyStore.java:411 - Initializing reaper_db.running_repairs INFO [Native-Transport-Requests-2] 2025-08-11 05:03:58,231 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@4829f8f2[cfId=9686b270-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=percent_repaired_by_schedule,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [keyspace_name percent_repaired table_name ts]],partitionKeyColumns=[cluster_name, repair_schedule_id, time_bucket],clusteringColumns=[node],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UUIDType,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[time_bucket, node, ts, keyspace_name, percent_repaired, repair_schedule_id, table_name, cluster_name],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:03:58,363 ColumnFamilyStore.java:411 - Initializing reaper_db.percent_repaired_by_schedule INFO [Native-Transport-Requests-3] 2025-08-11 05:03:59,247 MigrationManager.java:454 - Update table 'reaper_db/repair_unit_v1' From org.apache.cassandra.config.CFMetaData@5a8a4751[cfId=8ca73b30-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@5a486393[cfId=8ca73b30-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_unit_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster_name incremental_repair keyspace_name repair_thread_count timeout blacklisted_tables column_families datacenters nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[blacklisted_tables, datacenters, repair_thread_count, id, keyspace_name, timeout, nodes, cluster_name, incremental_repair, column_families],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-7] 2025-08-11 05:04:00,721 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@61273d56[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, segment_count, last_event, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3573c15c[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[adaptive_schedule cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, segment_count, last_event, adaptive_schedule, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-08-11 05:04:01,346 MigrationManager.java:454 - Update table 'reaper_db/repair_schedule_v1' From org.apache.cassandra.config.CFMetaData@4e2a23cf[cfId=8e5c1720-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3e4f584e[cfId=8e5c1720-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_schedule_v1,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [adaptive creation_time days_between intensity next_activation owner pause_time percent_unrepaired_threshold repair_parallelism repair_unit_id segment_count segment_count_per_node state run_history]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, segment_count, days_between, percent_unrepaired_threshold, id, state, run_history, creation_time, adaptive, owner, repair_parallelism, segment_count_per_node, pause_time, repair_unit_id, next_activation],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-08-11 05:04:05,241 MigrationManager.java:454 - Update table 'reaper_db/running_reapers' From org.apache.cassandra.config.CFMetaData@58a02ef7[cfId=8c43e580-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@3562bf4f[cfId=8c43e580-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=running_reapers,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=180, default_time_to_live=180, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [last_heartbeat reaper_instance_host]],partitionKeyColumns=[reaper_instance_id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[reaper_instance_id, last_heartbeat, reaper_instance_host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-7] 2025-08-11 05:04:05,241 MigrationManager.java:454 - Update table 'reaper_db/node_metrics_v3' From org.apache.cassandra.config.CFMetaData@4ba8186d[cfId=9557cf60-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@744ecb67[cfId=9557cf60-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_metrics_v3,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=3600, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.ReversedType(org.apache.cassandra.db.marshal.TimestampType), org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type, org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [value]],partitionKeyColumns=[cluster, metric_domain, metric_type, time_bucket, host],clusteringColumns=[ts, metric_scope, metric_name, metric_attribute],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, metric_domain, metric_attribute, time_bucket, ts, metric_type, metric_name, metric_scope, value, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-08-11 05:04:05,241 MigrationManager.java:454 - Update table 'reaper_db/diagnostic_event_subscription' From org.apache.cassandra.config.CFMetaData@5373e096[cfId=942bab70-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=diagnostic_event_subscription,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster description export_file_logger export_http_endpoint export_sse events nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[cluster, export_http_endpoint, events, id, export_sse, nodes, export_file_logger, description],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@64732204[cfId=942bab70-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=diagnostic_event_subscription,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(),partitionColumns=[[] | [cluster description export_file_logger export_http_endpoint export_sse events nodes]],partitionKeyColumns=[id],clusteringColumns=[],keyValidator=org.apache.cassandra.db.marshal.UUIDType,columnMetadata=[cluster, export_http_endpoint, events, id, export_sse, nodes, export_file_logger, description],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-5] 2025-08-11 05:04:05,241 MigrationManager.java:454 - Update table 'reaper_db/repair_run' From org.apache.cassandra.config.CFMetaData@61273d56[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[adaptive_schedule cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count host_id replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, host_id, segment_count, last_event, adaptive_schedule, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@13ef41fc[cfId=8ffea390-7670-11f0-8f47-2597d7221e0e,ksName=reaper_db,cfName=repair_run,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : '5000'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[adaptive_schedule cause cluster_name creation_time end_time intensity last_event owner pause_time repair_parallelism repair_unit_id segment_count start_time state tables] | [coordinator_host end_token fail_count host_id replicas segment_end_time segment_start_time segment_state start_token token_ranges]],partitionKeyColumns=[id],clusteringColumns=[segment_id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[intensity, id, segment_end_time, state, end_token, start_token, start_time, token_ranges, tables, pause_time, repair_unit_id, host_id, segment_count, last_event, adaptive_schedule, cluster_name, end_time, segment_start_time, segment_state, cause, creation_time, coordinator_host, replicas, owner, repair_parallelism, segment_id, fail_count],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-4] 2025-08-11 05:04:05,243 MigrationManager.java:454 - Update table 'reaper_db/schema_migration' From org.apache.cassandra.config.CFMetaData@6d4e38e2[cfId=8b9507e0-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@1a9975f8[cfId=8b9507e0-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=schema_migration,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.Int32Type),partitionColumns=[[] | [executed_at script script_name]],partitionKeyColumns=[applied_successful],clusteringColumns=[version],keyValidator=org.apache.cassandra.db.marshal.BooleanType,columnMetadata=[script_name, version, applied_successful, executed_at, script],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-3] 2025-08-11 05:04:05,243 MigrationManager.java:454 - Update table 'reaper_db/repair_run_by_unit' From org.apache.cassandra.config.CFMetaData@d7faf57[cfId=90340b70-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@58f6762c[cfId=90340b70-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=repair_run_by_unit,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.1, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'ALL'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.LeveledCompactionStrategy, options={}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.TimeUUIDType),partitionColumns=[[] | []],partitionKeyColumns=[repair_unit_id],clusteringColumns=[id],keyValidator=org.apache.cassandra.db.marshal.TimeUUIDType,columnMetadata=[repair_unit_id, id],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-08-11 05:04:15,697 MigrationManager.java:454 - Update table 'reaper_db/node_operations' From org.apache.cassandra.config.CFMetaData@42d57269[cfId=92571280-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@1ea7fe74[cfId=92571280-7670-11f0-bff1-ef813b7a18da,ksName=reaper_db,cfName=node_operations,flags=[COMPOUND],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.0, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=300, default_time_to_live=300, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy, options={min_threshold=4, max_threshold=32, compaction_window_size=30, compaction_window_unit=MINUTES, unchecked_tombstone_compaction=true}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.UTF8Type),partitionColumns=[[] | [data ts]],partitionKeyColumns=[cluster, type, time_bucket],clusteringColumns=[host],keyValidator=org.apache.cassandra.db.marshal.CompositeType(org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type,org.apache.cassandra.db.marshal.UTF8Type),columnMetadata=[cluster, type, time_bucket, ts, data, host],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-2] 2025-08-11 05:04:19,217 MigrationManager.java:427 - Update Keyspace 'svc_monitor_keyspace' From KeyspaceMetadata{name=svc_monitor_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=svc_monitor_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-3] 2025-08-11 05:04:20,509 MigrationManager.java:427 - Update Keyspace 'config_db_uuid' From KeyspaceMetadata{name=config_db_uuid, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=config_db_uuid, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-08-11 05:04:21,669 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_uuid_table INFO [Native-Transport-Requests-1] 2025-08-11 05:04:23,351 MigrationManager.java:331 - Create new Keyspace: KeyspaceMetadata{name=useragent, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [MigrationStage:1] 2025-08-11 05:04:24,818 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.service_instance_table INFO [Native-Transport-Requests-1] 2025-08-11 05:04:29,219 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@37c96ac2[cfId=a8fef020-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:04:29,669 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_fq_name_table INFO [Native-Transport-Requests-1] 2025-08-11 05:04:30,321 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@6cdfffd3[cfId=a9a73e10-7670-11f0-bff1-ef813b7a18da,ksName=svc_monitor_keyspace,cfName=pool_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:04:30,422 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.pool_table INFO [Native-Transport-Requests-1] 2025-08-11 05:04:34,751 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@ada29a1[cfId=ac4b34f0-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:04:34,872 ColumnFamilyStore.java:411 - Initializing config_db_uuid.obj_shared_table INFO [MigrationStage:1] 2025-08-11 05:04:35,592 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.loadbalancer_table INFO [Native-Transport-Requests-1] 2025-08-11 05:04:38,207 MigrationManager.java:454 - Update table 'config_db_uuid/obj_shared_table' From org.apache.cassandra.config.CFMetaData@2009acb4[cfId=ac4b34f0-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@69742278[cfId=ac4b34f0-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_shared_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-5] 2025-08-11 05:04:39,588 MigrationManager.java:454 - Update table 'svc_monitor_keyspace/loadbalancer_table' From org.apache.cassandra.config.CFMetaData@1cf0ad9f[cfId=acbeb740-7670-11f0-8f47-2597d7221e0e,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@42a87cb2[cfId=acbeb740-7670-11f0-8f47-2597d7221e0e,ksName=svc_monitor_keyspace,cfName=loadbalancer_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:04:40,601 ColumnFamilyStore.java:411 - Initializing useragent.useragent_keyval_table INFO [MigrationStage:1] 2025-08-11 05:04:41,804 ColumnFamilyStore.java:411 - Initializing svc_monitor_keyspace.healthmonitor_table INFO [Native-Transport-Requests-4] 2025-08-11 05:04:42,706 MigrationManager.java:454 - Update table 'useragent/useragent_keyval_table' From org.apache.cassandra.config.CFMetaData@5c771505[cfId=afba9220-7670-11f0-993e-6d60ca3e2f4e,ksName=useragent,cfName=useragent_keyval_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@4249c159[cfId=afba9220-7670-11f0-993e-6d60ca3e2f4e,ksName=useragent,cfName=useragent_keyval_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-08-11 05:04:45,810 MigrationManager.java:454 - Update table 'config_db_uuid/obj_fq_name_table' From org.apache.cassandra.config.CFMetaData@1e60beb4[cfId=a8fef020-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@422fc16d[cfId=a8fef020-7670-11f0-bff1-ef813b7a18da,ksName=config_db_uuid,cfName=obj_fq_name_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [Native-Transport-Requests-1] 2025-08-11 05:04:47,374 MigrationManager.java:427 - Update Keyspace 'to_bgp_keyspace' From KeyspaceMetadata{name=to_bgp_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=to_bgp_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-2] 2025-08-11 05:04:50,693 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@4006a736[cfId=b5cbc350-7670-11f0-bff1-ef813b7a18da,ksName=to_bgp_keyspace,cfName=route_target_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:04:50,799 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.route_target_table INFO [MigrationStage:1] 2025-08-11 05:04:57,683 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_ip_address_table INFO [Native-Transport-Requests-2] 2025-08-11 05:04:59,219 MigrationManager.java:454 - Update table 'to_bgp_keyspace/service_chain_ip_address_table' From org.apache.cassandra.config.CFMetaData@84428f9[cfId=b9e80250-7670-11f0-8f47-2597d7221e0e,ksName=to_bgp_keyspace,cfName=service_chain_ip_address_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@678e2cac[cfId=b9e80250-7670-11f0-8f47-2597d7221e0e,ksName=to_bgp_keyspace,cfName=service_chain_ip_address_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:05:00,341 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_table INFO [Native-Transport-Requests-1] 2025-08-11 05:05:01,208 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@5d772f7e[cfId=bc103980-7670-11f0-bff1-ef813b7a18da,ksName=to_bgp_keyspace,cfName=service_chain_uuid_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:05:01,297 ColumnFamilyStore.java:411 - Initializing to_bgp_keyspace.service_chain_uuid_table INFO [Native-Transport-Requests-1] 2025-08-11 05:06:03,386 MigrationManager.java:427 - Update Keyspace 'dm_keyspace' From KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} To KeyspaceMetadata{name=dm_keyspace, params=KeyspaceParams{durable_writes=true, replication=ReplicationParams{class=org.apache.cassandra.locator.SimpleStrategy, replication_factor=3}}, tables=[], views=[], functions=[], types=[]} INFO [Native-Transport-Requests-1] 2025-08-11 05:06:04,726 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@39a0a99d[cfId=e1ec4d60-7670-11f0-bff1-ef813b7a18da,ksName=dm_keyspace,cfName=dm_pr_vn_ip_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:06:04,814 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_vn_ip_table INFO [MigrationStage:1] 2025-08-11 05:06:07,136 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pr_asn_table INFO [Native-Transport-Requests-1] 2025-08-11 05:06:08,328 MigrationManager.java:454 - Update table 'dm_keyspace/dm_pr_asn_table' From org.apache.cassandra.config.CFMetaData@11ce3c0a[cfId=e3267b10-7670-11f0-8f47-2597d7221e0e,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] To org.apache.cassandra.config.CFMetaData@119c5ce8[cfId=e3267b10-7670-11f0-8f47-2597d7221e0e,ksName=dm_keyspace,cfName=dm_pr_asn_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={max_threshold=32, min_threshold=4}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:06:11,964 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_ni_ipv6_ll_table INFO [Native-Transport-Requests-3] 2025-08-11 05:06:14,443 MigrationManager.java:376 - Create new table: org.apache.cassandra.config.CFMetaData@46fa1b46[cfId=e7b6ffb0-7670-11f0-bff1-ef813b7a18da,ksName=dm_keyspace,cfName=dm_pnf_resource_table,flags=[DENSE],params=TableParams{comment=, read_repair_chance=0.0, dclocal_read_repair_chance=0.1, bloom_filter_fp_chance=0.01, crc_check_chance=1.0, gc_grace_seconds=864000, default_time_to_live=0, memtable_flush_period_in_ms=0, min_index_interval=128, max_index_interval=2048, speculative_retry=99PERCENTILE, caching={'keys' : 'ALL', 'rows_per_partition' : 'NONE'}, compaction=CompactionParams{class=org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy, options={min_threshold=4, max_threshold=32}}, compression=org.apache.cassandra.schema.CompressionParams@f60f75cc, extensions={}, cdc=false},comparator=comparator(org.apache.cassandra.db.marshal.BytesType),partitionColumns=[[] | [value]],partitionKeyColumns=[key],clusteringColumns=[column1],keyValidator=org.apache.cassandra.db.marshal.BytesType,columnMetadata=[key, column1, value],droppedColumns={},triggers=[],indexes=[]] INFO [MigrationStage:1] 2025-08-11 05:06:14,550 ColumnFamilyStore.java:411 - Initializing dm_keyspace.dm_pnf_resource_table INFO [AntiEntropyStage:1] 2025-08-11 05:11:23,733 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,057 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,079 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,153 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,283 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,298 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,315 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,341 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,356 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,384 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,643 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,934 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-08-11 05:11:24,946 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-08-11 05:11:25,214 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-08-11 05:11:25,288 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-08-11 05:11:25,541 Validator.java:281 - [repair #9fff5c20-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-08-11 05:11:25,567 ActiveRepairService.java:452 - [repair #9ff57110-7671-11f0-993e-6d60ca3e2f4e] Not a global repair, will not do anticompaction INFO [Repair-Task-2] 2025-08-11 05:11:29,994 RepairRunnable.java:139 - Starting repair command #1 (a3cc1b90-7671-11f0-bff1-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 2, pull repair: false) INFO [Repair-Task-2] 2025-08-11 05:11:30,051 RepairSession.java:228 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.48, /10.0.0.254 on range [(-6460304998359728668,-6427479982938010142], (-1662474736743062124,-1608583484966736162]] for reaper_db.[repair_run_by_cluster, running_reapers, schema_migration_leader, repair_schedule_v1, leader, repair_run_by_unit, running_repairs, cluster, repair_unit_v1, percent_repaired_by_schedule, diagnostic_event_subscription, repair_schedule_by_cluster_and_keyspace, snapshot, repair_run, schema_migration, repair_run_by_cluster_v2] INFO [RepairJobTask:2] 2025-08-11 05:11:30,090 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:30,091 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,096 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,097 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,101 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,101 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,103 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:30,107 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-08-11 05:11:30,108 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-08-11 05:11:30,108 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:2] 2025-08-11 05:11:30,108 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_run_by_cluster is fully synced INFO [RepairJobTask:2] 2025-08-11 05:11:30,110 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:30,110 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,123 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,123 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,127 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,127 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,130 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:5] 2025-08-11 05:11:30,130 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:5] 2025-08-11 05:11:30,132 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:6] 2025-08-11 05:11:30,132 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:5] 2025-08-11 05:11:30,133 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:30,134 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:30,136 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,143 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,143 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,144 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,144 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,146 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:30,146 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-08-11 05:11:30,147 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-08-11 05:11:30,147 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:3] 2025-08-11 05:11:30,147 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:3] 2025-08-11 05:11:30,150 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:30,151 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,163 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,163 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,167 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,167 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,169 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:30,170 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-08-11 05:11:30,170 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-08-11 05:11:30,170 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:5] 2025-08-11 05:11:30,170 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:30,173 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:30,173 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,176 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,177 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,179 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,179 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,181 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:30,185 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:3] 2025-08-11 05:11:30,185 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-08-11 05:11:30,185 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:30,185 RepairJob.java:257 - Validating /10.0.0.48 INFO [RepairJobTask:3] 2025-08-11 05:11:30,186 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:3] 2025-08-11 05:11:30,186 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] leader is fully synced INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,190 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,190 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,194 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,194 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,196 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:30,196 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-08-11 05:11:30,196 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-08-11 05:11:30,196 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:1] 2025-08-11 05:11:30,197 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:30,437 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:30,437 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,440 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,440 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,442 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,442 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,445 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:30,447 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:1] 2025-08-11 05:11:30,447 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:1] 2025-08-11 05:11:30,448 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:4] 2025-08-11 05:11:30,449 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:1] 2025-08-11 05:11:30,449 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] running_repairs is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:30,449 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,455 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,455 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,457 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,457 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,458 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:6] 2025-08-11 05:11:30,458 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:3] 2025-08-11 05:11:30,458 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:7] 2025-08-11 05:11:30,459 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:3] 2025-08-11 05:11:30,459 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] cluster is fully synced INFO [RepairJobTask:3] 2025-08-11 05:11:30,463 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:30,464 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,473 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,474 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,476 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,476 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,478 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:30,478 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-08-11 05:11:30,478 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:5] 2025-08-11 05:11:30,478 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-08-11 05:11:30,478 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_unit_v1 is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:30,481 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:6] 2025-08-11 05:11:30,481 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,483 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,483 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,484 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,484 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,486 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:30,486 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:5] 2025-08-11 05:11:30,486 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-08-11 05:11:30,490 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-08-11 05:11:30,491 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:7] 2025-08-11 05:11:30,492 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:7] 2025-08-11 05:11:30,492 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,494 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,495 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,496 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,496 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,497 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:30,498 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-08-11 05:11:30,498 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:5] 2025-08-11 05:11:30,498 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:7] 2025-08-11 05:11:30,498 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] diagnostic_event_subscription is fully synced INFO [RepairJobTask:7] 2025-08-11 05:11:30,501 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:7] 2025-08-11 05:11:30,501 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,506 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,506 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,508 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,508 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,511 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:30,511 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:4] 2025-08-11 05:11:30,511 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:4] 2025-08-11 05:11:30,511 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:4] 2025-08-11 05:11:30,512 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:30,516 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for snapshot (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:30,516 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,519 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,519 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,521 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,521 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,523 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:30,523 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:6] 2025-08-11 05:11:30,523 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:7] 2025-08-11 05:11:30,524 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:6] 2025-08-11 05:11:30,524 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] snapshot is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:30,577 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:6] 2025-08-11 05:11:30,577 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,580 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,580 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,582 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,582 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,584 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:30,585 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:5] 2025-08-11 05:11:30,585 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:6] 2025-08-11 05:11:30,585 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:5] 2025-08-11 05:11:30,585 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_run is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:30,587 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for schema_migration (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:30,588 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,590 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,591 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,593 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,593 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,594 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:30,595 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:2] 2025-08-11 05:11:30,595 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:3] 2025-08-11 05:11:30,595 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:2] 2025-08-11 05:11:30,595 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] schema_migration is fully synced INFO [RepairJobTask:2] 2025-08-11 05:11:30,598 RepairJob.java:234 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:30,598 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,600 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,602 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,605 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,606 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:30,607 RepairSession.java:180 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:30,607 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:7] 2025-08-11 05:11:30,607 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:1] 2025-08-11 05:11:30,607 SyncTask.java:66 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:7] 2025-08-11 05:11:30,607 RepairJob.java:143 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:7] 2025-08-11 05:11:30,609 RepairSession.java:270 - [repair #a3d4ce20-7671-11f0-bff1-ef813b7a18da] Session completed successfully INFO [RepairJobTask:7] 2025-08-11 05:11:30,609 RepairRunnable.java:261 - Repair session a3d4ce20-7671-11f0-bff1-ef813b7a18da for range [(-6460304998359728668,-6427479982938010142], (-1662474736743062124,-1608583484966736162]] finished INFO [RepairJobTask:7] 2025-08-11 05:11:30,611 ActiveRepairService.java:452 - [repair #a3cc1b90-7671-11f0-bff1-ef813b7a18da] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-08-11 05:11:30,616 RepairRunnable.java:343 - Repair command #1 finished in 0 seconds INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,778 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.running_repairs INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,796 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.cluster INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,821 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.leader INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,861 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_unit_v1 INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,879 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_cluster_v2 INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,899 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.snapshot INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,911 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_cluster INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,938 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.percent_repaired_by_schedule INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,950 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.schema_migration INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,961 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.diagnostic_event_subscription INFO [AntiEntropyStage:1] 2025-08-11 05:11:33,975 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.running_reapers INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,032 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,054 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.schema_migration_leader INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,073 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_schedule_by_cluster_and_keyspace INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,086 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_run_by_unit INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,103 Validator.java:281 - [repair #a5fd0550-7671-11f0-993e-6d60ca3e2f4e] Sending completed merkle tree to /10.0.0.48 for reaper_db.repair_schedule_v1 INFO [AntiEntropyStage:1] 2025-08-11 05:11:34,118 ActiveRepairService.java:452 - [repair #a5f84a60-7671-11f0-993e-6d60ca3e2f4e] Not a global repair, will not do anticompaction INFO [Repair-Task-3] 2025-08-11 05:11:40,157 RepairRunnable.java:139 - Starting repair command #2 (a9db02d0-7671-11f0-bff1-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 2, pull repair: false) INFO [Repair-Task-3] 2025-08-11 05:11:40,192 RepairSession.java:228 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.48, /10.0.0.254 on range [(-5338176710596373213,-5287898245960497758], (3981848638898681815,3983533382731384201]] for reaper_db.[repair_run_by_cluster, running_reapers, schema_migration_leader, repair_schedule_v1, leader, repair_run_by_unit, running_repairs, cluster, repair_unit_v1, percent_repaired_by_schedule, diagnostic_event_subscription, repair_schedule_by_cluster_and_keyspace, snapshot, repair_run, schema_migration, repair_run_by_cluster_v2] INFO [RepairJobTask:2] 2025-08-11 05:11:40,276 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:40,277 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,283 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,283 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,287 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,287 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,292 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:40,292 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-08-11 05:11:40,292 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-08-11 05:11:40,292 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-08-11 05:11:40,292 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_run_by_cluster is fully synced INFO [RepairJobTask:1] 2025-08-11 05:11:40,300 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:1] 2025-08-11 05:11:40,301 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,318 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,318 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,322 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,323 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,331 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:40,331 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:4] 2025-08-11 05:11:40,332 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:3] 2025-08-11 05:11:40,332 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:4] 2025-08-11 05:11:40,332 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:40,341 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:4] 2025-08-11 05:11:40,341 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,350 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,351 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,353 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,353 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,357 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:40,361 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:2] 2025-08-11 05:11:40,361 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:5] 2025-08-11 05:11:40,361 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:2] 2025-08-11 05:11:40,361 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:40,364 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:40,367 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,391 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,392 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,393 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,393 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,399 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:40,400 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-08-11 05:11:40,400 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-08-11 05:11:40,401 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:2] 2025-08-11 05:11:40,401 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:40,406 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:40,406 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,409 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,409 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,410 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,410 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,413 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:40,413 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:2] 2025-08-11 05:11:40,413 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-08-11 05:11:40,413 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-08-11 05:11:40,414 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] leader is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:40,419 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:40,419 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,425 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,425 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,427 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,427 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,429 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,429 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-08-11 05:11:40,430 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:6] 2025-08-11 05:11:40,432 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-08-11 05:11:40,432 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:40,477 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:4] 2025-08-11 05:11:40,477 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,481 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,481 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,482 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,483 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,485 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,486 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:2] 2025-08-11 05:11:40,486 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:5] 2025-08-11 05:11:40,486 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:2] 2025-08-11 05:11:40,486 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] running_repairs is fully synced INFO [RepairJobTask:2] 2025-08-11 05:11:40,489 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:40,497 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,500 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,500 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,502 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,502 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,504 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,504 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:4] 2025-08-11 05:11:40,504 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:4] 2025-08-11 05:11:40,505 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:4] 2025-08-11 05:11:40,505 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] cluster is fully synced INFO [RepairJobTask:2] 2025-08-11 05:11:40,510 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_unit_v1 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:40,510 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,513 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,513 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,515 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,515 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,516 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_unit_v1 from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:40,516 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:6] 2025-08-11 05:11:40,517 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_unit_v1 INFO [RepairJobTask:7] 2025-08-11 05:11:40,517 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_unit_v1 INFO [RepairJobTask:4] 2025-08-11 05:11:40,517 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_unit_v1 is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:40,519 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for percent_repaired_by_schedule (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:4] 2025-08-11 05:11:40,519 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,522 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,523 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,525 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,525 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,526 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for percent_repaired_by_schedule from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,526 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-08-11 05:11:40,526 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-08-11 05:11:40,526 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for percent_repaired_by_schedule INFO [RepairJobTask:4] 2025-08-11 05:11:40,526 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] percent_repaired_by_schedule is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:40,532 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for diagnostic_event_subscription (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:6] 2025-08-11 05:11:40,532 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,536 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,536 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,538 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,556 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,558 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for diagnostic_event_subscription from /10.0.0.38 INFO [RepairJobTask:1] 2025-08-11 05:11:40,571 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for diagnostic_event_subscription INFO [RepairJobTask:3] 2025-08-11 05:11:40,572 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:2] 2025-08-11 05:11:40,572 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for diagnostic_event_subscription INFO [RepairJobTask:3] 2025-08-11 05:11:40,572 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] diagnostic_event_subscription is fully synced INFO [RepairJobTask:3] 2025-08-11 05:11:40,581 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_schedule_by_cluster_and_keyspace (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:40,585 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,588 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,589 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,591 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,591 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,594 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_by_cluster_and_keyspace from /10.0.0.38 INFO [RepairJobTask:6] 2025-08-11 05:11:40,594 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:7] 2025-08-11 05:11:40,594 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:5] 2025-08-11 05:11:40,594 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_schedule_by_cluster_and_keyspace INFO [RepairJobTask:7] 2025-08-11 05:11:40,594 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_schedule_by_cluster_and_keyspace is fully synced INFO [RepairJobTask:7] 2025-08-11 05:11:40,598 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for snapshot (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:7] 2025-08-11 05:11:40,599 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,602 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,602 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,604 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,604 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,610 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for snapshot from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,610 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for snapshot INFO [RepairJobTask:4] 2025-08-11 05:11:40,610 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:6] 2025-08-11 05:11:40,611 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for snapshot INFO [RepairJobTask:4] 2025-08-11 05:11:40,611 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] snapshot is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:40,671 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:4] 2025-08-11 05:11:40,671 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,674 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,674 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,678 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,679 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,686 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:40,686 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run INFO [RepairJobTask:5] 2025-08-11 05:11:40,686 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:7] 2025-08-11 05:11:40,688 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run INFO [RepairJobTask:6] 2025-08-11 05:11:40,689 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_run is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:40,691 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for schema_migration (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:6] 2025-08-11 05:11:40,691 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,696 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,697 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,698 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,698 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,700 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration from /10.0.0.38 INFO [RepairJobTask:6] 2025-08-11 05:11:40,703 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:5] 2025-08-11 05:11:40,703 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for schema_migration INFO [RepairJobTask:4] 2025-08-11 05:11:40,703 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for schema_migration INFO [RepairJobTask:5] 2025-08-11 05:11:40,703 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] schema_migration is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:40,706 RepairJob.java:234 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster_v2 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:40,706 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,712 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,712 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,714 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,717 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:40,728 RepairSession.java:180 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster_v2 from /10.0.0.38 INFO [RepairJobTask:5] 2025-08-11 05:11:40,729 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:6] 2025-08-11 05:11:40,730 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:7] 2025-08-11 05:11:40,730 SyncTask.java:66 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_cluster_v2 INFO [RepairJobTask:6] 2025-08-11 05:11:40,730 RepairJob.java:143 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] repair_run_by_cluster_v2 is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:40,731 RepairSession.java:270 - [repair #a9e05a00-7671-11f0-bff1-ef813b7a18da] Session completed successfully INFO [RepairJobTask:6] 2025-08-11 05:11:40,731 RepairRunnable.java:261 - Repair session a9e05a00-7671-11f0-bff1-ef813b7a18da for range [(-5338176710596373213,-5287898245960497758], (3981848638898681815,3983533382731384201]] finished INFO [RepairJobTask:6] 2025-08-11 05:11:40,732 ActiveRepairService.java:452 - [repair #a9db02d0-7671-11f0-bff1-ef813b7a18da] Not a global repair, will not do anticompaction INFO [InternalResponseStage:8] 2025-08-11 05:11:40,736 RepairRunnable.java:343 - Repair command #2 finished in 0 seconds INFO [Repair-Task-4] 2025-08-11 05:11:43,734 RepairRunnable.java:139 - Starting repair command #3 (abfcd160-7671-11f0-bff1-ef813b7a18da), repairing keyspace reaper_db with repair options (parallelism: dc_parallel, primary range: false, incremental: false, job threads: 1, ColumnFamilies: [cluster, leader, diagnostic_event_subscription, schema_migration_leader, running_repairs, repair_run, repair_unit_v1, repair_run_by_cluster_v2, repair_schedule_by_cluster_and_keyspace, repair_run_by_cluster, running_reapers, repair_schedule_v1, percent_repaired_by_schedule, repair_run_by_unit, schema_migration, snapshot], dataCenters: [], hosts: [], # of ranges: 1, pull repair: false) INFO [Repair-Task-4] 2025-08-11 05:11:43,741 RepairSession.java:228 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] new session: will sync /10.0.0.38, /10.0.0.48, /10.0.0.254 on range [(-3233960077794428066,-3166033446281240082]] for reaper_db.[repair_run_by_cluster, running_reapers, schema_migration_leader, repair_schedule_v1, leader, repair_run_by_unit, running_repairs, cluster, repair_unit_v1, percent_repaired_by_schedule, diagnostic_event_subscription, repair_schedule_by_cluster_and_keyspace, snapshot, repair_run, schema_migration, repair_run_by_cluster_v2] INFO [RepairJobTask:3] 2025-08-11 05:11:43,783 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:43,784 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,789 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,792 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,802 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,803 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,807 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_cluster from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:43,808 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:5] 2025-08-11 05:11:43,808 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_cluster INFO [RepairJobTask:1] 2025-08-11 05:11:43,808 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_cluster INFO [RepairJobTask:4] 2025-08-11 05:11:43,812 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] repair_run_by_cluster is fully synced INFO [RepairJobTask:5] 2025-08-11 05:11:43,812 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_reapers (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:5] 2025-08-11 05:11:43,815 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,823 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,824 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,834 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,834 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,836 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_reapers from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:43,836 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_reapers INFO [RepairJobTask:2] 2025-08-11 05:11:43,837 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:2] 2025-08-11 05:11:43,837 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_reapers INFO [RepairJobTask:2] 2025-08-11 05:11:43,837 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] running_reapers is fully synced INFO [RepairJobTask:2] 2025-08-11 05:11:43,844 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for schema_migration_leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:2] 2025-08-11 05:11:43,844 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,853 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,856 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,863 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,864 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,866 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for schema_migration_leader from /10.0.0.38 INFO [RepairJobTask:2] 2025-08-11 05:11:43,869 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-08-11 05:11:43,870 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for schema_migration_leader INFO [RepairJobTask:1] 2025-08-11 05:11:43,870 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for schema_migration_leader INFO [RepairJobTask:4] 2025-08-11 05:11:43,870 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] schema_migration_leader is fully synced INFO [RepairJobTask:1] 2025-08-11 05:11:43,876 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_schedule_v1 (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:1] 2025-08-11 05:11:43,877 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,883 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,883 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,889 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,889 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,891 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_schedule_v1 from /10.0.0.38 INFO [RepairJobTask:5] 2025-08-11 05:11:43,892 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_schedule_v1 INFO [RepairJobTask:4] 2025-08-11 05:11:43,892 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:4] 2025-08-11 05:11:43,892 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_schedule_v1 INFO [RepairJobTask:4] 2025-08-11 05:11:43,892 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] repair_schedule_v1 is fully synced INFO [RepairJobTask:4] 2025-08-11 05:11:43,911 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for leader (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:4] 2025-08-11 05:11:43,911 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,915 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,915 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,937 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,937 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,940 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for leader from /10.0.0.38 INFO [RepairJobTask:5] 2025-08-11 05:11:43,940 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:2] 2025-08-11 05:11:43,940 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for leader INFO [RepairJobTask:4] 2025-08-11 05:11:43,940 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for leader INFO [RepairJobTask:3] 2025-08-11 05:11:43,941 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] leader is fully synced INFO [RepairJobTask:3] 2025-08-11 05:11:43,945 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for repair_run_by_unit (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:43,946 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,957 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,958 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,959 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,960 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:43,965 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for repair_run_by_unit from /10.0.0.38 INFO [RepairJobTask:3] 2025-08-11 05:11:43,966 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-08-11 05:11:43,966 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:2] 2025-08-11 05:11:43,966 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for repair_run_by_unit INFO [RepairJobTask:4] 2025-08-11 05:11:43,967 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] repair_run_by_unit is fully synced INFO [RepairJobTask:3] 2025-08-11 05:11:44,037 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for running_repairs (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:3] 2025-08-11 05:11:44,037 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,040 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,041 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,044 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,044 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,050 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for running_repairs from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:44,052 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:6] 2025-08-11 05:11:44,053 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for running_repairs INFO [RepairJobTask:3] 2025-08-11 05:11:44,054 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for running_repairs INFO [RepairJobTask:6] 2025-08-11 05:11:44,055 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] running_repairs is fully synced INFO [RepairJobTask:6] 2025-08-11 05:11:44,062 RepairJob.java:234 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Requesting merkle trees for cluster (to [/10.0.0.48, /10.0.0.254, /10.0.0.38]) INFO [RepairJobTask:6] 2025-08-11 05:11:44,062 RepairJob.java:257 - Validating /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,066 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.48 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,066 RepairJob.java:270 - Validating /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,068 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.254 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,068 RepairJob.java:270 - Validating /10.0.0.38 INFO [AntiEntropyStage:1] 2025-08-11 05:11:44,072 RepairSession.java:180 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Received merkle tree for cluster from /10.0.0.38 INFO [RepairJobTask:4] 2025-08-11 05:11:44,074 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.254 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:2] 2025-08-11 05:11:44,074 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.38 are consistent for cluster INFO [RepairJobTask:6] 2025-08-11 05:11:44,074 SyncTask.java:66 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] Endpoints /10.0.0.48 and /10.0.0.254 are consistent for cluster INFO [RepairJobTask:2] 2025-08-11 05:11:44,074 RepairJob.java:143 - [repair #abfde2d0-7671-11f0-bff1-ef813b7a18da] cluster is fully synced