"""Fault tolerance test (small cluster, non-elastic training) In this run, two training actors will die after some time. It is expected that in both cases lightgbm_ray stops training, restarts the dead actors, and continues training with all four actors. Test owner: Yard1 (primary), krfricke Acceptance criteria: Should run through and report final results. Intermediate output should show that training halts wenn an actor dies and continues only when all four actors are available again. The test will fail if fault tolerance did not work correctly. Notes: This test seems to be somewhat flaky. This might be due to race conditions in handling dead actors. This is likely a problem of the lightgbm_ray implementation and not of this test. """ import os import ray from lightgbm_ray import RayParams from release_test_util import ( train_ray, FailureState, FailureInjection, TrackingCallback, ) if __name__ == "__main__": ray.init(address="auto", runtime_env={"working_dir": os.path.dirname(__file__)}) failure_state = FailureState.remote() ray_params = RayParams( max_actor_restarts=2, num_actors=4, cpus_per_actor=4, gpus_per_actor=0 ) _, additional_results, _ = train_ray( path="/data/classification.parquet", num_workers=None, num_boost_rounds=100, num_files=200, regression=False, use_gpu=False, ray_params=ray_params, lightgbm_params=None, callbacks=[ TrackingCallback(), FailureInjection( id="first_fail", state=failure_state, ranks=[1], iteration=14 ), FailureInjection( id="second_fail", state=failure_state, ranks=[0], iteration=34 ), ], ) print("PASSED.")