==> ./dashboard.log <==
2021-05-06 11:10:06,525 INFO stats_collector_head.py:186 -- Getting all actor info from GCS.
2021-05-06 11:10:06,526 INFO stats_collector_head.py:284 -- Subscribed to <_Sender name:b'RAY_LOG_CHANNEL', is_pattern:False, receiver:<Receiver is_active:True, senders:1, qsize:0>>
2021-05-06 11:10:06,526 INFO stats_collector_head.py:307 -- Subscribed to b'ERROR_INFO:*'
2021-05-06 11:10:06,526 INFO stats_collector_head.py:212 -- Received 0 actor info from GCS.
2021-05-06 11:10:06,527 INFO job_head.py:89 -- Received 0 job info from GCS.
2021-05-06 11:10:06,963 INFO stats_collector_head.py:296 -- Received a log for 172.20.201.74 and autoscaler
2021-05-06 11:10:08,124 INFO stats_collector_head.py:296 -- Received a log for 172.20.201.74 and raylet
2021-05-06 11:20:06,529 INFO datacenter.py:65 -- Purge data.
2021-05-06 11:30:06,535 INFO datacenter.py:65 -- Purge data.
2021-05-06 11:40:06,537 INFO datacenter.py:65 -- Purge data.
2021-05-06 11:50:06,539 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:00:06,541 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:10:06,545 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:20:06,552 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:30:06,557 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:40:06,562 INFO datacenter.py:65 -- Purge data.
2021-05-06 12:50:06,569 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:00:06,571 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:10:06,572 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:20:06,573 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:30:06,576 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:40:06,578 INFO datacenter.py:65 -- Purge data.
2021-05-06 13:50:06,588 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:00:06,594 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:10:06,602 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:20:06,602 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:30:06,604 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:40:06,605 INFO datacenter.py:65 -- Purge data.
2021-05-06 14:50:06,606 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:00:06,610 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:10:06,619 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:20:06,623 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:30:06,630 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:40:06,634 INFO datacenter.py:65 -- Purge data.
2021-05-06 15:50:06,642 INFO datacenter.py:65 -- Purge data.
2021-05-06 16:00:06,646 INFO datacenter.py:65 -- Purge data.
2021-05-06 16:10:06,647 INFO datacenter.py:65 -- Purge data.
2021-05-06 16:20:06,652 INFO datacenter.py:65 -- Purge data.
2021-05-06 16:30:06,657 INFO datacenter.py:65 -- Purge data.
2021-05-06 16:33:23,681 ERROR stats_collector_head.py:276 -- Error updating node stats of 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb.
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 269, in _update_node_stats
reply = await stub.GetNodeStats(
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/grpc/aio/_call.py", line 285, in __await__
raise _create_rpc_error(self._cython_call._initial_metadata,
grpc.aio._call.AioRpcError: <AioRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "failed to connect to all addresses"
debug_error_string = "{"created":"@1620336803.681321755","description":"Failed to pick subchannel","file":"src/core/ext/filters/client_channel/client_channel.cc","file_line":5419,"referenced_errors":[{"created":"@1620336803.681318637","description":"failed to connect to all addresses","file":"src/core/ext/filters/client_channel/lb_policy/pick_first/pick_first.cc","file_line":397,"grpc_status":14}]}"
>
==> ./gcs_server.err <==
==> ./gcs_server.out <==
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:29:06,552 I 23823 23823] gcs_server.cc:379: GcsNodeManager: {RegisterNode request count: 1, UnregisterNode request count: 0, GetAllNodeInfo request count: 7618, GetInternalConfig request count: 1}
GcsActorManager: {RegisterActor request count: 0, CreateActor request count: 0, GetActorInfo request count: 0, GetNamedActorInfo request count: 0, KillActor request count: 0, Registered actors count: 0, Destroyed actors count: 0, Named actors count: 0, Unresolved actors count: 0, Pending actors count: 0, Created actors count: 0}
GcsObjectManager: {GetObjectLocations request count: 0, GetAllObjectLocations request count: 0, AddObjectLocation request count: 0, RemoveObjectLocation request count: 0, Object count: 0}
GcsPlacementGroupManager: {CreatePlacementGroup request count: 0, RemovePlacementGroup request count: 0, GetPlacementGroup request count: 0, GetAllPlacementGroup request count: 0, WaitPlacementGroupUntilReady request count: 0, Registered placement groups count: 0, Named placement group count: 0, Pending placement groups count: 0}
GcsPubSub:
- num channels subscribed to: 0
- total commands queued: 0
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:30:06,556 I 23823 23823] gcs_server.cc:379: GcsNodeManager: {RegisterNode request count: 1, UnregisterNode request count: 0, GetAllNodeInfo request count: 7642, GetInternalConfig request count: 1}
GcsActorManager: {RegisterActor request count: 0, CreateActor request count: 0, GetActorInfo request count: 0, GetNamedActorInfo request count: 0, KillActor request count: 0, Registered actors count: 0, Destroyed actors count: 0, Named actors count: 0, Unresolved actors count: 0, Pending actors count: 0, Created actors count: 0}
GcsObjectManager: {GetObjectLocations request count: 0, GetAllObjectLocations request count: 0, AddObjectLocation request count: 0, RemoveObjectLocation request count: 0, Object count: 0}
GcsPlacementGroupManager: {CreatePlacementGroup request count: 0, RemovePlacementGroup request count: 0, GetPlacementGroup request count: 0, GetAllPlacementGroup request count: 0, WaitPlacementGroupUntilReady request count: 0, Registered placement groups count: 0, Named placement group count: 0, Pending placement groups count: 0}
GcsPubSub:
- num channels subscribed to: 0
- total commands queued: 0
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:31:06,556 I 23823 23823] gcs_server.cc:379: GcsNodeManager: {RegisterNode request count: 1, UnregisterNode request count: 0, GetAllNodeInfo request count: 7666, GetInternalConfig request count: 1}
GcsActorManager: {RegisterActor request count: 0, CreateActor request count: 0, GetActorInfo request count: 0, GetNamedActorInfo request count: 0, KillActor request count: 0, Registered actors count: 0, Destroyed actors count: 0, Named actors count: 0, Unresolved actors count: 0, Pending actors count: 0, Created actors count: 0}
GcsObjectManager: {GetObjectLocations request count: 0, GetAllObjectLocations request count: 0, AddObjectLocation request count: 0, RemoveObjectLocation request count: 0, Object count: 0}
GcsPlacementGroupManager: {CreatePlacementGroup request count: 0, RemovePlacementGroup request count: 0, GetPlacementGroup request count: 0, GetAllPlacementGroup request count: 0, WaitPlacementGroupUntilReady request count: 0, Registered placement groups count: 0, Named placement group count: 0, Pending placement groups count: 0}
GcsPubSub:
- num channels subscribed to: 0
- total commands queued: 0
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:32:06,556 I 23823 23823] gcs_server.cc:379: GcsNodeManager: {RegisterNode request count: 1, UnregisterNode request count: 0, GetAllNodeInfo request count: 7690, GetInternalConfig request count: 1}
GcsActorManager: {RegisterActor request count: 0, CreateActor request count: 0, GetActorInfo request count: 0, GetNamedActorInfo request count: 0, KillActor request count: 0, Registered actors count: 0, Destroyed actors count: 0, Named actors count: 0, Unresolved actors count: 0, Pending actors count: 0, Created actors count: 0}
GcsObjectManager: {GetObjectLocations request count: 0, GetAllObjectLocations request count: 0, AddObjectLocation request count: 0, RemoveObjectLocation request count: 0, Object count: 0}
GcsPlacementGroupManager: {CreatePlacementGroup request count: 0, RemovePlacementGroup request count: 0, GetPlacementGroup request count: 0, GetAllPlacementGroup request count: 0, WaitPlacementGroupUntilReady request count: 0, Registered placement groups count: 0, Named placement group count: 0, Pending placement groups count: 0}
GcsPubSub:
- num channels subscribed to: 0
- total commands queued: 0
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:33:06,556 I 23823 23823] gcs_server.cc:379: GcsNodeManager: {RegisterNode request count: 1, UnregisterNode request count: 0, GetAllNodeInfo request count: 7714, GetInternalConfig request count: 1}
GcsActorManager: {RegisterActor request count: 0, CreateActor request count: 0, GetActorInfo request count: 0, GetNamedActorInfo request count: 0, KillActor request count: 0, Registered actors count: 0, Destroyed actors count: 0, Named actors count: 0, Unresolved actors count: 0, Pending actors count: 0, Created actors count: 0}
GcsObjectManager: {GetObjectLocations request count: 0, GetAllObjectLocations request count: 0, AddObjectLocation request count: 0, RemoveObjectLocation request count: 0, Object count: 0}
GcsPlacementGroupManager: {CreatePlacementGroup request count: 0, RemovePlacementGroup request count: 0, GetPlacementGroup request count: 0, GetAllPlacementGroup request count: 0, WaitPlacementGroupUntilReady request count: 0, Registered placement groups count: 0, Named placement group count: 0, Pending placement groups count: 0}
GcsPubSub:
- num channels subscribed to: 0
- total commands queued: 0
DefaultTaskInfoHandler: {AddTask request count: 0, GetTask request count: 0, AddTaskLease request count: 0, GetTaskLease request count: 0, AttemptTaskReconstruction request count: 0}
[2021-05-06 16:33:22,931 I 23823 23823] gcs_node_manager.cc:55: Unregistering node info, node id = 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb
[2021-05-06 16:33:22,931 I 23823 23823] gcs_node_manager.cc:136: Removing node, node id = 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb
[2021-05-06 16:33:22,931 I 23823 23823] gcs_placement_group_manager.cc:532: Node 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb failed, rescheduling the placement groups on the dead node.
[2021-05-06 16:33:22,931 I 23823 23823] gcs_actor_manager.cc:606: Node 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb failed, reconstructing actors.
[2021-05-06 16:33:22,931 I 23823 23823] gcs_node_manager.cc:72: Finished unregistering node info, node id = 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb
[2021-05-06 16:33:23,044 I 23823 23823] gcs_server_main.cc:111: GCS server received SIGTERM, shutting down...
[2021-05-06 16:33:23,044 I 23823 23823] gcs_server.cc:135: Stopping GCS server.
[2021-05-06 16:33:23,049 I 23823 23823] gcs_server.cc:142: GCS server stopped.
[2021-05-06 16:33:23,049 I 23823 23823] io_service_pool.cc:47: IOServicePool is stopped.
==> ./log_monitor.log <==
2021-05-06 11:10:06,960 INFO log_monitor.py:162 -- Beginning to track file raylet.err
2021-05-06 11:10:06,961 INFO log_monitor.py:162 -- Beginning to track file gcs_server.err
2021-05-06 11:10:06,961 INFO log_monitor.py:162 -- Beginning to track file monitor.log
==> ./monitor.err <==
return self._handle_failure(f"Terminated with signal {sig}\n" +
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 264, in _handle_failure
_internal_kv_put(DEBUG_AUTOSCALING_ERROR, message, overwrite=True)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/client_mode_hook.py", line 47, in wrapper
return func(*args, **kwargs)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/experimental/internal_kv.py", line 56, in _internal_kv_put
updated = ray.worker.global_worker.redis_client.hset(
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 3050, in hset
return self.execute_command('HSET', name, *items)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 898, in execute_command
conn = self.connection or pool.get_connection(command_name, **options)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 1202, in get_connection
connection.connect()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 563, in connect
raise ConnectionError(self._error_message(e))
redis.exceptions.ConnectionError: Error 111 connecting to 172.20.201.74:6379. Connection refused.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 559, in connect
sock = self._connect()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 615, in _connect
raise err
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 603, in _connect
sock.connect(socket_address)
ConnectionRefusedError: [Errno 111] Connection refused
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 376, in <module>
monitor.run()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 286, in run
self._handle_failure(traceback.format_exc())
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 264, in _handle_failure
_internal_kv_put(DEBUG_AUTOSCALING_ERROR, message, overwrite=True)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/client_mode_hook.py", line 47, in wrapper
return func(*args, **kwargs)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/experimental/internal_kv.py", line 56, in _internal_kv_put
updated = ray.worker.global_worker.redis_client.hset(
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 3050, in hset
return self.execute_command('HSET', name, *items)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 898, in execute_command
conn = self.connection or pool.get_connection(command_name, **options)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 1192, in get_connection
connection.connect()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 563, in connect
raise ConnectionError(self._error_message(e))
redis.exceptions.ConnectionError: Error 111 connecting to 172.20.201.74:6379. Connection refused.
==> ./monitor.log <==
2021-05-06 11:10:05,831 INFO monitor.py:122 -- Monitor: Started
2021-05-06 16:33:23,337 ERROR monitor.py:253 -- Error in monitor loop
NoneType: None
2021-05-06 16:33:23,340 ERROR monitor.py:253 -- Error in monitor loop
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 1198, in get_connection
if connection.can_read():
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 734, in can_read
return self._parser.can_read(timeout)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 416, in can_read
return self.read_from_socket(timeout=timeout,
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 429, in read_from_socket
raise ConnectionError(SERVER_CLOSED_CONNECTION_ERROR)
redis.exceptions.ConnectionError: Connection closed by server.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 559, in connect
sock = self._connect()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 615, in _connect
raise err
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 603, in _connect
sock.connect(socket_address)
ConnectionRefusedError: [Errno 111] Connection refused
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 284, in run
self._run()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 202, in _run
time.sleep(AUTOSCALER_UPDATE_INTERVAL_S)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 272, in _signal_handler
return self._handle_failure(f"Terminated with signal {sig}\n" +
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/monitor.py", line 264, in _handle_failure
_internal_kv_put(DEBUG_AUTOSCALING_ERROR, message, overwrite=True)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/_private/client_mode_hook.py", line 47, in wrapper
return func(*args, **kwargs)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/experimental/internal_kv.py", line 56, in _internal_kv_put
updated = ray.worker.global_worker.redis_client.hset(
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 3050, in hset
return self.execute_command('HSET', name, *items)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/client.py", line 898, in execute_command
conn = self.connection or pool.get_connection(command_name, **options)
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 1202, in get_connection
connection.connect()
File "/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/redis/connection.py", line 563, in connect
raise ConnectionError(self._error_message(e))
redis.exceptions.ConnectionError: Error 111 connecting to 172.20.201.74:6379. Connection refused.
==> ./monitor.out <==
==> ./old <==
tail: error reading ‘./old’: Is a directory
==> ./plasma_store.err <==
[2021-05-06 16:33:24,128 E 23867 23867] logging.cc:435: *** Aborted at 1620336804 (unix time) try "date -d @1620336804" if you are using GNU date ***
[2021-05-06 16:33:24,129 E 23867 23867] logging.cc:435: PC: @ 0x0 (unknown)
[2021-05-06 16:33:24,129 E 23867 23867] logging.cc:435: *** SIGTERM (@0x2d2800005cff) received by PID 23867 (TID 0x2aaaaaaf9b00) from PID 23807; stack trace: ***
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x5555555f677f google::(anonymous namespace)::FailureSignalHandler()
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x2aaaaacde630 (unknown)
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x2aaaaacdde80 __nanosleep_nocancel
[2021-05-06 16:33:24,135 E 23867 23867] logging.cc:435: @ 0x55555557049b main
[2021-05-06 16:33:24,135 E 23867 23867] logging.cc:435: @ 0x2aaaab7ab555 __libc_start_main
[2021-05-06 16:33:24,136 E 23867 23867] logging.cc:435: @ 0x555555572d35 (unknown)
==> ./plasma_store.out <==
[2021-05-06 11:10:06,619 I 23867 23867] store_exec.cc:81: The Plasma Store is started with the '-z' flag, and it will run idle as a placeholder.
[2021-05-06 16:33:24,128 E 23867 23867] logging.cc:435: *** Aborted at 1620336804 (unix time) try "date -d @1620336804" if you are using GNU date ***
[2021-05-06 16:33:24,129 E 23867 23867] logging.cc:435: PC: @ 0x0 (unknown)
[2021-05-06 16:33:24,129 E 23867 23867] logging.cc:435: *** SIGTERM (@0x2d2800005cff) received by PID 23867 (TID 0x2aaaaaaf9b00) from PID 23807; stack trace: ***
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x5555555f677f google::(anonymous namespace)::FailureSignalHandler()
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x2aaaaacde630 (unknown)
[2021-05-06 16:33:24,134 E 23867 23867] logging.cc:435: @ 0x2aaaaacdde80 __nanosleep_nocancel
[2021-05-06 16:33:24,135 E 23867 23867] logging.cc:435: @ 0x55555557049b main
[2021-05-06 16:33:24,135 E 23867 23867] logging.cc:435: @ 0x2aaaab7ab555 __libc_start_main
[2021-05-06 16:33:24,136 E 23867 23867] logging.cc:435: @ 0x555555572d35 (unknown)
==> ./ray_client_server.err <==
/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/autoscaler/_private/cli_logger.py:57: FutureWarning: Not all Ray CLI dependencies were found. In Ray 1.4+, the Ray CLI, autoscaler, and dashboard will only be usable via `pip install 'ray[default]'`. Please update your install command.
warnings.warn(
INFO:ray.util.client.server.server:Starting Ray Client server on 0.0.0.0:10001
==> ./ray_client_server.out <==
==> ./raylet.err <==
/home/tmamidi/.conda/envs/training/lib/python3.8/site-packages/ray/autoscaler/_private/cli_logger.py:57: FutureWarning: Not all Ray CLI dependencies were found. In Ray 1.4+, the Ray CLI, autoscaler, and dashboard will only be usable via `pip install 'ray[default]'`. Please update your install command.
warnings.warn(
==> ./raylet.out <==
[2021-05-06 11:10:06,629 I 23868 23868] io_service_pool.cc:35: IOServicePool is running with 1 io_service.
[2021-05-06 11:10:06,638 I 23868 23868] store_runner.cc:29: Allowing the Plasma store to use up to 79.4956GB of memory.
[2021-05-06 11:10:06,638 I 23868 23868] store_runner.cc:42: Starting object store with directory /dev/shm and huge page support disabled
[2021-05-06 11:10:07,639 I 23868 23868] grpc_server.cc:71: ObjectManager server started, listening on port 34945.
[2021-05-06 11:10:07,656 I 23868 23868] node_manager.cc:230: Initializing NodeManager with ID 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb
[2021-05-06 11:10:07,656 I 23868 23868] grpc_server.cc:71: NodeManager server started, listening on port 45147.
[2021-05-06 11:10:07,660 I 23868 23917] agent_manager.cc:76: Monitor agent process with pid 23916, register timeout 30000ms.
[2021-05-06 11:10:07,661 I 23868 23868] raylet.cc:146: Raylet of id, 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb started. Raylet consists of node_manager and object_manager. node_manager address: 172.20.201.74:45147 object_manager address: 172.20.201.74:34945 hostname: 172.20.201.74
[2021-05-06 11:10:07,664 I 23868 23868] service_based_accessor.cc:579: Received notification for node id = 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb, IsAlive = 1
[2021-05-06 11:10:08,562 I 23868 23868] agent_manager.cc:32: HandleRegisterAgent, ip: 172.20.201.74, port: 61122, pid: 23916
[2021-05-06 11:20:07,684 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 11:30:07,771 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 11:40:07,847 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 11:50:07,885 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:00:07,969 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:10:08,002 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:20:08,084 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:30:08,182 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:40:08,217 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 12:50:08,223 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:00:08,285 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:10:08,330 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:20:08,429 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:30:08,458 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:40:08,555 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 13:50:08,603 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:00:08,605 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:10:08,687 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:20:08,742 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:30:08,796 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:40:08,855 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 14:50:08,929 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:00:08,934 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:10:09,009 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:20:09,056 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:30:09,071 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:40:09,109 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 15:50:09,119 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 16:00:09,132 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 16:10:09,202 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 16:20:09,301 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 16:30:09,370 I 23868 23868] node_manager.cc:541: Sending Python GC request to 0 local workers to clean up Python cyclic references.
[2021-05-06 16:33:22,931 I 23868 23868] main.cc:254: Raylet received SIGTERM, shutting down...
[2021-05-06 16:33:22,931 I 23868 23868] service_based_accessor.cc:403: Unregistering node info, node id = 2b0737044534403624d8ebdf5244d07d18da84381b03741115f514bb
[2021-05-06 16:33:22,931 I 23868 23868] io_service_pool.cc:47: IOServicePool is stopped.
==> ./redis.err <==
==> ./redis.out <==
23813:C 06 May 2021 11:10:05.180 # oO0OoO0OoO0Oo Redis is starting oO0OoO0OoO0Oo
23813:C 06 May 2021 11:10:05.181 # Redis version=6.0.10, bits=64, commit=00000000, modified=0, pid=23813, just started
23813:C 06 May 2021 11:10:05.181 # Configuration loaded
23813:M 06 May 2021 11:10:05.181 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.
23813:M 06 May 2021 11:10:05.182 # Server initialized
23813:M 06 May 2021 11:10:05.182 # WARNING overcommit_memory is set to 0! Background save may fail under low memory condition. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect.
23813:M 06 May 2021 11:10:05.182 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo madvise > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled (set to 'madvise' or 'never').
23813:signal-handler (1620336803) Received SIGTERM scheduling shutdown...
23813:M 06 May 2021 16:33:23.174 # User requested shutdown...
23813:M 06 May 2021 16:33:23.174 # Redis is now ready to exit, bye bye...