Details
-
Bug
-
Status: Closed (View Workflow)
-
Major
-
Resolution: Duplicate
-
5.4.3
-
None
Description
After a primary node failover, non-primary nodes are occasionally failing to loadbrm.
The mcs-loadbrm.service outputs the following:
Mar 22 21:22:04 pm3 mcs-loadbrm.py[29596]: Loading BRM snapshot failed (/tmp/columnstore_tmp_files/rdwrscratch/BRM_saves)
Mar 22 21:22:04 pm3 mcs-loadbrm.py[29596]: ExtentMap::load(): That file is not a valid ExtentMap image
There are no other indications that there was an error, until queries are attempted on the system.
The workaround is to restart the cluster via cmapi cluster/stop, cluster/start
Attachments
Issue Links
- relates to
-
MCOL-4440 Multi-Node CS 5.4 with S3 storage failover gets stuck, requires API call to restart cluster
- Closed