Issue with PBM for sharded cluster

Hello i encountered an error with PBM, i have two shard replica set and one configsvr
when i do pbm status i got the following
Cluster:

rsconfig:

  • rsconfig/otvme061s:27019: pbm-agent v1.4.1 OK
  • rsconfig/otvme063s:27019: pbm-agent v1.4.1 OK
    RS2:
  • RS2/otvme061s:27017: pbm-agent v1.4.1 OK
  • RS2/otvme062s:27017: pbm-agent v1.4.1 OK
    RS1:
  • RS1/otvme063s:27017: pbm-agent v1.4.1 OK
  • RS1/otvme064s:27017: pbm-agent v1.4.1 OK

So all is good
but in backup part i got this
Backups:

FS /mnt
Snapshots:
2021-05-18T14:56:19Z 0.00B [ERROR: get file 2021-05-18T14:56:19Z_rsconfig.dump.s2: no such file] [2021-05-18T14:56:39]

What should i do, cause it seems that my issue occured cause i didnā€™t create filesystem what should i do ? and where should i put something, i follow all the percona docs and run all step one by one but i didnā€™t see anything about this
Thanks you !

2 Likes

Hi.

The details are a little thin at the moment, but it sounds like you might not have configured a shared remote storage (an object store, or a shared filesystem mounted at exactly the same path on all servers) yet.

Akira

1 Like

Hi.
Thank you for the answer!!
I understand the issue and iā€™m searching to shared a filesystem mounted at exactly the same path on all my server, but i didnā€™t success yet have you got a tuto or something, to do it works with percona ?
Thanks for time

1 Like

If youā€™ve mounted the same shared filesystem on all the servers at the same path then that should address the error situation I was thinking of.

Which server did the following error happen on?

2021-05-18T14:56:19Z 0.00B [ERROR: get file 2021-05-18T14:56:19Z_rsconfig.dump.s2: no such file]

Does this server have a file ā€œ2021-05-18T14:56:19Z_rsconfig.dump.s2ā€ in the configured filesystem backup path (/mnt/?) but it reported that ā€œno such fileā€ error nonetheless?

1 Like

Hello, thanks, i think i solved my issue but now i get this error :
ERROR with storage: storage: no init file, attempt to create failed: create destination file </data/.pbm.init>: open /data/.pbm.init: permission denied.
it seems that i had the a problem with storage but i insert the config with pbm command and i got all my remote filesystem on the same local path and i also allow pbm users authentification, the problems occurs on all my server.

1 Like

ā€œpbm.initā€ is an empty test file written into the remote store that will be used by status-checking functions to quickly test that the storage is up.

Iā€™m guessing the /data directory is neither owned by the process running pbm-agent, nor has write permission on it been given to ā€˜otherā€™ users. So the write of pbm.init, the last time storage was synced, failed then.

Please consider well the unix directory permissions. Thereā€™s a lot of unix admin consideration required to make a shared filesystem identical mounted and with the correct permissions on every server. This is why I recommend an object store as the remote backup storage. People donā€™t want to set up a bucket and configure its connection credentials just to test PBM, I get that, but setting up a shared filesystem is probably harder.

1 Like

systemctl start pbm-agentļ¼ŒIt will automatically create a Backup Storage directory if you want to use a local file system as your storage solution. This error will occur if you create the backup storage directory yourself. The correct owner and group should be mongod and mongod. You should not create this directory manually, let Systemctl start pbm-agent create it itself.
In addition, you should deploy an NFS first and see if you can solve the problem.

Sorry, the problem persists even though I have mounted the NFS service provided on Mongos in all sharding cluster members.

However, this ā€œno such fileā€ problem does not affect the backup and restore at all. It is just strange to see the ERROR in the PBM status output.