- Troubleshooting PGs — Ceph Documentation
List stuck placement groups by running one of the following commands: Stuck stale placement groups usually indicate that key ceph-osd daemons are not running Stuck inactive placement groups usually indicate a peering problem (see Placement Group Down - Peering Failure)
- Most common Ceph placement groups errors - IBM
Placement groups that are stuck in a state that is not optimal can be listed, as described in Listing placement groups stuck in stale, inactive, or unclean states Note: When troubleshooting Ceph placement group errors, be sure you have both a running IBM Storage Ceph cluster and Ceph Object Gateway
- Placement Groups — Ceph Documentation
Each pool has a pg_autoscale_mode property that can be set to off, on, or warn: off: Disable autoscaling for this pool It is up to the administrator to choose an appropriate pg_num for each pool For more information, see Choosing the Number of PGs on: Enable automated adjustments of the PG count for the given pool
- Chapter 6. Troubleshooting Placement Groups - Red Hat
Insufficient Placement Group (PG) count impacts the performance of the Ceph cluster and data distribution It is one of the main causes of the nearfull osds error messages
- Troubleshooting placement groups (PGs) | SES 7
If you are trying to set up a 1-node cluster and osd crush chooseleaf type is greater than 0, Ceph tries to pair the PGs of one OSD with the PGs of another OSD on another node, chassis, rack, row, or even datacenter depending on the setting
- [ceph-users] Cluster does not report which objects are unfound for stuck PG
Specifically, the output of the command is this: # ceph pg 5 658 list_missing { "offset": { "oid": "", "key": "", "snapid": 0, "hash": 0, "max": 0, "pool": -1, "namespace": "" }, "num_missing": 0, "num_unfound": 1, "objects": [], "more": 0 } I took a look on ceph's official docs and on older threads on this list, but on every case that I found
- Chapter 9. Troubleshooting Ceph placement groups - Red Hat
Most common Ceph placement groups errors The following table lists the most common errors messages that are returned by the ceph health detail command The table provides links to corresponding sections that explain the errors and point to specific procedures to fix the problems
- Repairing PG Inconsistencies — Ceph Documentation
To see a list of inconsistent snapsets in a specific PG, run the following commands: The form of the command to repair a broken PG is as follows: Here {pgid} represents the id of the affected PG For example: PG IDs have the form N xxxxx, where N is the number of the pool that contains the PG
|