TEXT   98
ceph osd pool get
Guest on 8th February 2023 12:55:47 PM


  1. ceph osd pool get default.rgw.buckets.data all
  2. size: 3
  3. min_size: 2
  4. pg_num: 32
  5. pgp_num: 32
  6. crush_rule: ectest
  7. hashpspool: true
  8. allow_ec_overwrites: true
  9. nodelete: false
  10. nopgchange: false
  11. nosizechange: false
  12. write_fadvise_dontneed: false
  13. noscrub: false
  14. nodeep-scrub: false
  15. use_gmt_hitset: 1
  16. erasure_code_profile: hdd-raid5
  17. fast_read: 0
  18. pg_autoscale_mode: warn
  19. root@kube:~# ceph osd erasure-code-profile get hdd-raid5
  20. crush-device-class=hdd
  21. crush-failure-domain=host
  22. crush-root=default
  23. jerasure-per-chunk-alignment=false
  24. k=2
  25. m=1
  26. plugin=jerasure
  27. technique=reed_sol_van
  28. w=6

Raw Paste

Login or Register to edit or fork this paste. It's free.