Proxmox ZFS Pool Wear Level very high (?)!

I have changed my Proxmox setup recently to a ZFS Mirror as Boot Device and VM storage consisting of 2x1TB WD Red SN700 NVMEs. I know that using ZFS with consumer grade SSDs is not the best solution but the wear levels of the two SSDs is rising so fast that I think I have misconfigured something.

Currently 125GB of the 1TB are in use and the pool has a fragmentation of 15%.

Output of smartctl for one of the new disks I installed 17.01.2025 (same for the other / mirror):

  • Percentage Used: 4%
  • Data Units Read: 2,004,613 [1.02 TB]
  • Data Units Written: 5,641,590 [2.88 TB]
  • Host Read Commands: 35,675,701
  • Host Write Commands: 109,642,925

I have applied the following changes to the ZFS config:

  • Compression to lz4: zfs set compression=lz4 <POOL>
  • Use internal SSD Cache for all kind of Data: zfs set primarycache=all <POOL>
  • Disable Secondary Cache on the SSD: zfs set secondarycache=none <POOL>
  • Only Write Data when necessary: zfs set logbias=throughput <POOL>
  • Disable Write Timestamp: zfs set atime=off <POOL>
  • Activate Autotrim: zpool set autotrim=on <POOL>
  • Increase Record Size: zfs set recordsize=128k <POOL>
  • Deactivate Sync Writes: zfs set sync=disabled <POOL>
  • Deactivate Deduplication (Off by Default): zfs set dedup=off <POOL>
  • Increase ARC and data size kept in RAM before writing (UPS):
  • echo "options zfs zfs_arc_max=34359738368" | tee -a /etc/modprobe.d/zfs.conf
  • echo "options zfs zfs_arc_min=8589934592" | tee -a /etc/modprobe.d/zfs.conf
  • echo "options zfs zfs_dirty_data_max=1073741824" | tee -a etc/modprobe.d/zfs.conf

Can someone maybe point me in the right direction where I messed up my setup? Thanks in advance!

Right now I think about going back the a standard lvm installation without ZFS or a Mirror but I'm playing around with Cluster and Replication which is only possible on ZFS isn't it?.

EDIT:

  • Added some info to storage use
  • Added my goals