So as it was explained to me before, the vm specific ioacct file resets when the vm is restarted. Which seemed to work and make sence until today. I have a vm that according to that file has made 403G of writes since it was restarted yesterday.
(1 day) VE WWWW 403 GB in average daily writes and 403 GB in total since it was last rebooted (51 days) VE XXXX 9 GB in average daily writes and 487 GB in total since it was last rebooted (11 days) VE YYYY 5 GB in average daily writes and 60 GB in total since it was last rebooted (51 days) VE ZZZZ 2 GB in average daily writes and 132 GB in total since it was last rebooted *above day listings are from me just checking on their uptimes. My Script runs as: for veid in `vzlist -Hoveid`; do #run in all running vms a=`vzctl exec $veid uptime | awk '{print $3}'` #grab uptime from vm b=`grep write /proc/bc/$veid/ioacct | awk '{print $2}'` #grab write information from ioacct file for vm c=`echo "$b/$a" | bc` #divide writes by days up d=`echo "$c/1073741824" | bc` #daily writes in GB e=`echo "$b/1073741824" | bc` #total wirtes since uptime in GB echo "VE $veid $d GB in average daily writes and $e GB in total since it was last rebooted" #spit out results per vm done The ioacct file for this vm contains: read 121361661952 write 433471070208 dirty 2439431475200 cancel 2005960404992 missed 0 syncs_total 5 fsyncs_total 179645 fdatasyncs_total 48312 range_syncs_total 0 syncs_active 0 fsyncs_active 0 fdatasyncs_active 0 range_syncs_active 0 io_pbs 0 fuse_requests 0 fuse_bytes 0 So based on the write numbers (I do understand that dirty is unwritten and canceled is data not written during a flush) it comes out to 403G in writes since its reboot Mon Jul 20 08:46. Am I completely off on my math? Seemed to be working just fine until this vm spit out a 403G number. Am I missing something else? _______________________________________________ Users mailing list Users@openvz.org https://lists.openvz.org/mailman/listinfo/users