• 1 Post
  • 37 Comments
Joined 3 years ago
cake
Cake day: June 22nd, 2023

help-circle

  • It works, but the memory allocator implementation is way too slower compared to glibc. This especially becomes a performance bottleneck if application does a lot of heap allocation/deallocation.

    I think Musl is a better choice when you work on embedded, low-end devices, or statically linked/self-contained applications. For high performance workstation usage, I still prefer glibc.








  • SuperiorOne@lemmy.mltoSelfhosted@lemmy.worldUPS input load
    link
    fedilink
    English
    arrow-up
    2
    ·
    9 months ago

    I don’t think its rpi or network switch, unless you’ve overclocked rpi with liquid nitrogen 😅. So, I assume its TrueNas device.

    If it were a significant power difference, say 20-30 watts, you could easily find the process using htop/iotop. However, 6 watt difference is a relatively small value for a device with ~25 watts of idle power . It might be a process using just 1% system resources. That’s why I would look for systemd timers, cronjobs etc. to find scheduled tasks on specific times. Another possibility is automated S.M.A.R.T. self-tests. Those tests don’t show up in htop or iotop.



  • SuperiorOne@lemmy.mltoSelfhosted@lemmy.worldUPS input load
    link
    fedilink
    English
    arrow-up
    4
    ·
    9 months ago

    UPS devices normally uses wall (input) power, and switches to battery when input voltage is out of the target thresholds. So, input.load should represent the percentage of current wall power (in VA) relative to UPS’s max rated input power (VA). If your devices uses more power, input power from wall should increase as well.

    If it’s peaking in certain times, it could be due some scheduled job temporarily increase CPU frequency, or automated tasks like file system snapshot might power-up/spin drives longer than regular usage.


    • 2x18k - mirrored ZFS pool.
    • 1x47k - 2.5" drive from an old laptop used for torrents, temp data, non-critical pod volumes, application logs etc.
    • 1x32k - automated backups from ZFS pool. It’s kinda partial mirror of the main pool.
    • 1x18k - (NVME) OS drive, cache volumes for pods.

    Instead of single pool, I simply split my drives into tiers: cache, storage, and trash due to limited drive counts. Most R/W goes to the cheap trash and cache disks instead of relatively new and expensive NAS drives.





  • I have a APC Back-UPS 1600VA. It powers two desktop PC/Server, a monitor, and router. So far, it gets the job done.

    The biggest downside is; battery is not user replaceable, at least it’s not straight forward like the other models. If possible, prefer a UPS with the easy battery replacement option.