• 0 Posts
  • 19 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle
  • I had a few ideas, I’m suspicious that handbrake is falling back to CPU, maybe check the logs of the container to make sure it isn’t falling back to CPU decoding. Otherwise here are a few things I would check next:

    • If you are not using docker locally so you are already doing this, you will need to configure the docker container to pass through the GPU for quicksync to work inside the container.
    • If you are already doing that then I would make sure the device is the same name on the synology, it probably is but just to be sure.
    • you will likely need to add your user to the video and/or render group on the synology if you haven’t, especially if you are running the container as your user instead of root
    • make sure you are reading and writing to volumes that use bind mounts and not docker volumes, overlayfs is not what I would call fast and writing especially.

  • So I had a few thoughts. I’m not sure that you can use the docker device flag with a directory as you have there, I think it expects a device node, you can pass that directory as a volume (-v) though.

    If that doesn’t work you might also try running the VM with host-passthrough mode set on the CPU as well if it isn’t set that way already, sometimes that is also required for pass through to work from my experience. Also, make sure you passed through the whole device node, sometimes there are audio devices you have to pass through with the GPU device or you will get odd errors like those initialization ones you had. I’m not sure if this is the case for Intel iGPU though offhand though. Are you able to use intel_gpu_top on the VM to access the GPU? None of that is necessarily specific to proxmox though (but probably applies to anything libvirt powered) so YMMV.

    Edit: I realized you may not know what a “device node” is, that is the full path to the device, like /dev/dri/renderD128 vs /dev/dri which is actually a directory.




  • Okay so when you say “unplug the power” do you mean shut it down first or just pull the plug? The latter is a great way to corrupt your storage pools as ZFS uses memory for read and write cache etc by default. You definitely need to do a graceful shutdown especially if there is data that was recently written to disk, that’s why a UPS is so recommended. That said you can usually import an existing pool when that happens, I think there is a UI menu for it now.






  • So if I understand this right you will need to change the network on the port attached to the synology in your UniFi configuration or set the vlan tag in the synology OS, I would do the former. It sounds like you just added a second network/vlan to the existing interface which means you actually created a trunk and are getting the old network untagged and the new network with vlan tags which the synology is dropping. Synology OS also doesn’t really support trunked ports through the UI (even though it does support a port that only uses a vlan tag) so it’s much easier to just leave them untagged.



  • Well 5ghz requires more power, has less range, and needs its own antenna so for microcontrollers this makes it pretty pointless for devices that need range and low bandwidth for sending sensor updates, especially those that are battery powered. 5ghz can also have its own issues in cities if you have a lot of use of the DFS bands as well as being worse at traversing reinforced concrete.

    Also, a 2.4ghz radio can also sometimes support other things like zigbee, BT, and BLE which can be used for other functions.

    For what it’s worth, I have probably 50 WiFi devices and the majority of them are 2.4ghz sensors or switches and other low bandwidth tasks and I don’t have any issues, even when living in an apartment complex. If you are having issues you might need different hardware or more access points or something.

    Anyway, all that to say that 2.4ghz definitely still has a lot of utility today.



  • Yeah I’m with you, I read most of it but I just don’t know where the disdain comes from. At most scales of infrastructure anymore you can use them interchangeably because the difference is immaterial in practical applications.

    Like if I am going to provision 2TB I don’t really care if it’s 2000 or 2048GB, I’ll be resizing it when it gets to 1800 either way, and if I needed to actually store 2TB I would create a 3TB volume, storage is cheap and my time calculating the difference is not.

    Wait until you learn about how different fields use different precision levels of pi.





  • I assume you are talking about the Plex client or the server too? Client wise, the shield is still excellent as other folks have mentioned but I have been using Apple TVs lately, the client is snappy and works well. I have one of the first gen 4k ones and it plays everything I have tried to play. I got tired of the adds on the android devices and the Rokus have even gotten annoying lately so that’s what prompted me to switch.

    ETA: in case someone is wondering, I also use the Home Assistant integration with my Apple TVs and it is superb.