• 15 Posts
  • 538 Comments
Joined 1 year ago
cake
Cake day: July 5th, 2023

help-circle





  • The thing is that I’m already at the last couple of leaves in the investigation tree and I’m not willing to change anything upwards of the USB driver level. That’s why there isn’t much point in getting people to spin their wheels for solutions I can’t or won’t apply. If I was completely unable to get the data corruption and disconnects under control, I’d trash the system and replace it with Intel. Fortunately, a PCIe add-in USB controller seems to work well so I avoided the most costly solution. At this point I don’t actually need to get the motherboard ports to work well but I’m curious to follow down the signalling rabbit hole because I’m not the only one who’s having this problem and the problem doesn’t affect just this one use case. If I find a solution like an in-line 5Gb USB hub (reduces data rate), or just using USB-C ports instead of USB-A (reduces noise), or using this kind of cable instead of that kind, I could throw that as a cheaper workaround in this ZFS thread and elsewhere. The PCIe cards work but aren’t cheap.


  • Unfortunately it won’t because the transfers are happening between ZFS and the hardware storing the data so I can’t control the data rate at the application level (there are many different applications) or even at the ZFS level. This is why in this particular case I’m stuck with a potential hardware-related workaround. I mean I could do something stupid like configuring a suboptimal recordsize in ZFS but I’d prefer to get the hardware to stop losing bits and hoping ZFS would catch that. Decreasing data rates is a generally acceptable strategy to deal with signalling issues, if the decreased rate is usable for the application at hand. In my case it is.


  • I am trying to transfer data via USB at high speed without data corruption, silent resets and occasional device disconnects. Those are things that happen because the USB controllers on my motherboard made by AMD with some help from ASMedia do not function correctly at the speed they advertise. So given the problem the right solution is to get a firmware or hardware fix for these USB controllers, however that’s unlikely to happen. So I’m trying to find a workaround. I already have one (PCIe add-in card) but now I’m also testing running the bad controllers at half-speed which seems successful so far but I was wondering if there’s a way to do it in software. I’m currently bottlenecking the links by using 5Gb hubs between the controllers and the devices.



  • Great question. In short, garbagy AMD USB controllers. I recently switched to a newer AMD board and have been hit with the same issues faced by these poor sods. I’ve been conducting testing over the last week, different combinations of ports, cables, loads, add-in PCIe USB controllers. The add-in cards seem to behave well, which is one way the folks from that thread solved their problems. The other being changing to Intel-based systems. Yesterday however I was watching an intro about USB redrivers by TI and they were discussing various signalling issues that could occur and how redrivers help. That led me to form the hypothesis that what I’m experiencing might be signalling related. E.g. that the combination of controllers/ports/cables simply can’t handle 10Gbps. That might be noise from some of those devices or surrounding ones that causes signal loss when operating at 10Gbps, speeds this setup can actually achieve. In order to test that I tried placing the DAS boxes behind a 5Gb hub plugged in a port that has previously shown a failure. So far it’s stable. This is why I was wondering whether there’s some magic in the kernel that could allow configuring 10Gb ports to operate at 5Gb.





  • Try to use the standard virtual desktop functionality. It’s not going to behave the exact same way but you might find a workable adaptation.

    You can pass through USB to VMs.

    You can use Premiere in a VM. Depending on what you want to do and how fast your CPU is, you might be able to use it without any special config. If you need to pass through a GPU, you can buy the config isn’t trivial. Definitely doable though.

    Generally you’d want to use KVM (with virt-manager) for virtualization but it doesn’t support any 3D acceleration in Windows yet. The result is that the Windows UI in a VM is not “smooth.” It’s usable but not smooth. If you need acceleration you’d have to do GPU passthrough. There’s some ongoing work to get basic acceleration without passthrough but it isn’t done yet. Both VirtualBox and VMware have basic 3D acceleration for Windows VMs. They have other pros and cons but if you find that one of them works better than KVM for your use case, go ahead and use that. We use both VirtualBox and VMware for different purposes at work. I know of people who use all sorts of engineering CAD software in VMware.


  • Avid Amoeba@lemmy.catoMemes@lemmy.mlCapitalist logix
    link
    fedilink
    arrow-up
    4
    ·
    edit-2
    6 days ago

    You already trust these people today. They run/own many large corporations today which dramatically affect our lives in multitude of ways. Except today we can’t get remove them from these positions of power under the current system.

    It’s thanks to this in part that your aunt keeps indulging her imaginary pain when she thinks about your lifestyle.




  • I think for games, people need newer kernels and drivers to support the newer hardware needed to play newer games, and they’re willing to put up with the bugs that come along with thay. Ubuntu and Debian (stable) aren’t strong at that by definition. I always use an older GPU that supported well by the Ubuntu LTS I run. If it doesn’t play something, I’ll wait till a new driver lands in that LTS or the next.