While I truly have no desire to get involved in this back-and-forth, I've still decided to post for the potential benefit of people who might genuinely be interested in listening test protocols.
The issue with this specific listening test is that the protocol lacks the basic mechanisms to ensure that the listener can actually tell the difference between any of the network switches at all. It is just assumed the perceived differences are related to the network switches being exchanged, and some pretty severe biasing factors are completely disregarded.
While the listener in this test doesn't know which network switch is connected, he knows exactly when a change is made as well as that each time a different device would be connected - so there's still a very real chance of cognitive bias influencing the results.
Not to even mention the fact that the person making the device exchanges was really quite suggestive in his comments and mannerism during the test (e.g. see how he behaves between 44min to 48min in while discussing switch 2) - this is a very strong biasing factor as well.
Imagine this: what if a few times they
didn't actually exchange the network switch model when they said they did, but did so at other times?
Would the listener notice this every time?
If the listener noticed this (and did so often enough that the results were statistically significant) then that would be a good indication that he can indeed tell the network switches apart just by listening. Then the qualitative impressions of the sound might have some weight.
Nothing like this was attempted in this video, however.
To compare with a common test protocol, this is why in e.g. the ABX protocol we have the "X", and we need to say whether "X" is the same as "A" or "B". If the listener is able to do this consistently (and under strict blind conditions), then that is good evidence that the listener can actually tell "A" and "B" apart.
I.e. just having a
preference for "A" or "B" (without establishing audibility first) doesn't necessarily mean that the perceived difference between "A" and "B" is real. Blind ABX may not the only protocol that can be used to determine this, but it is a fairly common one.
To summarize, doing the test blind is for sure important (and absolutely a step in the right direction), but that alone is not enough to make the test robust in a scientific sense. So while I absolutely appreciate whenever people put in effort to make their testing more reliable, the test in that YT video definitely can't be used as evidence that there are audible differences between different makes and models of network switches used for streaming.
EDIT: For those interested to learn more about listening test protocols, I can recommend
this nice article from Linear Audio Volume 2 by Stuart Yaniger.