|
From: | Marcus Müller |
Subject: | Re: [Discuss-gnuradio] On the convolutional code performance of gr-ieee802-11 |
Date: | Tue, 15 Sep 2015 10:44:34 +0200 |
User-agent: | Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Thunderbird/38.1.0 |
Hi Jeon, speed depends on your hardware and the implementation of the decoder. As a rule of thumb: the more "generally applicable" a decoder is, the slower it gets. Jan wrote a set of highly SIMD-optimized decoders, but these are (pretty common) special cases, so they don't cover all the cases that gr-trellis works in, or even more general, the ITPP decoders application range. I'd assume you could get a significant speed boost if you replaced the IT++ implementation with your own, highly specialized decoder if you know what you're doing, but honestly, implementing, leave alone optimizing, decoders is a non-trivial task and one should definitely not start a project like gr-ieee802-11 trying to write one's own decoder if there's an existing decoder out there that's usable (IT++ can be a pain to use, still). Generally, I'd frown upon using a VM to benchmark decoders: Good decoders might make substantial use advanced SIMD instructions, but these might not be enabled in your virtualizer. Furthermore, if you want to do real-world gr-ieee802-11 usage, *don't* work in a VM, unless you're super knowledgable about how to configure VMs; latency and CPU overhead is critical, so the default "NAT" network configuration will not work well for network-attached USRPs, and USB3 support in VMs is ranging between bad and horrible, so B2x0 aren't really the best thing to be used in VMs, either. Run "volk-config-info --avail-machines" and check whether the output contains: generic_orc;sse2_64_mmx_orc;sse3_64_orc;ssse3_64_orc;sse4_1_64_orc;sse4_2_64_orc;avx_64_mmx_orc; If that's the case, your VMWare does allow AVX/SSE4 inside your VM. Best regards, Marcus On 15.09.2015 09:47, Jeon wrote:
|
[Prev in Thread] | Current Thread | [Next in Thread] |