[Top][All Lists]
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Discuss-gnuradio] Problem with thread scheduler
From: |
cdong8812 |
Subject: |
Re: [Discuss-gnuradio] Problem with thread scheduler |
Date: |
Tue, 31 Jul 2012 09:27:04 -0700 (PDT) |
Hi Josh,
Thank you for the quick reply. I have went through previous post and find
one that had been fixed by your latest UHD build then. However I found this
annoying problem again with latest master build and N210. The structure of
my code is as follows:
Class Wimax_Receiver(gr.top_block):
...
def add_fft_cal(self, fft_cal_block):
# Get new block
self.FFT_calculate = fft_cal_block
# Add FFT to output of USRP
self.connect(self.u, self.FFT_calculate)
class FFT_Calculation(gr.top_block):
...
self.wr = Wimax_Receiver()
self.wr.add_fft_cal(self.s2v)
self.connect(self.s2v, self.fft_cal, self.c2mag, self.log,
self.stats)
self.connect(self.wr)
Below is the error message:
http://old.nabble.com/file/p34235614/QQ%25E6%2588%25AA%25E5%259B%25BE20120731121754.jpg
Josh Blum-3 wrote:
>
>
>
> On 07/29/2012 03:57 PM, cdong8812 wrote:
>>
>> I have a problem with the thread scheduler in GNU Radio. I am using USRP
>> N210
>> + RFX 2400 daughterboards with UHD. When I run my spectrum sensing codes,
>> if
>> I stop the flowgraph it will show the thread-per-block number mismatch
>> error
>> message on screen. I wonder if there is any way to get rid of this
>> problem.
>
>
> Can you post the error message?
>
> -josh
>
>> My BOOST version is 1.40 and I am using Ubuntu 10.04. UHD information is
>> shown below:
>>
>> address@hidden:/usr/local/src/gnuradio/build# uhd_find_devices
>> linux; GNU C++ version 4.4.3; Boost_104000; UHD_003.004.002-181-g25649781
>>
>
> _______________________________________________
> Discuss-gnuradio mailing list
> address@hidden
> https://lists.gnu.org/mailman/listinfo/discuss-gnuradio
>
>
--
View this message in context:
http://old.nabble.com/Problem-with-thread-scheduler-tp34213396p34235614.html
Sent from the GnuRadio mailing list archive at Nabble.com.