07-24-2013 12:07 PM
I'm using Labview 2012 to control an Agilent 33521A function generator. For my application, I need to burst the wave with a burst period of 1us. The issue is setting the wave programatically; if I manually set up a 1us burst period, everything works fine, however if I wire a 1E -6 constant to the relevant input terminal, the VI throws an error suggesting my burst period is too small. Wiring 2E -6 also works fine, and if I save the manually created waveform to a state and load it programatically, once again, the VI works fine.
The only conclusion I can come to is that there is some sort of floating point error that results in the device receiving a number just shy of 1us. I tried using 1.1E -6 but that didn't work either. On the offchance that someone reading this has the relevant equipment to test the VI out, I'm attaching it, along with a screenshot of the VI throwing an error.
If anyone has any suggestions for a workaround, it would be appreciated!
07-28-2013 10:11 PM
Something to try that might provide some insight would to setup the parameter for a 1us burst and then read the value back programatically and see how the instrument formats the value.
Mike...
07-29-2013 07:29 AM
It says "Burst period extended", so what's the form of the burst? Are you sending X number of waves of a certain frequency that takes longer than 1 us?
/Y