+xDear all,
I have a question concerning programming the duration of a trigger, which is sent to a parallel port.
This is my script:
<trial T1>
/ stimulustimes = [0 = Rahmen1, Rahmen2, Situation1, Frage1; 4000 = TriggerKSatz; 4100 = Frage1.1]
/ timeout = 5000
</trial>
Does this mean, that the trigger is sent after 4 seconds and is presented for 100ms to the parallel port?
<trial T1.1>
/ stimulustimes = [1 = TriggerControl; 101 = Erk1.1; 4900 = TriggerControlE]
/ timeout = 5000
</trial>
Does this mean, that after 1 second "triggerControl" is sent, which lasts 100ms?
And after 4900 "triggerControlE" is send, which is also presented for 100ms?
Thank you a lot for your help.
Best whishes!
> Does this mean, that the trigger is sent after 4 seconds and is presented for 100ms to the parallel port?
No. It means the trigger starts at 4000ms into the stimulus presentation sequence. It will *remain* raised for the remainder of the trial, not presented for only 100ms. If you wanted to end the trigger after 100ms, you would have to explicitly zero the bits as in
/ stimulustimes = [0 = Rahmen1, Rahmen2, Situation1, Frage1; 4000 = TriggerKSatz; 4100 = Frage1.1,
allbitstolow]
with
<port allbitstolow>
/ items = ("00000000")
...
</port>
> Does this mean, that after 1 second "triggerControl" is sent, which lasts 100ms?
> And after 4900 "triggerControlE" is send, which is also presented for 100ms?
Again no, as per the above. You need to explicitly lower the bits after 100ms.