2017-06-11, 10:49 
		
	
	
		Does someone tell me why the system graphic of my detector show always a offset of + 50/60mV from zero? Can I adjust this anywhere? In the web interface, a can`t find any adjustment for this! It`s seem any wrong value in the ADC.  
ST ID 1846
	

- Sampling per InputResolution: 12bits, Period: 666ns = 500.0kSPS, Time: 2000ns
 - ADC1
- Input1 Amp-Channel 3 , Offset 26mV, no Trigger
 - Input2 Amp-Channel 4 , Offset 30mV, Trigger -90mV/90mV
 
 - Input1 Amp-Channel 3 , Offset 26mV, no Trigger
 - ADC2
- Input1 Amp-Channel 2 , Offset 30mV, Trigger -95mV/95mV
 - Input2 Amp-Channel 3 , Offset 25mV, no Trigger
 
 - Input1 Amp-Channel 2 , Offset 30mV, Trigger -95mV/95mV
 - ADC3
- Input1 Amp-Channel 1 , Offset 39mV, Trigger -95mV/95mV
 - Input2 Amp-Channel 2 , Offset -151mV, no Trigger
 
 - Input1 Amp-Channel 1 , Offset 39mV, Trigger -95mV/95mV
 - Thanks for your help!
 
ST ID 1846
