2017-10-31, 10:36 (This post was last modified: 2017-10-31, 10:37 by cutty.)
(2017-10-31, 10:02)pasense Wrote: Cutty, would it not be reasonable to discuss which tools are missing to enable the improvement of the station's perrformance?
Why of course it's reasonable! More than reasonable, in fact. Advanced tools may be necessary in many situations! I'm sorry if there was an inference NOT to!... I was merely suggesting the 'similar analog' of the compression artifacts, and suggesting that the basic 'computation' possibilities not be over analyzed for a specific isolated situations... that'll always be the case. I just like to keep in mind that not every body will understand 'waterfalls', or know how to use a scope or spectrum analyzer, etc, and the controller doesn't have a whole lot of memory available to 'install' certain 'tools' on board...
(2017-10-31, 07:39)pasense Wrote: ...
Currently, only a few percent of the signals my station sends are used in strokes by the network, but looking at the signals I am quite sure that almost all of them are from real sferics. ...
I have a similar feeling but I'm not so sure what would be the necessary measure. I see many signals that are very sharp and I monitored the live view when a lightning was close by. While having the audio beep to announce a signal and the map with the lightning circles in parallel, I think that a lot of signals should have been from the lightning. Nevertheless the map did not draw a line or dot for my station and the statistic told it was not used.
So my problem was not only to reduce the number but to find the reason why obviously some good signals were not used. So, I repeat my idea that restarted all the discussion after a long time. I would like to see WHICH signal was used and WHICH NOT. For signals that were not used, it would be great to know how far away they were from a good signal. Was it some µs e.g. due to filters timeshift or reflections or was it at least some ms because it was not releated to a sferic at all. Even for signals associated to a sferic the accuracy in µs or meters would be helpful.
And if the data is available additionally a histogram would be a good statistical method to sum up the behavior of one station. That should show the time deviation in µs and the amount of signals used or not used. An ideal station would then have a very sharp peak of signals used and nothing around it. A station with reflections e.g. from hils or buildings close by would have a broader curve, where only some signals were used. And a station with a slow filter would have something like a shift by some µs leading to many not used signals.
Supporting the first station in India 1974 Stations:
2017-10-31, 21:35 (This post was last modified: 2017-10-31, 21:39 by mwaters.
Edit Reason: formatting
)
In order to register a stroke, both H-field antennas must sense it above the thresholds.
I see many signals here --that are obviously lightning-- that are not registered as a stroke for that reason. Only red, or only green, (and not both) pick it up the signal.
And it does not matter what any third H antenna (mine is a loop parallel with the earth) OR what the E-field antenna receives.
BOTH H-field antennas must sense it! Otherwise, it will only be a signal, and not a stroke. Some of the most successful stations in region 3 only have two H-field antennas.
(Note that stations with one H and one E antenna all have negative efficiencies! And there are many of them.)
2017-11-01, 07:25 (This post was last modified: 2017-11-01, 07:45 by pasense.)
@mwaters,
I tend to disagree with your statement. If the two H antennas are oriented at 90 degrees and one sferic arrives that gives a maximum signal in one channel, then the signal in the other, perpendicular channel would be zero, but a stroke should be registered. Just at the moment, my station is receiving a lot of signals from a direction such that only one H channel gets a signal, and looking in the list of registered strokes a lot turn up with only one H channel. Just have a look at the image below.
Of course there are others (second image), where all channels are involved, but both H channels are not required.
@micha.d,
I just checked the strokes where the signals from your station were involved, and it seem that it is working well. I did this by looking at the list of strokes in the statistics page. I do not know of any other way to check what strokes used the signals of a particular station. As I said above, some more information would be helpful, but it is also possible with what its available, but it takes some time.
The only comment I have regarding your station: it seems you have connected the two H field antennas to channel B and C; it would be better to shift to A and B. You can see that from the color of the traces in the signal plots below. I copied these from the list of strokes where your station was involved and they show typical signals.
Hi Pasense, I agree, I had my station, whilst I was in the initial set up, with only the "E" field functioning. I also set up the "H" channels individually.
I was still registering strokes and was able to optimise each channel, without the confusing array of squiggly lines that are present, when all the channels are activated at once.
Having said that, whilst in your first example, the Green signal is very clean and well defined, in your second example you have far too much gain!
The Red channel is flat topped and chopped and even the Green and Magenta channels have signs of distortion products.
Even in low noise situations, I believe that gains over 10 are very rarely needed and for signals within 2,000 km range, gains of 8*5 or even 5*5 will produce excellent results.
Reducing the gain actually makes the system more sensitive, because you are not amplifying the noise floor as much and therefore have a better signal to noise ratio.
I have made this suggestion to several stations now and they will confirm that their stroke rates and their locating rates are now much more efficient and their effectivity is more constant.
I know that you are in a special situation due to your location, but I would like to suggest that you try this for a few days and compare the results between various levels of lower gain and higher threshold levels.
You may be surprised and if not, it is only the matter of a few seconds to return to your present settings.
Respectfully yours,
Brian,
you are right, the signals go in saturation for these strokes, but these were actually the strongest signals I have seen so far, so I copied the signal traces. I am just experimenting in order to see if and how I can get also polarity information from strokes over the ocean (which is supposed to be difficult or impossible), but yes, I could reduced the gain of the H channels. Coming back to useful plots we would like to have: a histogram of maximum signal strength, both for signals sent and for signals used in strokes would make it easy to spot such situations: these would just go into the "overflow" bin. So far, I have to look at the list of strokes in order to find out if the amplification and thresholds are OK. Such histograms are standard in all situations where one has to adjust such parameters and I am sorely missing these here!
Heh .... and here's what happens when you have 33 strokes >500 km but a sporadic noise maker active,
which does NOT trigger interference mode,.... 83335 0 0% 0 0% 33 0%
...sigh.. sta 1439 1300UTC
Looking at the signals on this page there is obviously a lot of variability in signals even for the same stroke. If you look at the stations chosen for location nine of them are greens, six are reds, and five are blues. Some of the signals seem vastly over-driven, being cropped flat topped and distorted, some with various amounts of noise and some that are barely recognisable as lightning at all?
After looking at several pages like this I came to the conclusion that the criteria for being chosen for the location of strokes has nothing to do with quality of signals, but is just based on the first to arrive at the server. Greens send less information, so they predominate. Blues with H.O. or E.O. Send less info but more than greens or reds so they bring up the rear. there are very few 3 or 4 channel blues represented in this lottery?
Yes, this is all just for one stroke! I have not edited it in any way. The chosen stations are underlined.
My question involves the timing in micro seconds at the top of each graph at a quick scan this varies from a few + microseconds up to several + hundred microseconds, the average being +50-+60 microseconds! Only the fastest are chosen!
Does this timing have anything to do with the decision about whether your signal is chosen for the location calculations? Based on my short observation it seem to.
Secondary question. Is there anything that an individual station could improve on this time delay? Reduce channels and get a better result in the lists
Are there things that we should avoid doing that might increase this delay? Send good quality signals on all channels?
Last question, at least for the moment. Is this timing of great importance or is it the time that has the seconds to nine decimal places that is the important time?
I wish I had studied statistics more when I was young, and I wish there was a quicker way to get all this data from the archives.
Regards,
2017-11-01, 18:16 (This post was last modified: 2017-11-01, 19:52 by micha.d.)
(2017-11-01, 07:52)pasense Wrote: @micha.d,
I just checked the strokes where the signals from your station were involved, and it seem that it is working well. I did this by looking at the list of strokes in the statistics page. I do not know of any other way to check what strokes used the signals of a particular station. As I said above, some more information would be helpful, but it is also possible with what its available, but it takes some time.
The only comment I have regarding your station: it seems you have connected the two H field antennas to channel B and C; it would be better to shift to A and B. You can see that from the color of the traces in the signal plots below. I copied these from the list of strokes where your station was involved and they show typical signals.
Hi pasense,
I have read that channel A has more noise due to the GPS being close by. So I switched to B and C for testing. I now feel that it does not make much difference, at least with my antenna getting some interference from outside and NOT over the PCB itself. Even soldering a shield into the device did not make a difference:
But regarding the evaluation of separate channels I have some doubts. I'm also sure that single channels are used for strokes but signals coming in the direction of one of the two antennas seem to be prefered because there is a better chance to exceed the threshold. I think I can see that effect in one of the diagrams (https://www.lightningmaps.org/blitzortun...n_id=14731), but it's most probably still to few data. My antennas are oriented in direction north-east <-> south-west and south-east <-> north-west and I think the green ratio is better in those two directions. To remove this effect we would need some info on the direction of the single antennas to calculate the direction of the signal source and effective signal of that direction. In my next live I'll go for 3 antennas to reduce this side effect.
EDIT: Before someone complains. I mixed up some numbers. The link above is not my station. But I stick to 3 antennas in the next live. Back to efficiency ...
Supporting the first station in India 1974 Stations:
I would like to bring the attention back to the title: Efficiency
I have one more very very general observation that I would like to present. Having a value in % in my eyes must mean that the value is in the range 0 - 100. This is the case for the "Effectivity" on http://de.blitzortung.org. It is not the case for the "Efficiency" on https://www.lightningmaps.org. Values like "< -999%" are obviously a bit "special". Formulas like a/b - c/d or a/(b-c) cannot generate a % value.
--> So either use % and stick to the range 0 - 100 or please just name it differently.
Supporting the first station in India 1974 Stations:
(2017-10-26, 21:03)Egon Wrote: I would like to change the measure for the efficiency of the stations.
I quite like Cutty's:
Efficiency = Strokes Detected / Signals Sent Effectivity = Strokes Detected / Signals Sent - Strokes Detected
Being based purely on the numbers from a given station it removes the influence of the total region count which generally has the affect of pushing a station that sends few but high quality signals down the list.
Not sure how to combine those with a distance. "signals sent" doesn't have any distance information but does that matter? I don't think it does.
Actually, it does... in a 'thought experiment'...
Hmm, whilst I agree that a stations enviroment and location affect the settings (threshold/gain etc) of that station my gut feeling about the above formula is that distance "cancels out".
ARGH! I hate web based editors that drop your carefully crafted message in the bit bucket if you accidentaly load another page in the same tab/window. I'd typed in a load of thought experiment showing that distance doesn't come into the formulas. It also showed that gain governed range and that threshold is best set after the gain/range value to the point where local noise only rarely triggers a signal. Note this was a thought experiment in a nice friendly world not the hostile, real, one. Finding a setting for the threshold with variable local noise levels might be tricky.
The range a station is optimised for is a choice for the operator but for most benefit to the network ought to follow that relatively isolated stations go for range and ones with nearby neighbours restrict their range.
EDIT: Before someone complains. I mixed up some numbers. The link above is not my station. But I stick to 3 antennas in the next live. Back to efficiency ...
ATTENTION!!!
Hi All, Whoever the above Station does belong to. The gain settings are wrong, there appears to be an extra 0 on the 10 making it 16*100 on both channels shown!
My views on high gains are well enough known.
Whilst this is obviously a mistake, I think this station is still trying to use far too much gain!
I have yet to see a situation where anything over 10 was useful, except for generating problems and extra signals that are not lightning.
If you are one of these stations that insists on High gains and low thresholds, Please, Please try out very reduced gains, these boards are extremely sensitive and already well below the ambient noise floor in most situations, therefore increased gain only equals reduced signal to noise ratio, splatter caused by overload mixing products and other problems.
Spread the word! Make someone your Buddy, you help them and they help you, and when not much is happening you can "chat" via email to someone that understand what you are talking about.
Regards,
Efficiency = Strokes Detected / Signals Sent
Effectivity = Strokes Detected / Signals Sent - Strokes Detected
Has anyone used the above on real numbers?
Recent grab from the stations list of my station (1627) and the three closest others:
Sta Sent Used L Eff Eft
1690 1240 184 14% 15% 17%
1627 274 71 11% 26% 35%
1645 1714 180 10% 11% 12%
1620 1395 140 10% 10% 11%
L : Current Effectivity L (S and M where zero).
Eff : Used/Sent
Eft : Used/(Sent - Used)
Hey, I like this new measure. B-)
Hi All, I have yet to see any figures for Effectivity "S", ever, at my station and rarely get figures for Effectivity "M," even though some of the strokes are listed in the archives at only 25 km away and many more in the 100 to 300 km range.
I would like to see these numbers extended over a much greater time period, maybe several time options, 1 hr, 4 hrs, 12 hrs, one day, one week, one month, Total of all!
Some people have not yet realised, that if their station is sending many signals to the server and if there is no lightning in their vicinity, it is a sure sign that your station has a problem.
Yes, we all probably take more attention, when there is a lot of action on the screen, but it is also important to see what your station is doing during the "quiet" times.
Just another idea, may be we should not be looking so much at Individual station Efficiency/Effectivity, but at Overall Network Efficiency.
What are the effects of varying the number of stations needed to validate a stroke, on the description it says the more the better, but how many strokes are not recorded, because there were too few station to validate them, maybe by quite a small margin, i.e, because only ten, or even less, stations registered the stroke!
Is all this data lost?
Why are so few stations actually contributing to this forum. This is an important question for us all and I am sure there are many people brighter than me that would have perfectly valid ideas on this subject, if only they would state them, instead of keeping them to themselves.
Maybe, we need a "Banner Headline" Page, that appears every time we open the maps in our browsers, that gives a summary of the days events and the present situation of the network?
Regards,
Hi All
I'm very new to this but it seems that there are probably very few stations with the same characteristics which makes a comparison difficult if not impossible.
The only thing constant for all stations is a particular stroke.
So what if we could analyse by looking at who is watching what. In other words use only the signals from the same cluster of storms and then further analyse by range bands from the storms. If only one (or clustered) storm is used the distortion of data from those stations able to pick up signals from two or more storms would be eliminated. I suggest a formula something like:- Stations in range of storm A 0 - 500K "My detections / Mean of all detections" and the same for increasing range bands. It doesn't eliminate false detections but at least it tries to demonstrate how a station is performing in relation to the same events compared to its peers. Stations below the mean should perhaps adjust whilst those very much above the mean could expect false detections.
I'm no statistician so maybe this is too simplistic approach but I hope it is thinking out of the box.
Regards
Alan
Let me bring the general 'Cutty' suggestion back into focus, since it was generally outlined in two posts buried up above....
"
(2017-10-26, 21:03)Egon Wrote: Hi Folks,
I would like to change the measure for the efficiency of the stations. It may not be appropriate to calculate only one numerical value, but rather to introduce different possibly even competing measures for the efficiency of the systems.
The first value could indicate how many of the transmitted signals are involved in the calculated strikes. The second value would be distance dependent and indicates how many impacts with a certain distance to the own detector the station was involved.
The current combination of these values (thate what we are doing now) seems to be not usefull for adjusting a detektor.
Any ideas for further efficiency measures?
/Egon
Just brainstorming:
Suppose we explore this, for each unique station, as opposed to, or in addition to, station's relative "Network / Region" performance. This suggestion is for 'Typical" installations. It's my belief that a second group might be considered, for example, those stat1ons that are 'purposefully' employed across regions or oceans. However they, I believe, should be
"qualified" on Efficiency and Effectivity overall before being placed into that group. And 'disqualified' if performance degrades. If makes little sense to me to have a station with a noisy environment attempting to serve 'cross region' or ocean, for example. We define: Goal: detect maximum number of strokes accurately with minimum number of signals Efficiency: Ability to accomplish something with the least amount of time and effort. Effectivity: Actual production of the intended result.
Efficiency = Strokes Detected / Signals Sent Effectivity = Strokes Detected / Signals Sent - Strokes Detected
That is, a station which sends a lot of signals, but few strikes is not very efficient.
For Effectivity; some of the 'signals' sent ARE strikes, and they are removed from the Total Signals, then strikes compared to 'remaining' (unused" signals.
The thought experiment runs sort of like this:
Briefly, divide the stations according to 'intent'... say, for example, three groups.
Group ONE > operators who haven't bothered to define or optimize well. Anything 'in'.,.. Anything 'Out'.
Group TWO > Stations that have established a 'relatively' quiet operational invironment, normally function with a 'High" "Strokes to signals", and have optimized for 'distance'... would include many of the 'dual' region / fringe stations perhaps. Since they already operate with 'low' 'excess' signals. They can operate for example as the 'group greater than 1200 km'. Their signals would also include the signals of a nearby 'group three', of course.
Group THREE> Stations optimized for <1200 km... those that run many excess signals when using high gains and / or low thresholds, or with frequent 'sporadics' that could be eliminated, or minimized with lower gains / and or higher thresholds. They now optimize for reduced distance. Optimizing in this fashon also lessens the typical "excess signals' relative to 'detections within 1200 km'.
And you could have perhaps a smaller group, perhaps operating E field only, or C Horizontal only, operating at distances <600km or any other subdivision similar...
The decrease in 'out of distance' signals, with an optimizing within distance should allow the same similar formulas too be used... though a group 2 station might show a 'higher effieciency' or 'lower efficiency' than a station in group 3, or vice versa, the rules are the same, and the 'comparisons' similar, within the group. not necessarily outside the group.
Both groups should be much more efficient and effective than MOST of the 'group ones'... and my 'gut feeling' is that watching numbers between groups TWO and THREE would actually show similar efficiency / effectivity numbers... IF the group TWOs had strokes <1200 km subtracted from both signals and strokes ,,,
Again, an operator shouldn't compare 'outside his Group" , ...
But... Especially on group TWOs... a station should establish, over time, that it BELONGS in group TWO ! To do so, they'd have to operate as a GROUP one until the "Higher Effeciency / Effectivity" numbers qualified them as a 'group TWO'... if a station degrades over some time frame, than they should be 'demoted' to group ONE, unless they opted to go for Group 3.
All stations would begin as group ONE, and then optimize for, and qualify for group TWO or THREE. Actually, I think the server can do that automatically, since it already is capable of breaking strokes down by distance.d A 'quiet' group TWO station might not have as many 'signals' at high latitudes seasonally, but the 'noise signals' may remain essentially the same.... for a Group THREE, some adjustmen might need to be made seasonally, since 'theoretically' their 'local noise' soruces would remain at similar levels year round, while signal numbers rise and fall.
In those cases a comparison with 'degradation' over time of total 'signals' might be applied.
2017-11-03, 13:09 (This post was last modified: 2017-11-03, 13:16 by readbueno.)
Hi Alan, I am not so sure about "false strokes" maybe it would be better to call them mis-classified, as we measure mostly CG lightning, but often record CC lightning as much less powerful strokes, especially in near storms.
If our detection was better, as I am sure it will be, we would perhaps have better discrimination between different types of lightning, or interference that mimics lightning, from the point of view of our stations.
I am sure that most people running a station are more aware of how lightning behaves in its various forms, but here are a few basic sites were lightning is explained:-
I know there are many more, some very complicated, but these do give some basic facts.
Alan, As storms are a moving phenomena I am not sure how your cluster theory would work in practice.
Could you be more definite in describing how individual and groups of stations would organise their data, to make more usable.
With over 1.4 billion strokes a year world-wide, we are still missing a lot of data in the strokes that we do archive, and even for the data that we do collect, I am not sure what analysis or research is being done, in relation to classifying strokes, propagation characteristics, even diurnal variation.
Is there even a good way that this huge amount of data could be indexed and worked on, unless we have access to super computing or maybe distributed computing, like "SETI" or "Folding @ Home", or even something like "Zooniverse" where people organise vast amounts of data as a volunteer service and pastime?
Certainly even organising the data from my own station appears to be beyond my meagre capabilities, and I would welcome ideas and programs from those that are more expert in these matters.
Sorry, I am probably off topic again, but these ideas do come to mind whilst reading the forum and may add to the overall mill of information.
Regards,
2017-11-03, 13:13 (This post was last modified: 2017-11-03, 13:20 by cutty.)
Let me bring the general 'Cutty' suggestion back into focus, since it was generally outlined in two posts buried up above....
"
(2017-10-26, 21:03)Egon Wrote: Hi Folks,
I would like to change the measure for the efficiency of the stations. It may not be appropriate to calculate only one numerical value, but rather to introduce different possibly even competing measures for the efficiency of the systems.
The first value could indicate how many of the transmitted signals are involved in the calculated strikes. The second value would be distance dependent and indicates how many impacts with a certain distance to the own detector the station was involved.
The current combination of these values (thate what we are doing now) seems to be not usefull for adjusting a detektor.
Any ideas for further efficiency measures?
/Egon
Just brainstorming:
Suppose we explore this, for each unique station, as opposed to, or in addition to, station's relative "Network / Region" performance. This suggestion is for 'Typical" installations. It's my belief that a second group might be considered, for example, those stat1ons that are 'purposefully' employed across regions or oceans. However they, I believe, should be
"qualified" on Efficiency and Effectivity overall before being placed into that group. And 'disqualified' if performance degrades. If makes little sense to me to have a station with a noisy environment attempting to serve 'cross region' or ocean, for example.
We define: Goal: detect maximum number of strokes accurately with minimum number of signals Efficiency: Ability to accomplish something with the least amount of time and effort. Effectivity: Actual production of the intended result.
Efficiency = Strokes Detected / Signals Sent Effectivity = Strokes Detected / Signals Sent - Strokes Detected
That is, a station which sends a lot of signals, but few strikes is not very efficient.
For Effectivity; some of the 'signals' sent ARE strikes, and they are removed from the Total Signals, then strikes compared to 'remaining' (unused" signals.
The thought experiment runs sort of like this:
Briefly, divide the stations according to 'intent'... say, for example, three groups.
Group ONE > operators who haven't bothered to define or optimize well. Anything 'in'.,.. Anything 'Out'.
Group TWO > Stations that have established a 'relatively' quiet operational invironment, normally function with a 'High" "Strokes to signals", and have optimized for 'distance'... would include many of the 'dual' region / fringe stations perhaps. Since they already operate with 'low' 'excess' signals. They can operate for example as the 'group greater than 1200 km'. Their signals would also include the signals of a nearby 'group three', of course.
Group THREE> Stations optimized for <1200 km... those that run many excess signals when using high gains and / or low thresholds, or with frequent 'sporadics' that could be eliminated, or minimized with lower gains / and or higher thresholds. They now optimize for reduced distance. Optimizing in this fashon also lessens the typical "excess signals' relative to 'detections within 1200 km'.
And you could have perhaps a smaller group, perhaps operating E field only, or C Horizontal only, operating at distances <600km or any other subdivision similar...
The decrease in 'out of distance' signals, with an optimizing within distance should allow the same similar formulas too be used... though a group 2 station might show a 'higher effieciency' or 'lower efficiency' than a station in group 3, or vice versa, the rules are the same, and the 'comparisons' similar, within the group. not necessarily outside the group.
Both groups should be much more efficient and effective than MOST of the 'group ones'... and my 'gut feeling' is that watching numbers between groups TWO and THREE would actually show similar efficiency / effectivity numbers... IF the group TWOs had strokes <1200 km subtracted from both signals and strokes ,,,
Again, an operator shouldn't compare 'outside his Group" , ...
But... Especially on group TWOs... a station should establish, over time, that it BELONGS in group TWO ! To do so, they'd have to operate as a GROUP one until the "Higher Effeciency / Effectivity" numbers qualified them as a 'group TWO'... if a station degrades over some time frame, than they should be 'demoted' to group ONE, unless they opted to go for Group 3.
All stations would begin as group ONE, and then optimize for, and qualify for group TWO or THREE. Actually, I think the server can do that automatically, since it already is capable of breaking strokes down by distance.d A 'quiet' group TWO station might not have as many 'signals' at high latitudes seasonally, but the 'noise signals' may remain essentially the same.... for a Group THREE, some adjustmen might need to be made seasonally, since 'theoretically' their 'local noise' soruces would remain at similar levels year round, while signal numbers rise and fall.
In those cases a comparison with 'degradation' over time of total 'signals' might be applied.
Hi All
I realise that storms are moving phenomena but the movement is relatively slow and my thought was that there might be a simple way of tuning a detector. Well tuned detectors must benefit the network as a whole. Not everyone would have the time to analyse a lot of historical data and, like me, many have probably not learnt enough to use it effectively.
Maybe something along the lines I suggest would be a sort of 'Quick Start Guide' for those with limited time and/or knowledge run in parallel with the 'proper tools' if the server capacity would allow it. Brian has pointed me to the Strikes versus Signals graphic which gives a guide as to what I am doing but not what I should be achieving in comparison to stations which should be showing similar results.
As recently we have has storm clusters distant from each other I have been able to improve my performance somewhat by watching other stations fairly close to me. What I am suggesting is a formalisation of that.
I see the point of Egon's grading system but the spread of the system must mean there are some who 'don't have the time or knowledge/lost interest' who be be stuck in group one but might be motivated by a simple method of improving.
Regards
Alan