femr,
A lot of this is fine detail stuff.
But when you're pushing the limits of what is possible (0.2 pixel), all the tiny details really matter.
Fine, though 0.2 pixels is far from pushing the limits in technical terms.
2/1000ths of a pixel? Where did you get that number?
Thought you might question that
It's from the blob trace data. I posted the first few samples a while back. First few y values increment in 0.001 pixel increments as expected. I'll post the full dataset if y'like.
My problem with this is that it is unrealistic when compared to real video info.
You've missed the purpose. It's purpose is to highlight the level of accuracy
possible with SE. Call it the limiting case.
Most significantly, it doesn't capture any of the features of how video cameras digitize continuous spatial info into discrete pixels.
Not entirely true, as it's still pixelated. And it's not supposed to. Limiting case test video
I'll be more impressed if you
That's fine, but as I've indicated the blob test has a purpose. I can perform additional tests, and was going to suggest that you produce a video within which you already know the sub-pixel movements of a feature, then it can be blind-tested, yeah

(A video from W.D.Clinger would be trusted more personally though tbh) But it's the attitude that grates a bit. Ho hum...
take a non-symmetric color shape
Okay
What do you mean ? Trace only one side/section of it ?
Hmm. How about adding random noise instead ? Or I suppose I could blend some smoke over the top ?
place it over a variable background
Okay.
mimic how video CCDs digitize that info
Doesn't make much sense.
pull out every other line
Makes no sense at all. You mean replicate the result of deinterlacing an interlaced video of similar format ?
and then watch SE do its stuff on THAT digitized info.
Okay.
I'll be more impressed if you do some actual, proper calibration: using a test set-up where you KNOW what the motion is
Already have. Blob test. I assume you mean some video footage of something ? If so, it can be done, but it's all manner of hassle ensuring movement is *known* and I'm not at all sure you'd trust the results anyway.
But I supposed that's out of the question
Not out of the question, but all getting a bit ridiculous for what is really rather simple. The problem as I see it is still about trust and the fact that this is far from your arena than impartial skepticism. But ho hum...
and we're stuck with trying to pull what info we can out of this video.
Nope.
The answer for me is going to reside within the static point data. Precisely because we KNOW that those points are not moving. You cannot tell for certain if the points on the WTC7 roof line are changing because the building is moving or if they are moving because the algorithm is giving you variable results.
The purpose of the blob test is to show that SE is capable of detecting position change far finer than +/- 0.2 pixels. There's always going to be some noise after static point subtraction, of course. I can look at the blob test in more detail of course, and provide a clearer picture of the higher levels of accuracy attainable. I've suggest a figure, which is maximum sensitivity. There are a few obvious deviations from that during the blob test, reasons for which should be fairly obvious. I'll explain in more detail, but I do request that you have a stab at suggesting what those deviation may be caused by first. I'm not really here to teach you from the ground up about all this stuff.
You also can't tell with a single reference point. That's why I asked for multiple ones.
I have provided you with multiple static point data.
There are issues using multiple static points, but in principle I have no issue including further static points in the resultant data. Indeed I already have for certain tasks not being discussed here. The data has been provided to you with a single static point to keep the spreadsheet manageable. If I provided you with a spreadsheet with 50 static points and 16 NW corner traces, you'd have a hissy, complaining about being swamped with data. Sigh.
Let me give you a couple of other things to consider:
You are too gracious...
1. You use a reference point to compensate for camera motion. The absolute error in your position is going to now be twice the error of any given point: the errors add.
And the variance is STILL only +/- 0.2 pixels. Good, innit.
2. You have mentioned only a single static reference point from which you subtract all your data.
See above.
Using one reference point, you cannot compensate for rotational variations. You'd need to use at least two reference points.
True, though cutting through the noise between separated static point traces to determine a rotational variable would MOST probably introduce more error than leaving it as it is, which is clearly not rotating to any great extent, and assuming it's on a tripod (which it really does appear to be) the likelyhood of rotation is very slim, and verly slight otherwise. By all means have a go at totally negating noise from two separate static point traces and determining rotation

(It's much easier when using to
moving points, for obvious reasons

)
3. Let's assume that you use only one field of data. If you have a really well defined edge (assume that it transitions within a single pixel, with no edge blur), then as the edge moves down in 0.25 pixel steps thru the field, this is how it appears to the video output (assuming a perfect input/output transform:
Actual position / Apparent position
0.00 0.00
0.25 0.25
0.50 0.50
0.75 0.75
1.00 1.00 (this scan line is missing from this field)
1.25 1.00 |
1.50 1.00 |
1.75 1.00 V
2.00 1.00 (reaches, but doesn't encroach into pixel)
2.25 2.25
2.50 2.50
2.75 2.75
3.00 3.00
3.25 3.00 (pixel not in this field)
3.50 3.00
3.75 3.00
4.00 3.00
4.25 4.25
This is the transform out of which you're trying to get sub-pixel location.
I'm not using one field. There's a valid point in there, sure, but there are no *jumps* in the output data, as expected. Are you sure that each interlace field discarded alternate lines ?
Of course, this just applies to a point, or a line parallel to the pixel layout. In an unusual twist, points that have edge blurs that extend at least 2 pixels can actually improve your resolution.
See prior post on regions.
As I said at the beginning, all of the above are fine points. But when you're claiming 0.2 pixel accuracy, then every possible source of error matters. And watching things like, for example the camera panning back & forth (not hard mounted), leave me very skeptical.
Being skeptical is fine, but a lot of this comes from this not being a field you are experienced in. That's fine, but it would be appreaciated if you chill out a bit. I'm willing to do the leg work, but there's a limit
One thing that I got to do several days ago was to watch some motion tracking programs attempt to track points in videos. I was underwhelmed. The people who were doing the video work had to manually adjust the reference point (establish key points) every (it appeared) 20 frames or so.
Perhaps they need some tips

I perform zero manual adjustment.
Perhaps SE has much better algorithms than those programs. Again, something that needs to be proven.
SynthEyes should be paying me for this. lol.
You'd have to prove this correlation.
Okay. First intention would be variance from the perfect known movement.
You're asserting that the accuracy doesn't depend on the local "video data quality". I find that hard to believe. Proof would help.
No I'm not. I'm saying that if you accept tha SE is capable of detecting movement of +/- 10 pixels, then if the noise level in a trace is +/- 100 pixels, then that additional noise is caused by noise in the video, not noise in the detection alrogithms.
Impossible to say for sure. Probably the edge contrast. Makes me think you don't know what I'm talking about though.
Care to elaborate on what halo might do to SE's interpolation algorithms. Care to prove it?

Read up on what halo effects in video are. It shouldn't affect trace whose region includes such data as it's consistent across frames.
I don't know the various compression algorithms that may (or may not) be applied to standard TV broadcasts. But I'd be really surprised if there are none. Care to elaborate?
There are definitely compression artefacts. I don't think the feature you mentioned is such, as I said.
femr, femr, femr. The words are how we attempt to communicate. And using the same word to mean two completely different things doesn't help.
My usage of the words has been correct.
And you've used the same word, in the same discussion, to mean 1) an assemblage of 2 deinterlaced fields, 525 lines and 2) a single 262.5 lines.
Correctly.
We weren't talking about camera film, etc. We were talking about video frames in THIS application.
In context, we were talking about you not understanding the variable correct contextual use of the word frame. As I've said, to make things easier on ya, I'll try and use the word field more often. I might also say, sod it, I'm tired of the attitude, and call them all images instead.
I'm not retracting anything. I never said that I'd do an error analysis for you.
Incorrect, but no matter.
I said that I thought that I could demonstrate that you weren't getting accuracies down in the fractional pixel range.
Have you ?
That I thought that they were up close to a full pixel.
And now ?
I'll continue that discussion.
Okay.
What are you talking about? Replicating your study is not on my agenda.
You said *there are far too many details about your analysis that are obscure & not well explained*. I just provided you with the procedure to perform *my analysis* in full.
How about the original data (not the 2 point averaged) on those static points.
Again, no probs.