Adam Reuter
March 9th, 2008, 05:21 PM
I was originally going to reply to this thread http://dvinfo.net/conf/showthread.php?t=116196 but I think my post deserves it's own place. This camera is tack sharp...at least tackier sharpier than what most of us have been used to. And it introduces FULL RASTER footage into the prosumer HD equation that I think needs to be addressed.
Note: I am not a computer science wizard. I just have enough years with Photoshop and encoding footage under my belt to understand some concepts I will discuss here.
A good mathematical reason that HDV (SD) footage shot with the EX1 looks better than HQ footage is that computers/algorithms are more efficient (i.e. better quality output) with even integers. This is why in Photoshop (or pick your photo editing/viewing application of choice) if you scale (aka downres) your photo to 50% it looks perceptibly better (less jaggies, smoother) than when you zoom out to 33.3% or any uneven number. 25% also looks nice because it’s half of 50%.
It’s even easier to see at 66.7% vs. 50%...the jaggies are clearly evident! Look for fine lines like hair in portraits to REALLY see it. Put that image in motion and that’s the twitter/flickering you’re seeing! I can clearly see this on my HP w2207 LCD screen but if you are on a lower resolution monitor you may not see much difference. And keep in mind that a still image is like "progressive scan". Add interlacing into the equation (which you will even if you shot your footage at 720p) and you have more issues.
HDV resolution is 1440x1080. Standard definition resolution (in NTSC land) is 720x480. 720 is exactly half of 1440. And although half of 1080 is 540 the lines that are downrezzed are still even. And many SD cameras capture at 540 lines (DV I believe has a theoretical max of 525 lines) so it’s VERY close to the optimal number.
Full raster res on the other hand is 1920x1080. If you divide 1920 (horizontal HD res) by 720 (horizontal NTSC SD res) you get 2.666666666… Or approximately 2 2/3. For PAL it’s about the same for PAL (3 1/3). Again remember the Photoshop analogy. Half the res of your photo looks good (no matter what the resolution of your original photo) but 1/3 of that image or 2/3 of that image (66% zoomed out) does not look as nice.
So, why do computers…err…compute better with even integers? 1’s and 0’s baby!
With all of this said, the better your downressing system (be it software or hardware) the better math it can do and the better techniques it has to work around the rounding off of the integers. I guess in the computer world that’s called “error correction”, although I may be wrong about that. This same reasoning is why some recording studios choose to record at 88.2kHz sampling rate instead of the 96kHz that their hardware is capable of. 88.2 = double that of 44.1kHz, the sampling rate of music CDs. Now if the final output of their recording is for video, which is usually 48kHz then there they would use 96kHz or 192kHz if they have it. 96/2 = 48 and 192/4 = 48.
It only makes sense that outputting with a Kona card would look better than most software solutions. These cards cost many times more than the software and are DEDICATED for tasks such as up and downversion from HD to SD and back. If you have that hardware…USE IT!
One last side note: It's because of this downres/upressing battle that I absolutely HATE IT when movies or documentaries that were shot with standard definition cameras upres their footage to film they then downres it from the film back to standard def. All of this scaling creates a worse image than if they had just used the standard def footage to begin with! Supersize Me comes to mind (I know PD150's look better than what was in that film!)
And gosh darn it can these camera makers make up their minds on interlacing fields. Lower field first, Upper field first. This mixing of interlacing options is why a lot of footage you see in documentaries where they use multiple sources looks like crap (motion-wise). Someone didn't set the proper flag for the footage and all hell broke loose. Once 1080/60p is out none of this will be a problem (or you can shoot at 720/60p for now).
Note: I am not a computer science wizard. I just have enough years with Photoshop and encoding footage under my belt to understand some concepts I will discuss here.
A good mathematical reason that HDV (SD) footage shot with the EX1 looks better than HQ footage is that computers/algorithms are more efficient (i.e. better quality output) with even integers. This is why in Photoshop (or pick your photo editing/viewing application of choice) if you scale (aka downres) your photo to 50% it looks perceptibly better (less jaggies, smoother) than when you zoom out to 33.3% or any uneven number. 25% also looks nice because it’s half of 50%.
It’s even easier to see at 66.7% vs. 50%...the jaggies are clearly evident! Look for fine lines like hair in portraits to REALLY see it. Put that image in motion and that’s the twitter/flickering you’re seeing! I can clearly see this on my HP w2207 LCD screen but if you are on a lower resolution monitor you may not see much difference. And keep in mind that a still image is like "progressive scan". Add interlacing into the equation (which you will even if you shot your footage at 720p) and you have more issues.
HDV resolution is 1440x1080. Standard definition resolution (in NTSC land) is 720x480. 720 is exactly half of 1440. And although half of 1080 is 540 the lines that are downrezzed are still even. And many SD cameras capture at 540 lines (DV I believe has a theoretical max of 525 lines) so it’s VERY close to the optimal number.
Full raster res on the other hand is 1920x1080. If you divide 1920 (horizontal HD res) by 720 (horizontal NTSC SD res) you get 2.666666666… Or approximately 2 2/3. For PAL it’s about the same for PAL (3 1/3). Again remember the Photoshop analogy. Half the res of your photo looks good (no matter what the resolution of your original photo) but 1/3 of that image or 2/3 of that image (66% zoomed out) does not look as nice.
So, why do computers…err…compute better with even integers? 1’s and 0’s baby!
With all of this said, the better your downressing system (be it software or hardware) the better math it can do and the better techniques it has to work around the rounding off of the integers. I guess in the computer world that’s called “error correction”, although I may be wrong about that. This same reasoning is why some recording studios choose to record at 88.2kHz sampling rate instead of the 96kHz that their hardware is capable of. 88.2 = double that of 44.1kHz, the sampling rate of music CDs. Now if the final output of their recording is for video, which is usually 48kHz then there they would use 96kHz or 192kHz if they have it. 96/2 = 48 and 192/4 = 48.
It only makes sense that outputting with a Kona card would look better than most software solutions. These cards cost many times more than the software and are DEDICATED for tasks such as up and downversion from HD to SD and back. If you have that hardware…USE IT!
One last side note: It's because of this downres/upressing battle that I absolutely HATE IT when movies or documentaries that were shot with standard definition cameras upres their footage to film they then downres it from the film back to standard def. All of this scaling creates a worse image than if they had just used the standard def footage to begin with! Supersize Me comes to mind (I know PD150's look better than what was in that film!)
And gosh darn it can these camera makers make up their minds on interlacing fields. Lower field first, Upper field first. This mixing of interlacing options is why a lot of footage you see in documentaries where they use multiple sources looks like crap (motion-wise). Someone didn't set the proper flag for the footage and all hell broke loose. Once 1080/60p is out none of this will be a problem (or you can shoot at 720/60p for now).