[mythtv-users] NVidia non-interlaced video output quality-loss?

Thomas Börkel thomas at boerkel.de
Sun Feb 15 06:11:52 EST 2004


steve at nexusuk.org wrote:

>>What are the exact consequences of the non-interlaced video output with 
>>the NVidia cards?
> A loss of picture sharpness - basically it's slightly blurry in the 
> vertical direction and high motion scenes will look a bit blurry as it 
> combines the 2 fields (which will be significantly different from 
> eachother in high motion scenes).  Its basically the same as turning on 
> deinterlacing in Myth.

Don't I have to turn deinterlacing on in Myth, if I have a NVidia card 
and my captured MPEG2 is interlaced (from PVR-250)?

> Horizontal resolution should be unaffected.
>>Assuming that analog TV is 352x380 native, would I then only have 
>>352x240? Isn't that a big loss in quality?
> Where are you getting 352x380 from?  PAL is 768x576 and NTSC is 720x480.  

Sorry, typo! I meant 352x480. I thought that analog TV only has half the 
possible horizontal resolution.

> Deinterlacing only affects the vertical resolution, and it doesn't just 
> throw away one field, it blends them together, so the output won't be 
> nearly as bad as just throwing away one field (nor will it be anywhere 
> near as good as keeping all the fields and watching an interlaced 
> picture).

OK, thanks. I am curious, how it will look (buying my MythTV box 
components next week).


More information about the mythtv-users mailing list