[mythtv-users] VGA -SCART and interlace (2x)

John reidjr at btconnect.com
Wed Sep 23 18:33:56 UTC 2009


Paul Gardiner wrote:
> John wrote:
>> Paul Gardiner wrote:
>>> John wrote:
>>>> Thanks for the reply.  What I was doing was moving between 
>>>> autodetect, interlaced and progressive under the menu. The picture 
>>>> looks the same with interlace (x2) and Progressive (no -interlacer 
>>>> ) set. Same affect when I choose none as the interlacer under the 
>>>> TV settings menu.
>>>>
>>>> Looks like I need another graphics card,
>>>
>>> I don't think you've quite got the point I was making. 
>> You too :-)
>>
>> What I see is perfect "still" quality. The problem is interlacing 
>> artefact's on  movement, just as you have clearly described. What I 
>> was trying to say was that there was absolutely no reduction in 
>> interlacing artefact's between interlaced 2x, and none. So doubling 
>> the field rate, and repeating the fields does not lead to a constant 
>> synchronization between recorded and displayed interlacing.  The 
>> effect is disconcerting, and is clearly noticable on all scene 
>> changes, as well as movement.
>>
>> VGA to scart gives excellent quality, and the quality is also 
>> acceptable on movement using high quality deinterlacers, but the 
>> whole point is that we shouldn't need to de-interlace to display it :-)
>
> Ok, that makes more sense now I understand what you are seeing. You
> need to get "none" working intermittently well before using 
> "Interlaced x2".
>
> Do you see clear fine-lined combs, or a strange larger-scale sine-wave
> type pattern on the edge of moving objects?
>
> P.
>
>

I see "strange larger-scale sine-wave type pattern on the edge of moving 
objects"  ..... is there hope for me Doctor ;-)

john



More information about the mythtv-users mailing list