[mythtv-users] Nvidia TV Encoder not listing any support HD modes

Bonj mythtv at bigfathooker.com
Sun Mar 23 06:33:26 UTC 2008


Justin Nolan wrote:
> I'm giving up trying to output 1080i via DVI->HDMI without all types of 
> weird judder, so I've turned to tv-out/component video. I've never seen 
> any different between Component and DVI on my Sony 1080i RPTV, so it 
> shouldn't be any loss. Although I've ran into a major hurdle: I can't 
> get any HD resolutions working w/ TV out. My graphics card is a XFX 
> 7200GS, and I'm using a nvidia 7-pin HDTV breakout cable plugged into 
> the TV out port (gfx card manual says the port supports 4, 7, and 9 pin 
> cables/adapters).
> 
> I'm using the latest driver from nvidia, and my xorg.conf is as such:
> 
>> Section "Monitor"
>>     Identifier     "Generic Monitor"
>>     HorizSync       30.0 - 130.0   
>>     VertRefresh     50.0 - 160.0
>>     Option         "DPMS"
>> EndSection
>>
>> Section "Device"
>>     Identifier     "nVidia Corporation G72 [GeForce 7300 SE]"
>>     Driver         "nvidia"
>> EndSection
>>     
>> Section "Screen"
>>     Identifier     "Default Screen"
>>     Device         "nVidia Corporation G72 [GeForce 7300 SE]"
>>     Monitor        "Generic Monitor"
>>     DefaultDepth    24
>>     Option         "UseDisplayDevice" "TV"
>>     Option         "TVOutFormat" "COMPONENT"
>>     Option         "TVStandard" "HD1080i"
>>     SubSection     "Display"
>>         Depth       24
>>         Modes      "1920x1080"
>>     EndSubSection
>> EndSection
> 
> According to the xorg log, it doesn't seem to complain about the HD1080i 
> mode. But then:
> 
>> (WW) NVIDIA(0): No valid modes for "1920x1080"; removing.
>> (WW) NVIDIA(0):
>> (WW) NVIDIA(0): Unable to validate any modes; falling back to the 
>> default mode
>> (WW) NVIDIA(0): "nvidia-auto-select".
>> (WW) NVIDIA(0):
>> (II) NVIDIA(0): Validated modes:
>> (II) NVIDIA(0): "nvidia-auto-select"
>> (II) NVIDIA(0): Virtual screen size determined to be 800 x 600
> 
> What actually happens on the screen is that for the first few minutes 
> its a garbled mess of flickering red lines. After a few minutes the 
> screen appears in a small 800x600 window in the center of the screen 
> surrounded by black. My tv displays grey bars on the left and right on 
> everything except an HD signal. That tells me it actually is receiving a 
> 1080i signal, but with only a 800x600 pixel image.
> 
> And here's the part that is really making me pull out my hair. The 
> nvidia tv encoder on the card isn't reporting any supported HD modes:
> 
>> (--) NVIDIA(0): Connected display device(s) on GeForce 7300 SE/7200 GS at
>> (--) NVIDIA(0): PCI:1:0:0:
>> (--) NVIDIA(0): NVIDIA TV Encoder (TV-0)
>> (--) NVIDIA(0): NVIDIA TV Encoder (TV-0): 400.0 MHz maximum pixel clock
>> (--) NVIDIA(0): TV encoder: NVIDIA
>> (II) NVIDIA(0): TV modes supported by this encoder:
>> (II) NVIDIA(0): 1024x768; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI,
>> (II) NVIDIA(0): PAL-N, PAL-NC
>> (II) NVIDIA(0): 800x600; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC
>> (II) NVIDIA(0): 720x576; Standards: PAL-BDGHI, PAL-N, PAL-NC
>> (II) NVIDIA(0): 720x480; Standards: NTSC-M, NTSC-J, PAL-M
>> (II) NVIDIA(0): 640x480; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC
>> (II) NVIDIA(0): 640x400; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC
>> (II) NVIDIA(0): 400x300; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC
>> (II) NVIDIA(0): 320x240; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC 
>> (II) NVIDIA(0): 320x200; Standards: NTSC-M, NTSC-J, PAL-M, PAL-BDGHI, 
>> PAL-N,
>> (II) NVIDIA(0): PAL-NC
> 
> What would cause no HDTV modes to be listed? I've been struggling with 
> this for going on 8 hours, so any input is much appreciated. Thanks!

I use a 7200GS outputting via component breakout
cable to a 76cm widescreen Sony CRT at 1080i. At 1080i I experience 
combing effects without deinterlacing turned on. This is a side effect 
of using an interlaced mode and cannot be avoided without deinterlacing.
I have only recently upgraded my box to this hardware, so I'm still 
playing with the deinterlacers. The current setting is the "one field" 
deinterlacer, which gets rid of the combing, but isn't the best visually 
IMHO. I didn't have much time to play with it before it got commandeered 
for actual TV watching, so I have yet to experiment further, but I can 
say that XvMC with Bobx2 didn't look real flash either... must be a 
problem of either my XvMC setup, or the driver/card itself, but it 
seemed to have excessive buffering pauses.

Anyway, below is my xorg.conf:

Section "Monitor"
     Identifier     "Generic Monitor"
     Option         "DPMS"
EndSection

Section "Device"
     Identifier     "Generic Video Card"
     Driver         "nvidia"
     Option         "NvAGP" "1"
     Option         "DPI" "100x100"
     Option         "UseEvents" "1"
     Option         "AddARGBVisuals" "1"
     Option         "AddARGBGLXVisuals" "1"
     Option         "NoLogo" "1"
     Option         "UseDisplayDevice" "TV"
     Option         "TVOutFormat" "COMPONENT"
     Option         "TVStandard" "HD1080i"
EndSection

Section "Screen"
     Identifier     "Default Screen"
     Device         "Generic Video Card"
     Monitor        "Generic Monitor"
     DefaultDepth    24
     SubSection     "Display"
         Depth       24
         Modes      "1920x1080" "1280x720" "1024x768" "720x480" 
"800x600" "640x480"
     EndSubSection
EndSection


More information about the mythtv-users mailing list