Welcome to Doom9's Forum, THE in-place to be for everyone interested in DVD conversion. Before you start posting please read the forum rules. By posting to this forum you agree to abide by the rules. |
8th February 2020, 15:35 | #58561 | Link | |
Registered User
Join Date: May 2004
Posts: 5,351
|
Quote:
Sent from my SM-G975U using Tapatalk
__________________
HTPC: Windows 11, AMD 5900X, RTX 3080, Pioneer Elite VSX-LX303, LG G2 77" OLED |
|
8th February 2020, 16:00 | #58562 | Link |
QB the Slayer
Join Date: Feb 2011
Location: Toronto
Posts: 697
|
Agreed.
I just watched Zombieland (the original yes I hadn't watched it yet) and madVR OSD said the movie was 10000 nits... there is no way it would have looked as good as it did with 113 test build if I had just left passthrough and used the LG "Dynamic" tonemapping. Another movie I watched recently was Gemini Man (60fps) and for that I needed to turn off madVR's tonemapping and use passthrough. There is no way my GPU could handle it. There were many scenes that I know HSTM would have made better since after watching many movies now with it, I can easily spot where it would have been an improvement with that particular title. It's well worth the minor effort to setup. And like said above, many users post their settings as a starting point. QB
__________________
|
8th February 2020, 16:25 | #58563 | Link |
Registered User
Join Date: Dec 2018
Posts: 207
|
SamuriHL
Could you clarify whether it is necessary to activate the option "output video in HDR format" in HSTM in case of viewing on C8 for example?
__________________
R3 3200G / Vega8 / Samsung UE40NU7100 Win11Pro 21H2 / 4K RGB 59Hz / AMD last driver MPC-HC 1.9.17 / madVR 0.92.17 / FSW / SM / 8bit |
8th February 2020, 16:36 | #58564 | Link |
Registered User
Join Date: May 2004
Posts: 5,351
|
I would. If you convert to sdr on an lg OLED, you're limiting the max nits you can output. By outputting HDR you can set the real display nits to ~700 which gives you a really good range to play with. Sdr will max out ~400 nits if you max the brightness and OLED settings but you risk running into abl doing that. Make sure to disable lg dynamic tone mapping in the settings for hdr.
Sent from my SM-G975U using Tapatalk
__________________
HTPC: Windows 11, AMD 5900X, RTX 3080, Pioneer Elite VSX-LX303, LG G2 77" OLED |
8th February 2020, 16:36 | #58565 | Link |
Registered User
Join Date: Oct 2018
Posts: 319
|
MadVR continues (and will continue, since this is what it considers best quality) giving priority to detail in the highlights at the cost of luminance. So it does all kind of fancy stuff to give the highlights the maximum possible brightness, but that maximum is always limited by the need to preserve detail.
Personally, I can live perfectly without so much level of detail, especially when the highlights are most of the time flashes that one doesn't have time to analyze, and for which the luminance largely determines their impact |
8th February 2020, 16:40 | #58566 | Link |
Registered User
Join Date: May 2004
Posts: 5,351
|
The latest builds contain curves that somewhat balance that equation if you choose the right one.
Sent from my SM-G975U using Tapatalk
__________________
HTPC: Windows 11, AMD 5900X, RTX 3080, Pioneer Elite VSX-LX303, LG G2 77" OLED |
8th February 2020, 18:04 | #58567 | Link |
Soul Seeker
Join Date: Sep 2013
Posts: 711
|
I'm wondering about the nvidia power management options in the control panel. Just tested the 3 modes, I read you guys recommend using "adaptive" for madvr, but in my test it acts like optimal power. With the 3 modes I get best results with prefer maximum performance - ~ 42% GPU usage and rendering times ~ 17 ms. With the other two modes - optimal power and adaptive I get ~ 54% GPU usage and rendering times ~ 22,30 ms...Why adaptive is still recommended?
|
8th February 2020, 18:59 | #58569 | Link |
Soul Seeker
Join Date: Sep 2013
Posts: 711
|
OK I understand about the wasting power, but adaptive doesn't make that much of a sense. Why raise the rendering times and the GPU usage, when you can use your card more efficiently. And I guess in adaptive mode it won't drop any frames like optimal power, right?
|
9th February 2020, 03:12 | #58572 | Link |
Registered User
Join Date: Jul 2016
Posts: 130
|
Actually, for me personally madVR has saved a lot of money. Instead of HDR display, I bought a 4K SDR monitor with much less price and bigger screen . And my SDR projector becomes refreshed again for HDR movies. Many thanks to Madshi.
|
9th February 2020, 04:31 | #58573 | Link |
Registered User
Join Date: Oct 2016
Posts: 896
|
Because the clocks are lower while the GPU load is still sufficiently low that you don't get frame drops. Lower clocks with higher GPU usage is more efficient.
__________________
HTPC: Windows 10 22H2, MediaPortal 1, LAV Filters/ReClock/madVR. DVB-C TV, Panasonic GT60, Denon 2310, Core 2 Duo E7400 oc'd, GeForce 1050 Ti 536.40 |
9th February 2020, 13:53 | #58579 | Link |
Registered User
Join Date: Oct 2012
Location: Akron, OH
Posts: 491
|
I haven’t been following the avsforum thread, so I have no idea how madshi is going about getting the results I see in the current test build. Maybe someone here who has been following that thread can provide more information.
|
9th February 2020, 14:40 | #58580 | Link |
Registered User
Join Date: Dec 2019
Posts: 36
|
Hello again. Is there a place where I can ask a question about madvr, like some forum maybe? I have a serious problem, which many people have probably had as well, so if there is such a place, someone should help me, I think.
__________________
System: i5 3570k; 8 Gb ddr3; SW: Windows 10 x64, latest updates; k-lite codec pack mega, latest updates; Monitor: HP x2301, Receiver: YAMAHA RX-V371 |
Tags |
direct compute, dithering, error diffusion, madvr, ngu, nnedi3, quality, renderer, scaling, uhd upscaling, upsampling |
Thread Tools | Search this Thread |
Display Modes | |
|
|