Message ID | de7375affe908506e072ec505b12c133b8b43945.1517868388.git.luto@kernel.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Hi Andy, thanks for getting involved with PSR and sorry for not replying sooner. I first saw this patch on that bugzilla entry but only now I stop to really think why I have written the code that way. So some clarity below. On Mon, Feb 05, 2018 at 10:07:09PM +0000, Andy Lutomirski wrote: > The current PSR code has a two call sites that each schedule delayed > work to activate PSR. As far as I can tell, each call site intends > to keep PSR inactive for the given amount of time and then allow it > to be activated. > > The call sites are: > > - intel_psr_enable(), which explicitly states in a comment that > it's trying to keep PSR off a short time after the dispay is > initialized as a workaround. First of all I really want to kill this call here and remove the FIXME. It was an ugly hack that I added to solve a corner case that was leaving me with blank screens when activating so sooner. > > - intel_psr_flush(). There isn't an explcit explanation, but the > intent is presumably to keep PSR off until the display has been > idle for 100ms. The reason for 100 is kind of ugly-nonsense-empirical value I concluded from VLV/CHV experience. On platforms with HW tracking HW waits few identical frames until really activating PSR. VLV/CHV activation is immediate. But HW is also different and there it seemed that hw needed a few more time before starting the transitions. Furthermore I didn't want to add that so quickly because I didn't want to take the risk of killing battery with software tracking when doing transitions so quickly using software tracking. > > The current code doesn't actually accomplish either of these goals. > Rather than keeping PSR inactive for the given amount of time, it > will schedule PSR for activation after the given time, with the > earliest target time in such a request winning. Putting that way I was asking myself how that hack had ever fixed my issue. Because the way you explained here seems obvious that it wouldn't ever fix my bug or any other. So I applied your patch and it made even more sense (without considering the fact I want to kill the first call anyways). So I came back, removed your patch and tried to understand how did it ever worked. So, the thing is that intel_psr_flush will never be really executed if intel_psr_enable wasn't executed. That is guaranteed by: mutex_lock(&dev_priv->psr.lock); if (!dev_priv->psr.enabled) { So, intel_psr_enable will be for sure the first one to schedule the work delayed to the ugly higher delay. > > In other words, if intel_psr_enable() is immediately followed by > intel_psr_flush(), then PSR will be activated after 100ms even if > intel_psr_enable() wanted a longer delay. And, if the screen is > being constantly updated so that intel_psr_flush() is called once > per frame at 60Hz, PSR will still be activated once every 100ms. During this time you are right, many calls of intel_psr_exit coming from flush functions can be called... But none of them will schedule the work with 100 delay. they will skip on if (!work_busy(&dev_priv->psr.work.work)) So, the higher delayed *hack* will be respected and PSR won't get activated before that. On the other hand you might ask what if, for some strange reason, (intel_dp->panel_power_cycle_delay * 5) is lesser than 100. Well, on this case this would be ok, because it happens only once and only on gen > 9 where hw tracking will wait the minimal number of frames before the actual transition to PSR. In either cases I believe we are safe. Thanks, Rodrigo. > > Rewrite the code so that it does what was intended. This adds > a new function intel_psr_schedule(), which will enable PSR after > the requested time but no sooner. > > Signed-off-by: Andy Lutomirski <luto@kernel.org> > --- > drivers/gpu/drm/i915/i915_debugfs.c | 9 +++-- > drivers/gpu/drm/i915/i915_drv.h | 4 ++- > drivers/gpu/drm/i915/intel_psr.c | 69 ++++++++++++++++++++++++++++++++----- > 3 files changed, 71 insertions(+), 11 deletions(-) > > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index c65e381b85f3..b67db93f905d 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -2663,8 +2663,13 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) > seq_printf(m, "Active: %s\n", yesno(dev_priv->psr.active)); > seq_printf(m, "Busy frontbuffer bits: 0x%03x\n", > dev_priv->psr.busy_frontbuffer_bits); > - seq_printf(m, "Re-enable work scheduled: %s\n", > - yesno(work_busy(&dev_priv->psr.work.work))); > + > + if (timer_pending(&dev_priv->psr.activate_timer)) > + seq_printf(m, "Activate scheduled: yes, in %ldms\n", > + (long)(dev_priv->psr.earliest_activate - jiffies) * > + 1000 / HZ); > + else > + seq_printf(m, "Re-enable scheduled: no\n"); > > if (HAS_DDI(dev_priv)) { > if (dev_priv->psr.psr2_support) > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index 46eb729b367d..c0fb7d65cda6 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -1192,7 +1192,9 @@ struct i915_psr { > bool source_ok; > struct intel_dp *enabled; > bool active; > - struct delayed_work work; > + struct timer_list activate_timer; > + struct work_struct activate_work; > + unsigned long earliest_activate; > unsigned busy_frontbuffer_bits; > bool psr2_support; > bool aux_frame_sync; > diff --git a/drivers/gpu/drm/i915/intel_psr.c b/drivers/gpu/drm/i915/intel_psr.c > index 55ea5eb3b7df..333d90d4e5af 100644 > --- a/drivers/gpu/drm/i915/intel_psr.c > +++ b/drivers/gpu/drm/i915/intel_psr.c > @@ -461,6 +461,30 @@ static void intel_psr_activate(struct intel_dp *intel_dp) > dev_priv->psr.active = true; > } > > +static void intel_psr_schedule(struct drm_i915_private *dev_priv, > + unsigned long min_wait_ms) > +{ > + unsigned long next; > + > + lockdep_assert_held(&dev_priv->psr.lock); > + > + /* > + * We update next_enable *and* call mod_timer() because it's > + * possible that intel_psr_work() has already been called and is > + * waiting for psr.lock. If that's the case, we don't want it > + * to immediately enable PSR. > + * > + * We also need to make sure that PSR is never activated earlier > + * than requested to avoid breaking intel_psr_enable()'s workaround > + * for pre-gen9 hardware. > + */ > + next = jiffies + msecs_to_jiffies(min_wait_ms); > + if (time_after(next, dev_priv->psr.earliest_activate)) { > + dev_priv->psr.earliest_activate = next; > + mod_timer(&dev_priv->psr.activate_timer, next); > + } > +} > + > static void hsw_psr_enable_source(struct intel_dp *intel_dp, > const struct intel_crtc_state *crtc_state) > { > @@ -544,8 +568,7 @@ void intel_psr_enable(struct intel_dp *intel_dp, > * - On HSW/BDW we get a recoverable frozen screen until > * next exit-activate sequence. > */ > - schedule_delayed_work(&dev_priv->psr.work, > - msecs_to_jiffies(intel_dp->panel_power_cycle_delay * 5)); > + intel_psr_schedule(dev_priv, intel_dp->panel_power_cycle_delay * 5); > } > > unlock: > @@ -660,13 +683,14 @@ void intel_psr_disable(struct intel_dp *intel_dp, > dev_priv->psr.enabled = NULL; > mutex_unlock(&dev_priv->psr.lock); > > - cancel_delayed_work_sync(&dev_priv->psr.work); > + cancel_work_sync(&dev_priv->psr.activate_work); > + del_timer_sync(&dev_priv->psr.activate_timer); > } > > static void intel_psr_work(struct work_struct *work) > { > struct drm_i915_private *dev_priv = > - container_of(work, typeof(*dev_priv), psr.work.work); > + container_of(work, typeof(*dev_priv), psr.activate_work); > struct intel_dp *intel_dp = dev_priv->psr.enabled; > struct drm_crtc *crtc = dp_to_dig_port(intel_dp)->base.base.crtc; > enum pipe pipe = to_intel_crtc(crtc)->pipe; > @@ -676,6 +700,7 @@ static void intel_psr_work(struct work_struct *work) > * PSR might take some time to get fully disabled > * and be ready for re-enable. > */ > + > if (HAS_DDI(dev_priv)) { > if (dev_priv->psr.psr2_support) { > if (intel_wait_for_register(dev_priv, > @@ -712,6 +737,18 @@ static void intel_psr_work(struct work_struct *work) > if (!intel_dp) > goto unlock; > > + > + if (!time_after_eq(jiffies, dev_priv->psr.earliest_activate)) { > + /* > + * We raced: intel_psr_schedule() tried to delay us, but > + * we were already in intel_psr_timer_fn() or already in > + * the workqueue. We can safely return -- the > + * intel_psr_schedule() call that put earliest_activeate > + * in the future will have called mod_timer(). > + */ > + goto unlock; > + } > + > /* > * The delayed work can race with an invalidate hence we need to > * recheck. Since psr_flush first clears this and then reschedules we > @@ -725,6 +762,20 @@ static void intel_psr_work(struct work_struct *work) > mutex_unlock(&dev_priv->psr.lock); > } > > +static void intel_psr_timer_fn(struct timer_list *timer) > +{ > + struct drm_i915_private *dev_priv = > + container_of(timer, typeof(*dev_priv), psr.activate_timer); > + > + /* > + * We need a non-atomic context to activate PSR. Using > + * delayed_work wouldn't be an improvement -- delayed_work is > + * just a timer that schedules work when it fires, but there's > + * no equivalent of mod_timer() for delayed_work. > + */ > + schedule_work(&dev_priv->psr.activate_work); > +} > + > static void intel_psr_exit(struct drm_i915_private *dev_priv) > { > struct intel_dp *intel_dp = dev_priv->psr.enabled; > @@ -905,9 +956,8 @@ void intel_psr_flush(struct drm_i915_private *dev_priv, > intel_psr_exit(dev_priv); > > if (!dev_priv->psr.active && !dev_priv->psr.busy_frontbuffer_bits) > - if (!work_busy(&dev_priv->psr.work.work)) > - schedule_delayed_work(&dev_priv->psr.work, > - msecs_to_jiffies(100)); > + intel_psr_schedule(dev_priv, 100); > + > mutex_unlock(&dev_priv->psr.lock); > } > > @@ -951,7 +1001,8 @@ void intel_psr_init(struct drm_i915_private *dev_priv) > dev_priv->psr.link_standby = false; > } > > - INIT_DELAYED_WORK(&dev_priv->psr.work, intel_psr_work); > + timer_setup(&dev_priv->psr.activate_timer, intel_psr_timer_fn, 0); > + INIT_WORK(&dev_priv->psr.activate_work, intel_psr_work); > mutex_init(&dev_priv->psr.lock); > > if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > @@ -967,4 +1018,6 @@ void intel_psr_init(struct drm_i915_private *dev_priv) > dev_priv->psr.activate = hsw_psr_activate; > dev_priv->psr.setup_vsc = hsw_psr_setup_vsc; > } > + > + dev_priv->psr.earliest_activate = jiffies; > } > -- > 2.14.3 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/intel-gfx
On Thu, 2018-02-08 at 14:48 -0800, Rodrigo Vivi wrote: > Hi Andy, > > thanks for getting involved with PSR and sorry for not replying sooner. > > I first saw this patch on that bugzilla entry but only now I stop to > really think why I have written the code that way. > > So some clarity below. > > On Mon, Feb 05, 2018 at 10:07:09PM +0000, Andy Lutomirski wrote: > > The current PSR code has a two call sites that each schedule delayed > > work to activate PSR. As far as I can tell, each call site intends > > to keep PSR inactive for the given amount of time and then allow it > > to be activated. > > > > The call sites are: > > > > - intel_psr_enable(), which explicitly states in a comment that > > it's trying to keep PSR off a short time after the dispay is > > initialized as a workaround. > > First of all I really want to kill this call here and remove the > FIXME. It was an ugly hack that I added to solve a corner case > that was leaving me with blank screens when activating so sooner. > > > > > - intel_psr_flush(). There isn't an explcit explanation, but the > > intent is presumably to keep PSR off until the display has been > > idle for 100ms. > > The reason for 100 is kind of ugly-nonsense-empirical value > I concluded from VLV/CHV experience. > On platforms with HW tracking HW waits few identical frames > until really activating PSR. VLV/CHV activation is immediate. > But HW is also different and there it seemed that hw needed a > few more time before starting the transitions. > Furthermore I didn't want to add that so quickly because I didn't > want to take the risk of killing battery with software tracking > when doing transitions so quickly using software tracking. > > > > > The current code doesn't actually accomplish either of these goals. > > Rather than keeping PSR inactive for the given amount of time, it > > will schedule PSR for activation after the given time, with the > > earliest target time in such a request winning. > > Putting that way I was asking myself how that hack had ever fixed > my issue. Because the way you explained here seems obvious that it > wouldn't ever fix my bug or any other. > > So I applied your patch and it made even more sense (without considering > the fact I want to kill the first call anyways). > > So I came back, removed your patch and tried to understand how did > it ever worked. > > So, the thing is that intel_psr_flush will never be really executed > if intel_psr_enable wasn't executed. That is guaranteed by: > > mutex_lock(&dev_priv->psr.lock); > if (!dev_priv->psr.enabled) { > > So, intel_psr_enable will be for sure the first one to schedule the > work delayed to the ugly higher delay. > > > > > In other words, if intel_psr_enable() is immediately followed by > > intel_psr_flush(), then PSR will be activated after 100ms even if > > intel_psr_enable() wanted a longer delay. And, if the screen is > > being constantly updated so that intel_psr_flush() is called once > > per frame at 60Hz, PSR will still be activated once every 100ms. > > During this time you are right, many calls of intel_psr_exit > coming from flush functions can be called... But none of > them will schedule the work with 100 delay. > > they will skip on > if (!work_busy(&dev_priv->psr.work.work)) Wouldn't work_busy() return false until the work is actually queued which is 100ms after calling schedule_delayed_work()? For e.g, flushes at 0, 16, 32...96 will have work_busy() returning false until 100ms. The first psr_work will end up getting scheduled at 100ms, which I believe is not what we want. However, I think if (dev_priv->psr.busy_frontbuffer_bits) goto unlock; intel_psr_activate(intel_dp); in psr_work might prevent activate being called at 100ms if an invalidate happened to be called before that. > > So, the higher delayed *hack* will be respected and PSR won't get > activated before that. > > On the other hand you might ask what if, > for some strange reason, > (intel_dp->panel_power_cycle_delay * 5) is lesser than 100. > Well, on this case this would be ok, because it happens only > once and only on gen > 9 where hw tracking will wait the minimal > number of frames before the actual transition to PSR. > > In either cases I believe we are safe. > > Thanks, > Rodrigo.
> On Feb 8, 2018, at 4:39 PM, Pandiyan, Dhinakaran <dhinakaran.pandiyan@intel.com> wrote: > > >> On Thu, 2018-02-08 at 14:48 -0800, Rodrigo Vivi wrote: >> Hi Andy, >> >> thanks for getting involved with PSR and sorry for not replying sooner. >> >> I first saw this patch on that bugzilla entry but only now I stop to >> really think why I have written the code that way. >> >> So some clarity below. >> >>> On Mon, Feb 05, 2018 at 10:07:09PM +0000, Andy Lutomirski wrote: >>> The current PSR code has a two call sites that each schedule delayed >>> work to activate PSR. As far as I can tell, each call site intends >>> to keep PSR inactive for the given amount of time and then allow it >>> to be activated. >>> >>> The call sites are: >>> >>> - intel_psr_enable(), which explicitly states in a comment that >>> it's trying to keep PSR off a short time after the dispay is >>> initialized as a workaround. >> >> First of all I really want to kill this call here and remove the >> FIXME. It was an ugly hack that I added to solve a corner case >> that was leaving me with blank screens when activating so sooner. >> >>> >>> - intel_psr_flush(). There isn't an explcit explanation, but the >>> intent is presumably to keep PSR off until the display has been >>> idle for 100ms. >> >> The reason for 100 is kind of ugly-nonsense-empirical value >> I concluded from VLV/CHV experience. >> On platforms with HW tracking HW waits few identical frames >> until really activating PSR. VLV/CHV activation is immediate. >> But HW is also different and there it seemed that hw needed a >> few more time before starting the transitions. >> Furthermore I didn't want to add that so quickly because I didn't >> want to take the risk of killing battery with software tracking >> when doing transitions so quickly using software tracking. >> >>> >>> The current code doesn't actually accomplish either of these goals. >>> Rather than keeping PSR inactive for the given amount of time, it >>> will schedule PSR for activation after the given time, with the >>> earliest target time in such a request winning. >> >> Putting that way I was asking myself how that hack had ever fixed >> my issue. Because the way you explained here seems obvious that it >> wouldn't ever fix my bug or any other. >> >> So I applied your patch and it made even more sense (without considering >> the fact I want to kill the first call anyways). >> >> So I came back, removed your patch and tried to understand how did >> it ever worked. >> >> So, the thing is that intel_psr_flush will never be really executed >> if intel_psr_enable wasn't executed. That is guaranteed by: >> >> mutex_lock(&dev_priv->psr.lock); >> if (!dev_priv->psr.enabled) { >> >> So, intel_psr_enable will be for sure the first one to schedule the >> work delayed to the ugly higher delay. >> >>> >>> In other words, if intel_psr_enable() is immediately followed by >>> intel_psr_flush(), then PSR will be activated after 100ms even if >>> intel_psr_enable() wanted a longer delay. And, if the screen is >>> being constantly updated so that intel_psr_flush() is called once >>> per frame at 60Hz, PSR will still be activated once every 100ms. >> >> During this time you are right, many calls of intel_psr_exit >> coming from flush functions can be called... But none of >> them will schedule the work with 100 delay. >> >> they will skip on >> if (!work_busy(&dev_priv->psr.work.work)) As below, the first call will. Then, 100ms later, the work will fire. Then the next flush will schedule it again, etc. > > Wouldn't work_busy() return false until the work is actually queued > which is 100ms after calling schedule_delayed_work()? > > For e.g, flushes at 0, 16, 32...96 will have work_busy() returning false > until 100ms. > > The first psr_work will end up getting scheduled at 100ms, which I > believe is not what we want. Indeed. I stuck some printks in and this seems to be what happens. > > > However, I think > > if (dev_priv->psr.busy_frontbuffer_bits) > goto unlock; > > intel_psr_activate(intel_dp); > > in psr_work might prevent activate being called at 100ms if an > invalidate happened to be called before that. > On my system, invalidate is never called. Even if it were called, that check would only help if we got lucky and the work fired after invalidate and before the subsequent flush. > > > >> >> So, the higher delayed *hack* will be respected and PSR won't get >> activated before that. >> >> On the other hand you might ask what if, >> for some strange reason, >> (intel_dp->panel_power_cycle_delay * 5) is lesser than 100. >> Well, on this case this would be ok, because it happens only >> once and only on gen > 9 where hw tracking will wait the minimal >> number of frames before the actual transition to PSR. >> >> In either cases I believe we are safe. >> >> Thanks, >> Rodrigo.
On Thu, 2018-02-08 at 17:01 -0800, Andy Lutomirski wrote: > > > > On Feb 8, 2018, at 4:39 PM, Pandiyan, Dhinakaran <dhinakaran.pandiyan@intel.com> wrote: > > > > > >> On Thu, 2018-02-08 at 14:48 -0800, Rodrigo Vivi wrote: > >> Hi Andy, > >> > >> thanks for getting involved with PSR and sorry for not replying sooner. > >> > >> I first saw this patch on that bugzilla entry but only now I stop to > >> really think why I have written the code that way. > >> > >> So some clarity below. > >> > >>> On Mon, Feb 05, 2018 at 10:07:09PM +0000, Andy Lutomirski wrote: > >>> The current PSR code has a two call sites that each schedule delayed > >>> work to activate PSR. As far as I can tell, each call site intends > >>> to keep PSR inactive for the given amount of time and then allow it > >>> to be activated. > >>> > >>> The call sites are: > >>> > >>> - intel_psr_enable(), which explicitly states in a comment that > >>> it's trying to keep PSR off a short time after the dispay is > >>> initialized as a workaround. > >> > >> First of all I really want to kill this call here and remove the > >> FIXME. It was an ugly hack that I added to solve a corner case > >> that was leaving me with blank screens when activating so sooner. > >> > >>> > >>> - intel_psr_flush(). There isn't an explcit explanation, but the > >>> intent is presumably to keep PSR off until the display has been > >>> idle for 100ms. > >> > >> The reason for 100 is kind of ugly-nonsense-empirical value > >> I concluded from VLV/CHV experience. > >> On platforms with HW tracking HW waits few identical frames > >> until really activating PSR. VLV/CHV activation is immediate. > >> But HW is also different and there it seemed that hw needed a > >> few more time before starting the transitions. > >> Furthermore I didn't want to add that so quickly because I didn't > >> want to take the risk of killing battery with software tracking > >> when doing transitions so quickly using software tracking. > >> > >>> > >>> The current code doesn't actually accomplish either of these goals. > >>> Rather than keeping PSR inactive for the given amount of time, it > >>> will schedule PSR for activation after the given time, with the > >>> earliest target time in such a request winning. > >> > >> Putting that way I was asking myself how that hack had ever fixed > >> my issue. Because the way you explained here seems obvious that it > >> wouldn't ever fix my bug or any other. > >> > >> So I applied your patch and it made even more sense (without considering > >> the fact I want to kill the first call anyways). > >> > >> So I came back, removed your patch and tried to understand how did > >> it ever worked. > >> > >> So, the thing is that intel_psr_flush will never be really executed > >> if intel_psr_enable wasn't executed. That is guaranteed by: > >> > >> mutex_lock(&dev_priv->psr.lock); > >> if (!dev_priv->psr.enabled) { > >> > >> So, intel_psr_enable will be for sure the first one to schedule the > >> work delayed to the ugly higher delay. > >> > >>> > >>> In other words, if intel_psr_enable() is immediately followed by > >>> intel_psr_flush(), then PSR will be activated after 100ms even if > >>> intel_psr_enable() wanted a longer delay. And, if the screen is > >>> being constantly updated so that intel_psr_flush() is called once > >>> per frame at 60Hz, PSR will still be activated once every 100ms. > >> > >> During this time you are right, many calls of intel_psr_exit > >> coming from flush functions can be called... But none of > >> them will schedule the work with 100 delay. > >> > >> they will skip on > >> if (!work_busy(&dev_priv->psr.work.work)) > > As below, the first call will. Then, 100ms later, the work will fire. Then the next flush will schedule it again, etc. > > > > > Wouldn't work_busy() return false until the work is actually queued > > which is 100ms after calling schedule_delayed_work()? > > > > For e.g, flushes at 0, 16, 32...96 will have work_busy() returning false > > until 100ms. > > > > The first psr_work will end up getting scheduled at 100ms, which I > > believe is not what we want. > > Indeed. I stuck some printks in and this seems to be what happens. > > > > > > > However, I think > > > > if (dev_priv->psr.busy_frontbuffer_bits) > > goto unlock; > > > > intel_psr_activate(intel_dp); > > > > in psr_work might prevent activate being called at 100ms if an > > invalidate happened to be called before that. > > > > On my system, invalidate is never called. I noticed the same in console mode with fbdev, there are no invalidates. > Even if it were called, that check would only help if we got lucky and the work fired after invalidate and before the subsequent flush. > Agreed. This dependency on invalidate sure looks wrong. > > > > > > > >> > >> So, the higher delayed *hack* will be respected and PSR won't get > >> activated before that. > >> > >> On the other hand you might ask what if, > >> for some strange reason, > >> (intel_dp->panel_power_cycle_delay * 5) is lesser than 100. > >> Well, on this case this would be ok, because it happens only > >> once and only on gen > 9 where hw tracking will wait the minimal > >> number of frames before the actual transition to PSR. > >> > >> In either cases I believe we are safe. > >> > >> Thanks, > >> Rodrigo.
"Pandiyan, Dhinakaran" <dhinakaran.pandiyan@intel.com> writes: > On Thu, 2018-02-08 at 14:48 -0800, Rodrigo Vivi wrote: >> Hi Andy, >> >> thanks for getting involved with PSR and sorry for not replying sooner. >> >> I first saw this patch on that bugzilla entry but only now I stop to >> really think why I have written the code that way. >> >> So some clarity below. >> >> On Mon, Feb 05, 2018 at 10:07:09PM +0000, Andy Lutomirski wrote: >> > The current PSR code has a two call sites that each schedule delayed >> > work to activate PSR. As far as I can tell, each call site intends >> > to keep PSR inactive for the given amount of time and then allow it >> > to be activated. >> > >> > The call sites are: >> > >> > - intel_psr_enable(), which explicitly states in a comment that >> > it's trying to keep PSR off a short time after the dispay is >> > initialized as a workaround. >> >> First of all I really want to kill this call here and remove the >> FIXME. It was an ugly hack that I added to solve a corner case >> that was leaving me with blank screens when activating so sooner. >> >> > >> > - intel_psr_flush(). There isn't an explcit explanation, but the >> > intent is presumably to keep PSR off until the display has been >> > idle for 100ms. >> >> The reason for 100 is kind of ugly-nonsense-empirical value >> I concluded from VLV/CHV experience. >> On platforms with HW tracking HW waits few identical frames >> until really activating PSR. VLV/CHV activation is immediate. >> But HW is also different and there it seemed that hw needed a >> few more time before starting the transitions. >> Furthermore I didn't want to add that so quickly because I didn't >> want to take the risk of killing battery with software tracking >> when doing transitions so quickly using software tracking. >> >> > >> > The current code doesn't actually accomplish either of these goals. >> > Rather than keeping PSR inactive for the given amount of time, it >> > will schedule PSR for activation after the given time, with the >> > earliest target time in such a request winning. >> >> Putting that way I was asking myself how that hack had ever fixed >> my issue. Because the way you explained here seems obvious that it >> wouldn't ever fix my bug or any other. >> >> So I applied your patch and it made even more sense (without considering >> the fact I want to kill the first call anyways). >> >> So I came back, removed your patch and tried to understand how did >> it ever worked. >> >> So, the thing is that intel_psr_flush will never be really executed >> if intel_psr_enable wasn't executed. That is guaranteed by: >> >> mutex_lock(&dev_priv->psr.lock); >> if (!dev_priv->psr.enabled) { >> >> So, intel_psr_enable will be for sure the first one to schedule the >> work delayed to the ugly higher delay. >> >> > >> > In other words, if intel_psr_enable() is immediately followed by >> > intel_psr_flush(), then PSR will be activated after 100ms even if >> > intel_psr_enable() wanted a longer delay. And, if the screen is >> > being constantly updated so that intel_psr_flush() is called once >> > per frame at 60Hz, PSR will still be activated once every 100ms. >> >> During this time you are right, many calls of intel_psr_exit >> coming from flush functions can be called... But none of >> them will schedule the work with 100 delay. >> >> they will skip on >> if (!work_busy(&dev_priv->psr.work.work)) > > Wouldn't work_busy() return false until the work is actually queued > which is 100ms after calling schedule_delayed_work()? That's not my understanding of work_busy man. "work_busy - test whether a work is currently pending or running" I consider it as pending one... But yeap... it was a long time ago that I did this so I'm not sure... > > For e.g, flushes at 0, 16, 32...96 will have work_busy() returning false > until 100ms. > > The first psr_work will end up getting scheduled at 100ms, which I > believe is not what we want. > > > However, I think > > if (dev_priv->psr.busy_frontbuffer_bits) > goto unlock; > > intel_psr_activate(intel_dp); > > in psr_work might prevent activate being called at 100ms if an > invalidate happened to be called before that. > > > > >> >> So, the higher delayed *hack* will be respected and PSR won't get >> activated before that. >> >> On the other hand you might ask what if, >> for some strange reason, >> (intel_dp->panel_power_cycle_delay * 5) is lesser than 100. >> Well, on this case this would be ok, because it happens only >> once and only on gen > 9 where hw tracking will wait the minimal >> number of frames before the actual transition to PSR. >> >> In either cases I believe we are safe. >> >> Thanks, >> Rodrigo. > _______________________________________________ > dri-devel mailing list > dri-devel@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/dri-devel
diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c index c65e381b85f3..b67db93f905d 100644 --- a/drivers/gpu/drm/i915/i915_debugfs.c +++ b/drivers/gpu/drm/i915/i915_debugfs.c @@ -2663,8 +2663,13 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) seq_printf(m, "Active: %s\n", yesno(dev_priv->psr.active)); seq_printf(m, "Busy frontbuffer bits: 0x%03x\n", dev_priv->psr.busy_frontbuffer_bits); - seq_printf(m, "Re-enable work scheduled: %s\n", - yesno(work_busy(&dev_priv->psr.work.work))); + + if (timer_pending(&dev_priv->psr.activate_timer)) + seq_printf(m, "Activate scheduled: yes, in %ldms\n", + (long)(dev_priv->psr.earliest_activate - jiffies) * + 1000 / HZ); + else + seq_printf(m, "Re-enable scheduled: no\n"); if (HAS_DDI(dev_priv)) { if (dev_priv->psr.psr2_support) diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index 46eb729b367d..c0fb7d65cda6 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h @@ -1192,7 +1192,9 @@ struct i915_psr { bool source_ok; struct intel_dp *enabled; bool active; - struct delayed_work work; + struct timer_list activate_timer; + struct work_struct activate_work; + unsigned long earliest_activate; unsigned busy_frontbuffer_bits; bool psr2_support; bool aux_frame_sync; diff --git a/drivers/gpu/drm/i915/intel_psr.c b/drivers/gpu/drm/i915/intel_psr.c index 55ea5eb3b7df..333d90d4e5af 100644 --- a/drivers/gpu/drm/i915/intel_psr.c +++ b/drivers/gpu/drm/i915/intel_psr.c @@ -461,6 +461,30 @@ static void intel_psr_activate(struct intel_dp *intel_dp) dev_priv->psr.active = true; } +static void intel_psr_schedule(struct drm_i915_private *dev_priv, + unsigned long min_wait_ms) +{ + unsigned long next; + + lockdep_assert_held(&dev_priv->psr.lock); + + /* + * We update next_enable *and* call mod_timer() because it's + * possible that intel_psr_work() has already been called and is + * waiting for psr.lock. If that's the case, we don't want it + * to immediately enable PSR. + * + * We also need to make sure that PSR is never activated earlier + * than requested to avoid breaking intel_psr_enable()'s workaround + * for pre-gen9 hardware. + */ + next = jiffies + msecs_to_jiffies(min_wait_ms); + if (time_after(next, dev_priv->psr.earliest_activate)) { + dev_priv->psr.earliest_activate = next; + mod_timer(&dev_priv->psr.activate_timer, next); + } +} + static void hsw_psr_enable_source(struct intel_dp *intel_dp, const struct intel_crtc_state *crtc_state) { @@ -544,8 +568,7 @@ void intel_psr_enable(struct intel_dp *intel_dp, * - On HSW/BDW we get a recoverable frozen screen until * next exit-activate sequence. */ - schedule_delayed_work(&dev_priv->psr.work, - msecs_to_jiffies(intel_dp->panel_power_cycle_delay * 5)); + intel_psr_schedule(dev_priv, intel_dp->panel_power_cycle_delay * 5); } unlock: @@ -660,13 +683,14 @@ void intel_psr_disable(struct intel_dp *intel_dp, dev_priv->psr.enabled = NULL; mutex_unlock(&dev_priv->psr.lock); - cancel_delayed_work_sync(&dev_priv->psr.work); + cancel_work_sync(&dev_priv->psr.activate_work); + del_timer_sync(&dev_priv->psr.activate_timer); } static void intel_psr_work(struct work_struct *work) { struct drm_i915_private *dev_priv = - container_of(work, typeof(*dev_priv), psr.work.work); + container_of(work, typeof(*dev_priv), psr.activate_work); struct intel_dp *intel_dp = dev_priv->psr.enabled; struct drm_crtc *crtc = dp_to_dig_port(intel_dp)->base.base.crtc; enum pipe pipe = to_intel_crtc(crtc)->pipe; @@ -676,6 +700,7 @@ static void intel_psr_work(struct work_struct *work) * PSR might take some time to get fully disabled * and be ready for re-enable. */ + if (HAS_DDI(dev_priv)) { if (dev_priv->psr.psr2_support) { if (intel_wait_for_register(dev_priv, @@ -712,6 +737,18 @@ static void intel_psr_work(struct work_struct *work) if (!intel_dp) goto unlock; + + if (!time_after_eq(jiffies, dev_priv->psr.earliest_activate)) { + /* + * We raced: intel_psr_schedule() tried to delay us, but + * we were already in intel_psr_timer_fn() or already in + * the workqueue. We can safely return -- the + * intel_psr_schedule() call that put earliest_activeate + * in the future will have called mod_timer(). + */ + goto unlock; + } + /* * The delayed work can race with an invalidate hence we need to * recheck. Since psr_flush first clears this and then reschedules we @@ -725,6 +762,20 @@ static void intel_psr_work(struct work_struct *work) mutex_unlock(&dev_priv->psr.lock); } +static void intel_psr_timer_fn(struct timer_list *timer) +{ + struct drm_i915_private *dev_priv = + container_of(timer, typeof(*dev_priv), psr.activate_timer); + + /* + * We need a non-atomic context to activate PSR. Using + * delayed_work wouldn't be an improvement -- delayed_work is + * just a timer that schedules work when it fires, but there's + * no equivalent of mod_timer() for delayed_work. + */ + schedule_work(&dev_priv->psr.activate_work); +} + static void intel_psr_exit(struct drm_i915_private *dev_priv) { struct intel_dp *intel_dp = dev_priv->psr.enabled; @@ -905,9 +956,8 @@ void intel_psr_flush(struct drm_i915_private *dev_priv, intel_psr_exit(dev_priv); if (!dev_priv->psr.active && !dev_priv->psr.busy_frontbuffer_bits) - if (!work_busy(&dev_priv->psr.work.work)) - schedule_delayed_work(&dev_priv->psr.work, - msecs_to_jiffies(100)); + intel_psr_schedule(dev_priv, 100); + mutex_unlock(&dev_priv->psr.lock); } @@ -951,7 +1001,8 @@ void intel_psr_init(struct drm_i915_private *dev_priv) dev_priv->psr.link_standby = false; } - INIT_DELAYED_WORK(&dev_priv->psr.work, intel_psr_work); + timer_setup(&dev_priv->psr.activate_timer, intel_psr_timer_fn, 0); + INIT_WORK(&dev_priv->psr.activate_work, intel_psr_work); mutex_init(&dev_priv->psr.lock); if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { @@ -967,4 +1018,6 @@ void intel_psr_init(struct drm_i915_private *dev_priv) dev_priv->psr.activate = hsw_psr_activate; dev_priv->psr.setup_vsc = hsw_psr_setup_vsc; } + + dev_priv->psr.earliest_activate = jiffies; }
The current PSR code has a two call sites that each schedule delayed work to activate PSR. As far as I can tell, each call site intends to keep PSR inactive for the given amount of time and then allow it to be activated. The call sites are: - intel_psr_enable(), which explicitly states in a comment that it's trying to keep PSR off a short time after the dispay is initialized as a workaround. - intel_psr_flush(). There isn't an explcit explanation, but the intent is presumably to keep PSR off until the display has been idle for 100ms. The current code doesn't actually accomplish either of these goals. Rather than keeping PSR inactive for the given amount of time, it will schedule PSR for activation after the given time, with the earliest target time in such a request winning. In other words, if intel_psr_enable() is immediately followed by intel_psr_flush(), then PSR will be activated after 100ms even if intel_psr_enable() wanted a longer delay. And, if the screen is being constantly updated so that intel_psr_flush() is called once per frame at 60Hz, PSR will still be activated once every 100ms. Rewrite the code so that it does what was intended. This adds a new function intel_psr_schedule(), which will enable PSR after the requested time but no sooner. Signed-off-by: Andy Lutomirski <luto@kernel.org> --- drivers/gpu/drm/i915/i915_debugfs.c | 9 +++-- drivers/gpu/drm/i915/i915_drv.h | 4 ++- drivers/gpu/drm/i915/intel_psr.c | 69 ++++++++++++++++++++++++++++++++----- 3 files changed, 71 insertions(+), 11 deletions(-)