1. 21 6月, 2019 3 次提交
  2. 18 6月, 2019 1 次提交
  3. 14 6月, 2019 1 次提交
  4. 12 6月, 2019 2 次提交
  5. 06 6月, 2019 2 次提交
  6. 29 5月, 2019 1 次提交
  7. 25 5月, 2019 10 次提交
  8. 03 5月, 2019 2 次提交
  9. 24 4月, 2019 1 次提交
  10. 13 4月, 2019 1 次提交
  11. 11 4月, 2019 1 次提交
  12. 04 4月, 2019 1 次提交
  13. 28 3月, 2019 3 次提交
  14. 22 3月, 2019 1 次提交
  15. 21 3月, 2019 2 次提交
  16. 20 3月, 2019 6 次提交
  17. 08 2月, 2019 1 次提交
  18. 26 1月, 2019 1 次提交
    • A
      drm/sched: Refactor ring mirror list handling. · 222b5f04
      Andrey Grodzovsky 提交于
      Decauple sched threads stop and start and ring mirror
      list handling from the policy of what to do about the
      guilty jobs.
      When stoppping the sched thread and detaching sched fences
      from non signaled HW fenes wait for all signaled HW fences
      to complete before rerunning the jobs.
      
      v2: Fix resubmission of guilty job into HW after refactoring.
      
      v4:
      Full restart for all the jobs, not only from guilty ring.
      Extract karma increase into standalone function.
      
      v5:
      Rework waiting for signaled jobs without relying on the job
      struct itself as those might already be freed for non 'guilty'
      job's schedulers.
      Expose karma increase to drivers.
      
      v6:
      Use list_for_each_entry_safe_continue and drm_sched_process_job
      in case fence already signaled.
      Call drm_sched_increase_karma only once for amdgpu and add documentation.
      
      v7:
      Wait only for the latest job's fence.
      Suggested-by: NChristian Koenig <Christian.Koenig@amd.com>
      Signed-off-by: NAndrey Grodzovsky <andrey.grodzovsky@amd.com>
      Reviewed-by: NChristian König <christian.koenig@amd.com>
      Signed-off-by: NAlex Deucher <alexander.deucher@amd.com>
      222b5f04