1. 27 5月, 2022 1 次提交
  2. 16 5月, 2022 1 次提交
  3. 21 4月, 2022 1 次提交
  4. 05 3月, 2022 1 次提交
  5. 03 3月, 2022 2 次提交
  6. 17 2月, 2022 1 次提交
  7. 12 2月, 2022 1 次提交
    • K
      drm/amdgpu: remove ctx->lock · 461fa7b0
      Ken Xue 提交于
      KMD reports a warning on holding a lock from drm_syncobj_find_fence,
      when running amdgpu_test case “syncobj timeline test”.
      
      ctx->lock was designed to prevent concurrent "amdgpu_ctx_wait_prev_fence"
      calls and avoid dead reservation lock from GPU reset. since no reservation
      lock is held in latest GPU reset any more, ctx->lock can be simply removed
      and concurrent "amdgpu_ctx_wait_prev_fence" call also can be prevented by
      PD root bo reservation lock.
      
      call stacks:
      =================
      //hold lock
      amdgpu_cs_ioctl->amdgpu_cs_parser_init->mutex_lock(&parser->ctx->lock);
      …
      //report warning
      amdgpu_cs_dependencies->amdgpu_cs_process_syncobj_timeline_in_dep \
      ->amdgpu_syncobj_lookup_and_add_to_sync -> drm_syncobj_find_fence \
      -> lockdep_assert_none_held_once
      …
      amdgpu_cs_ioctl->amdgpu_cs_parser_fini->mutex_unlock(&parser->ctx->lock);
      Signed-off-by: NKen Xue <Ken.Xue@amd.com>
      Reviewed-by: NChristian König <christian.koenig@amd.com>
      Signed-off-by: NAlex Deucher <alexander.deucher@amd.com>
      461fa7b0
  8. 28 1月, 2022 1 次提交
    • A
      drm/amdgpu/UAPI: add new CTX OP to get/set stable pstates · 8cda7a4f
      Alex Deucher 提交于
      Add a new CTX ioctl operation to set stable pstates for profiling.
      When creating traces for tools like RGP or using SPM or doing
      performance profiling, it's required to enable a special
      stable profiling power state on the GPU.  These profiling
      states set fixed clocks and disable certain other power
      features like powergating which may impact the results.
      
      Historically, these profiling pstates were enabled via sysfs,
      but this adds an interface to enable it via the CTX ioctl
      from the application.  Since the power state is global
      only one application can set it at a time, so if multiple
      applications try and use it only the first will get it,
      the ioctl will return -EBUSY for others.  The sysfs interface
      will override whatever has been set by this interface.
      
      Mesa MR: https://gitlab.freedesktop.org/mesa/drm/-/merge_requests/207
      
      v2: don't default r = 0;
      v3: rebase on Evan's PM cleanup
      Reviewed-by: NEvan Quan <evan.quan@amd.com>
      Signed-off-by: NAlex Deucher <alexander.deucher@amd.com>
      8cda7a4f
  9. 02 9月, 2021 3 次提交
  10. 03 6月, 2021 1 次提交
  11. 28 5月, 2021 2 次提交
  12. 14 5月, 2021 2 次提交
  13. 05 5月, 2021 1 次提交
  14. 03 11月, 2020 1 次提交
  15. 10 10月, 2020 1 次提交
  16. 04 9月, 2020 1 次提交
  17. 25 8月, 2020 1 次提交
  18. 19 8月, 2020 2 次提交
  19. 15 8月, 2020 1 次提交
  20. 28 7月, 2020 1 次提交
    • D
      drm/amdgpu: fix system hang issue during GPU reset · df9c8d1a
      Dennis Li 提交于
      when GPU hang, driver has multi-paths to enter amdgpu_device_gpu_recover,
      the atomic adev->in_gpu_reset and hive->in_reset are used to avoid
      re-entering GPU recovery.
      
      During GPU reset and resume, it is unsafe that other threads access GPU,
      which maybe cause GPU reset failed. Therefore the new rw_semaphore
      adev->reset_sem is introduced, which protect GPU from being accessed by
      external threads during recovery.
      
      v2:
      1. add rwlock for some ioctls, debugfs and file-close function.
      2. change to use dqm->is_resetting and dqm_lock for protection in kfd
      driver.
      3. remove try_lock and change adev->in_gpu_reset as atomic, to avoid
      re-enter GPU recovery for the same GPU hang.
      
      v3:
      1. change back to use adev->reset_sem to protect kfd callback
      functions, because dqm_lock couldn't protect all codes, for example:
      free_mqd must be called outside of dqm_lock;
      
      [ 1230.176199] Hardware name: Supermicro SYS-7049GP-TRT/X11DPG-QT, BIOS 3.1 05/23/2019
      [ 1230.177221] Call Trace:
      [ 1230.178249]  dump_stack+0x98/0xd5
      [ 1230.179443]  amdgpu_virt_kiq_reg_write_reg_wait+0x181/0x190 [amdgpu]
      [ 1230.180673]  gmc_v9_0_flush_gpu_tlb+0xcc/0x310 [amdgpu]
      [ 1230.181882]  amdgpu_gart_unbind+0xa9/0xe0 [amdgpu]
      [ 1230.183098]  amdgpu_ttm_backend_unbind+0x46/0x180 [amdgpu]
      [ 1230.184239]  ? ttm_bo_put+0x171/0x5f0 [ttm]
      [ 1230.185394]  ttm_tt_unbind+0x21/0x40 [ttm]
      [ 1230.186558]  ttm_tt_destroy.part.12+0x12/0x60 [ttm]
      [ 1230.187707]  ttm_tt_destroy+0x13/0x20 [ttm]
      [ 1230.188832]  ttm_bo_cleanup_memtype_use+0x36/0x80 [ttm]
      [ 1230.189979]  ttm_bo_put+0x1be/0x5f0 [ttm]
      [ 1230.191230]  amdgpu_bo_unref+0x1e/0x30 [amdgpu]
      [ 1230.192522]  amdgpu_amdkfd_free_gtt_mem+0xaf/0x140 [amdgpu]
      [ 1230.193833]  free_mqd+0x25/0x40 [amdgpu]
      [ 1230.195143]  destroy_queue_cpsch+0x1a7/0x270 [amdgpu]
      [ 1230.196475]  pqm_destroy_queue+0x105/0x260 [amdgpu]
      [ 1230.197819]  kfd_ioctl_destroy_queue+0x37/0x70 [amdgpu]
      [ 1230.199154]  kfd_ioctl+0x277/0x500 [amdgpu]
      [ 1230.200458]  ? kfd_ioctl_get_clock_counters+0x60/0x60 [amdgpu]
      [ 1230.201656]  ? tomoyo_file_ioctl+0x19/0x20
      [ 1230.202831]  ksys_ioctl+0x98/0xb0
      [ 1230.204004]  __x64_sys_ioctl+0x1a/0x20
      [ 1230.205174]  do_syscall_64+0x5f/0x250
      [ 1230.206339]  entry_SYSCALL_64_after_hwframe+0x49/0xbe
      
      2. remove try_lock and introduce atomic hive->in_reset, to avoid
      re-enter GPU recovery.
      
      v4:
      1. remove an unnecessary whitespace change in kfd_chardev.c
      2. remove comment codes in amdgpu_device.c
      3. add more detailed comment in commit message
      4. define a wrap function amdgpu_in_reset
      
      v5:
      1. Fix some style issues.
      Reviewed-by: NHawking Zhang <Hawking.Zhang@amd.com>
      Suggested-by: NAndrey Grodzovsky <andrey.grodzovsky@amd.com>
      Suggested-by: NChristian König <christian.koenig@amd.com>
      Suggested-by: NFelix Kuehling <Felix.Kuehling@amd.com>
      Suggested-by: NLijo Lazar <Lijo.Lazar@amd.com>
      Suggested-by: NLuben Tukov <luben.tuikov@amd.com>
      Signed-off-by: NDennis Li <Dennis.Li@amd.com>
      Signed-off-by: NAlex Deucher <alexander.deucher@amd.com>
      df9c8d1a
  21. 09 4月, 2020 1 次提交
    • N
      drm/amdgpu: rework sched_list generation · 1c6d567b
      Nirmoy Das 提交于
      Generate HW IP's sched_list in amdgpu_ring_init() instead of
      amdgpu_ctx.c. This makes amdgpu_ctx_init_compute_sched(),
      ring.has_high_prio and amdgpu_ctx_init_sched() unnecessary.
      This patch also stores sched_list for all HW IPs in one big
      array in struct amdgpu_device which makes amdgpu_ctx_init_entity()
      much more leaner.
      
      v2:
      fix a coding style issue
      do not use drm hw_ip const to populate amdgpu_ring_type enum
      
      v3:
      remove ctx reference and move sched array and num_sched to a struct
      use num_scheds to detect uninitialized scheduler list
      
      v4:
      use array_index_nospec for user space controlled variables
      fix possible checkpatch.pl warnings
      Signed-off-by: NNirmoy Das <nirmoy.das@amd.com>
      Reviewed-by: NChristian König <christian.koenig@amd.com>
      Signed-off-by: NAlex Deucher <alexander.deucher@amd.com>
      1c6d567b
  22. 17 3月, 2020 2 次提交
  23. 11 3月, 2020 1 次提交
  24. 10 3月, 2020 2 次提交
  25. 31 1月, 2020 1 次提交
  26. 23 1月, 2020 1 次提交
  27. 24 12月, 2019 1 次提交
  28. 19 12月, 2019 2 次提交
  29. 12 12月, 2019 2 次提交
  30. 19 11月, 2019 1 次提交