Contributors: 9
| Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
| Hawking Zhang |
116 |
33.05% |
3 |
18.75% |
| Asad kamal |
109 |
31.05% |
1 |
6.25% |
| Likun Gao |
71 |
20.23% |
3 |
18.75% |
| Alex Deucher |
30 |
8.55% |
4 |
25.00% |
| Junwei (Martin) Zhang |
18 |
5.13% |
1 |
6.25% |
| Tao Zhou |
3 |
0.85% |
1 |
6.25% |
| Lijo Lazar |
2 |
0.57% |
1 |
6.25% |
| Victor Zhao |
1 |
0.28% |
1 |
6.25% |
| yipechai |
1 |
0.28% |
1 |
6.25% |
| Total |
351 |
|
16 |
|
/*
* Copyright 2023 Advanced Micro Devices, Inc.
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
* OTHER DEALINGS IN THE SOFTWARE.
*
*/
#include "amdgpu.h"
#include "amdgpu_ras.h"
#include <uapi/linux/kfd_ioctl.h>
int amdgpu_hdp_ras_sw_init(struct amdgpu_device *adev)
{
int err;
struct amdgpu_hdp_ras *ras;
if (!adev->hdp.ras)
return 0;
ras = adev->hdp.ras;
err = amdgpu_ras_register_ras_block(adev, &ras->ras_block);
if (err) {
dev_err(adev->dev, "Failed to register hdp ras block!\n");
return err;
}
strcpy(ras->ras_block.ras_comm.name, "hdp");
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__HDP;
ras->ras_block.ras_comm.type = AMDGPU_RAS_ERROR__MULTI_UNCORRECTABLE;
adev->hdp.ras_if = &ras->ras_block.ras_comm;
/* hdp ras follows amdgpu_ras_block_late_init_default for late init */
return 0;
}
void amdgpu_hdp_generic_flush(struct amdgpu_device *adev,
struct amdgpu_ring *ring)
{
if (!ring || !ring->funcs->emit_wreg) {
WREG32((adev->rmmio_remap.reg_offset +
KFD_MMIO_REMAP_HDP_MEM_FLUSH_CNTL) >>
2,
0);
if (adev->nbio.funcs->get_memsize)
adev->nbio.funcs->get_memsize(adev);
} else {
amdgpu_ring_emit_wreg(ring,
(adev->rmmio_remap.reg_offset +
KFD_MMIO_REMAP_HDP_MEM_FLUSH_CNTL) >>
2,
0);
}
}
void amdgpu_hdp_invalidate(struct amdgpu_device *adev, struct amdgpu_ring *ring)
{
if (adev->asic_funcs && adev->asic_funcs->invalidate_hdp)
adev->asic_funcs->invalidate_hdp(adev, ring);
else if (adev->hdp.funcs && adev->hdp.funcs->invalidate_hdp)
adev->hdp.funcs->invalidate_hdp(adev, ring);
}
void amdgpu_hdp_flush(struct amdgpu_device *adev, struct amdgpu_ring *ring)
{
if (adev->asic_funcs && adev->asic_funcs->flush_hdp)
adev->asic_funcs->flush_hdp(adev, ring);
else if (adev->hdp.funcs && adev->hdp.funcs->flush_hdp)
adev->hdp.funcs->flush_hdp(adev, ring);
}