diff options
author | Chuck Lever <chuck.lever@oracle.com> | 2024-04-19 11:51:12 -0400 |
---|---|---|
committer | Chuck Lever <chuck.lever@oracle.com> | 2024-04-20 11:20:41 -0400 |
commit | 32cf5a4eda464d76d553ee3f1b06c4d33d796c52 (patch) | |
tree | aabc832c30c8176a3675573bc4157163e869e119 /drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c | |
parent | f488138b526715c6d2568d7329c4477911be4210 (diff) |
Revert "svcrdma: Add Write chunk WRs to the RPC's Send WR chain"
Performance regression reported with NFS/RDMA using Omnipath,
bisected to commit e084ee673c77 ("svcrdma: Add Write chunk WRs to
the RPC's Send WR chain").
Tracing on the server reports:
nfsd-7771 [060] 1758.891809: svcrdma_sq_post_err:
cq.id=205 cid=226 sc_sq_avail=13643/851 status=-12
sq_post_err reports ENOMEM, and the rdma->sc_sq_avail (13643) is
larger than rdma->sc_sq_depth (851). The number of available Send
Queue entries is always supposed to be smaller than the Send Queue
depth. That seems like a Send Queue accounting bug in svcrdma.
As it's getting to be late in the 6.9-rc cycle, revert this commit.
It can be revisited in a subsequent kernel release.
Link: https://bugzilla.kernel.org/show_bug.cgi?id=218743
Fixes: e084ee673c77 ("svcrdma: Add Write chunk WRs to the RPC's Send WR chain")
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Diffstat (limited to 'drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c')
0 files changed, 0 insertions, 0 deletions