On Fri, Oct 9, 2020 at 1:50 AM Sagi Grimberg <s...@grimberg.me> wrote:
> > @@ -412,8 +413,6 @@ int nvme_tcp_offload_limits(struct nvme_tcp_queue > > *queue, > > queue->ctrl->ctrl.max_segments = limits->max_ddp_sgl_len; > > queue->ctrl->ctrl.max_hw_sectors = > > limits->max_ddp_sgl_len << (ilog2(SZ_4K) - 9); > > - } else { > > - queue->ctrl->offloading_netdev = NULL; > > Squash this change to the patch that introduced it. OK, will look on that and I guess it should be fine to make this as you suggested > > + case NETDEV_GOING_DOWN: > > + mutex_lock(&nvme_tcp_ctrl_mutex); > > + list_for_each_entry(ctrl, &nvme_tcp_ctrl_list, list) { > > + if (ndev != ctrl->offloading_netdev) > > + continue; > > + nvme_tcp_error_recovery(&ctrl->ctrl); > > + } > > + mutex_unlock(&nvme_tcp_ctrl_mutex); > > + flush_workqueue(nvme_reset_wq); > > Worth a small comment that this we want the err_work to complete > here. So if someone changes workqueues he may see this. ack