From patchwork Thu Jul 29 18:04:42 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jeff Layton X-Patchwork-Id: 488780 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-19.7 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED, DKIM_VALID, DKIM_VALID_AU, INCLUDES_CR_TRAILER, INCLUDES_PATCH, MAILING_LIST_MULTI, SPF_HELO_NONE, SPF_PASS, URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 2A35CC4338F for ; Thu, 29 Jul 2021 18:04:45 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 083176054E for ; Thu, 29 Jul 2021 18:04:45 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229947AbhG2SEr (ORCPT ); Thu, 29 Jul 2021 14:04:47 -0400 Received: from mail.kernel.org ([198.145.29.99]:58576 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229731AbhG2SEr (ORCPT ); Thu, 29 Jul 2021 14:04:47 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 581D360F48; Thu, 29 Jul 2021 18:04:43 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1627581883; bh=svDqoy4I4UEPoVjXyx5yc4e8EwnyP0te/OIdBTiyi6A=; h=From:To:Cc:Subject:Date:From; b=vNzqdCak0XlbYrCqXWF4P/xKasTLg0IzbybL3oKSLfgQKz3r4GUj4GaKLnV/5jTaW lcVpCPHHu7A4f3M8loEirrOr1AYfoJxjhVdEcykOAVtcsUoM4Bt46/Z/xk36uL1qAt ikEhQWJydB/YVVRTkV7TnSRy3qN2jt3kXLKbQr06cPf4jsaLMRDNfqUi2y+yriaRK5 iWJAZJSViykNKbHltQO0agIWV1nbxdA60WERczBOFxfJBysP/sPBSdCzQRv7MJlzZe +Wlfei6vPdd8WKsz4ylsDGaGiK2B+5Nk2N/u5KLrO6WiQj7UKBByno75rX7u544O9C kiyCnbSEnWLGQ== From: Jeff Layton To: ceph-devel@vger.kernel.org Cc: pdonnell@redhat.com, idryomov@gmail.com Subject: [PATCH v3] ceph: dump info about cap flushes when we're waiting too long for them Date: Thu, 29 Jul 2021 14:04:42 -0400 Message-Id: <20210729180442.177399-1-jlayton@kernel.org> X-Mailer: git-send-email 2.31.1 MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: ceph-devel@vger.kernel.org We've had some cases of hung umounts in teuthology testing. It looks like client is waiting for cap flushes to complete, but they aren't. Add a field to the inode to track the highest cap flush tid seen for that inode. Also, add a backpointer to the inode to the ceph_cap_flush struct. Change wait_caps_flush to wait 60s, and then dump info about the condition of the list. Also, print pr_info messages if we end up dropping a FLUSH_ACK for an inode onto the floor. Reported-by: Patrick Donnelly URL: https://tracker.ceph.com/issues/51279 Signed-off-by: Jeff Layton --- fs/ceph/caps.c | 17 +++++++++++++++-- fs/ceph/inode.c | 1 + fs/ceph/mds_client.c | 31 +++++++++++++++++++++++++++++-- fs/ceph/super.h | 2 ++ 4 files changed, 47 insertions(+), 4 deletions(-) v3: more debugging has shown the client waiting on FLUSH_ACK messages that seem to never have come. Add some new printks if we end up dropping a FLUSH_ACK onto the floor. diff --git a/fs/ceph/caps.c b/fs/ceph/caps.c index 7ae83d06d48c..cb551c9e5867 100644 --- a/fs/ceph/caps.c +++ b/fs/ceph/caps.c @@ -1829,6 +1829,7 @@ static u64 __mark_caps_flushing(struct inode *inode, swap(cf, ci->i_prealloc_cap_flush); cf->caps = flushing; cf->wake = wake; + cf->ci = ci; spin_lock(&mdsc->cap_dirty_lock); list_del_init(&ci->i_dirty_item); @@ -3588,6 +3589,10 @@ static void handle_cap_flush_ack(struct inode *inode, u64 flush_tid, bool wake_ci = false; bool wake_mdsc = false; + /* track latest cap flush ack seen for this inode */ + if (flush_tid > ci->i_last_cap_flush_ack) + ci->i_last_cap_flush_ack = flush_tid; + list_for_each_entry_safe(cf, tmp_cf, &ci->i_cap_flush_list, i_list) { /* Is this the one that was flushed? */ if (cf->tid == flush_tid) @@ -4116,7 +4121,11 @@ void ceph_handle_caps(struct ceph_mds_session *session, (unsigned)seq); if (!inode) { - dout(" i don't have ino %llx\n", vino.ino); + if (op == CEPH_CAP_OP_FLUSH_ACK) + pr_info("%s: can't find ino %llx:%llx for flush_ack!\n", + __func__, vino.snap, vino.ino); + else + dout(" i don't have ino %llx\n", vino.ino); if (op == CEPH_CAP_OP_IMPORT) { cap = ceph_get_cap(mdsc, NULL); @@ -4169,10 +4178,14 @@ void ceph_handle_caps(struct ceph_mds_session *session, spin_lock(&ci->i_ceph_lock); cap = __get_cap_for_mds(ceph_inode(inode), session->s_mds); if (!cap) { - dout(" no cap on %p ino %llx.%llx from mds%d\n", + dout(" no cap on %p ino %llx:%llx from mds%d\n", inode, ceph_ino(inode), ceph_snap(inode), session->s_mds); spin_unlock(&ci->i_ceph_lock); + if (op == CEPH_CAP_OP_FLUSH_ACK) + pr_info("%s: no cap on %p ino %llx:%llx from mds%d for flush_ack!\n", + __func__, inode, ceph_ino(inode), + ceph_snap(inode), session->s_mds); goto flush_cap_releases; } diff --git a/fs/ceph/inode.c b/fs/ceph/inode.c index 1bd2cc015913..84e4f112fc45 100644 --- a/fs/ceph/inode.c +++ b/fs/ceph/inode.c @@ -499,6 +499,7 @@ struct inode *ceph_alloc_inode(struct super_block *sb) INIT_LIST_HEAD(&ci->i_cap_snaps); ci->i_head_snapc = NULL; ci->i_snap_caps = 0; + ci->i_last_cap_flush_ack = 0; ci->i_last_rd = ci->i_last_wr = jiffies - 3600 * HZ; for (i = 0; i < CEPH_FILE_MODE_BITS; i++) diff --git a/fs/ceph/mds_client.c b/fs/ceph/mds_client.c index c3fa0c0e4f6c..fc26527b8059 100644 --- a/fs/ceph/mds_client.c +++ b/fs/ceph/mds_client.c @@ -2064,6 +2064,24 @@ static int check_caps_flush(struct ceph_mds_client *mdsc, return ret; } +static void dump_cap_flushes(struct ceph_mds_client *mdsc, u64 want_tid) +{ + struct ceph_cap_flush *cf; + + pr_info("%s: still waiting for cap flushes through %llu\n:\n", + __func__, want_tid); + spin_lock(&mdsc->cap_dirty_lock); + list_for_each_entry(cf, &mdsc->cap_flush_list, g_list) { + if (cf->tid > want_tid) + break; + pr_info("%llx:%llx %s %llu %llu %d\n", + ceph_vinop(&cf->ci->vfs_inode), + ceph_cap_string(cf->caps), cf->tid, + cf->ci->i_last_cap_flush_ack, cf->wake); + } + spin_unlock(&mdsc->cap_dirty_lock); +} + /* * flush all dirty inode data to disk. * @@ -2072,10 +2090,19 @@ static int check_caps_flush(struct ceph_mds_client *mdsc, static void wait_caps_flush(struct ceph_mds_client *mdsc, u64 want_flush_tid) { + long ret; + dout("check_caps_flush want %llu\n", want_flush_tid); - wait_event(mdsc->cap_flushing_wq, - check_caps_flush(mdsc, want_flush_tid)); + do { + ret = wait_event_timeout(mdsc->cap_flushing_wq, + check_caps_flush(mdsc, want_flush_tid), 60 * HZ); + if (ret == 0) + dump_cap_flushes(mdsc, want_flush_tid); + else if (ret == 1) + pr_info("%s: condition evaluated to true after timeout!\n", + __func__); + } while (ret == 0); dout("check_caps_flush ok, flushed thru %llu\n", want_flush_tid); } diff --git a/fs/ceph/super.h b/fs/ceph/super.h index 07eb542efa1d..d51d42a00f33 100644 --- a/fs/ceph/super.h +++ b/fs/ceph/super.h @@ -189,6 +189,7 @@ struct ceph_cap_flush { bool wake; /* wake up flush waiters when finish ? */ struct list_head g_list; // global struct list_head i_list; // per inode + struct ceph_inode_info *ci; }; /* @@ -388,6 +389,7 @@ struct ceph_inode_info { struct ceph_snap_context *i_head_snapc; /* set if wr_buffer_head > 0 or dirty|flushing caps */ unsigned i_snap_caps; /* cap bits for snapped files */ + u64 i_last_cap_flush_ack; /* latest cap flush_ack tid for this inode */ unsigned long i_last_rd; unsigned long i_last_wr;