From patchwork Thu May 14 14:24:13 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186725 Delivered-To: patch@linaro.org Received: by 2002:a92:5b0a:0:0:0:0:0 with SMTP id p10csp1452081ilb; Thu, 14 May 2020 07:27:03 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxMkdAEqdPIlZB1CB5s/RSKaOEGkTsmbAavUHEPxk9rwMAaaFI9Dyf9Q1l/l4c+Xg9j+ibK X-Received: by 2002:a5d:5490:: with SMTP id h16mr5693381wrv.250.1589466423509; Thu, 14 May 2020 07:27:03 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1589466423; cv=none; d=google.com; s=arc-20160816; b=WKpHqUDrfNdvBybGEO4AeLWQoy5djqWYYTlJVq1i9zcHjpn5LnsKL8And9SIW4UyVu OAlioglWyZYmDUNMB/lNtX+c+Y5+7p5Riumpn1om1+6s1YNus9VhALwJbz6rqR0y4kg3 tNPPY4BhXUn4u6xcxdH0L39EvcxIBXqPXzizXM8qdNDm5OaEKOc5h9Wjq7owq/X4mJiB Y3Y8eHadVy+mUFBJrIzbS43wtkUsUaZFfRCRkLA+1RtVR9mp/e1+IRZbRlXQpy4wiHcL PnVbOvPeD7LZYOWgi/869ANPF8HRr2hEmuGNbAcmA05WoDnp1pQW4kNOI4jK49LmzMpT lzGQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=S0sWlNNT4UcKTZwyYNDt4nwZsDYgWE0Uu0b1p41EbQs=; b=uTRTWS3h4IiK0lxJkMch5Zt496AlkvgPJvpEVfNH2qrbatpperzBvGgzyaLetkNB+C cWd7wOQy0BFKlzI51DmKEcIXoBddoDGa6+ywz0aRcr3Sy/PSJCYJZDCsv0dyejD/Derw P8boriKGHf84OWer44lefwOk/YfBfaxrbXDmR8F5Ji1MiJoVsIdQQZdGehzzS5Si6zat FAXU3vQO4qpMizylAgkyQbAp9PMDDtXTutRR0OiRJYb++87dHkfkxSPbgcRtDEM6LX91 X6IxQ9NwfXPHD3ZJVB76Rpcn9KTFifHiKus33lH8/WQv6JQliWIzvFobpOLUmLFUSLwh BR3g== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id w16si2759956wro.394.2020.05.14.07.27.03; Thu, 14 May 2020 07:27:03 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 2D8701D97D; Thu, 14 May 2020 16:26:57 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id A7C151D972 for ; Thu, 14 May 2020 16:26:53 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 7929C2003B8; Thu, 14 May 2020 16:26:53 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 3F5AE200394; Thu, 14 May 2020 16:26:51 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 4F85B402BB; Thu, 14 May 2020 22:26:48 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com, mdr@ashroe.eu Cc: Hemant Agrawal Date: Thu, 14 May 2020 19:54:13 +0530 Message-Id: <20200514142426.31583-2-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200514142426.31583-1-hemant.agrawal@nxp.com> References: <20200514132533.13752-1-hemant.agrawal@nxp.com> <20200514142426.31583-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH v5 01/13] common/dpaax: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- devtools/libabigail.abignore | 3 +++ drivers/common/dpaax/dpaa_of.h | 15 +++++++++++++++ drivers/common/dpaax/dpaax_iova_table.h | 4 ++++ drivers/common/dpaax/rte_common_dpaax_version.map | 2 +- 4 files changed, 23 insertions(+), 1 deletion(-) -- 2.17.1 diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index c9ee73cb3c..b1488d5549 100644 --- a/devtools/libabigail.abignore +++ b/devtools/libabigail.abignore @@ -48,3 +48,6 @@ changed_enumerators = RTE_CRYPTO_AEAD_LIST_END [suppress_variable] name = rte_crypto_aead_algorithm_strings +; Ignore moving DPAAx stable functions to INTERNAL tag +[suppress_file] + file_name_regexp = ^librte_common_dpaax\. diff --git a/drivers/common/dpaax/dpaa_of.h b/drivers/common/dpaax/dpaa_of.h index 960b421766..38d91a1afe 100644 --- a/drivers/common/dpaax/dpaa_of.h +++ b/drivers/common/dpaax/dpaa_of.h @@ -24,6 +24,7 @@ #include #include #include +#include #ifndef OF_INIT_DEFAULT_PATH #define OF_INIT_DEFAULT_PATH "/proc/device-tree" @@ -102,6 +103,7 @@ struct dt_file { uint64_t buf[OF_FILE_BUF_MAX >> 3]; }; +__rte_internal const struct device_node *of_find_compatible_node( const struct device_node *from, const char *type __rte_unused, @@ -113,32 +115,44 @@ const struct device_node *of_find_compatible_node( dev_node != NULL; \ dev_node = of_find_compatible_node(dev_node, type, compatible)) +__rte_internal const void *of_get_property(const struct device_node *from, const char *name, size_t *lenp) __attribute__((nonnull(2))); +__rte_internal bool of_device_is_available(const struct device_node *dev_node); + +__rte_internal const struct device_node *of_find_node_by_phandle(uint64_t ph); +__rte_internal const struct device_node *of_get_parent(const struct device_node *dev_node); +__rte_internal const struct device_node *of_get_next_child(const struct device_node *dev_node, const struct device_node *prev); +__rte_internal const void *of_get_mac_address(const struct device_node *np); #define for_each_child_node(parent, child) \ for (child = of_get_next_child(parent, NULL); child != NULL; \ child = of_get_next_child(parent, child)) + +__rte_internal uint32_t of_n_addr_cells(const struct device_node *dev_node); uint32_t of_n_size_cells(const struct device_node *dev_node); +__rte_internal const uint32_t *of_get_address(const struct device_node *dev_node, size_t idx, uint64_t *size, uint32_t *flags); +__rte_internal uint64_t of_translate_address(const struct device_node *dev_node, const uint32_t *addr) __attribute__((nonnull)); +__rte_internal bool of_device_is_compatible(const struct device_node *dev_node, const char *compatible); @@ -146,6 +160,7 @@ bool of_device_is_compatible(const struct device_node *dev_node, * subsystem that is device-tree-dependent. Eg. Qman/Bman, config layers, etc. * The path should usually be "/proc/device-tree". */ +__rte_internal int of_init_path(const char *dt_path); /* of_finish() allows a controlled tear-down of the device-tree layer, eg. if a diff --git a/drivers/common/dpaax/dpaax_iova_table.h b/drivers/common/dpaax/dpaax_iova_table.h index fc3b9e7a8f..230fba8ba0 100644 --- a/drivers/common/dpaax/dpaax_iova_table.h +++ b/drivers/common/dpaax/dpaax_iova_table.h @@ -61,9 +61,13 @@ extern struct dpaax_iova_table *dpaax_iova_table_p; #define DPAAX_MEM_SPLIT_MASK_OFF (DPAAX_MEM_SPLIT - 1) /**< Offset */ /* APIs exposed */ +__rte_internal int dpaax_iova_table_populate(void); +__rte_internal void dpaax_iova_table_depopulate(void); +__rte_internal int dpaax_iova_table_update(phys_addr_t paddr, void *vaddr, size_t length); +__rte_internal void dpaax_iova_table_dump(void); static inline void *dpaax_iova_table_get_va(phys_addr_t paddr) __rte_hot; diff --git a/drivers/common/dpaax/rte_common_dpaax_version.map b/drivers/common/dpaax/rte_common_dpaax_version.map index f72eba761d..ad2b2b3fec 100644 --- a/drivers/common/dpaax/rte_common_dpaax_version.map +++ b/drivers/common/dpaax/rte_common_dpaax_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaax_iova_table_depopulate; From patchwork Thu May 14 14:24:15 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186727 Delivered-To: patch@linaro.org Received: by 2002:a92:5b0a:0:0:0:0:0 with SMTP id p10csp1452337ilb; Thu, 14 May 2020 07:27:23 -0700 (PDT) X-Google-Smtp-Source: APiQypLJ4PnFJdd8MRCVfblvL8pyGnfFqH7m2iV7JgZ+pwxvtkOG2i70EQYRN8SY0qBP49PEl+zO X-Received: by 2002:a1c:1983:: with SMTP id 125mr48077543wmz.43.1589466442637; Thu, 14 May 2020 07:27:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1589466442; cv=none; d=google.com; s=arc-20160816; b=o9byVmA/uFJdFmQpStYxH/JcMXzIpHa6Smmio4+zROTFPJoikGPOgbsaud8L7RGJlY 9gVLvwry1CL6frLMn+EzCAj8tuExCGcdsIQpQPit+p/AeJ9ZhRg7cxuMevNzSkMTACCO TSbx1lvX3ZlSraOCA6vFxBiVlEi++TtX14rSRvCe0ySzUo85kMdp5kjH7YCcavYAPvLp kLgJLVT4fuHvmNE6kiP15y85x756sIVMP8dpJj2wX4Yuta5Ng0aDj26FkYFB/NFqZvX3 9E6OOJ/jaTRlAKacG/Uo3InuCVi1GaPV/Yxj+vTIWoZIf1G5ScIMxCzMuMkqDUJms2j5 QCKw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=qfbvdc2RT3CPhs1S2XrFILN0ARRzKyAlxZR4VF0X+Uw=; b=eJqE0YEIgbR/6u/HlAQIfLelivLJYOaB8rgTi8jpMvc1bD1FlgeCZhucj7Ifl+Y8gu 7ylD8QRhNmVRJP2JHTuxC1DqhEZ/N5Nk4XADHJUiqvROnI27IIWHfmgS0BscxjbauisH y/8bzgoime1lu6TNPFUbMlOPv4tOdLFHHA4rTPc/NbSLi+MxgC8Vyt9Dmn9ATEDXvR+Z iBvHEDV5bUYTULMeNmKYx/r9QokbPc1KkmFbV9uFb6r/Ou3hWQv2P9h60I5s0CHJVDx/ TC8vipHSj3/xStD1PJNC+ejIfGrnUhD49fWgtNOgmFFMMPAn0l/jYT7inSiWid5XOB5/ cJJQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id b188si18930769wmd.120.2020.05.14.07.27.22; Thu, 14 May 2020 07:27:22 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id DED171D99D; Thu, 14 May 2020 16:26:59 +0200 (CEST) Received: from inva020.nxp.com (inva020.nxp.com [92.121.34.13]) by dpdk.org (Postfix) with ESMTP id CD2491D97B for ; Thu, 14 May 2020 16:26:55 +0200 (CEST) Received: from inva020.nxp.com (localhost [127.0.0.1]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id 203D81A0391; Thu, 14 May 2020 16:26:55 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id B5BB11A0367; Thu, 14 May 2020 16:26:52 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id B1E77402E0; Thu, 14 May 2020 22:26:49 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com, mdr@ashroe.eu Cc: Hemant Agrawal Date: Thu, 14 May 2020 19:54:15 +0530 Message-Id: <20200514142426.31583-4-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200514142426.31583-1-hemant.agrawal@nxp.com> References: <20200514132533.13752-1-hemant.agrawal@nxp.com> <20200514142426.31583-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH v5 02/13] bus/fslmc: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. This patch also removes two symbols, which were not used anywhere else i.e. rte_fslmc_vfio_dmamap & dpaa2_get_qbman_swp Signed-off-by: Hemant Agrawal --- devtools/libabigail.abignore | 2 + drivers/bus/fslmc/fslmc_vfio.h | 4 ++ drivers/bus/fslmc/mc/fsl_dpbp.h | 6 +++ drivers/bus/fslmc/mc/fsl_dpci.h | 3 ++ drivers/bus/fslmc/mc/fsl_dpcon.h | 2 + drivers/bus/fslmc/mc/fsl_dpdmai.h | 8 ++++ drivers/bus/fslmc/mc/fsl_dpio.h | 9 ++++ drivers/bus/fslmc/mc/fsl_dpmng.h | 2 + drivers/bus/fslmc/mc/fsl_mc_cmd.h | 1 + drivers/bus/fslmc/portal/dpaa2_hw_dpio.h | 5 +++ drivers/bus/fslmc/portal/dpaa2_hw_pvt.h | 8 ++++ .../bus/fslmc/qbman/include/fsl_qbman_debug.h | 3 ++ .../fslmc/qbman/include/fsl_qbman_portal.h | 41 +++++++++++++++++++ drivers/bus/fslmc/rte_bus_fslmc_version.map | 4 +- drivers/bus/fslmc/rte_fslmc.h | 4 ++ 15 files changed, 99 insertions(+), 3 deletions(-) -- 2.17.1 diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index b1488d5549..877c6d5be8 100644 --- a/devtools/libabigail.abignore +++ b/devtools/libabigail.abignore @@ -51,3 +51,5 @@ ; Ignore moving DPAAx stable functions to INTERNAL tag [suppress_file] file_name_regexp = ^librte_common_dpaax\. +[suppress_file] + file_name_regexp = ^librte_bus_fslmc\. diff --git a/drivers/bus/fslmc/fslmc_vfio.h b/drivers/bus/fslmc/fslmc_vfio.h index c988121294..609e48aea3 100644 --- a/drivers/bus/fslmc/fslmc_vfio.h +++ b/drivers/bus/fslmc/fslmc_vfio.h @@ -41,7 +41,11 @@ typedef struct fslmc_vfio_container { } fslmc_vfio_container; extern char *fslmc_container; + +__rte_internal int rte_dpaa2_intr_enable(struct rte_intr_handle *intr_handle, int index); + +__rte_internal int rte_dpaa2_intr_disable(struct rte_intr_handle *intr_handle, int index); int rte_dpaa2_vfio_setup_intr(struct rte_intr_handle *intr_handle, diff --git a/drivers/bus/fslmc/mc/fsl_dpbp.h b/drivers/bus/fslmc/mc/fsl_dpbp.h index 9d405b42c4..7b537a21be 100644 --- a/drivers/bus/fslmc/mc/fsl_dpbp.h +++ b/drivers/bus/fslmc/mc/fsl_dpbp.h @@ -14,6 +14,7 @@ struct fsl_mc_io; +__rte_internal int dpbp_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpbp_id, @@ -42,10 +43,12 @@ int dpbp_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t obj_id); +__rte_internal int dpbp_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpbp_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -55,6 +58,7 @@ int dpbp_is_enabled(struct fsl_mc_io *mc_io, uint16_t token, int *en); +__rte_internal int dpbp_reset(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -70,6 +74,7 @@ struct dpbp_attr { uint16_t bpid; }; +__rte_internal int dpbp_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -88,6 +93,7 @@ int dpbp_get_api_version(struct fsl_mc_io *mc_io, uint16_t *major_ver, uint16_t *minor_ver); +__rte_internal int dpbp_get_num_free_bufs(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpci.h b/drivers/bus/fslmc/mc/fsl_dpci.h index a0ee5bfe69..81fd3438aa 100644 --- a/drivers/bus/fslmc/mc/fsl_dpci.h +++ b/drivers/bus/fslmc/mc/fsl_dpci.h @@ -181,6 +181,7 @@ struct dpci_rx_queue_cfg { int order_preservation_en; }; +__rte_internal int dpci_set_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -228,6 +229,7 @@ int dpci_get_api_version(struct fsl_mc_io *mc_io, uint16_t *major_ver, uint16_t *minor_ver); +__rte_internal int dpci_set_opr(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -235,6 +237,7 @@ int dpci_set_opr(struct fsl_mc_io *mc_io, uint8_t options, struct opr_cfg *cfg); +__rte_internal int dpci_get_opr(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpcon.h b/drivers/bus/fslmc/mc/fsl_dpcon.h index af81d51195..7caa6c68a1 100644 --- a/drivers/bus/fslmc/mc/fsl_dpcon.h +++ b/drivers/bus/fslmc/mc/fsl_dpcon.h @@ -20,6 +20,7 @@ struct fsl_mc_io; */ #define DPCON_INVALID_DPIO_ID (int)(-1) +__rte_internal int dpcon_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpcon_id, @@ -77,6 +78,7 @@ struct dpcon_attr { uint8_t num_priorities; }; +__rte_internal int dpcon_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpdmai.h b/drivers/bus/fslmc/mc/fsl_dpdmai.h index 40469cc139..e7e8a5dda9 100644 --- a/drivers/bus/fslmc/mc/fsl_dpdmai.h +++ b/drivers/bus/fslmc/mc/fsl_dpdmai.h @@ -23,11 +23,13 @@ struct fsl_mc_io; */ #define DPDMAI_ALL_QUEUES (uint8_t)(-1) +__rte_internal int dpdmai_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpdmai_id, uint16_t *token); +__rte_internal int dpdmai_close(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -54,10 +56,12 @@ int dpdmai_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t object_id); +__rte_internal int dpdmai_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpdmai_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -82,6 +86,7 @@ struct dpdmai_attr { uint8_t num_of_queues; }; +__rte_internal int dpdmai_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -148,6 +153,7 @@ struct dpdmai_rx_queue_cfg { }; +__rte_internal int dpdmai_set_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -168,6 +174,7 @@ struct dpdmai_rx_queue_attr { uint32_t fqid; }; +__rte_internal int dpdmai_get_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -184,6 +191,7 @@ struct dpdmai_tx_queue_attr { uint32_t fqid; }; +__rte_internal int dpdmai_get_tx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpio.h b/drivers/bus/fslmc/mc/fsl_dpio.h index 3158f53191..92e97db94b 100644 --- a/drivers/bus/fslmc/mc/fsl_dpio.h +++ b/drivers/bus/fslmc/mc/fsl_dpio.h @@ -13,11 +13,13 @@ struct fsl_mc_io; +__rte_internal int dpio_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpio_id, uint16_t *token); +__rte_internal int dpio_close(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -57,10 +59,12 @@ int dpio_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t object_id); +__rte_internal int dpio_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpio_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -70,10 +74,12 @@ int dpio_is_enabled(struct fsl_mc_io *mc_io, uint16_t token, int *en); +__rte_internal int dpio_reset(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpio_set_stashing_destination(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -84,12 +90,14 @@ int dpio_get_stashing_destination(struct fsl_mc_io *mc_io, uint16_t token, uint8_t *sdest); +__rte_internal int dpio_add_static_dequeue_channel(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, int dpcon_id, uint8_t *channel_index); +__rte_internal int dpio_remove_static_dequeue_channel(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -119,6 +127,7 @@ struct dpio_attr { uint32_t clk; }; +__rte_internal int dpio_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpmng.h b/drivers/bus/fslmc/mc/fsl_dpmng.h index 36c387af27..cdd8506625 100644 --- a/drivers/bus/fslmc/mc/fsl_dpmng.h +++ b/drivers/bus/fslmc/mc/fsl_dpmng.h @@ -34,6 +34,7 @@ struct mc_version { uint32_t revision; }; +__rte_internal int mc_get_version(struct fsl_mc_io *mc_io, uint32_t cmd_flags, struct mc_version *mc_ver_info); @@ -48,6 +49,7 @@ struct mc_soc_version { uint32_t pvr; }; +__rte_internal int mc_get_soc_version(struct fsl_mc_io *mc_io, uint32_t cmd_flags, struct mc_soc_version *mc_platform_info); diff --git a/drivers/bus/fslmc/mc/fsl_mc_cmd.h b/drivers/bus/fslmc/mc/fsl_mc_cmd.h index ac919610cf..06ea41a3b2 100644 --- a/drivers/bus/fslmc/mc/fsl_mc_cmd.h +++ b/drivers/bus/fslmc/mc/fsl_mc_cmd.h @@ -80,6 +80,7 @@ enum mc_cmd_status { #define MC_CMD_HDR_FLAGS_MASK 0xFF00FF00 +__rte_internal int mc_send_command(struct fsl_mc_io *mc_io, struct mc_command *cmd); static inline uint64_t mc_encode_cmd_header(uint16_t cmd_id, diff --git a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h index 2829c93806..7c5966241a 100644 --- a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h +++ b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h @@ -36,20 +36,25 @@ extern uint8_t dpaa2_eqcr_size; extern struct dpaa2_io_portal_t dpaa2_io_portal[RTE_MAX_LCORE]; /* Affine a DPIO portal to current processing thread */ +__rte_internal int dpaa2_affine_qbman_swp(void); /* Affine additional DPIO portal to current crypto processing thread */ +__rte_internal int dpaa2_affine_qbman_ethrx_swp(void); /* allocate memory for FQ - dq storage */ +__rte_internal int dpaa2_alloc_dq_storage(struct queue_storage_info_t *q_storage); /* free memory for FQ- dq storage */ +__rte_internal void dpaa2_free_dq_storage(struct queue_storage_info_t *q_storage); /* free the enqueue response descriptors */ +__rte_internal uint32_t dpaa2_free_eq_descriptors(void); diff --git a/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h b/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h index 368fe7c688..33b191f823 100644 --- a/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h +++ b/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h @@ -426,11 +426,19 @@ void set_swp_active_dqs(uint16_t dpio_index, struct qbman_result *dqs) { rte_global_active_dqs_list[dpio_index].global_active_dqs = dqs; } +__rte_internal struct dpaa2_dpbp_dev *dpaa2_alloc_dpbp_dev(void); + +__rte_internal void dpaa2_free_dpbp_dev(struct dpaa2_dpbp_dev *dpbp); + +__rte_internal int dpaa2_dpbp_supported(void); +__rte_internal struct dpaa2_dpci_dev *rte_dpaa2_alloc_dpci_dev(void); + +__rte_internal void rte_dpaa2_free_dpci_dev(struct dpaa2_dpci_dev *dpci); #endif diff --git a/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h b/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h index e010b1b6ae..328f2022fc 100644 --- a/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h +++ b/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h @@ -24,7 +24,10 @@ uint8_t verb; uint8_t reserved2[29]; }; +__rte_internal int qbman_fq_query_state(struct qbman_swp *s, uint32_t fqid, struct qbman_fq_query_np_rslt *r); + +__rte_internal uint32_t qbman_fq_state_frame_count(const struct qbman_fq_query_np_rslt *r); uint32_t qbman_fq_state_byte_count(const struct qbman_fq_query_np_rslt *r); diff --git a/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h b/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h index 88f0a99686..7ac0f82106 100644 --- a/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h +++ b/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h @@ -117,6 +117,7 @@ uint32_t qbman_swp_interrupt_read_status(struct qbman_swp *p); * @p: the given software portal object. * @mask: The value to set in SWP_ISR register. */ +__rte_internal void qbman_swp_interrupt_clear_status(struct qbman_swp *p, uint32_t mask); /** @@ -286,6 +287,7 @@ void qbman_swp_push_get(struct qbman_swp *s, uint8_t channel_idx, int *enabled); * rather by specifying the index (from 0 to 15) that has been mapped to the * desired channel. */ +__rte_internal void qbman_swp_push_set(struct qbman_swp *s, uint8_t channel_idx, int enable); /* ------------------- */ @@ -325,6 +327,7 @@ enum qbman_pull_type_e { * default/starting state. * @d: the pull dequeue descriptor to be cleared. */ +__rte_internal void qbman_pull_desc_clear(struct qbman_pull_desc *d); /** @@ -340,6 +343,7 @@ void qbman_pull_desc_clear(struct qbman_pull_desc *d); * the caller provides in 'storage_phys'), and 'stash' controls whether or not * those writes to main-memory express a cache-warming attribute. */ +__rte_internal void qbman_pull_desc_set_storage(struct qbman_pull_desc *d, struct qbman_result *storage, uint64_t storage_phys, @@ -349,6 +353,7 @@ void qbman_pull_desc_set_storage(struct qbman_pull_desc *d, * @d: the pull dequeue descriptor to be set. * @numframes: number of frames to be set, must be between 1 and 16, inclusive. */ +__rte_internal void qbman_pull_desc_set_numframes(struct qbman_pull_desc *d, uint8_t numframes); /** @@ -372,6 +377,7 @@ void qbman_pull_desc_set_token(struct qbman_pull_desc *d, uint8_t token); * qbman_pull_desc_set_fq() - Set fqid from which the dequeue command dequeues. * @fqid: the frame queue index of the given FQ. */ +__rte_internal void qbman_pull_desc_set_fq(struct qbman_pull_desc *d, uint32_t fqid); /** @@ -407,6 +413,7 @@ void qbman_pull_desc_set_rad(struct qbman_pull_desc *d, int rad); * Return 0 for success, and -EBUSY if the software portal is not ready * to do pull dequeue. */ +__rte_internal int qbman_swp_pull(struct qbman_swp *s, struct qbman_pull_desc *d); /* -------------------------------- */ @@ -421,12 +428,14 @@ int qbman_swp_pull(struct qbman_swp *s, struct qbman_pull_desc *d); * only once, so repeated calls can return a sequence of DQRR entries, without * requiring they be consumed immediately or in any particular order. */ +__rte_internal const struct qbman_result *qbman_swp_dqrr_next(struct qbman_swp *p); /** * qbman_swp_prefetch_dqrr_next() - prefetch the next DQRR entry. * @s: the software portal object. */ +__rte_internal void qbman_swp_prefetch_dqrr_next(struct qbman_swp *s); /** @@ -435,6 +444,7 @@ void qbman_swp_prefetch_dqrr_next(struct qbman_swp *s); * @s: the software portal object. * @dq: the DQRR entry to be consumed. */ +__rte_internal void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct qbman_result *dq); /** @@ -442,6 +452,7 @@ void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct qbman_result *dq); * @s: the software portal object. * @dqrr_index: the DQRR index entry to be consumed. */ +__rte_internal void qbman_swp_dqrr_idx_consume(struct qbman_swp *s, uint8_t dqrr_index); /** @@ -450,6 +461,7 @@ void qbman_swp_dqrr_idx_consume(struct qbman_swp *s, uint8_t dqrr_index); * * Return dqrr index. */ +__rte_internal uint8_t qbman_get_dqrr_idx(const struct qbman_result *dqrr); /** @@ -460,6 +472,7 @@ uint8_t qbman_get_dqrr_idx(const struct qbman_result *dqrr); * * Return dqrr entry object. */ +__rte_internal struct qbman_result *qbman_get_dqrr_from_idx(struct qbman_swp *s, uint8_t idx); /* ------------------------------------------------- */ @@ -485,6 +498,7 @@ struct qbman_result *qbman_get_dqrr_from_idx(struct qbman_swp *s, uint8_t idx); * Return 1 for getting a valid dequeue result, or 0 for not getting a valid * dequeue result. */ +__rte_internal int qbman_result_has_new_result(struct qbman_swp *s, struct qbman_result *dq); @@ -497,8 +511,10 @@ int qbman_result_has_new_result(struct qbman_swp *s, * Return 1 for getting a valid dequeue result, or 0 for not getting a valid * dequeue result. */ +__rte_internal int qbman_check_command_complete(struct qbman_result *dq); +__rte_internal int qbman_check_new_result(struct qbman_result *dq); /* -------------------------------------------------------- */ @@ -624,6 +640,7 @@ int qbman_result_is_FQPN(const struct qbman_result *dq); * * Return the state field. */ +__rte_internal uint8_t qbman_result_DQ_flags(const struct qbman_result *dq); /** @@ -658,6 +675,7 @@ static inline int qbman_result_DQ_is_pull_complete( * * Return seqnum. */ +__rte_internal uint16_t qbman_result_DQ_seqnum(const struct qbman_result *dq); /** @@ -667,6 +685,7 @@ uint16_t qbman_result_DQ_seqnum(const struct qbman_result *dq); * * Return odpid. */ +__rte_internal uint16_t qbman_result_DQ_odpid(const struct qbman_result *dq); /** @@ -699,6 +718,7 @@ uint32_t qbman_result_DQ_frame_count(const struct qbman_result *dq); * * Return the frame queue context. */ +__rte_internal uint64_t qbman_result_DQ_fqd_ctx(const struct qbman_result *dq); /** @@ -707,6 +727,7 @@ uint64_t qbman_result_DQ_fqd_ctx(const struct qbman_result *dq); * * Return the frame descriptor. */ +__rte_internal const struct qbman_fd *qbman_result_DQ_fd(const struct qbman_result *dq); /* State-change notifications (FQDAN/CDAN/CSCN/...). */ @@ -717,6 +738,7 @@ const struct qbman_fd *qbman_result_DQ_fd(const struct qbman_result *dq); * * Return the state in the notifiation. */ +__rte_internal uint8_t qbman_result_SCN_state(const struct qbman_result *scn); /** @@ -850,6 +872,7 @@ struct qbman_eq_response { * default/starting state. * @d: the given enqueue descriptor. */ +__rte_internal void qbman_eq_desc_clear(struct qbman_eq_desc *d); /* Exactly one of the following descriptor "actions" should be set. (Calling @@ -870,6 +893,7 @@ void qbman_eq_desc_clear(struct qbman_eq_desc *d); * @response_success: 1 = enqueue with response always; 0 = enqueue with * rejections returned on a FQ. */ +__rte_internal void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success); /** * qbman_eq_desc_set_orp() - Set order-resotration in the enqueue descriptor @@ -881,6 +905,7 @@ void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success); * @incomplete: indiates whether this is the last fragments using the same * sequeue number. */ +__rte_internal void qbman_eq_desc_set_orp(struct qbman_eq_desc *d, int respond_success, uint16_t opr_id, uint16_t seqnum, int incomplete); @@ -915,6 +940,7 @@ void qbman_eq_desc_set_orp_nesn(struct qbman_eq_desc *d, uint16_t opr_id, * data structure.) 'stash' controls whether or not the write to main-memory * expresses a cache-warming attribute. */ +__rte_internal void qbman_eq_desc_set_response(struct qbman_eq_desc *d, uint64_t storage_phys, int stash); @@ -929,6 +955,7 @@ void qbman_eq_desc_set_response(struct qbman_eq_desc *d, * result "storage" before issuing an enqueue, and use any non-zero 'token' * value. */ +__rte_internal void qbman_eq_desc_set_token(struct qbman_eq_desc *d, uint8_t token); /** @@ -944,6 +971,7 @@ void qbman_eq_desc_set_token(struct qbman_eq_desc *d, uint8_t token); * @d: the enqueue descriptor * @fqid: the id of the frame queue to be enqueued. */ +__rte_internal void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, uint32_t fqid); /** @@ -953,6 +981,7 @@ void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, uint32_t fqid); * @qd_bin: the queuing destination bin * @qd_prio: the queuing destination priority. */ +__rte_internal void qbman_eq_desc_set_qd(struct qbman_eq_desc *d, uint32_t qdid, uint16_t qd_bin, uint8_t qd_prio); @@ -978,6 +1007,7 @@ void qbman_eq_desc_set_eqdi(struct qbman_eq_desc *d, int enable); * held-active (order-preserving) FQ, whether the FQ should be parked instead of * being rescheduled.) */ +__rte_internal void qbman_eq_desc_set_dca(struct qbman_eq_desc *d, int enable, uint8_t dqrr_idx, int park); @@ -987,6 +1017,7 @@ void qbman_eq_desc_set_dca(struct qbman_eq_desc *d, int enable, * * Return the fd pointer. */ +__rte_internal struct qbman_fd *qbman_result_eqresp_fd(struct qbman_result *eqresp); /** @@ -997,6 +1028,7 @@ struct qbman_fd *qbman_result_eqresp_fd(struct qbman_result *eqresp); * This value is set into the response id before the enqueue command, which, * get overwritten by qbman once the enqueue command is complete. */ +__rte_internal void qbman_result_eqresp_set_rspid(struct qbman_result *eqresp, uint8_t val); /** @@ -1009,6 +1041,7 @@ void qbman_result_eqresp_set_rspid(struct qbman_result *eqresp, uint8_t val); * copied into the enqueue response to determine if the command has been * completed, and response has been updated. */ +__rte_internal uint8_t qbman_result_eqresp_rspid(struct qbman_result *eqresp); /** @@ -1017,6 +1050,7 @@ uint8_t qbman_result_eqresp_rspid(struct qbman_result *eqresp); * * Return 0 when command is sucessful. */ +__rte_internal uint8_t qbman_result_eqresp_rc(struct qbman_result *eqresp); /** @@ -1043,6 +1077,7 @@ int qbman_swp_enqueue(struct qbman_swp *s, const struct qbman_eq_desc *d, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple(struct qbman_swp *s, const struct qbman_eq_desc *d, const struct qbman_fd *fd, @@ -1060,6 +1095,7 @@ int qbman_swp_enqueue_multiple(struct qbman_swp *s, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple_fd(struct qbman_swp *s, const struct qbman_eq_desc *d, struct qbman_fd **fd, @@ -1076,6 +1112,7 @@ int qbman_swp_enqueue_multiple_fd(struct qbman_swp *s, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple_desc(struct qbman_swp *s, const struct qbman_eq_desc *d, const struct qbman_fd *fd, @@ -1117,12 +1154,14 @@ struct qbman_release_desc { * default/starting state. * @d: the qbman release descriptor. */ +__rte_internal void qbman_release_desc_clear(struct qbman_release_desc *d); /** * qbman_release_desc_set_bpid() - Set the ID of the buffer pool to release to * @d: the qbman release descriptor. */ +__rte_internal void qbman_release_desc_set_bpid(struct qbman_release_desc *d, uint16_t bpid); /** @@ -1141,6 +1180,7 @@ void qbman_release_desc_set_rcdi(struct qbman_release_desc *d, int enable); * * Return 0 for success, -EBUSY if the release command ring is not ready. */ +__rte_internal int qbman_swp_release(struct qbman_swp *s, const struct qbman_release_desc *d, const uint64_t *buffers, unsigned int num_buffers); @@ -1166,6 +1206,7 @@ int qbman_swp_release_thresh(struct qbman_swp *s, unsigned int thresh); * Return 0 for success, or negative error code if the acquire command * fails. */ +__rte_internal int qbman_swp_acquire(struct qbman_swp *s, uint16_t bpid, uint64_t *buffers, unsigned int num_buffers); diff --git a/drivers/bus/fslmc/rte_bus_fslmc_version.map b/drivers/bus/fslmc/rte_bus_fslmc_version.map index fe45575046..04e61156c3 100644 --- a/drivers/bus/fslmc/rte_bus_fslmc_version.map +++ b/drivers/bus/fslmc/rte_bus_fslmc_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaa2_affine_qbman_ethrx_swp; @@ -11,7 +11,6 @@ DPDK_20.0 { dpaa2_free_dpbp_dev; dpaa2_free_dq_storage; dpaa2_free_eq_descriptors; - dpaa2_get_qbman_swp; dpaa2_io_portal; dpaa2_svr_family; dpaa2_virt_mode; @@ -101,7 +100,6 @@ DPDK_20.0 { rte_fslmc_driver_unregister; rte_fslmc_get_device_count; rte_fslmc_object_register; - rte_fslmc_vfio_dmamap; rte_global_active_dqs_list; rte_mcp_ptr_list; diff --git a/drivers/bus/fslmc/rte_fslmc.h b/drivers/bus/fslmc/rte_fslmc.h index 96ba8dc259..5078b48ee1 100644 --- a/drivers/bus/fslmc/rte_fslmc.h +++ b/drivers/bus/fslmc/rte_fslmc.h @@ -162,6 +162,7 @@ RTE_DECLARE_PER_LCORE(struct dpaa2_portal_dqrr, dpaa2_held_bufs); * A pointer to a rte_dpaa2_driver structure describing the driver * to be registered. */ +__rte_internal void rte_fslmc_driver_register(struct rte_dpaa2_driver *driver); /** @@ -171,6 +172,7 @@ void rte_fslmc_driver_register(struct rte_dpaa2_driver *driver); * A pointer to a rte_dpaa2_driver structure describing the driver * to be unregistered. */ +__rte_internal void rte_fslmc_driver_unregister(struct rte_dpaa2_driver *driver); /** Helper for DPAA2 device registration from driver (eth, crypto) instance */ @@ -189,6 +191,7 @@ RTE_PMD_EXPORT_NAME(nm, __COUNTER__) * A pointer to a rte_dpaa_object structure describing the mc object * to be registered. */ +__rte_internal void rte_fslmc_object_register(struct rte_dpaa2_object *object); /** @@ -200,6 +203,7 @@ void rte_fslmc_object_register(struct rte_dpaa2_object *object); * >=0 for count; 0 indicates either no device of the said type scanned or * invalid device type. */ +__rte_internal uint32_t rte_fslmc_get_device_count(enum rte_dpaa2_dev_type device_type); /** Helper for DPAA2 object registration */ From patchwork Thu May 14 14:24:16 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186730 Delivered-To: patch@linaro.org Received: by 2002:a92:5b0a:0:0:0:0:0 with SMTP id p10csp1452622ilb; Thu, 14 May 2020 07:27:46 -0700 (PDT) X-Google-Smtp-Source: ABdhPJw0/rEtx2EDafqtstKnNt0e7fBUkQfVPwrdE8fQDz8SJwlYDrwCvCl70Enu0fu5eOhzfitN X-Received: by 2002:a1c:c38b:: with SMTP id t133mr2762340wmf.31.1589466466422; Thu, 14 May 2020 07:27:46 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1589466466; cv=none; d=google.com; s=arc-20160816; b=sc6BezAQ1CP/dA8i0SuPEJ8cIv197+DTnnyYuKyy0UUFN+sOxEQvNCGUKSCS9u0O6x 8BHMKpk7BdtbrFXsfQMS1+s5ldqTFkQtMNu8IHzE+8ASXDJY43W+jesqWmHfMwBPwVrw QKCjaky4p27MyCY6T/ES0tjI+yzlvberXADGXPC+YyUdn+6QyCDEish/+ValogHijcR+ 0YofTEq4Tv9fIsaAh5gQXqIn4jw5vLdhQlOx80b1EjpsLGZVYC/h0UEqK9TwmvPfIYLX QOat86WyvrEV68ZHaJszSIPtFWYRjODPdHjQUeht+TW5vIq/KOS9NFOYSnDZar4f19UQ zR7w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=VzjWVuPFTHeTK4/uialA45+i6KAN8R/rEApLzi6oqM0=; b=h7AjRZ90zxlZi/BE8hbW46PdKkmRokEKgaLRv+AxfOhRxcmvAxFRdgfuZkq8OxIJxC eXwyE5GW2mvWVpOdzJQKjzOWK7gv4qbZwi9FqvgptZiSfNFqYSFq1vJUpm6Fvdq7/hoo RX7PTHUW/DkLcm2m5i3/CLLY7WNs4HDBPkB5DHadsr2m0HU1tIe3E6Y8yO76ozk4d2FV 784jDAMBj1aQvSkjtdgzO+pVXasdfF2s6VFhp3gXNmeNNaMJHd+Oq3RCbeTafUgTIygg nk1J4FTMj1udX74iV4tfcfUOnOgyDOE+QS0tExmUo6dgGv94WeZYLIDQ/fSA1/612VLd bP6g== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id n16si2668767wro.455.2020.05.14.07.27.46; Thu, 14 May 2020 07:27:46 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 171581D9B8; Thu, 14 May 2020 16:27:06 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 04AA51D97B for ; Thu, 14 May 2020 16:26:57 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id DC6652003CB; Thu, 14 May 2020 16:26:56 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 7E56C200394; Thu, 14 May 2020 16:26:54 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 78750402E1; Thu, 14 May 2020 22:26:50 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com, mdr@ashroe.eu Cc: Hemant Agrawal Date: Thu, 14 May 2020 19:54:16 +0530 Message-Id: <20200514142426.31583-5-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200514142426.31583-1-hemant.agrawal@nxp.com> References: <20200514132533.13752-1-hemant.agrawal@nxp.com> <20200514142426.31583-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH v5 03/13] bus/dpaa: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- devtools/libabigail.abignore | 2 ++ drivers/bus/dpaa/include/fsl_bman.h | 6 +++++ drivers/bus/dpaa/include/fsl_fman.h | 27 +++++++++++++++++++ drivers/bus/dpaa/include/fsl_qman.h | 32 +++++++++++++++++++++++ drivers/bus/dpaa/include/fsl_usd.h | 6 +++++ drivers/bus/dpaa/include/netcfg.h | 2 ++ drivers/bus/dpaa/rte_bus_dpaa_version.map | 7 +---- drivers/bus/dpaa/rte_dpaa_bus.h | 5 ++++ 8 files changed, 81 insertions(+), 6 deletions(-) -- 2.17.1 diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index 877c6d5be8..ab34302d0c 100644 --- a/devtools/libabigail.abignore +++ b/devtools/libabigail.abignore @@ -53,3 +53,5 @@ file_name_regexp = ^librte_common_dpaax\. [suppress_file] file_name_regexp = ^librte_bus_fslmc\. +[suppress_file] + file_name_regexp = ^librte_bus_dpaa\. diff --git a/drivers/bus/dpaa/include/fsl_bman.h b/drivers/bus/dpaa/include/fsl_bman.h index f9cd972153..82da2fcfe0 100644 --- a/drivers/bus/dpaa/include/fsl_bman.h +++ b/drivers/bus/dpaa/include/fsl_bman.h @@ -264,12 +264,14 @@ int bman_shutdown_pool(u32 bpid); * the structure provided by the caller can be released or reused after the * function returns. */ +__rte_internal struct bman_pool *bman_new_pool(const struct bman_pool_params *params); /** * bman_free_pool - Deallocates a Buffer Pool object * @pool: the pool object to release */ +__rte_internal void bman_free_pool(struct bman_pool *pool); /** @@ -279,6 +281,7 @@ void bman_free_pool(struct bman_pool *pool); * The returned pointer refers to state within the pool object so must not be * modified and can no longer be read once the pool object is destroyed. */ +__rte_internal const struct bman_pool_params *bman_get_params(const struct bman_pool *pool); /** @@ -289,6 +292,7 @@ const struct bman_pool_params *bman_get_params(const struct bman_pool *pool); * @flags: bit-mask of BMAN_RELEASE_FLAG_*** options * */ +__rte_internal int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, u32 flags); @@ -302,6 +306,7 @@ int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, * The return value will be the number of buffers obtained from the pool, or a * negative error code if a h/w error or pool starvation was encountered. */ +__rte_internal int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num, u32 flags); @@ -317,6 +322,7 @@ int bman_query_pools(struct bm_pool_state *state); * * Return the number of the free buffers */ +__rte_internal u32 bman_query_free_buffers(struct bman_pool *pool); /** diff --git a/drivers/bus/dpaa/include/fsl_fman.h b/drivers/bus/dpaa/include/fsl_fman.h index 5705ebfdce..6c87c8db0d 100644 --- a/drivers/bus/dpaa/include/fsl_fman.h +++ b/drivers/bus/dpaa/include/fsl_fman.h @@ -7,6 +7,8 @@ #ifndef __FSL_FMAN_H #define __FSL_FMAN_H +#include + #ifdef __cplusplus extern "C" { #endif @@ -43,18 +45,23 @@ struct fm_status_t { } __rte_packed; /* Set MAC address for a particular interface */ +__rte_internal int fman_if_add_mac_addr(struct fman_if *p, uint8_t *eth, uint8_t addr_num); /* Remove a MAC address for a particular interface */ +__rte_internal void fman_if_clear_mac_addr(struct fman_if *p, uint8_t addr_num); /* Get the FMAN statistics */ +__rte_internal void fman_if_stats_get(struct fman_if *p, struct rte_eth_stats *stats); /* Reset the FMAN statistics */ +__rte_internal void fman_if_stats_reset(struct fman_if *p); /* Get all of the FMAN statistics */ +__rte_internal void fman_if_stats_get_all(struct fman_if *p, uint64_t *value, int n); /* Set ignore pause option for a specific interface */ @@ -64,32 +71,43 @@ void fman_if_set_rx_ignore_pause_frames(struct fman_if *p, bool enable); void fman_if_conf_max_frame_len(struct fman_if *p, unsigned int max_frame_len); /* Enable/disable Rx promiscuous mode on specified interface */ +__rte_internal void fman_if_promiscuous_enable(struct fman_if *p); +__rte_internal void fman_if_promiscuous_disable(struct fman_if *p); /* Enable/disable Rx on specific interfaces */ +__rte_internal void fman_if_enable_rx(struct fman_if *p); +__rte_internal void fman_if_disable_rx(struct fman_if *p); /* Enable/disable loopback on specific interfaces */ +__rte_internal void fman_if_loopback_enable(struct fman_if *p); +__rte_internal void fman_if_loopback_disable(struct fman_if *p); /* Set buffer pool on specific interface */ +__rte_internal void fman_if_set_bp(struct fman_if *fm_if, unsigned int num, int bpid, size_t bufsize); /* Get Flow Control threshold parameters on specific interface */ +__rte_internal int fman_if_get_fc_threshold(struct fman_if *fm_if); /* Enable and Set Flow Control threshold parameters on specific interface */ +__rte_internal int fman_if_set_fc_threshold(struct fman_if *fm_if, u32 high_water, u32 low_water, u32 bpid); /* Get Flow Control pause quanta on specific interface */ +__rte_internal int fman_if_get_fc_quanta(struct fman_if *fm_if); /* Set Flow Control pause quanta on specific interface */ +__rte_internal int fman_if_set_fc_quanta(struct fman_if *fm_if, u16 pause_quanta); /* Set default error fqid on specific interface */ @@ -99,35 +117,44 @@ void fman_if_set_err_fqid(struct fman_if *fm_if, uint32_t err_fqid); int fman_if_get_ic_params(struct fman_if *fm_if, struct fman_if_ic_params *icp); /* Set IC transfer params */ +__rte_internal int fman_if_set_ic_params(struct fman_if *fm_if, const struct fman_if_ic_params *icp); /* Get interface fd->offset value */ +__rte_internal int fman_if_get_fdoff(struct fman_if *fm_if); /* Set interface fd->offset value */ +__rte_internal void fman_if_set_fdoff(struct fman_if *fm_if, uint32_t fd_offset); /* Get interface SG enable status value */ +__rte_internal int fman_if_get_sg_enable(struct fman_if *fm_if); /* Set interface SG support mode */ +__rte_internal void fman_if_set_sg(struct fman_if *fm_if, int enable); /* Get interface Max Frame length (MTU) */ uint16_t fman_if_get_maxfrm(struct fman_if *fm_if); /* Set interface Max Frame length (MTU) */ +__rte_internal void fman_if_set_maxfrm(struct fman_if *fm_if, uint16_t max_frm); /* Set interface next invoked action for dequeue operation */ void fman_if_set_dnia(struct fman_if *fm_if, uint32_t nia); /* discard error packets on rx */ +__rte_internal void fman_if_discard_rx_errors(struct fman_if *fm_if); +__rte_internal void fman_if_set_mcast_filter_table(struct fman_if *p); +__rte_internal void fman_if_reset_mcast_filter_table(struct fman_if *p); int fman_if_add_hash_mac_addr(struct fman_if *p, uint8_t *eth); diff --git a/drivers/bus/dpaa/include/fsl_qman.h b/drivers/bus/dpaa/include/fsl_qman.h index 1b3342e7e6..4411bb0a79 100644 --- a/drivers/bus/dpaa/include/fsl_qman.h +++ b/drivers/bus/dpaa/include/fsl_qman.h @@ -1314,6 +1314,7 @@ struct qman_cgr { #define QMAN_CGR_MODE_FRAME 0x00000001 #ifdef CONFIG_FSL_QMAN_FQ_LOOKUP +__rte_internal void qman_set_fq_lookup_table(void **table); #endif @@ -1322,6 +1323,7 @@ void qman_set_fq_lookup_table(void **table); */ int qman_get_portal_index(void); +__rte_internal u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit, void **bufs); @@ -1333,6 +1335,7 @@ u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit, * processed via qman_poll_***() functions). Returns zero for success, or * -EINVAL if the current CPU is sharing a portal hosted on another CPU. */ +__rte_internal int qman_irqsource_add(u32 bits); /** @@ -1340,6 +1343,7 @@ int qman_irqsource_add(u32 bits); * takes portal (fq specific) as input rather than using the thread affined * portal. */ +__rte_internal int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits); /** @@ -1350,6 +1354,7 @@ int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits); * instead be processed via qman_poll_***() functions. Returns zero for success, * or -EINVAL if the current CPU is sharing a portal hosted on another CPU. */ +__rte_internal int qman_irqsource_remove(u32 bits); /** @@ -1357,6 +1362,7 @@ int qman_irqsource_remove(u32 bits); * takes portal (fq specific) as input rather than using the thread affined * portal. */ +__rte_internal int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits); /** @@ -1369,6 +1375,7 @@ int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits); */ u16 qman_affine_channel(int cpu); +__rte_internal unsigned int qman_portal_poll_rx(unsigned int poll_limit, void **bufs, struct qman_portal *q); @@ -1380,6 +1387,7 @@ unsigned int qman_portal_poll_rx(unsigned int poll_limit, * * This function will issue a volatile dequeue command to the QMAN. */ +__rte_internal int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags); /** @@ -1390,6 +1398,7 @@ int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags); * is issued. It will keep returning NULL until there is no packet available on * the DQRR. */ +__rte_internal struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq); /** @@ -1401,6 +1410,7 @@ struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq); * This will consume the DQRR enrey and make it available for next volatile * dequeue. */ +__rte_internal void qman_dqrr_consume(struct qman_fq *fq, struct qm_dqrr_entry *dq); @@ -1414,6 +1424,7 @@ void qman_dqrr_consume(struct qman_fq *fq, * this function will return -EINVAL, otherwise the return value is >=0 and * represents the number of DQRR entries processed. */ +__rte_internal int qman_poll_dqrr(unsigned int limit); /** @@ -1460,6 +1471,7 @@ void qman_start_dequeues(void); * (SDQCR). The requested pools are limited to those the portal has dequeue * access to. */ +__rte_internal void qman_static_dequeue_add(u32 pools, struct qman_portal *qm); /** @@ -1507,6 +1519,7 @@ void qman_dca(const struct qm_dqrr_entry *dq, int park_request); * function must be called from the same CPU as that which processed the DQRR * entry in the first place. */ +__rte_internal void qman_dca_index(u8 index, int park_request); /** @@ -1564,6 +1577,7 @@ void qman_set_dc_ern(qman_cb_dc_ern handler, int affine); * a frame queue object based on that, rather than assuming/requiring that it be * Out of Service. */ +__rte_internal int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq); /** @@ -1582,6 +1596,7 @@ void qman_destroy_fq(struct qman_fq *fq, u32 flags); * qman_fq_fqid - Queries the frame queue ID of a FQ object * @fq: the frame queue object to query */ +__rte_internal u32 qman_fq_fqid(struct qman_fq *fq); /** @@ -1594,6 +1609,7 @@ u32 qman_fq_fqid(struct qman_fq *fq); * This captures the state, as seen by the driver, at the time the function * executes. */ +__rte_internal void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags); /** @@ -1630,6 +1646,7 @@ void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags); * context_a.address fields and will leave the stashing fields provided by the * user alone, otherwise it will zero out the context_a.stashing fields. */ +__rte_internal int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts); /** @@ -1659,6 +1676,7 @@ int qman_schedule_fq(struct qman_fq *fq); * caller should be prepared to accept the callback as the function is called, * not only once it has returned. */ +__rte_internal int qman_retire_fq(struct qman_fq *fq, u32 *flags); /** @@ -1668,6 +1686,7 @@ int qman_retire_fq(struct qman_fq *fq, u32 *flags); * The frame queue must be retired and empty, and if any order restoration list * was released as ERNs at the time of retirement, they must all be consumed. */ +__rte_internal int qman_oos_fq(struct qman_fq *fq); /** @@ -1701,6 +1720,7 @@ int qman_query_fq_has_pkts(struct qman_fq *fq); * @fq: the frame queue object to be queried * @np: storage for the queried FQD fields */ +__rte_internal int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np); /** @@ -1708,6 +1728,7 @@ int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np); * @fq: the frame queue object to be queried * @frm_cnt: number of frames in the queue */ +__rte_internal int qman_query_fq_frm_cnt(struct qman_fq *fq, u32 *frm_cnt); /** @@ -1738,6 +1759,7 @@ int qman_query_wq(u8 query_dedicated, struct qm_mcr_querywq *wq); * callback, or by waiting for the QMAN_FQ_STATE_VDQCR bit to disappear from the * "flags" retrieved from qman_fq_state(). */ +__rte_internal int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr); /** @@ -1773,8 +1795,10 @@ int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr); * of an already busy hardware resource by throttling many of the to-be-dropped * enqueues "at the source". */ +__rte_internal int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags); +__rte_internal int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags, int frames_to_send); @@ -1788,6 +1812,7 @@ int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags, * This API is similar to qman_enqueue_multi(), but it takes fd which needs * to be processed by different frame queues. */ +__rte_internal int qman_enqueue_multi_fq(struct qman_fq *fq[], const struct qm_fd *fd, u32 *flags, int frames_to_send); @@ -1876,6 +1901,7 @@ int qman_shutdown_fq(u32 fqid); * @fqid: the base FQID of the range to deallocate * @count: the number of FQIDs in the range */ +__rte_internal int qman_reserve_fqid_range(u32 fqid, unsigned int count); static inline int qman_reserve_fqid(u32 fqid) { @@ -1895,6 +1921,7 @@ static inline int qman_reserve_fqid(u32 fqid) * than requested (though alignment will be as requested). If @partial is zero, * the return value will either be 'count' or negative. */ +__rte_internal int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial); static inline int qman_alloc_pool(u32 *result) { @@ -1942,6 +1969,7 @@ void qman_seed_pool_range(u32 id, unsigned int count); * any unspecified parameters) will be used rather than a modify hw hardware * (which only modifies the specified parameters). */ +__rte_internal int qman_create_cgr(struct qman_cgr *cgr, u32 flags, struct qm_mcc_initcgr *opts); @@ -1964,6 +1992,7 @@ int qman_create_cgr_to_dcp(struct qman_cgr *cgr, u32 flags, u16 dcp_portal, * is executed. This must be excuted on the same affine portal on which it was * created. */ +__rte_internal int qman_delete_cgr(struct qman_cgr *cgr); /** @@ -1980,6 +2009,7 @@ int qman_delete_cgr(struct qman_cgr *cgr); * unspecified parameters) will be used rather than a modify hw hardware (which * only modifies the specified parameters). */ +__rte_internal int qman_modify_cgr(struct qman_cgr *cgr, u32 flags, struct qm_mcc_initcgr *opts); @@ -2008,6 +2038,7 @@ int qman_query_congestion(struct qm_mcr_querycongestion *congestion); * than requested (though alignment will be as requested). If @partial is zero, * the return value will either be 'count' or negative. */ +__rte_internal int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial); static inline int qman_alloc_cgrid(u32 *result) { @@ -2021,6 +2052,7 @@ static inline int qman_alloc_cgrid(u32 *result) * @id: the base CGR ID of the range to deallocate * @count: the number of CGR IDs in the range */ +__rte_internal void qman_release_cgrid_range(u32 id, unsigned int count); static inline void qman_release_cgrid(u32 id) { diff --git a/drivers/bus/dpaa/include/fsl_usd.h b/drivers/bus/dpaa/include/fsl_usd.h index 263d9bb976..30ec63a09d 100644 --- a/drivers/bus/dpaa/include/fsl_usd.h +++ b/drivers/bus/dpaa/include/fsl_usd.h @@ -58,6 +58,7 @@ int bman_allocate_raw_portal(struct dpaa_raw_portal *portal); int bman_free_raw_portal(struct dpaa_raw_portal *portal); /* Obtain thread-local UIO file-descriptors */ +__rte_internal int qman_thread_fd(void); int bman_thread_fd(void); @@ -66,8 +67,12 @@ int bman_thread_fd(void); * processing is complete. As such, it is essential to call this before going * into another blocking read/select/poll. */ +__rte_internal void qman_thread_irq(void); + +__rte_internal void bman_thread_irq(void); +__rte_internal void qman_fq_portal_thread_irq(struct qman_portal *qp); void qman_clear_irq(void); @@ -77,6 +82,7 @@ int qman_global_init(void); int bman_global_init(void); /* Direct portal create and destroy */ +__rte_internal struct qman_portal *fsl_qman_fq_portal_create(int *fd); int fsl_qman_fq_portal_destroy(struct qman_portal *qp); int fsl_qman_fq_portal_init(struct qman_portal *qp); diff --git a/drivers/bus/dpaa/include/netcfg.h b/drivers/bus/dpaa/include/netcfg.h index bf7bfae8cb..d7d1befd24 100644 --- a/drivers/bus/dpaa/include/netcfg.h +++ b/drivers/bus/dpaa/include/netcfg.h @@ -46,11 +46,13 @@ struct netcfg_interface { * cfg_file: FMC config XML file * Returns the configuration information in newly allocated memory. */ +__rte_internal struct netcfg_info *netcfg_acquire(void); /* cfg_ptr: configuration information pointer. * Frees the resources allocated by the configuration layer. */ +__rte_internal void netcfg_release(struct netcfg_info *cfg_ptr); #ifdef RTE_LIBRTE_DPAA_DEBUG_DRIVER diff --git a/drivers/bus/dpaa/rte_bus_dpaa_version.map b/drivers/bus/dpaa/rte_bus_dpaa_version.map index e6ca4361e0..f4947fac41 100644 --- a/drivers/bus/dpaa/rte_bus_dpaa_version.map +++ b/drivers/bus/dpaa/rte_bus_dpaa_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: bman_acquire; @@ -13,7 +13,6 @@ DPDK_20.0 { dpaa_logtype_pmd; dpaa_netcfg; dpaa_svr_family; - fman_ccsr_map_fd; fman_dealloc_bufs_mask_hi; fman_dealloc_bufs_mask_lo; fman_if_add_mac_addr; @@ -51,7 +50,6 @@ DPDK_20.0 { qm_channel_pool1; qman_alloc_cgrid_range; qman_alloc_pool_range; - qman_clear_irq; qman_create_cgr; qman_create_fq; qman_dca_index; @@ -87,10 +85,7 @@ DPDK_20.0 { qman_volatile_dequeue; rte_dpaa_driver_register; rte_dpaa_driver_unregister; - rte_dpaa_mem_ptov; rte_dpaa_portal_fq_close; rte_dpaa_portal_fq_init; rte_dpaa_portal_init; - - local: *; }; diff --git a/drivers/bus/dpaa/rte_dpaa_bus.h b/drivers/bus/dpaa/rte_dpaa_bus.h index 373aca9785..d4aee132ef 100644 --- a/drivers/bus/dpaa/rte_dpaa_bus.h +++ b/drivers/bus/dpaa/rte_dpaa_bus.h @@ -158,6 +158,7 @@ rte_dpaa_mem_vtop(void *vaddr) * A pointer to a rte_dpaa_driver structure describing the driver * to be registered. */ +__rte_internal void rte_dpaa_driver_register(struct rte_dpaa_driver *driver); /** @@ -167,6 +168,7 @@ void rte_dpaa_driver_register(struct rte_dpaa_driver *driver); * A pointer to a rte_dpaa_driver structure describing the driver * to be unregistered. */ +__rte_internal void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver); /** @@ -178,10 +180,13 @@ void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver); * @return * 0 in case of success, error otherwise */ +__rte_internal int rte_dpaa_portal_init(void *arg); +__rte_internal int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq); +__rte_internal int rte_dpaa_portal_fq_close(struct qman_fq *fq); /** From patchwork Thu May 14 14:24:17 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186729 Delivered-To: patch@linaro.org Received: by 2002:a92:5b0a:0:0:0:0:0 with SMTP id p10csp1452486ilb; Thu, 14 May 2020 07:27:35 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwkS08FMPR3JMTdLcw3M8FGeYjsfL9YdTO3Ty1UYM8Y7EHb/STv13VzkPU/eLJYQ8WAL7th X-Received: by 2002:a05:6000:1241:: with SMTP id j1mr5826318wrx.42.1589466455784; Thu, 14 May 2020 07:27:35 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1589466455; cv=none; d=google.com; s=arc-20160816; b=qKAtglfvVFAxZi9OgW0GURE73wW64sDTIaO2f9VSQATtbOK336UwgO2ZY7BP8KTf2M IiSVXpR1A25MXGS1yR8+5s2JBaJOohR937pjEZSTGDeUTvDPP0jeaGFy75cm/o2Yan8s 0t4ydwszCnB7169fTrJaZr0nLFsNAB7qI5kUcmyRHDOdJehIK3weigwhAUwNGnYSPFoJ oRCRXnuOVdlFDnnACEjhxf6l5HDcfks7UyuEfvikA7itzro/dW/WfeitackRXcvDNn5z ILg9dkK95aV2D/+oVDewCc/tXw55XZ+Dj+VPbA72j1TlNyRskNhtJZAl0KoIx6xoCve2 mepA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=N+Rf6qfyA2mKoMjrILgnr7L/r+QXHTkV7gY6gHuehdM=; b=m9KwLn+d1KxZupIMNmU1hakmAWlbwnlTfmgP63+HwfBiMha+6iRJCimwgPhQGdXx4g Ac26HBIt2cPFsOTo8PcY2ZLVoTsMrsdVlkz2lV4e78gsPI0iknTpQyIg4mDV7Qhk58Tv WwtsVGwSIM4qyyNOCyqd1Y6aqbd52+PqorFYMWGDwLSsUuRq/k/vm/odOLs0f63hZ3L+ Or356I7R5dQ443iaw6ohYysNJt1b94+mOX81jxJMdcs/6o+81v6uQct8jFMKJcN8+dg/ eIp8I4XRut2dkq+LS9LQe26ih2pQKZ6a1EadqiabSlEE0C3SZwFfMGfpab9V2QDpf8Et YLKA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id c8si2925578wrp.414.2020.05.14.07.27.35; Thu, 14 May 2020 07:27:35 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id BB68A1D9B3; Thu, 14 May 2020 16:27:04 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 9D22C1D97B for ; Thu, 14 May 2020 16:26:56 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 82EA52003BE; Thu, 14 May 2020 16:26:56 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 7E53F200378; Thu, 14 May 2020 16:26:54 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 2A792402E3; Thu, 14 May 2020 22:26:51 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com, mdr@ashroe.eu Cc: Hemant Agrawal Date: Thu, 14 May 2020 19:54:17 +0530 Message-Id: <20200514142426.31583-6-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200514142426.31583-1-hemant.agrawal@nxp.com> References: <20200514132533.13752-1-hemant.agrawal@nxp.com> <20200514142426.31583-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH v5 04/13] crypto: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- devtools/libabigail.abignore | 4 ++++ drivers/crypto/dpaa2_sec/dpaa2_sec_event.h | 5 +++-- drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map | 2 +- drivers/crypto/dpaa_sec/dpaa_sec_event.h | 8 ++++---- drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map | 4 +--- 5 files changed, 13 insertions(+), 10 deletions(-) -- 2.17.1 diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index ab34302d0c..8db64f267d 100644 --- a/devtools/libabigail.abignore +++ b/devtools/libabigail.abignore @@ -55,3 +55,7 @@ file_name_regexp = ^librte_bus_fslmc\. [suppress_file] file_name_regexp = ^librte_bus_dpaa\. +[suppress_file] + file_name_regexp = ^librte_pmd_dpaa2_sec\. +[suppress_file] + file_name_regexp = ^librte_pmd_dpaa_sec\. diff --git a/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h b/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h index c779d5d837..675cbbb81d 100644 --- a/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h +++ b/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h @@ -6,12 +6,13 @@ #ifndef _DPAA2_SEC_EVENT_H_ #define _DPAA2_SEC_EVENT_H_ -int -dpaa2_sec_eventq_attach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa2_sec_eventq_attach(const struct rte_cryptodev *dev, int qp_id, struct dpaa2_dpcon_dev *dpcon, const struct rte_event *event); +__rte_internal int dpaa2_sec_eventq_detach(const struct rte_cryptodev *dev, int qp_id); diff --git a/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map b/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map index 5952d645fd..1352f576e5 100644 --- a/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map +++ b/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaa2_sec_eventq_attach; diff --git a/drivers/crypto/dpaa_sec/dpaa_sec_event.h b/drivers/crypto/dpaa_sec/dpaa_sec_event.h index 8d1a018096..0b09fa8f75 100644 --- a/drivers/crypto/dpaa_sec/dpaa_sec_event.h +++ b/drivers/crypto/dpaa_sec/dpaa_sec_event.h @@ -6,14 +6,14 @@ #ifndef _DPAA_SEC_EVENT_H_ #define _DPAA_SEC_EVENT_H_ -int -dpaa_sec_eventq_attach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa_sec_eventq_attach(const struct rte_cryptodev *dev, int qp_id, uint16_t ch_id, const struct rte_event *event); -int -dpaa_sec_eventq_detach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa_sec_eventq_detach(const struct rte_cryptodev *dev, int qp_id); #endif /* _DPAA_SEC_EVENT_H_ */ diff --git a/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map b/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map index 8580fa13db..aed07fb371 100644 --- a/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map +++ b/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map @@ -1,8 +1,6 @@ -DPDK_20.0 { +INTERNAL { global: dpaa_sec_eventq_attach; dpaa_sec_eventq_detach; - - local: *; };