From patchwork Thu Jul 16 09:20:10 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Suanming Mou X-Patchwork-Id: 74210 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 456A8A0540; Thu, 16 Jul 2020 11:20:36 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 1657E1BF7D; Thu, 16 Jul 2020 11:20:28 +0200 (CEST) Received: from git-send-mailer.rdmz.labs.mlnx (unknown [37.142.13.130]) by dpdk.org (Postfix) with ESMTP id 47D081BED9 for ; Thu, 16 Jul 2020 11:20:25 +0200 (CEST) From: Suanming Mou To: viacheslavo@mellanox.com, matan@mellanox.com Cc: orika@mellanox.com, rasland@mellanox.com, dev@dpdk.org Date: Thu, 16 Jul 2020 17:20:10 +0800 Message-Id: <1594891216-11778-2-git-send-email-suanmingm@mellanox.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1594891216-11778-1-git-send-email-suanmingm@mellanox.com> References: <1594785603-152773-1-git-send-email-suanmingm@mellanox.com> <1594891216-11778-1-git-send-email-suanmingm@mellanox.com> Subject: [dpdk-dev] [PATCH v2 1/7] common/mlx5: add mlx5 memory management functions X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Add the internal mlx5 memory management functions: mlx5_malloc_mem_select(); mlx5_memory_stat_dump(); mlx5_rellaocate(); mlx5_malloc(); mlx5_free(); User will be allowed to manage memory from system or from rte memory with the unified functions. In this case, for the system with limited memory which can not reserve lots of rte hugepage memory in advanced, will allocate the memory from system for some of not so important control path objects based on the sys_mem_en configuration. Signed-off-by: Suanming Mou Acked-by: Matan Azrad --- drivers/common/mlx5/Makefile | 1 + drivers/common/mlx5/meson.build | 1 + drivers/common/mlx5/mlx5_malloc.c | 306 ++++++++++++++++++++++++ drivers/common/mlx5/mlx5_malloc.h | 99 ++++++++ drivers/common/mlx5/rte_common_mlx5_version.map | 6 + 5 files changed, 413 insertions(+) create mode 100644 drivers/common/mlx5/mlx5_malloc.c create mode 100644 drivers/common/mlx5/mlx5_malloc.h diff --git a/drivers/common/mlx5/Makefile b/drivers/common/mlx5/Makefile index f6c762b..239d681 100644 --- a/drivers/common/mlx5/Makefile +++ b/drivers/common/mlx5/Makefile @@ -21,6 +21,7 @@ SRCS-y += linux/mlx5_nl.c SRCS-y += linux/mlx5_common_verbs.c SRCS-y += mlx5_common_mp.c SRCS-y += mlx5_common_mr.c +SRCS-y += mlx5_malloc.c ifeq ($(CONFIG_RTE_IBVERBS_LINK_DLOPEN),y) INSTALL-y-lib += $(LIB_GLUE) endif diff --git a/drivers/common/mlx5/meson.build b/drivers/common/mlx5/meson.build index ba43714..70e2c1c 100644 --- a/drivers/common/mlx5/meson.build +++ b/drivers/common/mlx5/meson.build @@ -13,6 +13,7 @@ sources += files( 'mlx5_common.c', 'mlx5_common_mp.c', 'mlx5_common_mr.c', + 'mlx5_malloc.c', ) cflags_options = [ diff --git a/drivers/common/mlx5/mlx5_malloc.c b/drivers/common/mlx5/mlx5_malloc.c new file mode 100644 index 0000000..316305d --- /dev/null +++ b/drivers/common/mlx5/mlx5_malloc.c @@ -0,0 +1,306 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright 2020 Mellanox Technologies, Ltd + */ + +#include +#include +#include +#include +#include + +#include + +#include "mlx5_common_utils.h" +#include "mlx5_malloc.h" + +struct mlx5_sys_mem { + uint32_t init:1; /* Memory allocator initialized. */ + uint32_t enable:1; /* System memory select. */ + uint32_t reserve:30; /* Reserve. */ + union { + struct rte_memseg_list *last_msl; + rte_atomic64_t a64_last_msl; + }; + /* last allocated rte memory memseg list. */ +#ifdef RTE_LIBRTE_MLX5_DEBUG + rte_atomic64_t malloc_sys; + /* Memory allocated from system count. */ + rte_atomic64_t malloc_rte; + /* Memory allocated from hugepage count. */ + rte_atomic64_t realloc_sys; + /* Memory reallocate from system count. */ + rte_atomic64_t realloc_rte; + /* Memory reallocate from hugepage count. */ + rte_atomic64_t free_sys; + /* Memory free to system count. */ + rte_atomic64_t free_rte; + /* Memory free to hugepage count. */ + rte_atomic64_t msl_miss; + /* MSL miss count. */ + rte_atomic64_t msl_update; + /* MSL update count. */ +#endif +}; + +/* Initialize default as not */ +static struct mlx5_sys_mem mlx5_sys_mem = { + .init = 0, + .enable = 0, +#ifdef RTE_LIBRTE_MLX5_DEBUG + .malloc_sys = RTE_ATOMIC64_INIT(0), + .malloc_rte = RTE_ATOMIC64_INIT(0), + .realloc_sys = RTE_ATOMIC64_INIT(0), + .realloc_rte = RTE_ATOMIC64_INIT(0), + .free_sys = RTE_ATOMIC64_INIT(0), + .free_rte = RTE_ATOMIC64_INIT(0), + .msl_miss = RTE_ATOMIC64_INIT(0), + .msl_update = RTE_ATOMIC64_INIT(0), +#endif +}; + +/** + * Check if the address belongs to memory seg list. + * + * @param addr + * Memory address to be ckeced. + * @param msl + * Memory seg list. + * + * @return + * True if it belongs, false otherwise. + */ +static bool +mlx5_mem_check_msl(void *addr, struct rte_memseg_list *msl) +{ + void *start, *end; + + if (!msl) + return false; + start = msl->base_va; + end = RTE_PTR_ADD(start, msl->len); + if (addr >= start && addr < end) + return true; + return false; +} + +/** + * Update the msl if memory belongs to new msl. + * + * @param addr + * Memory address. + */ +static void +mlx5_mem_update_msl(void *addr) +{ + /* + * Update the cache msl if the new addr comes from the new msl + * different with the cached msl. + */ + if (addr && !mlx5_mem_check_msl(addr, + (struct rte_memseg_list *)(uintptr_t)rte_atomic64_read + (&mlx5_sys_mem.a64_last_msl))) { + rte_atomic64_set(&mlx5_sys_mem.a64_last_msl, + (int64_t)(uintptr_t)rte_mem_virt2memseg_list(addr)); +#ifdef RTE_LIBRTE_MLX5_DEBUG + rte_atomic64_inc(&mlx5_sys_mem.msl_update); +#endif + } +} + +/** + * Check if the address belongs to rte memory. + * + * @param addr + * Memory address to be ckeced. + * + * @return + * True if it belongs, false otherwise. + */ +static bool +mlx5_mem_is_rte(void *addr) +{ + /* + * Check if the last cache msl matches. Drop to slow path + * to check if the memory belongs to rte memory. + */ + if (!mlx5_mem_check_msl(addr, (struct rte_memseg_list *)(uintptr_t) + rte_atomic64_read(&mlx5_sys_mem.a64_last_msl))) { + if (!rte_mem_virt2memseg_list(addr)) + return false; +#ifdef RTE_LIBRTE_MLX5_DEBUG + rte_atomic64_inc(&mlx5_sys_mem.msl_miss); +#endif + } + return true; +} + +/** + * Allocate memory with alignment. + * + * @param size + * Memory size to be allocated. + * @param align + * Memory alignment. + * @param zero + * Clear the allocated memory or not. + * + * @return + * Pointer of the allocated memory, NULL otherwise. + */ +static void * +mlx5_alloc_align(size_t size, unsigned int align, unsigned int zero) +{ + void *buf; + buf = memalign(align, size); + if (!buf) { + DRV_LOG(ERR, "Couldn't allocate buf.\n"); + return NULL; + } + if (zero) + memset(buf, 0, size); + return buf; +} + +void * +mlx5_malloc(uint32_t flags, size_t size, unsigned int align, int socket) +{ + void *addr; + bool rte_mem; + + /* + * If neither system memory nor rte memory is required, allocate + * memory according to mlx5_sys_mem.enable. + */ + if (flags & MLX5_MEM_RTE) + rte_mem = true; + else if (flags & MLX5_MEM_SYS) + rte_mem = false; + else + rte_mem = mlx5_sys_mem.enable ? false : true; + if (rte_mem) { + if (flags & MLX5_MEM_ZERO) + addr = rte_zmalloc_socket(NULL, size, align, socket); + else + addr = rte_malloc_socket(NULL, size, align, socket); + mlx5_mem_update_msl(addr); +#ifdef RTE_LIBRTE_MLX5_DEBUG + if (addr) + rte_atomic64_inc(&mlx5_sys_mem.malloc_rte); +#endif + return addr; + } + /* The memory will be allocated from system. */ + if (align) + addr = mlx5_alloc_align(size, align, !!(flags & MLX5_MEM_ZERO)); + else if (flags & MLX5_MEM_ZERO) + addr = calloc(1, size); + else + addr = malloc(size); +#ifdef RTE_LIBRTE_MLX5_DEBUG + if (addr) + rte_atomic64_inc(&mlx5_sys_mem.malloc_sys); +#endif + return addr; +} + +void * +mlx5_realloc(void *addr, uint32_t flags, size_t size, unsigned int align, + int socket) +{ + void *new_addr; + bool rte_mem; + + /* Allocate directly if old memory address is NULL. */ + if (!addr) + return mlx5_malloc(flags, size, align, socket); + /* Get the memory type. */ + if (flags & MLX5_MEM_RTE) + rte_mem = true; + else if (flags & MLX5_MEM_SYS) + rte_mem = false; + else + rte_mem = mlx5_sys_mem.enable ? false : true; + /* Check if old memory and to be allocated memory are the same type. */ + if (rte_mem != mlx5_mem_is_rte(addr)) { + DRV_LOG(ERR, "Couldn't reallocate to different memory type."); + return NULL; + } + /* Allocate memory from rte memory. */ + if (rte_mem) { + new_addr = rte_realloc_socket(addr, size, align, socket); + mlx5_mem_update_msl(new_addr); +#ifdef RTE_LIBRTE_MLX5_DEBUG + if (new_addr) + rte_atomic64_inc(&mlx5_sys_mem.realloc_rte); +#endif + return new_addr; + } + /* Align is not supported for system memory. */ + if (align) { + DRV_LOG(ERR, "Couldn't reallocate with alignment"); + return NULL; + } + new_addr = realloc(addr, size); +#ifdef RTE_LIBRTE_MLX5_DEBUG + if (new_addr) + rte_atomic64_inc(&mlx5_sys_mem.realloc_sys); +#endif + return new_addr; +} + +void +mlx5_free(void *addr) +{ + if (addr == NULL) + return; + if (!mlx5_mem_is_rte(addr)) { +#ifdef RTE_LIBRTE_MLX5_DEBUG + rte_atomic64_inc(&mlx5_sys_mem.free_sys); +#endif + free(addr); + } else { +#ifdef RTE_LIBRTE_MLX5_DEBUG + rte_atomic64_inc(&mlx5_sys_mem.free_rte); +#endif + rte_free(addr); + } +} + +void +mlx5_memory_stat_dump(void) +{ +#ifdef RTE_LIBRTE_MLX5_DEBUG + DRV_LOG(INFO, "System memory malloc:%"PRIi64", realloc:%"PRIi64"," + " free:%"PRIi64"\nRTE memory malloc:%"PRIi64"," + " realloc:%"PRIi64", free:%"PRIi64"\nMSL miss:%"PRIi64"," + " update:%"PRIi64"", + rte_atomic64_read(&mlx5_sys_mem.malloc_sys), + rte_atomic64_read(&mlx5_sys_mem.realloc_sys), + rte_atomic64_read(&mlx5_sys_mem.free_sys), + rte_atomic64_read(&mlx5_sys_mem.malloc_rte), + rte_atomic64_read(&mlx5_sys_mem.realloc_rte), + rte_atomic64_read(&mlx5_sys_mem.free_rte), + rte_atomic64_read(&mlx5_sys_mem.msl_miss), + rte_atomic64_read(&mlx5_sys_mem.msl_update)); +#endif +} + +void +mlx5_malloc_mem_select(uint32_t sys_mem_en) +{ + /* + * The initialization should be called only once and all devices + * should use the same memory type. Otherwise, when new device is + * being attached with some different memory allocation configuration, + * the memory will get wrong behavior or a failure will be raised. + */ + if (!mlx5_sys_mem.init) { + if (sys_mem_en) + mlx5_sys_mem.enable = 1; + mlx5_sys_mem.init = 1; + DRV_LOG(INFO, "%s is selected.", sys_mem_en ? "SYS_MEM" : "RTE_MEM"); + } else if (mlx5_sys_mem.enable != sys_mem_en) { + DRV_LOG(WARNING, "%s is already selected.", + mlx5_sys_mem.enable ? "SYS_MEM" : "RTE_MEM"); + } +} diff --git a/drivers/common/mlx5/mlx5_malloc.h b/drivers/common/mlx5/mlx5_malloc.h new file mode 100644 index 0000000..d3e5f5b --- /dev/null +++ b/drivers/common/mlx5/mlx5_malloc.h @@ -0,0 +1,99 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright 2020 Mellanox Technologies, Ltd + */ + +#ifndef MLX5_MALLOC_H_ +#define MLX5_MALLOC_H_ + +#ifdef __cplusplus +extern "C" { +#endif + +enum mlx5_mem_flags { + MLX5_MEM_ANY = 0, + /* Memory will be allocated dpends on sys_mem_en. */ + MLX5_MEM_SYS = 1 << 0, + /* Memory should be allocated from system. */ + MLX5_MEM_RTE = 1 << 1, + /* Memory should be allocated from rte hugepage. */ + MLX5_MEM_ZERO = 1 << 2, + /* Memory should be cleared to zero. */ +}; + +/** + * Select the PMD memory allocate preference. + * + * Once sys_mem_en is set, the default memory allocate will from + * system only if an explicitly flag is set to order the memory + * from rte hugepage memory. + * + * @param sys_mem_en + * Use system memory or not. + */ +__rte_internal +void mlx5_malloc_mem_select(uint32_t sys_mem_en); + +/** + * Dump the PMD memory usage statistic. + */ +__rte_internal +void mlx5_memory_stat_dump(void); + +/** + * Memory allocate function. + * + * @param flags + * The bits as enum mlx5_mem_flags defined. + * @param size + * Memory size to be allocated. + * @param align + * Memory alignment. + * @param socket + * The socket memory should allocated. + * Valid only when allocate the memory from rte hugepage. + * + * @return + * Pointer of the allocated memory, NULL otherwise. + */ +__rte_internal +void *mlx5_malloc(uint32_t flags, size_t size, unsigned int align, int socket); + +/** + * Memory reallocate function. + * + * + * + * @param addr + * The memory to be reallocated. + * @param flags + * The bits as enum mlx5_mem_flags defined. + * @param size + * Memory size to be allocated. + * @param align + * Memory alignment. + * @param socket + * The socket memory should allocated. + * Valid only when allocate the memory from rte hugepage. + * + * @return + * Pointer of the allocated memory, NULL otherwise. + */ + +__rte_internal +void *mlx5_realloc(void *addr, uint32_t flags, size_t size, unsigned int align, + int socket); + +/** + * Memory free function. + * + * @param addr + * The memory address to be freed.. + */ +__rte_internal +void mlx5_free(void *addr); + +#ifdef __cplusplus +} +#endif + +#endif diff --git a/drivers/common/mlx5/rte_common_mlx5_version.map b/drivers/common/mlx5/rte_common_mlx5_version.map index ae57ebd..381a455 100644 --- a/drivers/common/mlx5/rte_common_mlx5_version.map +++ b/drivers/common/mlx5/rte_common_mlx5_version.map @@ -81,5 +81,11 @@ INTERNAL { mlx5_release_dbr; mlx5_translate_port_name; + + mlx5_malloc_mem_select; + mlx5_memory_stat_dump; + mlx5_malloc; + mlx5_realloc; + mlx5_free; };