/* * Copyright (C) Igor Sysoev * Copyright (C) NGINX, Inc. */ #ifndef _NXT_UNIX_MALLOC_H_INCLUDED_ #define _NXT_UNIX_MALLOC_H_INCLUDED_ NXT_EXPORT void *nxt_malloc(size_t size) NXT_MALLOC_LIKE; NXT_EXPORT void *nxt_zalloc(size_t size) NXT_MALLOC_LIKE; NXT_EXPORT void *nxt_realloc(void *p, size_t size) NXT_MALLOC_LIKE; NXT_EXPORT void *nxt_memalign(size_t alignment, size_t size) NXT_MALLOC_LIKE; #if (NXT_DEBUG) NXT_EXPORT void nxt_free(void *p); #else #define nxt_free(p) \ free(p) #endif #if (NXT_HAVE_MALLOC_USABLE_SIZE) /* * Due to allocation strategies malloc() allocators may allocate more * memory than is requested, so malloc_usable_size() allows to use all * allocated memory. It is helpful for socket buffers or unaligned disk * file I/O. However, they may be suboptimal for aligned disk file I/O. */ #if (NXT_LINUX) /* * Linux glibc stores bookkeeping information together with allocated * memory itself. Size of the bookkeeping information is 12 or 24 bytes * on 32-bit and 64-bit platforms respectively. Due to alignment there * are usually 4 or 8 spare bytes respectively. However, if allocation * is larger than about 128K, spare size may be up to one page: glibc aligns * sum of allocation and bookkeeping size to a page. So if requirement * of the large allocation size is not strict it is better to allocate * with small cutback and then to adjust size with malloc_usable_size(). * Glibc malloc_usable_size() is fast operation. */ #define nxt_malloc_usable_size(p, size) \ size = malloc_usable_size(p) #define nxt_malloc_cutback(cutback, size) \ size = ((cutback) && size > 127 * 1024) ? size - 32 : size #elif (NXT_FREEBSD) /* * FreeBSD prior to 7.0 (phkmalloc) aligns sizes to * 16 - 2048 a power of two * 2049 - ... aligned to 4K * * FreeBSD 7.0 (jemalloc) aligns sizes to: * 2 - 8 a power of two * 9 - 512 aligned to 16 * 513 - 2048 a power of two, i.e. aligned to 1K * 2049 - 1M aligned to 4K * 1M- ... aligned to 1M * See table in src/lib/libc/stdlib/malloc.c * * FreeBSD 7.0 malloc_usable_size() is fast for allocations, which * are lesser than 1M. Larger allocations require mutex acquiring. */ #define nxt_malloc_usable_size(p, size) \ size = malloc_usable_size(p) #define nxt_malloc_cutback(cutback, size) #endif #elif (NXT_HAVE_MALLOC_GOOD_SIZE) /* * MacOSX aligns sizes to * 16 - 496 aligned to 16, 32-bit * 16 - 992 aligned to 16, 64-bit * 497/993 - 15K aligned to 512, if lesser than 1G RAM * 497/993 - 127K aligned to 512, otherwise * 15K/127K- ... aligned to 4K * * malloc_good_size() is faster than malloc_size() */ #define nxt_malloc_usable_size(p, size) \ size = malloc_good_size(size) #define nxt_malloc_cutback(cutback, size) #else #define nxt_malloc_usable_size(p, size) #define nxt_malloc_cutback(cutback, size) #endif #if (NXT_HAVE_POSIX_MEMALIGN || NXT_HAVE_MEMALIGN) #define NXT_MAX_MEMALIGN_SHIFT 32 #elif (NXT_FREEBSD) #define NXT_MAX_MEMALIGN_SHIFT 12 #else #define NXT_MAX_MEMALIGN_SHIFT 3 #endif #endif /* _NXT_UNIX_MALLOC_H_INCLUDED_ */