mirror of
https://github.com/php/php-src.git
synced 2024-09-24 19:37:26 +00:00
c0cb3a1dc4
during shutdown when the main thread frees all the other threads resources. Even otherwise, exact resource passed to the dtor should be freed, not the one from the current thread.
2549 lines
73 KiB
C
2549 lines
73 KiB
C
/*
|
|
+----------------------------------------------------------------------+
|
|
| Zend Engine |
|
|
+----------------------------------------------------------------------+
|
|
| Copyright (c) 1998-2014 Zend Technologies Ltd. (http://www.zend.com) |
|
|
+----------------------------------------------------------------------+
|
|
| This source file is subject to version 2.00 of the Zend license, |
|
|
| that is bundled with this package in the file LICENSE, and is |
|
|
| available through the world-wide-web at the following url: |
|
|
| http://www.zend.com/license/2_00.txt. |
|
|
| If you did not receive a copy of the Zend license and are unable to |
|
|
| obtain it through the world-wide-web, please send a note to |
|
|
| license@zend.com so we can mail you a copy immediately. |
|
|
+----------------------------------------------------------------------+
|
|
| Authors: Andi Gutmans <andi@zend.com> |
|
|
| Zeev Suraski <zeev@zend.com> |
|
|
| Dmitry Stogov <dmitry@zend.com> |
|
|
+----------------------------------------------------------------------+
|
|
*/
|
|
|
|
/* $Id$ */
|
|
|
|
/*
|
|
* zend_alloc is designed to be a modern CPU cache friendly memory manager
|
|
* for PHP. Most ideas are taken from jemalloc and tcmalloc implementations.
|
|
*
|
|
* All allocations are split into 3 categories:
|
|
*
|
|
* Huge - the size is greater than CHUNK size (~2M by default), allocation is
|
|
* performed using mmap(). The result is aligned on 2M boundary.
|
|
*
|
|
* Large - a number of 4096K pages inside a CHUNK. Large blocks
|
|
* are always aligned on page boundary.
|
|
*
|
|
* Small - less than 3/4 of page size. Small sizes are rounded up to nearest
|
|
* greater predefined small size (there are 30 predefined sizes:
|
|
* 8, 16, 24, 32, ... 3072). Small blocks are allocated from
|
|
* RUNs. Each RUN is allocated as a single or few following pages.
|
|
* Allocation inside RUNs implemented using linked list of free
|
|
* elements. The result is aligned to 8 bytes.
|
|
*
|
|
* zend_alloc allocates memory from OS by CHUNKs, these CHUNKs and huge memory
|
|
* blocks are always aligned to CHUNK boundary. So it's very easy to determine
|
|
* the CHUNK owning the certain pointer. Regular CHUNKs reserve a single
|
|
* page at start for special purpose. It contains bitset of free pages,
|
|
* few bitset for available runs of predefined small sizes, map of pages that
|
|
* keeps information about usage of each page in this CHUNK, etc.
|
|
*
|
|
* zend_alloc provides familiar emalloc/efree/erealloc API, but in addition it
|
|
* provides specialized and optimized routines to allocate blocks of predefined
|
|
* sizes (e.g. emalloc_2(), emallc_4(), ..., emalloc_large(), etc)
|
|
* The library uses C preprocessor tricks that substitute calls to emalloc()
|
|
* with more specialized routines when the requested size is known.
|
|
*/
|
|
|
|
#include "zend.h"
|
|
#include "zend_alloc.h"
|
|
#include "zend_globals.h"
|
|
#include "zend_operators.h"
|
|
#include "zend_multiply.h"
|
|
|
|
#ifdef HAVE_SIGNAL_H
|
|
# include <signal.h>
|
|
#endif
|
|
#ifdef HAVE_UNISTD_H
|
|
# include <unistd.h>
|
|
#endif
|
|
|
|
#ifdef ZEND_WIN32
|
|
# include <wincrypt.h>
|
|
# include <process.h>
|
|
#endif
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#include <sys/types.h>
|
|
#include <sys/stat.h>
|
|
#if HAVE_LIMITS_H
|
|
#include <limits.h>
|
|
#endif
|
|
#include <fcntl.h>
|
|
#include <errno.h>
|
|
|
|
#ifndef _WIN32
|
|
# ifdef HAVE_MREMAP
|
|
# ifndef _GNU_SOURCE
|
|
# define _GNU_SOURCE
|
|
# endif
|
|
# ifndef __USE_GNU
|
|
# define __USE_GNU
|
|
# endif
|
|
# endif
|
|
# include <sys/mman.h>
|
|
# ifndef MAP_ANON
|
|
# ifdef MAP_ANONYMOUS
|
|
# define MAP_ANON MAP_ANONYMOUS
|
|
# endif
|
|
# endif
|
|
# ifndef MREMAP_MAYMOVE
|
|
# define MREMAP_MAYMOVE 0
|
|
# endif
|
|
# ifndef MAP_FAILED
|
|
# define MAP_FAILED ((void*)-1)
|
|
# endif
|
|
# ifndef MAP_POPULATE
|
|
# define MAP_POPULATE 0
|
|
# endif
|
|
# if defined(_SC_PAGESIZE) || (_SC_PAGE_SIZE)
|
|
# define REAL_PAGE_SIZE _real_page_size
|
|
static size_t _real_page_size = ZEND_MM_PAGE_SIZE;
|
|
# endif
|
|
#endif
|
|
|
|
#ifndef REAL_PAGE_SIZE
|
|
# define REAL_PAGE_SIZE ZEND_MM_PAGE_SIZE
|
|
#endif
|
|
|
|
#ifndef ZEND_MM_STAT
|
|
# define ZEND_MM_STAT 1 /* track current and peak memory usage */
|
|
#endif
|
|
#ifndef ZEND_MM_LIMIT
|
|
# define ZEND_MM_LIMIT 1 /* support for user-defined memory limit */
|
|
#endif
|
|
#ifndef ZEND_MM_CUSTOM
|
|
# define ZEND_MM_CUSTOM 1 /* support for custom memory allocator */
|
|
/* USE_ZEND_ALLOC=0 may switch to system malloc() */
|
|
#endif
|
|
#ifndef ZEND_MM_STORAGE
|
|
# define ZEND_MM_STORAGE 1 /* support for custom memory storage */
|
|
#endif
|
|
#ifndef ZEND_MM_ERROR
|
|
# define ZEND_MM_ERROR 1 /* report system errors */
|
|
#endif
|
|
|
|
#ifndef ZEND_MM_CHECK
|
|
# define ZEND_MM_CHECK(condition, message) do { \
|
|
if (UNEXPECTED(!(condition))) { \
|
|
zend_mm_panic(message); \
|
|
} \
|
|
} while (0)
|
|
#endif
|
|
|
|
typedef uint32_t zend_mm_page_info; /* 4-byte integer */
|
|
typedef zend_ulong zend_mm_bitset; /* 4-byte or 8-byte integer */
|
|
|
|
#define ZEND_MM_ALIGNED_OFFSET(size, alignment) \
|
|
(((size_t)(size)) & ((alignment) - 1))
|
|
#define ZEND_MM_ALIGNED_BASE(size, alignment) \
|
|
(((size_t)(size)) & ~((alignment) - 1))
|
|
#define ZEND_MM_ALIGNED_SIZE_EX(size, alignment) \
|
|
(((size_t)(size) + ((alignment) - 1)) & ~((alignment) - 1))
|
|
#define ZEND_MM_SIZE_TO_NUM(size, alignment) \
|
|
(((size_t)(size) + ((alignment) - 1)) / (alignment))
|
|
|
|
#define ZEND_MM_BITSET_LEN (sizeof(zend_mm_bitset) * 8) /* 32 or 64 */
|
|
#define ZEND_MM_PAGE_MAP_LEN (ZEND_MM_PAGES / ZEND_MM_BITSET_LEN) /* 16 or 8 */
|
|
|
|
typedef zend_mm_bitset zend_mm_page_map[ZEND_MM_PAGE_MAP_LEN]; /* 64B */
|
|
|
|
#define ZEND_MM_IS_FRUN 0x00000000
|
|
#define ZEND_MM_IS_LRUN 0x40000000
|
|
#define ZEND_MM_IS_SRUN 0x80000000
|
|
|
|
#define ZEND_MM_LRUN_PAGES_MASK 0x000003ff
|
|
#define ZEND_MM_LRUN_PAGES_OFFSET 0
|
|
|
|
#define ZEND_MM_SRUN_BIN_NUM_MASK 0x0000001f
|
|
#define ZEND_MM_SRUN_BIN_NUM_OFFSET 0
|
|
|
|
#define ZEND_MM_LRUN_PAGES(info) (((info) & ZEND_MM_LRUN_PAGES_MASK) >> ZEND_MM_LRUN_PAGES_OFFSET)
|
|
#define ZEND_MM_SRUN_BIN_NUM(info) (((info) & ZEND_MM_SRUN_BIN_NUM_MASK) >> ZEND_MM_SRUN_BIN_NUM_OFFSET)
|
|
|
|
#define ZEND_MM_FRUN() ZEND_MM_IS_FRUN
|
|
#define ZEND_MM_LRUN(count) (ZEND_MM_IS_LRUN | ((count) << ZEND_MM_LRUN_PAGES_OFFSET))
|
|
#define ZEND_MM_SRUN(bin_num) (ZEND_MM_IS_SRUN | ((bin_num) << ZEND_MM_SRUN_BIN_NUM_OFFSET))
|
|
|
|
#define ZEND_MM_BINS 30
|
|
|
|
typedef struct _zend_mm_page zend_mm_page;
|
|
typedef struct _zend_mm_bin zend_mm_bin;
|
|
typedef struct _zend_mm_free_slot zend_mm_free_slot;
|
|
typedef struct _zend_mm_chunk zend_mm_chunk;
|
|
typedef struct _zend_mm_huge_list zend_mm_huge_list;
|
|
|
|
#ifdef _WIN64
|
|
# define PTR_FMT "0x%0.16I64x"
|
|
#elif SIZEOF_LONG == 8
|
|
# define PTR_FMT "0x%0.16lx"
|
|
#else
|
|
# define PTR_FMT "0x%0.8lx"
|
|
#endif
|
|
|
|
/*
|
|
* Memory is retrived from OS by chunks of fixed size 2MB.
|
|
* Inside chunk it's managed by pages of fixed size 4096B.
|
|
* So each chunk consists from 512 pages.
|
|
* The first page of each chunk is reseved for chunk header.
|
|
* It contains service information about all pages.
|
|
*
|
|
* free_pages - current number of free pages in this chunk
|
|
*
|
|
* free_tail - number of continuous free pages at the end of chunk
|
|
*
|
|
* free_map - bitset (a bit for each page). The bit is set if the corresponding
|
|
* page is allocated. Allocator for "lage sizes" may easily find a
|
|
* free page (or a continuous number of pages) searching for zero
|
|
* bits.
|
|
*
|
|
* map - contains service information for each page. (32-bits for each
|
|
* page).
|
|
* usage:
|
|
* (2 bits)
|
|
* FRUN - free page,
|
|
* LRUN - first page of "large" allocation
|
|
* SRUN - first page of a bin used for "small" allocation
|
|
*
|
|
* lrun_pages:
|
|
* (10 bits) number of allocated pages
|
|
*
|
|
* srun_bin_num:
|
|
* (5 bits) bin number (e.g. 0 for sizes 0-2, 1 for 3-4,
|
|
* 2 for 5-8, 3 for 9-16 etc) see zend_alloc_sizes.h
|
|
*/
|
|
|
|
struct _zend_mm_heap {
|
|
#if ZEND_MM_CUSTOM
|
|
int use_custom_heap;
|
|
#endif
|
|
#if ZEND_MM_STORAGE
|
|
zend_mm_storage *storage;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
size_t size; /* current memory usage */
|
|
size_t peak; /* peak memory usage */
|
|
#endif
|
|
zend_mm_free_slot *free_slot[ZEND_MM_BINS]; /* free lists for small sizes */
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
size_t real_size; /* current size of allocated pages */
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
size_t real_peak; /* peak size of allocated pages */
|
|
#endif
|
|
#if ZEND_MM_LIMIT
|
|
size_t limit; /* memory limit */
|
|
int overflow; /* memory overflow flag */
|
|
#endif
|
|
|
|
zend_mm_huge_list *huge_list; /* list of huge allocated blocks */
|
|
|
|
zend_mm_chunk *main_chunk;
|
|
zend_mm_chunk *cached_chunks; /* list of unused chunks */
|
|
int chunks_count; /* number of alocated chunks */
|
|
int peak_chunks_count; /* peak number of allocated chunks for current request */
|
|
int cached_chunks_count; /* number of cached chunks */
|
|
double avg_chunks_count; /* average number of chunks allocated per request */
|
|
#if ZEND_MM_CUSTOM
|
|
void *(*_malloc)(size_t);
|
|
void (*_free)(void*);
|
|
void *(*_realloc)(void*, size_t);
|
|
#endif
|
|
};
|
|
|
|
struct _zend_mm_chunk {
|
|
zend_mm_heap *heap;
|
|
zend_mm_chunk *next;
|
|
zend_mm_chunk *prev;
|
|
int free_pages; /* number of free pages */
|
|
int free_tail; /* number of free pages at the end of chunk */
|
|
int num;
|
|
char reserve[64 - (sizeof(void*) * 3 + sizeof(int) * 3)];
|
|
zend_mm_heap heap_slot; /* used only in main chunk */
|
|
zend_mm_page_map free_map; /* 512 bits or 64 bytes */
|
|
zend_mm_page_info map[ZEND_MM_PAGES]; /* 2 KB = 512 * 4 */
|
|
};
|
|
|
|
struct _zend_mm_page {
|
|
char bytes[ZEND_MM_PAGE_SIZE];
|
|
};
|
|
|
|
/*
|
|
* bin - is one or few continuous pages (up to 8) used for allocation of
|
|
* a particular "small size".
|
|
*/
|
|
struct _zend_mm_bin {
|
|
char bytes[ZEND_MM_PAGE_SIZE * 8];
|
|
};
|
|
|
|
#if ZEND_DEBUG
|
|
typedef struct _zend_mm_debug_info {
|
|
size_t size;
|
|
const char *filename;
|
|
const char *orig_filename;
|
|
uint lineno;
|
|
uint orig_lineno;
|
|
} zend_mm_debug_info;
|
|
#endif
|
|
|
|
struct _zend_mm_free_slot {
|
|
zend_mm_free_slot *next_free_slot;
|
|
};
|
|
|
|
struct _zend_mm_huge_list {
|
|
void *ptr;
|
|
size_t size;
|
|
zend_mm_huge_list *next;
|
|
#if ZEND_DEBUG
|
|
zend_mm_debug_info dbg;
|
|
#endif
|
|
};
|
|
|
|
#define ZEND_MM_PAGE_ADDR(chunk, page_num) \
|
|
((void*)(((zend_mm_page*)(chunk)) + (page_num)))
|
|
|
|
#define _BIN_DATA_SIZE(num, size, elements, pages, x, y) size,
|
|
static const unsigned int bin_data_size[] = {
|
|
ZEND_MM_BINS_INFO(_BIN_DATA_SIZE, x, y)
|
|
};
|
|
|
|
#define _BIN_DATA_ELEMENTS(num, size, elements, pages, x, y) elements,
|
|
static const int bin_elements[] = {
|
|
ZEND_MM_BINS_INFO(_BIN_DATA_ELEMENTS, x, y)
|
|
};
|
|
|
|
#define _BIN_DATA_PAGES(num, size, elements, pages, x, y) pages,
|
|
static const int bin_pages[] = {
|
|
ZEND_MM_BINS_INFO(_BIN_DATA_PAGES, x, y)
|
|
};
|
|
|
|
#if ZEND_DEBUG
|
|
void zend_debug_alloc_output(char *format, ...)
|
|
{
|
|
char output_buf[256];
|
|
va_list args;
|
|
|
|
va_start(args, format);
|
|
vsprintf(output_buf, format, args);
|
|
va_end(args);
|
|
|
|
#ifdef ZEND_WIN32
|
|
OutputDebugString(output_buf);
|
|
#else
|
|
fprintf(stderr, "%s", output_buf);
|
|
#endif
|
|
}
|
|
#endif
|
|
|
|
static ZEND_NORETURN void zend_mm_panic(const char *message)
|
|
{
|
|
fprintf(stderr, "%s\n", message);
|
|
/* See http://support.microsoft.com/kb/190351 */
|
|
#ifdef PHP_WIN32
|
|
fflush(stderr);
|
|
#endif
|
|
#if ZEND_DEBUG && defined(HAVE_KILL) && defined(HAVE_GETPID)
|
|
kill(getpid(), SIGSEGV);
|
|
#endif
|
|
exit(1);
|
|
}
|
|
|
|
static ZEND_NORETURN void zend_mm_safe_error(zend_mm_heap *heap,
|
|
const char *format,
|
|
size_t limit,
|
|
#if ZEND_DEBUG
|
|
const char *filename,
|
|
uint lineno,
|
|
#endif
|
|
size_t size)
|
|
{
|
|
|
|
heap->overflow = 1;
|
|
zend_try {
|
|
zend_error_noreturn(E_ERROR,
|
|
format,
|
|
limit,
|
|
#if ZEND_DEBUG
|
|
filename,
|
|
lineno,
|
|
#endif
|
|
size);
|
|
} zend_catch {
|
|
} zend_end_try();
|
|
heap->overflow = 0;
|
|
zend_bailout();
|
|
exit(1);
|
|
}
|
|
|
|
#ifdef _WIN32
|
|
void
|
|
stderr_last_error(char *msg)
|
|
{
|
|
LPSTR buf = NULL;
|
|
DWORD err = GetLastError();
|
|
|
|
if (!FormatMessage(
|
|
FORMAT_MESSAGE_ALLOCATE_BUFFER |
|
|
FORMAT_MESSAGE_FROM_SYSTEM |
|
|
FORMAT_MESSAGE_IGNORE_INSERTS,
|
|
NULL,
|
|
err,
|
|
MAKELANGID(LANG_NEUTRAL, SUBLANG_DEFAULT),
|
|
(LPSTR)&buf,
|
|
0, NULL)) {
|
|
fprintf(stderr, "\n%s: [0x%08lx]\n", msg, err);
|
|
}
|
|
else {
|
|
fprintf(stderr, "\n%s: [0x%08lx] %s\n", msg, err, buf);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/*****************/
|
|
/* OS Allocation */
|
|
/*****************/
|
|
|
|
static void *zend_mm_mmap_fixed(void *addr, size_t size)
|
|
{
|
|
#ifdef _WIN32
|
|
return VirtualAlloc(addr, size, MEM_COMMIT | MEM_RESERVE, PAGE_READWRITE);
|
|
#else
|
|
/* MAP_FIXED leads to discarding of the old mapping, so it can't be used. */
|
|
void *ptr = mmap(addr, size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON /*| MAP_POPULATE | MAP_HUGETLB*/, -1, 0);
|
|
|
|
if (ptr == MAP_FAILED) {
|
|
#if ZEND_MM_ERROR
|
|
fprintf(stderr, "\nmmap() failed: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
return NULL;
|
|
} else if (ptr != addr) {
|
|
if (munmap(ptr, size) != 0) {
|
|
#if ZEND_MM_ERROR
|
|
fprintf(stderr, "\nmunmap() failed: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
}
|
|
return NULL;
|
|
}
|
|
return ptr;
|
|
#endif
|
|
}
|
|
|
|
static void *zend_mm_mmap(size_t size)
|
|
{
|
|
#ifdef _WIN32
|
|
void *ptr = VirtualAlloc(NULL, size, MEM_COMMIT | MEM_RESERVE, PAGE_READWRITE);
|
|
|
|
if (ptr == NULL) {
|
|
#if ZEND_MM_ERROR
|
|
stderr_last_error("VirtualAlloc() failed");
|
|
#endif
|
|
return NULL;
|
|
}
|
|
return ptr;
|
|
#else
|
|
void *ptr = mmap(NULL, size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANON /*| MAP_POPULATE | MAP_HUGETLB*/, -1, 0);
|
|
|
|
if (ptr == MAP_FAILED) {
|
|
#if ZEND_MM_ERROR
|
|
fprintf(stderr, "\nmmap() failed: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
return NULL;
|
|
}
|
|
return ptr;
|
|
#endif
|
|
}
|
|
|
|
static void zend_mm_munmap(void *addr, size_t size)
|
|
{
|
|
#ifdef _WIN32
|
|
if (VirtualFree(addr, 0, MEM_RELEASE) == 0) {
|
|
#if ZEND_MM_ERROR
|
|
stderr_last_error("VirtualFree() failed");
|
|
#endif
|
|
}
|
|
#else
|
|
if (munmap(addr, size) != 0) {
|
|
#if ZEND_MM_ERROR
|
|
fprintf(stderr, "\nmunmap() failed: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/***********/
|
|
/* Bitmask */
|
|
/***********/
|
|
|
|
/* number of trailing set (1) bits */
|
|
static zend_always_inline int zend_mm_bitset_nts(zend_mm_bitset bitset)
|
|
{
|
|
#if defined(__GNUC__)
|
|
# if SIZEOF_ZEND_LONG == SIZEOF_LONG
|
|
return __builtin_ctzl(~bitset);
|
|
# else
|
|
return __builtin_ctzll(~bitset);
|
|
# endif
|
|
#elif defined(_WIN32)
|
|
unsigned long index;
|
|
|
|
#if defined(_WIN64)
|
|
if (!BitScanForward64(&index, ~bitset)) {
|
|
#else
|
|
if (!BitScanForward(&index, ~bitset)) {
|
|
#endif
|
|
/* undefined behavior */
|
|
return 32;
|
|
}
|
|
|
|
return (int)index;
|
|
#else
|
|
int n;
|
|
|
|
if (bitset == (zend_mm_bitset)-1) return ZEND_MM_BITSET_LEN;
|
|
|
|
n = 0;
|
|
#if SIZEOF_ZEND_LONG == 8
|
|
if (sizeof(zend_mm_bitset) == 8) {
|
|
if ((bitset & 0xffffffff) == 0xffffffff) {n += 32; bitset = bitset >> Z_UL(32);}
|
|
}
|
|
#endif
|
|
if ((bitset & 0x0000ffff) == 0x0000ffff) {n += 16; bitset = bitset >> 16;}
|
|
if ((bitset & 0x000000ff) == 0x000000ff) {n += 8; bitset = bitset >> 8;}
|
|
if ((bitset & 0x0000000f) == 0x0000000f) {n += 4; bitset = bitset >> 4;}
|
|
if ((bitset & 0x00000003) == 0x00000003) {n += 2; bitset = bitset >> 2;}
|
|
return n + (bitset & 1);
|
|
#endif
|
|
}
|
|
|
|
/* number of trailing zero bits (0x01 -> 1; 0x40 -> 6; 0x00 -> LEN) */
|
|
static zend_always_inline int zend_mm_bitset_ntz(zend_mm_bitset bitset)
|
|
{
|
|
#if defined(__GNUC__)
|
|
# if SIZEOF_ZEND_LONG == SIZEOF_LONG
|
|
return __builtin_ctzl(bitset);
|
|
# else
|
|
return __builtin_ctzll(bitset);
|
|
# endif
|
|
#elif defined(_WIN32)
|
|
unsigned long index;
|
|
|
|
#if defined(_WIN64)
|
|
if (!BitScanForward64(&index, bitset)) {
|
|
#else
|
|
if (!BitScanForward(&index, bitset)) {
|
|
#endif
|
|
/* undefined behavior */
|
|
return 32;
|
|
}
|
|
|
|
return (int)index;
|
|
#else
|
|
int n;
|
|
|
|
if (bitset == (zend_mm_bitset)0) return ZEND_MM_BITSET_LEN;
|
|
|
|
n = 1;
|
|
#if SIZEOF_ZEND_LONG == 8
|
|
if (sizeof(zend_mm_bitset) == 8) {
|
|
if ((bitset & 0xffffffff) == 0) {n += 32; bitset = bitset >> Z_UL(32);}
|
|
}
|
|
#endif
|
|
if ((bitset & 0x0000ffff) == 0) {n += 16; bitset = bitset >> 16;}
|
|
if ((bitset & 0x000000ff) == 0) {n += 8; bitset = bitset >> 8;}
|
|
if ((bitset & 0x0000000f) == 0) {n += 4; bitset = bitset >> 4;}
|
|
if ((bitset & 0x00000003) == 0) {n += 2; bitset = bitset >> 2;}
|
|
return n - (bitset & 1);
|
|
#endif
|
|
}
|
|
|
|
static zend_always_inline int zend_mm_bitset_find_zero(zend_mm_bitset *bitset, int size)
|
|
{
|
|
int i = 0;
|
|
|
|
do {
|
|
zend_mm_bitset tmp = bitset[i];
|
|
if (tmp != (zend_mm_bitset)-1) {
|
|
return i * ZEND_MM_BITSET_LEN + zend_mm_bitset_nts(tmp);
|
|
}
|
|
i++;
|
|
} while (i < size);
|
|
return -1;
|
|
}
|
|
|
|
static zend_always_inline int zend_mm_bitset_find_one(zend_mm_bitset *bitset, int size)
|
|
{
|
|
int i = 0;
|
|
|
|
do {
|
|
zend_mm_bitset tmp = bitset[i];
|
|
if (tmp != 0) {
|
|
return i * ZEND_MM_BITSET_LEN + zend_mm_bitset_ntz(tmp);
|
|
}
|
|
i++;
|
|
} while (i < size);
|
|
return -1;
|
|
}
|
|
|
|
static zend_always_inline int zend_mm_bitset_find_zero_and_set(zend_mm_bitset *bitset, int size)
|
|
{
|
|
int i = 0;
|
|
|
|
do {
|
|
zend_mm_bitset tmp = bitset[i];
|
|
if (tmp != (zend_mm_bitset)-1) {
|
|
int n = zend_mm_bitset_nts(tmp);
|
|
bitset[i] |= Z_UL(1) << n;
|
|
return i * ZEND_MM_BITSET_LEN + n;
|
|
}
|
|
i++;
|
|
} while (i < size);
|
|
return -1;
|
|
}
|
|
|
|
static zend_always_inline int zend_mm_bitset_is_set(zend_mm_bitset *bitset, int bit)
|
|
{
|
|
return (bitset[bit / ZEND_MM_BITSET_LEN] & (Z_L(1) << (bit & (ZEND_MM_BITSET_LEN-1)))) != 0;
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_bitset_set_bit(zend_mm_bitset *bitset, int bit)
|
|
{
|
|
bitset[bit / ZEND_MM_BITSET_LEN] |= (Z_L(1) << (bit & (ZEND_MM_BITSET_LEN-1)));
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_bitset_reset_bit(zend_mm_bitset *bitset, int bit)
|
|
{
|
|
bitset[bit / ZEND_MM_BITSET_LEN] &= ~(Z_L(1) << (bit & (ZEND_MM_BITSET_LEN-1)));
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_bitset_set_range(zend_mm_bitset *bitset, int start, int len)
|
|
{
|
|
if (len == 1) {
|
|
zend_mm_bitset_set_bit(bitset, start);
|
|
} else {
|
|
int pos = start / ZEND_MM_BITSET_LEN;
|
|
int end = (start + len - 1) / ZEND_MM_BITSET_LEN;
|
|
int bit = start & (ZEND_MM_BITSET_LEN - 1);
|
|
zend_mm_bitset tmp;
|
|
|
|
if (pos != end) {
|
|
/* set bits from "bit" to ZEND_MM_BITSET_LEN-1 */
|
|
tmp = (zend_mm_bitset)-1 << bit;
|
|
bitset[pos++] |= tmp;
|
|
while (pos != end) {
|
|
/* set all bits */
|
|
bitset[pos++] = (zend_mm_bitset)-1;
|
|
}
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* set bits from "0" to "end" */
|
|
tmp = (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
bitset[pos] |= tmp;
|
|
} else {
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* set bits from "bit" to "end" */
|
|
tmp = (zend_mm_bitset)-1 << bit;
|
|
tmp &= (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
bitset[pos] |= tmp;
|
|
}
|
|
}
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_bitset_reset_range(zend_mm_bitset *bitset, int start, int len)
|
|
{
|
|
if (len == 1) {
|
|
zend_mm_bitset_reset_bit(bitset, start);
|
|
} else {
|
|
int pos = start / ZEND_MM_BITSET_LEN;
|
|
int end = (start + len - 1) / ZEND_MM_BITSET_LEN;
|
|
int bit = start & (ZEND_MM_BITSET_LEN - 1);
|
|
zend_mm_bitset tmp;
|
|
|
|
if (pos != end) {
|
|
/* reset bits from "bit" to ZEND_MM_BITSET_LEN-1 */
|
|
tmp = ~((Z_L(1) << bit) - 1);
|
|
bitset[pos++] &= ~tmp;
|
|
while (pos != end) {
|
|
/* set all bits */
|
|
bitset[pos++] = 0;
|
|
}
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* reset bits from "0" to "end" */
|
|
tmp = (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
bitset[pos] &= ~tmp;
|
|
} else {
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* reset bits from "bit" to "end" */
|
|
tmp = (zend_mm_bitset)-1 << bit;
|
|
tmp &= (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
bitset[pos] &= ~tmp;
|
|
}
|
|
}
|
|
}
|
|
|
|
static zend_always_inline int zend_mm_bitset_is_free_range(zend_mm_bitset *bitset, int start, int len)
|
|
{
|
|
if (len == 1) {
|
|
return !zend_mm_bitset_is_set(bitset, start);
|
|
} else {
|
|
int pos = start / ZEND_MM_BITSET_LEN;
|
|
int end = (start + len - 1) / ZEND_MM_BITSET_LEN;
|
|
int bit = start & (ZEND_MM_BITSET_LEN - 1);
|
|
zend_mm_bitset tmp;
|
|
|
|
if (pos != end) {
|
|
/* set bits from "bit" to ZEND_MM_BITSET_LEN-1 */
|
|
tmp = (zend_mm_bitset)-1 << bit;
|
|
if ((bitset[pos++] & tmp) != 0) {
|
|
return 0;
|
|
}
|
|
while (pos != end) {
|
|
/* set all bits */
|
|
if (bitset[pos++] != 0) {
|
|
return 0;
|
|
}
|
|
}
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* set bits from "0" to "end" */
|
|
tmp = (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
return (bitset[pos] & tmp) == 0;
|
|
} else {
|
|
end = (start + len - 1) & (ZEND_MM_BITSET_LEN - 1);
|
|
/* set bits from "bit" to "end" */
|
|
tmp = (zend_mm_bitset)-1 << bit;
|
|
tmp &= (zend_mm_bitset)-1 >> ((ZEND_MM_BITSET_LEN - 1) - end);
|
|
return (bitset[pos] & tmp) == 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**********/
|
|
/* Chunks */
|
|
/**********/
|
|
|
|
static void *zend_mm_chunk_alloc_int(size_t size, size_t alignment)
|
|
{
|
|
void *ptr = zend_mm_mmap(size);
|
|
|
|
if (ptr == NULL) {
|
|
return NULL;
|
|
} else if (ZEND_MM_ALIGNED_OFFSET(ptr, alignment) == 0) {
|
|
#ifdef MADV_HUGEPAGE
|
|
madvise(ptr, size, MADV_HUGEPAGE);
|
|
#endif
|
|
return ptr;
|
|
} else {
|
|
size_t offset;
|
|
|
|
/* chunk has to be aligned */
|
|
zend_mm_munmap(ptr, size);
|
|
ptr = zend_mm_mmap(size + alignment - REAL_PAGE_SIZE);
|
|
#ifdef _WIN32
|
|
offset = ZEND_MM_ALIGNED_OFFSET(ptr, alignment);
|
|
zend_mm_munmap(ptr, size + alignment - REAL_PAGE_SIZE);
|
|
ptr = zend_mm_mmap_fixed((void*)((char*)ptr + (alignment - offset)), size);
|
|
offset = ZEND_MM_ALIGNED_OFFSET(ptr, alignment);
|
|
if (offset != 0) {
|
|
zend_mm_munmap(ptr, size);
|
|
return NULL;
|
|
}
|
|
return ptr;
|
|
#else
|
|
offset = ZEND_MM_ALIGNED_OFFSET(ptr, alignment);
|
|
if (offset != 0) {
|
|
offset = alignment - offset;
|
|
zend_mm_munmap(ptr, offset);
|
|
ptr = (char*)ptr + offset;
|
|
alignment -= offset;
|
|
}
|
|
if (alignment > REAL_PAGE_SIZE) {
|
|
zend_mm_munmap((char*)ptr + size, alignment - REAL_PAGE_SIZE);
|
|
}
|
|
# ifdef MADV_HUGEPAGE
|
|
madvise(ptr, size, MADV_HUGEPAGE);
|
|
# endif
|
|
#endif
|
|
return ptr;
|
|
}
|
|
}
|
|
|
|
static void *zend_mm_chunk_alloc(zend_mm_heap *heap, size_t size, size_t alignment)
|
|
{
|
|
#if ZEND_MM_STORAGE
|
|
if (UNEXPECTED(heap->storage)) {
|
|
void *ptr = heap->storage->chunk_alloc(heap->storage, size, alignment);
|
|
ZEND_ASSERT(((zend_uintptr_t)((char*)ptr + (alignment-1)) & (alignment-1)) == (zend_uintptr_t)ptr);
|
|
return ptr;
|
|
}
|
|
#endif
|
|
return zend_mm_chunk_alloc_int(size, alignment);
|
|
}
|
|
|
|
static void zend_mm_chunk_free(zend_mm_heap *heap, void *addr, size_t size)
|
|
{
|
|
#if ZEND_MM_STORAGE
|
|
if (UNEXPECTED(heap->storage)) {
|
|
heap->storage->chunk_free(heap->storage, addr, size);
|
|
return;
|
|
}
|
|
#endif
|
|
zend_mm_munmap(addr, size);
|
|
}
|
|
|
|
static void zend_mm_chunk_truncate(zend_mm_heap *heap, void *addr, size_t old_size, size_t new_size)
|
|
{
|
|
#if ZEND_MM_STORAGE
|
|
if (UNEXPECTED(heap->storage)) {
|
|
heap->storage->chunk_truncate(heap->storage, addr, old_size, new_size);
|
|
return;
|
|
}
|
|
#endif
|
|
zend_mm_munmap((char*)addr + new_size, old_size - new_size);
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_chunk_init(zend_mm_heap *heap, zend_mm_chunk *chunk)
|
|
{
|
|
chunk->heap = heap;
|
|
chunk->next = heap->main_chunk;
|
|
chunk->prev = heap->main_chunk->prev;
|
|
chunk->prev->next = chunk;
|
|
chunk->next->prev = chunk;
|
|
/* mark first pages as allocated */
|
|
chunk->free_pages = ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE;
|
|
chunk->free_tail = ZEND_MM_FIRST_PAGE;
|
|
/* the younger chunks have bigger number */
|
|
chunk->num = chunk->prev->num + 1;
|
|
/* mark first pages as allocated */
|
|
chunk->free_map[0] = (1L << ZEND_MM_FIRST_PAGE) - 1;
|
|
chunk->map[0] = ZEND_MM_LRUN(ZEND_MM_FIRST_PAGE);
|
|
}
|
|
|
|
/***********************/
|
|
/* Huge Runs (forward) */
|
|
/***********************/
|
|
|
|
static size_t zend_mm_get_huge_block_size(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC);
|
|
static void *zend_mm_alloc_huge(zend_mm_heap *heap, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC);
|
|
static void zend_mm_free_huge(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC);
|
|
|
|
#if ZEND_DEBUG
|
|
static void zend_mm_change_huge_block_size(zend_mm_heap *heap, void *ptr, size_t size, size_t dbg_size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC);
|
|
#else
|
|
static void zend_mm_change_huge_block_size(zend_mm_heap *heap, void *ptr, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC);
|
|
#endif
|
|
|
|
/**************/
|
|
/* Large Runs */
|
|
/**************/
|
|
|
|
#if ZEND_DEBUG
|
|
static void *zend_mm_alloc_pages(zend_mm_heap *heap, int pages_count, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#else
|
|
static void *zend_mm_alloc_pages(zend_mm_heap *heap, int pages_count ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#endif
|
|
{
|
|
zend_mm_chunk *chunk = heap->main_chunk;
|
|
int page_num, len;
|
|
|
|
while (1) {
|
|
if (UNEXPECTED(chunk->free_pages < pages_count)) {
|
|
goto not_found;
|
|
#if 0
|
|
} else if (UNEXPECTED(chunk->free_pages + chunk->free_tail == ZEND_MM_PAGES)) {
|
|
if (UNEXPECTED(ZEND_MM_PAGES - chunk->free_tail < pages_count)) {
|
|
goto not_found;
|
|
} else {
|
|
page_num = chunk->free_tail;
|
|
goto found;
|
|
}
|
|
} else if (0) {
|
|
/* First-Fit Search */
|
|
int free_tail = chunk->free_tail;
|
|
zend_mm_bitset *bitset = chunk->free_map;
|
|
zend_mm_bitset tmp = *(bitset++);
|
|
int i = 0;
|
|
|
|
while (1) {
|
|
/* skip allocated blocks */
|
|
while (tmp == (zend_mm_bitset)-1) {
|
|
i += ZEND_MM_BITSET_LEN;
|
|
if (i == ZEND_MM_PAGES) {
|
|
goto not_found;
|
|
}
|
|
tmp = *(bitset++);
|
|
}
|
|
/* find first 0 bit */
|
|
page_num = i + zend_mm_bitset_nts(tmp);
|
|
/* reset bits from 0 to "bit" */
|
|
tmp &= tmp + 1;
|
|
/* skip free blocks */
|
|
while (tmp == 0) {
|
|
i += ZEND_MM_BITSET_LEN;
|
|
len = i - page_num;
|
|
if (len >= pages_count) {
|
|
goto found;
|
|
} else if (i >= free_tail) {
|
|
goto not_found;
|
|
}
|
|
tmp = *(bitset++);
|
|
}
|
|
/* find first 1 bit */
|
|
len = (i + zend_mm_bitset_ntz(tmp)) - page_num;
|
|
if (len >= pages_count) {
|
|
goto found;
|
|
}
|
|
/* set bits from 0 to "bit" */
|
|
tmp |= tmp - 1;
|
|
}
|
|
#endif
|
|
} else {
|
|
/* Best-Fit Search */
|
|
int best = -1;
|
|
int best_len = ZEND_MM_PAGES;
|
|
int free_tail = chunk->free_tail;
|
|
zend_mm_bitset *bitset = chunk->free_map;
|
|
zend_mm_bitset tmp = *(bitset++);
|
|
int i = 0;
|
|
|
|
while (1) {
|
|
/* skip allocated blocks */
|
|
while (tmp == (zend_mm_bitset)-1) {
|
|
i += ZEND_MM_BITSET_LEN;
|
|
if (i == ZEND_MM_PAGES) {
|
|
if (best > 0) {
|
|
page_num = best;
|
|
goto found;
|
|
} else {
|
|
goto not_found;
|
|
}
|
|
}
|
|
tmp = *(bitset++);
|
|
}
|
|
/* find first 0 bit */
|
|
page_num = i + zend_mm_bitset_nts(tmp);
|
|
/* reset bits from 0 to "bit" */
|
|
tmp &= tmp + 1;
|
|
/* skip free blocks */
|
|
while (tmp == 0) {
|
|
i += ZEND_MM_BITSET_LEN;
|
|
if (i >= free_tail) {
|
|
len = ZEND_MM_PAGES - page_num;
|
|
if (len >= pages_count && len < best_len) {
|
|
chunk->free_tail = page_num + pages_count;
|
|
goto found;
|
|
} else {
|
|
/* set accurate value */
|
|
chunk->free_tail = page_num;
|
|
if (best > 0) {
|
|
page_num = best;
|
|
goto found;
|
|
} else {
|
|
goto not_found;
|
|
}
|
|
}
|
|
}
|
|
tmp = *(bitset++);
|
|
}
|
|
/* find first 1 bit */
|
|
len = i + zend_mm_bitset_ntz(tmp) - page_num;
|
|
if (len >= pages_count) {
|
|
if (len == pages_count) {
|
|
goto found;
|
|
} else if (len < best_len) {
|
|
best_len = len;
|
|
best = page_num;
|
|
}
|
|
}
|
|
/* set bits from 0 to "bit" */
|
|
tmp |= tmp - 1;
|
|
}
|
|
}
|
|
|
|
not_found:
|
|
if (chunk->next == heap->main_chunk) {
|
|
if (heap->cached_chunks) {
|
|
heap->cached_chunks_count--;
|
|
chunk = heap->cached_chunks;
|
|
heap->cached_chunks = chunk->next;
|
|
} else {
|
|
#if ZEND_MM_LIMIT
|
|
if (heap->real_size + ZEND_MM_CHUNK_SIZE > heap->limit) {
|
|
if (heap->overflow == 0) {
|
|
#if ZEND_DEBUG
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted at %s:%d (tried to allocate %zu bytes)", heap->limit, __zend_filename, __zend_lineno, size);
|
|
#else
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted (tried to allocate %zu bytes)", heap->limit, ZEND_MM_PAGE_SIZE * pages_count);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
}
|
|
#endif
|
|
chunk = (zend_mm_chunk*)zend_mm_chunk_alloc(heap, ZEND_MM_CHUNK_SIZE, ZEND_MM_CHUNK_SIZE);
|
|
if (UNEXPECTED(chunk == NULL)) {
|
|
/* insufficient memory */
|
|
#if !ZEND_MM_LIMIT
|
|
zend_mm_safe_error(heap, "Out of memory");
|
|
#elif ZEND_DEBUG
|
|
zend_mm_safe_error(heap, "Out of memory (allocated %zu) at %s:%d (tried to allocate %zu bytes)", heap->real_size, __zend_filename, __zend_lineno, size);
|
|
#else
|
|
zend_mm_safe_error(heap, "Out of memory (allocated %zu) (tried to allocate %zu bytes)", heap->real_size, ZEND_MM_PAGE_SIZE * pages_count);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
#if ZEND_MM_STAT
|
|
do {
|
|
size_t size = heap->real_size + ZEND_MM_CHUNK_SIZE;
|
|
size_t peak = MAX(heap->real_peak, size);
|
|
heap->real_size = size;
|
|
heap->real_peak = peak;
|
|
} while (0);
|
|
#elif ZEND_MM_LIMIT
|
|
heap->real_size += ZEND_MM_CHUNK_SIZE;
|
|
|
|
#endif
|
|
}
|
|
heap->chunks_count++;
|
|
if (heap->chunks_count > heap->peak_chunks_count) {
|
|
heap->peak_chunks_count = heap->chunks_count;
|
|
}
|
|
zend_mm_chunk_init(heap, chunk);
|
|
page_num = ZEND_MM_FIRST_PAGE;
|
|
len = ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE;
|
|
goto found;
|
|
} else {
|
|
chunk = chunk->next;
|
|
}
|
|
}
|
|
|
|
found:
|
|
/* mark run as allocated */
|
|
chunk->free_pages -= pages_count;
|
|
zend_mm_bitset_set_range(chunk->free_map, page_num, pages_count);
|
|
chunk->map[page_num] = ZEND_MM_LRUN(pages_count);
|
|
if (page_num == chunk->free_tail) {
|
|
chunk->free_tail = page_num + pages_count;
|
|
}
|
|
return ZEND_MM_PAGE_ADDR(chunk, page_num);
|
|
}
|
|
|
|
static zend_always_inline void *zend_mm_alloc_large(zend_mm_heap *heap, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
int pages_count = (int)ZEND_MM_SIZE_TO_NUM(size, ZEND_MM_PAGE_SIZE);
|
|
#if ZEND_DEBUG
|
|
void *ptr = zend_mm_alloc_pages(heap, pages_count, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
void *ptr = zend_mm_alloc_pages(heap, pages_count ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
do {
|
|
size_t size = heap->size + pages_count * ZEND_MM_PAGE_SIZE;
|
|
size_t peak = MAX(heap->peak, size);
|
|
heap->size = size;
|
|
heap->peak = peak;
|
|
} while (0);
|
|
#endif
|
|
return ptr;
|
|
}
|
|
|
|
static void zend_mm_free_pages(zend_mm_heap *heap, zend_mm_chunk *chunk, int page_num, int pages_count)
|
|
{
|
|
chunk->free_pages += pages_count;
|
|
zend_mm_bitset_reset_range(chunk->free_map, page_num, pages_count);
|
|
chunk->map[page_num] = 0;
|
|
if (chunk->free_tail == page_num + pages_count) {
|
|
/* this setting may be not accurate */
|
|
chunk->free_tail = page_num;
|
|
}
|
|
if (chunk->free_pages == ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE) {
|
|
/* delete chunk */
|
|
chunk->next->prev = chunk->prev;
|
|
chunk->prev->next = chunk->next;
|
|
heap->chunks_count--;
|
|
if (heap->chunks_count + heap->cached_chunks_count < heap->avg_chunks_count + 0.1) {
|
|
/* delay deletion */
|
|
heap->cached_chunks_count++;
|
|
chunk->next = heap->cached_chunks;
|
|
heap->cached_chunks = chunk;
|
|
} else {
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size -= ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
if (!heap->cached_chunks || chunk->num > heap->cached_chunks->num) {
|
|
zend_mm_chunk_free(heap, chunk, ZEND_MM_CHUNK_SIZE);
|
|
} else {
|
|
//TODO: select the best chunk to delete???
|
|
chunk->next = heap->cached_chunks->next;
|
|
zend_mm_chunk_free(heap, heap->cached_chunks, ZEND_MM_CHUNK_SIZE);
|
|
heap->cached_chunks = chunk;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_free_large(zend_mm_heap *heap, zend_mm_chunk *chunk, int page_num, int pages_count)
|
|
{
|
|
#if ZEND_MM_STAT
|
|
heap->size -= pages_count * ZEND_MM_PAGE_SIZE;
|
|
#endif
|
|
zend_mm_free_pages(heap, chunk, page_num, pages_count);
|
|
}
|
|
|
|
/**************/
|
|
/* Small Runs */
|
|
/**************/
|
|
|
|
/* higher set bit number (0->N/A, 1->1, 2->2, 4->3, 8->4, 127->7, 128->8 etc) */
|
|
static zend_always_inline int zend_mm_small_size_to_bit(int size)
|
|
{
|
|
#if defined(__GNUC__)
|
|
return (__builtin_clz(size) ^ 0x1f) + 1;
|
|
#elif defined(_WIN32)
|
|
unsigned long index;
|
|
|
|
if (!BitScanReverse(&index, (unsigned long)size)) {
|
|
/* undefined behavior */
|
|
return 64;
|
|
}
|
|
|
|
return (((31 - (int)index) ^ 0x1f) + 1);
|
|
#else
|
|
int n = 16;
|
|
if (size <= 0x00ff) {n -= 8; size = size << 8;}
|
|
if (size <= 0x0fff) {n -= 4; size = size << 4;}
|
|
if (size <= 0x3fff) {n -= 2; size = size << 2;}
|
|
if (size <= 0x7fff) {n -= 1;}
|
|
return n;
|
|
#endif
|
|
}
|
|
|
|
#ifndef MAX
|
|
# define MAX(a, b) (((a) > (b)) ? (a) : (b))
|
|
#endif
|
|
|
|
#ifndef MIN
|
|
# define MIN(a, b) (((a) < (b)) ? (a) : (b))
|
|
#endif
|
|
|
|
static zend_always_inline int zend_mm_small_size_to_bin(size_t size)
|
|
{
|
|
#if 0
|
|
int n;
|
|
/*0, 1, 2, 3, 4, 5, 6, 7, 8, 9 10, 11, 12*/
|
|
static const int f1[] = { 3, 3, 3, 3, 3, 3, 3, 4, 5, 6, 7, 8, 9};
|
|
static const int f2[] = { 0, 0, 0, 0, 0, 0, 0, 4, 8, 12, 16, 20, 24};
|
|
|
|
if (UNEXPECTED(size <= 2)) return 0;
|
|
n = zend_mm_small_size_to_bit(size - 1);
|
|
return ((size-1) >> f1[n]) + f2[n];
|
|
#else
|
|
int t1, t2, t3;
|
|
|
|
if (UNEXPECTED(size <= 8)) return 0;
|
|
t1 = (int)(size - 1);
|
|
t2 = zend_mm_small_size_to_bit(t1);
|
|
t3 = t2 - 6;
|
|
t3 = (t3 < 0) ? 0 : t3;
|
|
t2 = t3 + 3;
|
|
t1 = t1 >> t2;
|
|
t3 = t3 << 2;
|
|
return t1 + t3;
|
|
#endif
|
|
}
|
|
|
|
#define ZEND_MM_SMALL_SIZE_TO_BIN(size) zend_mm_small_size_to_bin(size)
|
|
|
|
static zend_never_inline void *zend_mm_alloc_small_slow(zend_mm_heap *heap, int bin_num ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
zend_mm_chunk *chunk;
|
|
int page_num;
|
|
zend_mm_bin *bin;
|
|
zend_mm_free_slot *p, *end;
|
|
|
|
#if ZEND_DEBUG
|
|
bin = (zend_mm_bin*)zend_mm_alloc_pages(heap, bin_pages[bin_num], bin_data_size[bin_num] ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
bin = (zend_mm_bin*)zend_mm_alloc_pages(heap, bin_pages[bin_num] ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
if (UNEXPECTED(bin == NULL)) {
|
|
/* insufficient memory */
|
|
return NULL;
|
|
}
|
|
|
|
chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(bin, ZEND_MM_CHUNK_SIZE);
|
|
page_num = ZEND_MM_ALIGNED_OFFSET(bin, ZEND_MM_CHUNK_SIZE) / ZEND_MM_PAGE_SIZE;
|
|
chunk->map[page_num] = ZEND_MM_SRUN(bin_num);
|
|
if (bin_pages[bin_num] > 1) {
|
|
int i = 1;
|
|
do {
|
|
chunk->map[page_num+i] = ZEND_MM_SRUN(bin_num);
|
|
i++;
|
|
} while (i < bin_pages[bin_num]);
|
|
}
|
|
|
|
/* create a linked list of elements from 1 to last */
|
|
end = (zend_mm_free_slot*)((char*)bin + (bin_data_size[bin_num] * (bin_elements[bin_num] - 1)));
|
|
heap->free_slot[bin_num] = p = (zend_mm_free_slot*)((char*)bin + bin_data_size[bin_num]);
|
|
do {
|
|
p->next_free_slot = (zend_mm_free_slot*)((char*)p + bin_data_size[bin_num]);;
|
|
#if ZEND_DEBUG
|
|
do {
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + bin_data_size[bin_num] - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
dbg->size = 0;
|
|
} while (0);
|
|
#endif
|
|
p = (zend_mm_free_slot*)((char*)p + bin_data_size[bin_num]);
|
|
} while (p != end);
|
|
|
|
/* terminate list using NULL */
|
|
p->next_free_slot = NULL;
|
|
#if ZEND_DEBUG
|
|
do {
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + bin_data_size[bin_num] - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
dbg->size = 0;
|
|
} while (0);
|
|
#endif
|
|
|
|
/* return first element */
|
|
return (char*)bin;
|
|
}
|
|
|
|
static zend_always_inline void *zend_mm_alloc_small(zend_mm_heap *heap, size_t size, int bin_num ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
#if ZEND_MM_STAT
|
|
do {
|
|
size_t size = heap->size + bin_data_size[bin_num];
|
|
size_t peak = MAX(heap->peak, size);
|
|
heap->size = size;
|
|
heap->peak = peak;
|
|
} while (0);
|
|
#endif
|
|
|
|
if (EXPECTED(heap->free_slot[bin_num] != NULL)) {
|
|
zend_mm_free_slot *p = heap->free_slot[bin_num];
|
|
heap->free_slot[bin_num] = p->next_free_slot;
|
|
return (void*)p;
|
|
} else {
|
|
return zend_mm_alloc_small_slow(heap, bin_num ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_free_small(zend_mm_heap *heap, void *ptr, int bin_num)
|
|
{
|
|
zend_mm_free_slot *p;
|
|
|
|
#if ZEND_MM_STAT
|
|
heap->size -= bin_data_size[bin_num];
|
|
#endif
|
|
|
|
#if ZEND_DEBUG
|
|
do {
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)ptr + bin_data_size[bin_num] - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
dbg->size = 0;
|
|
} while (0);
|
|
#endif
|
|
|
|
p = (zend_mm_free_slot*)ptr;
|
|
p->next_free_slot = heap->free_slot[bin_num];
|
|
heap->free_slot[bin_num] = p;
|
|
}
|
|
|
|
/********/
|
|
/* Heap */
|
|
/********/
|
|
|
|
#if ZEND_DEBUG
|
|
static zend_always_inline zend_mm_debug_info *zend_mm_get_debug_info(zend_mm_heap *heap, void *ptr)
|
|
{
|
|
size_t page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE);
|
|
zend_mm_chunk *chunk;
|
|
int page_num;
|
|
zend_mm_page_info info;
|
|
|
|
ZEND_MM_CHECK(page_offset != 0, "zend_mm_heap corrupted");
|
|
chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE);
|
|
page_num = (int)(page_offset / ZEND_MM_PAGE_SIZE);
|
|
info = chunk->map[page_num];
|
|
ZEND_MM_CHECK(chunk->heap == heap, "zend_mm_heap corrupted");
|
|
if (EXPECTED(info & ZEND_MM_IS_SRUN)) {
|
|
int bin_num = ZEND_MM_SRUN_BIN_NUM(info);
|
|
return (zend_mm_debug_info*)((char*)ptr + bin_data_size[bin_num] - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
} else /* if (info & ZEND_MM_IS_LRUN) */ {
|
|
int pages_count = ZEND_MM_LRUN_PAGES(info);
|
|
|
|
return (zend_mm_debug_info*)((char*)ptr + ZEND_MM_PAGE_SIZE * pages_count - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static zend_always_inline void *zend_mm_alloc_heap(zend_mm_heap *heap, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
void *ptr;
|
|
#if ZEND_DEBUG
|
|
size_t real_size = size;
|
|
zend_mm_debug_info *dbg;
|
|
|
|
/* special handling for zero-size allocation */
|
|
size = MAX(size, 1);
|
|
size = ZEND_MM_ALIGNED_SIZE(size) + ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info));
|
|
#endif
|
|
if (size <= ZEND_MM_MAX_SMALL_SIZE) {
|
|
ptr = zend_mm_alloc_small(heap, size, ZEND_MM_SMALL_SIZE_TO_BIN(size) ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
} else if (size <= ZEND_MM_MAX_LARGE_SIZE) {
|
|
ptr = zend_mm_alloc_large(heap, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
} else {
|
|
#if ZEND_DEBUG
|
|
size = real_size;
|
|
#endif
|
|
return zend_mm_alloc_huge(heap, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
}
|
|
|
|
static zend_always_inline void zend_mm_free_heap(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE);
|
|
|
|
if (UNEXPECTED(page_offset == 0)) {
|
|
if (ptr != NULL) {
|
|
zend_mm_free_huge(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
} else {
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE);
|
|
int page_num = (int)(page_offset / ZEND_MM_PAGE_SIZE);
|
|
zend_mm_page_info info = chunk->map[page_num];
|
|
|
|
ZEND_MM_CHECK(chunk->heap == heap, "zend_mm_heap corrupted");
|
|
if (EXPECTED(info & ZEND_MM_IS_SRUN)) {
|
|
zend_mm_free_small(heap, ptr, ZEND_MM_SRUN_BIN_NUM(info));
|
|
} else /* if (info & ZEND_MM_IS_LRUN) */ {
|
|
int pages_count = ZEND_MM_LRUN_PAGES(info);
|
|
|
|
ZEND_MM_CHECK(ZEND_MM_ALIGNED_OFFSET(page_offset, ZEND_MM_PAGE_SIZE) == 0, "zend_mm_heap corrupted");
|
|
zend_mm_free_large(heap, chunk, page_num, pages_count);
|
|
}
|
|
}
|
|
}
|
|
|
|
static size_t zend_mm_size(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE);
|
|
|
|
if (UNEXPECTED(page_offset == 0)) {
|
|
return zend_mm_get_huge_block_size(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
} else {
|
|
zend_mm_chunk *chunk;
|
|
#if 0 && ZEND_DEBUG
|
|
zend_mm_debug_info *dbg = zend_mm_get_debug_info(heap, ptr);
|
|
return dbg->size;
|
|
#else
|
|
int page_num;
|
|
zend_mm_page_info info;
|
|
|
|
chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE);
|
|
page_num = (int)(page_offset / ZEND_MM_PAGE_SIZE);
|
|
info = chunk->map[page_num];
|
|
ZEND_MM_CHECK(chunk->heap == heap, "zend_mm_heap corrupted");
|
|
if (EXPECTED(info & ZEND_MM_IS_SRUN)) {
|
|
return bin_data_size[ZEND_MM_SRUN_BIN_NUM(info)];
|
|
} else /* if (info & ZEND_MM_IS_LARGE_RUN) */ {
|
|
return ZEND_MM_LRUN_PAGES(info) * ZEND_MM_PAGE_SIZE;
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static void *zend_mm_realloc_heap(zend_mm_heap *heap, void *ptr, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t page_offset;
|
|
size_t old_size;
|
|
size_t new_size;
|
|
void *ret;
|
|
#if ZEND_DEBUG
|
|
size_t real_size;
|
|
zend_mm_debug_info *dbg;
|
|
#endif
|
|
|
|
page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE);
|
|
if (UNEXPECTED(page_offset == 0)) {
|
|
if (UNEXPECTED(ptr == NULL)) {
|
|
return zend_mm_alloc_heap(heap, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
old_size = zend_mm_get_huge_block_size(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#if ZEND_DEBUG
|
|
real_size = size;
|
|
size = ZEND_MM_ALIGNED_SIZE(size) + ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info));
|
|
#endif
|
|
if (size > ZEND_MM_MAX_LARGE_SIZE) {
|
|
#if ZEND_DEBUG
|
|
size = real_size;
|
|
#endif
|
|
new_size = ZEND_MM_ALIGNED_SIZE_EX(size, REAL_PAGE_SIZE);
|
|
if (new_size == old_size) {
|
|
#if ZEND_DEBUG
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size, real_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
return ptr;
|
|
#ifndef _WIN32
|
|
} else if (new_size < old_size) {
|
|
/* unmup tail */
|
|
zend_mm_chunk_truncate(heap, ptr, old_size, new_size);
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size -= old_size - new_size;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->size -= old_size - new_size;
|
|
#endif
|
|
#if ZEND_DEBUG
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size, real_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
return ptr;
|
|
} else /* if (new_size > old_size) */ {
|
|
#if ZEND_MM_LIMIT
|
|
if (heap->real_size + (new_size - old_size) > heap->limit) {
|
|
if (heap->overflow == 0) {
|
|
#if ZEND_DEBUG
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted at %s:%d (tried to allocate %zu bytes)", heap->limit, __zend_filename, __zend_lineno, size);
|
|
#else
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted (tried to allocate %zu bytes)", heap->limit, size);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
}
|
|
#endif
|
|
/* try to map tail right after this block */
|
|
if (zend_mm_mmap_fixed((char*)ptr + old_size, new_size - old_size)) {
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size += new_size - old_size;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->size += new_size - old_size;
|
|
#endif
|
|
#if ZEND_DEBUG
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size, real_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
zend_mm_change_huge_block_size(heap, ptr, new_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
return ptr;
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
} else {
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE);
|
|
int page_num = (int)(page_offset / ZEND_MM_PAGE_SIZE);
|
|
zend_mm_page_info info = chunk->map[page_num];
|
|
#if ZEND_DEBUG
|
|
size_t real_size = size;
|
|
|
|
size = ZEND_MM_ALIGNED_SIZE(size) + ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info));
|
|
#endif
|
|
|
|
ZEND_MM_CHECK(chunk->heap == heap, "zend_mm_heap corrupted");
|
|
if (info & ZEND_MM_IS_SRUN) {
|
|
int old_bin_num, bin_num;
|
|
|
|
old_bin_num = ZEND_MM_SRUN_BIN_NUM(info);
|
|
old_size = bin_data_size[old_bin_num];
|
|
bin_num = ZEND_MM_SMALL_SIZE_TO_BIN(size);
|
|
if (old_bin_num == bin_num) {
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
}
|
|
} else /* if (info & ZEND_MM_IS_LARGE_RUN) */ {
|
|
ZEND_MM_CHECK(ZEND_MM_ALIGNED_OFFSET(page_offset, ZEND_MM_PAGE_SIZE) == 0, "zend_mm_heap corrupted");
|
|
old_size = ZEND_MM_LRUN_PAGES(info) * ZEND_MM_PAGE_SIZE;
|
|
if (size > ZEND_MM_MAX_SMALL_SIZE && size <= ZEND_MM_MAX_LARGE_SIZE) {
|
|
new_size = ZEND_MM_ALIGNED_SIZE_EX(size, ZEND_MM_PAGE_SIZE);
|
|
if (new_size == old_size) {
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
} else if (new_size < old_size) {
|
|
/* free tail pages */
|
|
int new_pages_count = (int)(new_size / ZEND_MM_PAGE_SIZE);
|
|
int rest_pages_count = (int)((old_size - new_size) / ZEND_MM_PAGE_SIZE);
|
|
|
|
#if ZEND_MM_STAT
|
|
heap->size -= rest_pages_count * ZEND_MM_PAGE_SIZE;
|
|
#endif
|
|
chunk->map[page_num] = ZEND_MM_LRUN(new_pages_count);
|
|
chunk->free_pages += rest_pages_count;
|
|
zend_mm_bitset_reset_range(chunk->free_map, page_num + new_pages_count, rest_pages_count);
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
} else /* if (new_size > old_size) */ {
|
|
int new_pages_count = (int)(new_size / ZEND_MM_PAGE_SIZE);
|
|
int old_pages_count = (int)(old_size / ZEND_MM_PAGE_SIZE);
|
|
|
|
/* try to allocate tail pages after this block */
|
|
if (page_num + new_pages_count <= ZEND_MM_PAGES &&
|
|
zend_mm_bitset_is_free_range(chunk->free_map, page_num + old_pages_count, new_pages_count - old_pages_count)) {
|
|
#if ZEND_MM_STAT
|
|
do {
|
|
size_t size = heap->size + (new_size - old_size);
|
|
size_t peak = MAX(heap->peak, size);
|
|
heap->size = size;
|
|
heap->peak = peak;
|
|
} while (0);
|
|
#endif
|
|
chunk->free_pages -= new_pages_count - old_pages_count;
|
|
zend_mm_bitset_set_range(chunk->free_map, page_num + old_pages_count, new_pages_count - old_pages_count);
|
|
chunk->map[page_num] = ZEND_MM_LRUN(new_pages_count);
|
|
#if ZEND_DEBUG
|
|
dbg = zend_mm_get_debug_info(heap, ptr);
|
|
dbg->size = real_size;
|
|
dbg->filename = __zend_filename;
|
|
dbg->orig_filename = __zend_orig_filename;
|
|
dbg->lineno = __zend_lineno;
|
|
dbg->orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return ptr;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#if ZEND_DEBUG
|
|
size = real_size;
|
|
#endif
|
|
}
|
|
|
|
/* Naive reallocation */
|
|
old_size = zend_mm_size(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
ret = zend_mm_alloc_heap(heap, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
memcpy(ret, ptr, MIN(old_size, size));
|
|
zend_mm_free_heap(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
return ret;
|
|
}
|
|
|
|
/*********************/
|
|
/* Huge Runs (again) */
|
|
/*********************/
|
|
|
|
#if ZEND_DEBUG
|
|
static void zend_mm_add_huge_block(zend_mm_heap *heap, void *ptr, size_t size, size_t dbg_size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#else
|
|
static void zend_mm_add_huge_block(zend_mm_heap *heap, void *ptr, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#endif
|
|
{
|
|
zend_mm_huge_list *list = (zend_mm_huge_list*)zend_mm_alloc_heap(heap, sizeof(zend_mm_huge_list) ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
list->ptr = ptr;
|
|
list->size = size;
|
|
list->next = heap->huge_list;
|
|
#if ZEND_DEBUG
|
|
list->dbg.size = dbg_size;
|
|
list->dbg.filename = __zend_filename;
|
|
list->dbg.orig_filename = __zend_orig_filename;
|
|
list->dbg.lineno = __zend_lineno;
|
|
list->dbg.orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
heap->huge_list = list;
|
|
}
|
|
|
|
static size_t zend_mm_del_huge_block(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
zend_mm_huge_list *prev = NULL;
|
|
zend_mm_huge_list *list = heap->huge_list;
|
|
while (list != NULL) {
|
|
if (list->ptr == ptr) {
|
|
size_t size;
|
|
|
|
if (prev) {
|
|
prev->next = list->next;
|
|
} else {
|
|
heap->huge_list = list->next;
|
|
}
|
|
size = list->size;
|
|
zend_mm_free_heap(heap, list ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
return size;
|
|
}
|
|
prev = list;
|
|
list = list->next;
|
|
}
|
|
ZEND_MM_CHECK(0, "zend_mm_heap corrupted");
|
|
return 0;
|
|
}
|
|
|
|
static size_t zend_mm_get_huge_block_size(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
zend_mm_huge_list *list = heap->huge_list;
|
|
while (list != NULL) {
|
|
if (list->ptr == ptr) {
|
|
return list->size;
|
|
}
|
|
list = list->next;
|
|
}
|
|
ZEND_MM_CHECK(0, "zend_mm_heap corrupted");
|
|
return 0;
|
|
}
|
|
|
|
#if ZEND_DEBUG
|
|
static void zend_mm_change_huge_block_size(zend_mm_heap *heap, void *ptr, size_t size, size_t dbg_size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#else
|
|
static void zend_mm_change_huge_block_size(zend_mm_heap *heap, void *ptr, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
#endif
|
|
{
|
|
zend_mm_huge_list *list = heap->huge_list;
|
|
while (list != NULL) {
|
|
if (list->ptr == ptr) {
|
|
list->size = size;
|
|
#if ZEND_DEBUG
|
|
list->dbg.size = dbg_size;
|
|
list->dbg.filename = __zend_filename;
|
|
list->dbg.orig_filename = __zend_orig_filename;
|
|
list->dbg.lineno = __zend_lineno;
|
|
list->dbg.orig_lineno = __zend_orig_lineno;
|
|
#endif
|
|
return;
|
|
}
|
|
list = list->next;
|
|
}
|
|
}
|
|
|
|
static void *zend_mm_alloc_huge(zend_mm_heap *heap, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t new_size = ZEND_MM_ALIGNED_SIZE_EX(size, REAL_PAGE_SIZE);
|
|
void *ptr;
|
|
|
|
#if ZEND_MM_LIMIT
|
|
if (heap->real_size + new_size > heap->limit) {
|
|
if (heap->overflow == 0) {
|
|
#if ZEND_DEBUG
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted at %s:%d (tried to allocate %zu bytes)", heap->limit, __zend_filename, __zend_lineno, size);
|
|
#else
|
|
zend_mm_safe_error(heap, "Allowed memory size of %zu bytes exhausted (tried to allocate %zu bytes)", heap->limit, size);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
}
|
|
#endif
|
|
ptr = zend_mm_chunk_alloc(heap, new_size, ZEND_MM_CHUNK_SIZE);
|
|
if (UNEXPECTED(ptr == NULL)) {
|
|
/* insufficient memory */
|
|
#if !ZEND_MM_LIMIT
|
|
zend_mm_safe_error(heap, "Out of memory");
|
|
#elif ZEND_DEBUG
|
|
zend_mm_safe_error(heap, "Out of memory (allocated %zu) at %s:%d (tried to allocate %zu bytes)", heap->real_size, __zend_filename, __zend_lineno, size);
|
|
#else
|
|
zend_mm_safe_error(heap, "Out of memory (allocated %zu) (tried to allocate %zu bytes)", heap->real_size, size);
|
|
#endif
|
|
return NULL;
|
|
}
|
|
#if ZEND_DEBUG
|
|
zend_mm_add_huge_block(heap, ptr, new_size, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#else
|
|
zend_mm_add_huge_block(heap, ptr, new_size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
do {
|
|
size_t size = heap->real_size + new_size;
|
|
size_t peak = MAX(heap->real_peak, size);
|
|
heap->real_size = size;
|
|
heap->real_peak = peak;
|
|
} while (0);
|
|
do {
|
|
size_t size = heap->size + new_size;
|
|
size_t peak = MAX(heap->peak, size);
|
|
heap->size = size;
|
|
heap->peak = peak;
|
|
} while (0);
|
|
#elif ZEND_MM_LIMIT
|
|
heap->real_size += new_size;
|
|
#endif
|
|
return ptr;
|
|
}
|
|
|
|
static void zend_mm_free_huge(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t size;
|
|
|
|
ZEND_MM_CHECK(ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE) == 0, "zend_mm_heap corrupted");
|
|
size = zend_mm_del_huge_block(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
zend_mm_chunk_free(heap, ptr, size);
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size -= size;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->size -= size;
|
|
#endif
|
|
}
|
|
|
|
/******************/
|
|
/* Initialization */
|
|
/******************/
|
|
|
|
static zend_mm_heap *zend_mm_init(void)
|
|
{
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)zend_mm_chunk_alloc_int(ZEND_MM_CHUNK_SIZE, ZEND_MM_CHUNK_SIZE);
|
|
zend_mm_heap *heap;
|
|
|
|
if (UNEXPECTED(chunk == NULL)) {
|
|
#if ZEND_MM_ERROR
|
|
#ifdef _WIN32
|
|
stderr_last_error("Can't initialize heap");
|
|
#else
|
|
fprintf(stderr, "\nCan't initialize heap: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
#endif
|
|
return NULL;
|
|
}
|
|
heap = &chunk->heap_slot;
|
|
chunk->heap = heap;
|
|
chunk->next = chunk;
|
|
chunk->prev = chunk;
|
|
chunk->free_pages = ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE;
|
|
chunk->free_tail = ZEND_MM_FIRST_PAGE;
|
|
chunk->num = 0;
|
|
chunk->free_map[0] = (Z_L(1) << ZEND_MM_FIRST_PAGE) - 1;
|
|
chunk->map[0] = ZEND_MM_LRUN(ZEND_MM_FIRST_PAGE);
|
|
heap->main_chunk = chunk;
|
|
heap->cached_chunks = NULL;
|
|
heap->chunks_count = 1;
|
|
heap->peak_chunks_count = 1;
|
|
heap->cached_chunks_count = 0;
|
|
heap->avg_chunks_count = 1.0;
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size = ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->real_peak = ZEND_MM_CHUNK_SIZE;
|
|
heap->size = 0;
|
|
heap->peak = 0;
|
|
#endif
|
|
#if ZEND_MM_LIMIT
|
|
heap->limit = (Z_L(-1) >> Z_L(1));
|
|
heap->overflow = 0;
|
|
#endif
|
|
#if ZEND_MM_CUSTOM
|
|
heap->use_custom_heap = 0;
|
|
#endif
|
|
#if ZEND_MM_STORAGE
|
|
heap->storage = NULL;
|
|
#endif
|
|
heap->huge_list = NULL;
|
|
return heap;
|
|
}
|
|
|
|
#if ZEND_DEBUG
|
|
/******************/
|
|
/* Leak detection */
|
|
/******************/
|
|
|
|
static zend_long zend_mm_find_leaks_small(zend_mm_chunk *p, int i, int j, zend_leak_info *leak)
|
|
{
|
|
int empty = 1;
|
|
zend_long count = 0;
|
|
int bin_num = ZEND_MM_SRUN_BIN_NUM(p->map[i]);
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + ZEND_MM_PAGE_SIZE * i + bin_data_size[bin_num] * (j + 1) - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
|
|
while (j < bin_elements[bin_num]) {
|
|
if (dbg->size != 0) {
|
|
if (dbg->filename == leak->filename && dbg->lineno == leak->lineno) {
|
|
count++;
|
|
dbg->size = 0;
|
|
dbg->filename = NULL;
|
|
dbg->lineno = 0;
|
|
} else {
|
|
empty = 0;
|
|
}
|
|
}
|
|
j++;
|
|
dbg = (zend_mm_debug_info*)((char*)dbg + bin_data_size[bin_num]);
|
|
}
|
|
if (empty) {
|
|
zend_mm_bitset_reset_range(p->free_map, i, bin_pages[bin_num]);
|
|
}
|
|
return count;
|
|
}
|
|
|
|
static zend_long zend_mm_find_leaks(zend_mm_heap *heap, zend_mm_chunk *p, int i, zend_leak_info *leak)
|
|
{
|
|
zend_long count = 0;
|
|
|
|
do {
|
|
while (i < p->free_tail) {
|
|
if (zend_mm_bitset_is_set(p->free_map, i)) {
|
|
if (p->map[i] & ZEND_MM_IS_SRUN) {
|
|
int bin_num = ZEND_MM_SRUN_BIN_NUM(p->map[i]);
|
|
count += zend_mm_find_leaks_small(p, i, 0, leak);
|
|
i += bin_pages[bin_num];
|
|
} else /* if (p->map[i] & ZEND_MM_IS_LRUN) */ {
|
|
int pages_count = ZEND_MM_LRUN_PAGES(p->map[i]);
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + ZEND_MM_PAGE_SIZE * (i + pages_count) - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
|
|
if (dbg->filename == leak->filename && dbg->lineno == leak->lineno) {
|
|
count++;
|
|
}
|
|
zend_mm_bitset_reset_range(p->free_map, i, pages_count);
|
|
i += pages_count;
|
|
}
|
|
} else {
|
|
i++;
|
|
}
|
|
}
|
|
p = p->next;
|
|
} while (p != heap->main_chunk);
|
|
return count;
|
|
}
|
|
|
|
static void zend_mm_check_leaks(zend_mm_heap *heap)
|
|
{
|
|
zend_mm_huge_list *list;
|
|
zend_mm_chunk *p;
|
|
zend_leak_info leak;
|
|
zend_long repeated = 0;
|
|
uint32_t total = 0;
|
|
int i, j;
|
|
|
|
/* find leaked huge blocks and free them */
|
|
list = heap->huge_list;
|
|
while (list) {
|
|
zend_mm_huge_list *q = list;
|
|
|
|
heap->huge_list = list->next;
|
|
|
|
leak.addr = list->ptr;
|
|
leak.size = list->dbg.size;
|
|
leak.filename = list->dbg.filename;
|
|
leak.orig_filename = list->dbg.orig_filename;
|
|
leak.lineno = list->dbg.lineno;
|
|
leak.orig_lineno = list->dbg.orig_lineno;
|
|
|
|
zend_message_dispatcher(ZMSG_LOG_SCRIPT_NAME, NULL);
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_DETECTED, &leak);
|
|
//??? repeated = zend_mm_find_leaks_huge(segment, p);
|
|
total += 1 + repeated;
|
|
if (repeated) {
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_REPEATED, (void *)(zend_uintptr_t)repeated);
|
|
}
|
|
|
|
list = list->next;
|
|
zend_mm_chunk_free(heap, q->ptr, q->size);
|
|
zend_mm_free_heap(heap, q, NULL, 0, NULL, 0);
|
|
}
|
|
|
|
/* for each chunk */
|
|
p = heap->main_chunk;
|
|
do {
|
|
i = ZEND_MM_FIRST_PAGE;
|
|
while (i < p->free_tail) {
|
|
if (zend_mm_bitset_is_set(p->free_map, i)) {
|
|
if (p->map[i] & ZEND_MM_IS_SRUN) {
|
|
int bin_num = ZEND_MM_SRUN_BIN_NUM(p->map[i]);
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + ZEND_MM_PAGE_SIZE * i + bin_data_size[bin_num] - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
|
|
j = 0;
|
|
while (j < bin_elements[bin_num]) {
|
|
if (dbg->size != 0) {
|
|
leak.addr = (zend_mm_debug_info*)((char*)p + ZEND_MM_PAGE_SIZE * i + bin_data_size[bin_num] * j);
|
|
leak.size = dbg->size;
|
|
leak.filename = dbg->filename;
|
|
leak.orig_filename = dbg->orig_filename;
|
|
leak.lineno = dbg->lineno;
|
|
leak.orig_lineno = dbg->orig_lineno;
|
|
|
|
zend_message_dispatcher(ZMSG_LOG_SCRIPT_NAME, NULL);
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_DETECTED, &leak);
|
|
|
|
dbg->size = 0;
|
|
dbg->filename = NULL;
|
|
dbg->lineno = 0;
|
|
|
|
repeated = zend_mm_find_leaks_small(p, i, j + 1, &leak) +
|
|
zend_mm_find_leaks(heap, p, i + bin_pages[bin_num], &leak);
|
|
total += 1 + repeated;
|
|
if (repeated) {
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_REPEATED, (void *)(zend_uintptr_t)repeated);
|
|
}
|
|
}
|
|
dbg = (zend_mm_debug_info*)((char*)dbg + bin_data_size[bin_num]);
|
|
j++;
|
|
}
|
|
i += bin_pages[bin_num];
|
|
} else /* if (p->map[i] & ZEND_MM_IS_LRUN) */ {
|
|
int pages_count = ZEND_MM_LRUN_PAGES(p->map[i]);
|
|
zend_mm_debug_info *dbg = (zend_mm_debug_info*)((char*)p + ZEND_MM_PAGE_SIZE * (i + pages_count) - ZEND_MM_ALIGNED_SIZE(sizeof(zend_mm_debug_info)));
|
|
|
|
leak.addr = (void*)((char*)p + ZEND_MM_PAGE_SIZE * i);
|
|
leak.size = dbg->size;
|
|
leak.filename = dbg->filename;
|
|
leak.orig_filename = dbg->orig_filename;
|
|
leak.lineno = dbg->lineno;
|
|
leak.orig_lineno = dbg->orig_lineno;
|
|
|
|
zend_message_dispatcher(ZMSG_LOG_SCRIPT_NAME, NULL);
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_DETECTED, &leak);
|
|
|
|
zend_mm_bitset_reset_range(p->free_map, i, pages_count);
|
|
|
|
repeated = zend_mm_find_leaks(heap, p, i + pages_count, &leak);
|
|
total += 1 + repeated;
|
|
if (repeated) {
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAK_REPEATED, (void *)(zend_uintptr_t)repeated);
|
|
}
|
|
i += pages_count;
|
|
}
|
|
} else {
|
|
i++;
|
|
}
|
|
}
|
|
p = p->next;
|
|
} while (p != heap->main_chunk);
|
|
if (total) {
|
|
zend_message_dispatcher(ZMSG_MEMORY_LEAKS_GRAND_TOTAL, &total);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
void zend_mm_shutdown(zend_mm_heap *heap, int full, int silent)
|
|
{
|
|
zend_mm_chunk *p;
|
|
zend_mm_huge_list *list;
|
|
|
|
#if ZEND_MM_CUSTOM
|
|
if (heap->use_custom_heap) {
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
#if ZEND_DEBUG
|
|
if (!silent) {
|
|
zend_mm_check_leaks(heap);
|
|
}
|
|
#endif
|
|
|
|
/* free huge blocks */
|
|
list = heap->huge_list;
|
|
while (list) {
|
|
zend_mm_huge_list *q = list;
|
|
list = list->next;
|
|
zend_mm_chunk_free(heap, q->ptr, q->size);
|
|
}
|
|
|
|
/* move all chunks except of the first one into the cache */
|
|
p = heap->main_chunk->next;
|
|
while (p != heap->main_chunk) {
|
|
zend_mm_chunk *q = p->next;
|
|
p->next = heap->cached_chunks;
|
|
heap->cached_chunks = p;
|
|
p = q;
|
|
heap->chunks_count--;
|
|
heap->cached_chunks_count++;
|
|
}
|
|
|
|
if (full) {
|
|
/* free all cached chunks */
|
|
while (heap->cached_chunks) {
|
|
p = heap->cached_chunks;
|
|
heap->cached_chunks = p->next;
|
|
zend_mm_chunk_free(heap, p, ZEND_MM_CHUNK_SIZE);
|
|
}
|
|
/* free the first chunk */
|
|
#if ZEND_MM_STORAGE
|
|
if (UNEXPECTED(heap->storage)) {
|
|
zend_mm_storage *storage = heap->storage;
|
|
zend_mm_chunk_free(heap, heap->main_chunk, ZEND_MM_CHUNK_SIZE);
|
|
storage->dtor(storage);
|
|
} else {
|
|
zend_mm_chunk_free(heap, heap->main_chunk, ZEND_MM_CHUNK_SIZE);
|
|
}
|
|
#else
|
|
zend_mm_chunk_free(heap, heap->main_chunk, ZEND_MM_CHUNK_SIZE);
|
|
#endif
|
|
} else {
|
|
zend_mm_heap old_heap;
|
|
|
|
/* free some cached chunks to keep average count */
|
|
heap->avg_chunks_count = (heap->avg_chunks_count + (double)heap->peak_chunks_count) / 2.0;
|
|
while ((double)heap->cached_chunks_count + 0.9 > heap->avg_chunks_count &&
|
|
heap->cached_chunks) {
|
|
p = heap->cached_chunks;
|
|
heap->cached_chunks = p->next;
|
|
zend_mm_chunk_free(heap, p, ZEND_MM_CHUNK_SIZE);
|
|
heap->cached_chunks_count--;
|
|
}
|
|
/* clear cached chunks */
|
|
p = heap->cached_chunks;
|
|
while (p != NULL) {
|
|
zend_mm_chunk *q = p->next;
|
|
memset(p, 0, sizeof(zend_mm_chunk));
|
|
p->next = q;
|
|
p = q;
|
|
}
|
|
|
|
/* reinitialize the first chunk and heap */
|
|
old_heap = *heap;
|
|
p = heap->main_chunk;
|
|
memset(p, 0, ZEND_MM_FIRST_PAGE * ZEND_MM_PAGE_SIZE);
|
|
*heap = old_heap;
|
|
memset(heap->free_slot, 0, sizeof(heap->free_slot));
|
|
heap->main_chunk = p;
|
|
p->heap = &p->heap_slot;
|
|
p->next = p;
|
|
p->prev = p;
|
|
p->free_pages = ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE;
|
|
p->free_tail = ZEND_MM_FIRST_PAGE;
|
|
p->free_map[0] = (1L << ZEND_MM_FIRST_PAGE) - 1;
|
|
p->map[0] = ZEND_MM_LRUN(ZEND_MM_FIRST_PAGE);
|
|
heap->chunks_count = 1;
|
|
heap->peak_chunks_count = 1;
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size = ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->real_peak = ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
/**************/
|
|
/* PUBLIC API */
|
|
/**************/
|
|
|
|
ZEND_API void* ZEND_FASTCALL _zend_mm_alloc(zend_mm_heap *heap, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
return zend_mm_alloc_heap(heap, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API void ZEND_FASTCALL _zend_mm_free(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
zend_mm_free_heap(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
void* ZEND_FASTCALL _zend_mm_realloc(zend_mm_heap *heap, void *ptr, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
return zend_mm_realloc_heap(heap, ptr, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API size_t ZEND_FASTCALL _zend_mm_block_size(zend_mm_heap *heap, void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
return zend_mm_size(heap, ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
/**********************/
|
|
/* Allocation Manager */
|
|
/**********************/
|
|
|
|
typedef struct _zend_alloc_globals {
|
|
zend_mm_heap *mm_heap;
|
|
} zend_alloc_globals;
|
|
|
|
#ifdef ZTS
|
|
static int alloc_globals_id;
|
|
# define AG(v) ZEND_TSRMG(alloc_globals_id, zend_alloc_globals *, v)
|
|
#else
|
|
# define AG(v) (alloc_globals.v)
|
|
static zend_alloc_globals alloc_globals;
|
|
#endif
|
|
|
|
ZEND_API int is_zend_mm(void)
|
|
{
|
|
#if ZEND_MM_CUSTOM
|
|
return !AG(mm_heap)->use_custom_heap;
|
|
#else
|
|
return 1;
|
|
#endif
|
|
}
|
|
|
|
#if !ZEND_DEBUG && !defined(_WIN32)
|
|
#undef _emalloc
|
|
|
|
#if ZEND_MM_CUSTOM
|
|
# define ZEND_MM_CUSTOM_ALLOCATOR(size) do { \
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) { \
|
|
return AG(mm_heap)->_malloc(size); \
|
|
} \
|
|
} while (0)
|
|
# define ZEND_MM_CUSTOM_DEALLOCATOR(ptr) do { \
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) { \
|
|
AG(mm_heap)->_free(ptr); \
|
|
return; \
|
|
} \
|
|
} while (0)
|
|
#else
|
|
# define ZEND_MM_CUSTOM_ALLOCATOR(size)
|
|
# define ZEND_MM_CUSTOM_DEALLOCATOR(ptr)
|
|
#endif
|
|
|
|
# define _ZEND_BIN_ALLOCATOR(_num, _size, _elements, _pages, x, y) \
|
|
ZEND_API void* ZEND_FASTCALL _emalloc_ ## _size(void) { \
|
|
ZEND_MM_CUSTOM_ALLOCATOR(_size); \
|
|
return zend_mm_alloc_small(AG(mm_heap), _size, _num ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC); \
|
|
}
|
|
|
|
ZEND_MM_BINS_INFO(_ZEND_BIN_ALLOCATOR, x, y)
|
|
|
|
ZEND_API void* ZEND_FASTCALL _emalloc_large(size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
|
|
ZEND_MM_CUSTOM_ALLOCATOR(size);
|
|
return zend_mm_alloc_large(AG(mm_heap), size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API void* ZEND_FASTCALL _emalloc_huge(size_t size)
|
|
{
|
|
|
|
ZEND_MM_CUSTOM_ALLOCATOR(size);
|
|
return zend_mm_alloc_huge(AG(mm_heap), size);
|
|
}
|
|
|
|
#if ZEND_DEBUG
|
|
# define _ZEND_BIN_FREE(_num, _size, _elements, _pages, x, y) \
|
|
ZEND_API void ZEND_FASTCALL _efree_ ## _size(void *ptr) { \
|
|
ZEND_MM_CUSTOM_DEALLOCATOR(ptr); \
|
|
{ \
|
|
size_t page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE); \
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE); \
|
|
int page_num = page_offset / ZEND_MM_PAGE_SIZE; \
|
|
ZEND_MM_CHECK(chunk->heap == AG(mm_heap), "zend_mm_heap corrupted"); \
|
|
ZEND_ASSERT(chunk->map[page_num] & ZEND_MM_IS_SRUN); \
|
|
ZEND_ASSERT(ZEND_MM_SRUN_BIN_NUM(chunk->map[page_num]) == _num); \
|
|
zend_mm_free_small(AG(mm_heap), ptr, _num); \
|
|
} \
|
|
}
|
|
#else
|
|
# define _ZEND_BIN_FREE(_num, _size, _elements, _pages, x, y) \
|
|
ZEND_API void ZEND_FASTCALL _efree_ ## _size(void *ptr) { \
|
|
ZEND_MM_CUSTOM_DEALLOCATOR(ptr); \
|
|
{ \
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE); \
|
|
ZEND_MM_CHECK(chunk->heap == AG(mm_heap), "zend_mm_heap corrupted"); \
|
|
zend_mm_free_small(AG(mm_heap), ptr, _num); \
|
|
} \
|
|
}
|
|
#endif
|
|
|
|
ZEND_MM_BINS_INFO(_ZEND_BIN_FREE, x, y)
|
|
|
|
ZEND_API void ZEND_FASTCALL _efree_large(void *ptr, size_t size)
|
|
{
|
|
|
|
ZEND_MM_CUSTOM_DEALLOCATOR(ptr);
|
|
{
|
|
size_t page_offset = ZEND_MM_ALIGNED_OFFSET(ptr, ZEND_MM_CHUNK_SIZE);
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)ZEND_MM_ALIGNED_BASE(ptr, ZEND_MM_CHUNK_SIZE);
|
|
int page_num = page_offset / ZEND_MM_PAGE_SIZE;
|
|
int pages_count = ZEND_MM_ALIGNED_SIZE_EX(size, ZEND_MM_PAGE_SIZE) / ZEND_MM_PAGE_SIZE;
|
|
|
|
ZEND_MM_CHECK(chunk->heap == AG(mm_heap) && ZEND_MM_ALIGNED_OFFSET(page_offset, ZEND_MM_PAGE_SIZE) == 0, "zend_mm_heap corrupted");
|
|
ZEND_ASSERT(chunk->map[page_num] & ZEND_MM_IS_LRUN);
|
|
ZEND_ASSERT(ZEND_MM_LRUN_PAGES(chunk->map[page_num]) == pages_count);
|
|
zend_mm_free_large(AG(mm_heap), chunk, page_num, pages_count);
|
|
}
|
|
}
|
|
|
|
ZEND_API void ZEND_FASTCALL _efree_huge(void *ptr, size_t size)
|
|
{
|
|
|
|
ZEND_MM_CUSTOM_DEALLOCATOR(ptr);
|
|
// TODO: use size???
|
|
zend_mm_free_huge(AG(mm_heap), ptr);
|
|
}
|
|
#endif
|
|
|
|
ZEND_API void* ZEND_FASTCALL _emalloc(size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
|
|
#if ZEND_MM_CUSTOM
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) {
|
|
return AG(mm_heap)->_malloc(size);
|
|
}
|
|
#endif
|
|
return zend_mm_alloc_heap(AG(mm_heap), size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API void ZEND_FASTCALL _efree(void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
|
|
#if ZEND_MM_CUSTOM
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) {
|
|
AG(mm_heap)->_free(ptr);
|
|
return;
|
|
}
|
|
#endif
|
|
zend_mm_free_heap(AG(mm_heap), ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API void* ZEND_FASTCALL _erealloc(void *ptr, size_t size, int allow_failure ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) {
|
|
return AG(mm_heap)->_realloc(ptr, size);
|
|
}
|
|
return zend_mm_realloc_heap(AG(mm_heap), ptr, size ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
ZEND_API size_t ZEND_FASTCALL _zend_mem_block_size(void *ptr ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
if (UNEXPECTED(AG(mm_heap)->use_custom_heap)) {
|
|
return 0;
|
|
}
|
|
return zend_mm_size(AG(mm_heap), ptr ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
}
|
|
|
|
static zend_always_inline size_t safe_address(size_t nmemb, size_t size, size_t offset)
|
|
{
|
|
int overflow;
|
|
size_t ret = zend_safe_address(nmemb, size, offset, &overflow);
|
|
|
|
if (UNEXPECTED(overflow)) {
|
|
zend_error_noreturn(E_ERROR, "Possible integer overflow in memory allocation (%zu * %zu + %zu)", nmemb, size, offset);
|
|
return 0;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
|
|
ZEND_API void* ZEND_FASTCALL _safe_emalloc(size_t nmemb, size_t size, size_t offset ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
return emalloc_rel(safe_address(nmemb, size, offset));
|
|
}
|
|
|
|
ZEND_API void* ZEND_FASTCALL _safe_malloc(size_t nmemb, size_t size, size_t offset)
|
|
{
|
|
return pemalloc(safe_address(nmemb, size, offset), 1);
|
|
}
|
|
|
|
ZEND_API void* ZEND_FASTCALL _safe_erealloc(void *ptr, size_t nmemb, size_t size, size_t offset ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
return erealloc_rel(ptr, safe_address(nmemb, size, offset));
|
|
}
|
|
|
|
ZEND_API void* ZEND_FASTCALL _safe_realloc(void *ptr, size_t nmemb, size_t size, size_t offset)
|
|
{
|
|
return perealloc(ptr, safe_address(nmemb, size, offset), 1);
|
|
}
|
|
|
|
|
|
ZEND_API void* ZEND_FASTCALL _ecalloc(size_t nmemb, size_t size ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
void *p;
|
|
#ifdef ZEND_SIGNALS
|
|
#endif
|
|
HANDLE_BLOCK_INTERRUPTIONS();
|
|
|
|
p = _safe_emalloc(nmemb, size, 0 ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
if (UNEXPECTED(p == NULL)) {
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
memset(p, 0, size * nmemb);
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
|
|
ZEND_API char* ZEND_FASTCALL _estrdup(const char *s ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
size_t length;
|
|
char *p;
|
|
#ifdef ZEND_SIGNALS
|
|
#endif
|
|
|
|
HANDLE_BLOCK_INTERRUPTIONS();
|
|
|
|
length = strlen(s)+1;
|
|
p = (char *) _emalloc(length ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
if (UNEXPECTED(p == NULL)) {
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
memcpy(p, s, length);
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
|
|
ZEND_API char* ZEND_FASTCALL _estrndup(const char *s, size_t length ZEND_FILE_LINE_DC ZEND_FILE_LINE_ORIG_DC)
|
|
{
|
|
char *p;
|
|
#ifdef ZEND_SIGNALS
|
|
#endif
|
|
|
|
HANDLE_BLOCK_INTERRUPTIONS();
|
|
|
|
p = (char *) _emalloc(length+1 ZEND_FILE_LINE_RELAY_CC ZEND_FILE_LINE_ORIG_RELAY_CC);
|
|
if (UNEXPECTED(p == NULL)) {
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
memcpy(p, s, length);
|
|
p[length] = 0;
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
|
|
|
|
ZEND_API char* ZEND_FASTCALL zend_strndup(const char *s, size_t length)
|
|
{
|
|
char *p;
|
|
#ifdef ZEND_SIGNALS
|
|
#endif
|
|
|
|
HANDLE_BLOCK_INTERRUPTIONS();
|
|
|
|
p = (char *) malloc(length+1);
|
|
if (UNEXPECTED(p == NULL)) {
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
if (length) {
|
|
memcpy(p, s, length);
|
|
}
|
|
p[length] = 0;
|
|
HANDLE_UNBLOCK_INTERRUPTIONS();
|
|
return p;
|
|
}
|
|
|
|
|
|
ZEND_API int zend_set_memory_limit(size_t memory_limit)
|
|
{
|
|
#if ZEND_MM_LIMIT
|
|
AG(mm_heap)->limit = (memory_limit >= ZEND_MM_CHUNK_SIZE) ? memory_limit : ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
return SUCCESS;
|
|
}
|
|
|
|
ZEND_API size_t zend_memory_usage(int real_usage)
|
|
{
|
|
#if ZEND_MM_STAT
|
|
if (real_usage) {
|
|
return AG(mm_heap)->real_size;
|
|
} else {
|
|
size_t usage = AG(mm_heap)->size;
|
|
return usage;
|
|
}
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
ZEND_API size_t zend_memory_peak_usage(int real_usage)
|
|
{
|
|
#if ZEND_MM_STAT
|
|
if (real_usage) {
|
|
return AG(mm_heap)->real_peak;
|
|
} else {
|
|
return AG(mm_heap)->peak;
|
|
}
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
ZEND_API void shutdown_memory_manager(int silent, int full_shutdown)
|
|
{
|
|
zend_mm_shutdown(AG(mm_heap), full_shutdown, silent);
|
|
}
|
|
|
|
static void alloc_globals_ctor(zend_alloc_globals *alloc_globals)
|
|
{
|
|
#if ZEND_MM_CUSTOM
|
|
char *tmp = getenv("USE_ZEND_ALLOC");
|
|
|
|
if (tmp && !zend_atoi(tmp, 0)) {
|
|
alloc_globals->mm_heap = malloc(sizeof(zend_mm_heap));
|
|
memset(alloc_globals->mm_heap, 0, sizeof(zend_mm_heap));
|
|
alloc_globals->mm_heap->use_custom_heap = 1;
|
|
alloc_globals->mm_heap->_malloc = malloc;
|
|
alloc_globals->mm_heap->_free = free;
|
|
alloc_globals->mm_heap->_realloc = realloc;
|
|
return;
|
|
}
|
|
#endif
|
|
ZEND_TSRMLS_CACHE_UPDATE;
|
|
alloc_globals->mm_heap = zend_mm_init();
|
|
}
|
|
|
|
#ifdef ZTS
|
|
static void alloc_globals_dtor(zend_alloc_globals *alloc_globals)
|
|
{
|
|
zend_mm_shutdown(alloc_globals->mm_heap, 1, 1);
|
|
}
|
|
#endif
|
|
|
|
ZEND_API void start_memory_manager(void)
|
|
{
|
|
#ifdef ZTS
|
|
ts_allocate_id(&alloc_globals_id, sizeof(zend_alloc_globals), (ts_allocate_ctor) alloc_globals_ctor, (ts_allocate_dtor) alloc_globals_dtor);
|
|
#else
|
|
alloc_globals_ctor(&alloc_globals);
|
|
#endif
|
|
#ifndef _WIN32
|
|
# if defined(_SC_PAGESIZE)
|
|
REAL_PAGE_SIZE = sysconf(_SC_PAGESIZE);
|
|
# elif defined(_SC_PAGE_SIZE)
|
|
REAL_PAGE_SIZE = sysconf(_SC_PAGE_SIZE);
|
|
# endif
|
|
#endif
|
|
}
|
|
|
|
ZEND_API zend_mm_heap *zend_mm_set_heap(zend_mm_heap *new_heap)
|
|
{
|
|
zend_mm_heap *old_heap;
|
|
|
|
old_heap = AG(mm_heap);
|
|
AG(mm_heap) = (zend_mm_heap*)new_heap;
|
|
return (zend_mm_heap*)old_heap;
|
|
}
|
|
|
|
ZEND_API void zend_mm_set_custom_handlers(zend_mm_heap *heap,
|
|
void* (*_malloc)(size_t),
|
|
void (*_free)(void*),
|
|
void* (*_realloc)(void*, size_t))
|
|
{
|
|
#if ZEND_MM_CUSTOM
|
|
zend_mm_heap *_heap = (zend_mm_heap*)heap;
|
|
|
|
_heap->use_custom_heap = 1;
|
|
_heap->_malloc = _malloc;
|
|
_heap->_free = _free;
|
|
_heap->_realloc = _realloc;
|
|
#endif
|
|
}
|
|
|
|
ZEND_API void zend_mm_get_custom_handlers(zend_mm_heap *heap,
|
|
void* (**_malloc)(size_t),
|
|
void (**_free)(void*),
|
|
void* (**_realloc)(void*, size_t))
|
|
{
|
|
#if ZEND_MM_CUSTOM
|
|
zend_mm_heap *_heap = (zend_mm_heap*)heap;
|
|
|
|
if (heap->use_custom_heap) {
|
|
*_malloc = _heap->_malloc;
|
|
*_free = _heap->_free;
|
|
*_realloc = _heap->_realloc;
|
|
} else {
|
|
*_malloc = NULL;
|
|
*_free = NULL;
|
|
*_realloc = NULL;
|
|
}
|
|
#else
|
|
*_malloc = NULL;
|
|
*_free = NULL;
|
|
*_realloc = NULL;
|
|
#endif
|
|
}
|
|
|
|
ZEND_API zend_mm_storage *zend_mm_get_storage(zend_mm_heap *heap)
|
|
{
|
|
#if ZEND_MM_CUSTOM
|
|
return heap->storage;
|
|
#else
|
|
return NULL
|
|
#endif
|
|
}
|
|
|
|
ZEND_API zend_mm_heap *zend_mm_startup(void)
|
|
{
|
|
return zend_mm_init();
|
|
}
|
|
|
|
ZEND_API zend_mm_heap *zend_mm_startup_ex(zend_mm_storage *storage)
|
|
{
|
|
#if ZEND_MM_STORAGE
|
|
zend_mm_chunk *chunk = (zend_mm_chunk*)storage->chunk_alloc(storage, ZEND_MM_CHUNK_SIZE, ZEND_MM_CHUNK_SIZE);
|
|
zend_mm_heap *heap;
|
|
|
|
if (UNEXPECTED(chunk == NULL)) {
|
|
#if ZEND_MM_ERROR
|
|
#ifdef _WIN32
|
|
stderr_last_error("Can't initialize heap");
|
|
#else
|
|
fprintf(stderr, "\nCan't initialize heap: [%d] %s\n", errno, strerror(errno));
|
|
#endif
|
|
#endif
|
|
return NULL;
|
|
}
|
|
heap = &chunk->heap_slot;
|
|
chunk->heap = heap;
|
|
chunk->next = chunk;
|
|
chunk->prev = chunk;
|
|
chunk->free_pages = ZEND_MM_PAGES - ZEND_MM_FIRST_PAGE;
|
|
chunk->free_tail = ZEND_MM_FIRST_PAGE;
|
|
chunk->num = 0;
|
|
chunk->free_map[0] = (Z_L(1) << ZEND_MM_FIRST_PAGE) - 1;
|
|
chunk->map[0] = ZEND_MM_LRUN(ZEND_MM_FIRST_PAGE);
|
|
heap->main_chunk = chunk;
|
|
heap->cached_chunks = NULL;
|
|
heap->chunks_count = 1;
|
|
heap->peak_chunks_count = 1;
|
|
heap->cached_chunks_count = 0;
|
|
heap->avg_chunks_count = 1.0;
|
|
#if ZEND_MM_STAT || ZEND_MM_LIMIT
|
|
heap->real_size = ZEND_MM_CHUNK_SIZE;
|
|
#endif
|
|
#if ZEND_MM_STAT
|
|
heap->real_peak = ZEND_MM_CHUNK_SIZE;
|
|
heap->size = 0;
|
|
heap->peak = 0;
|
|
#endif
|
|
#if ZEND_MM_LIMIT
|
|
heap->limit = (Z_L(-1) >> Z_L(1));
|
|
heap->overflow = 0;
|
|
#endif
|
|
#if ZEND_MM_CUSTOM
|
|
heap->use_custom_heap = 0;
|
|
#endif
|
|
heap->storage = storage;
|
|
heap->huge_list = NULL;
|
|
return heap;
|
|
#else
|
|
return NULL;
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Local variables:
|
|
* tab-width: 4
|
|
* c-basic-offset: 4
|
|
* indent-tabs-mode: t
|
|
* End:
|
|
*/
|