| /* |
| * This file is from CPython's Modules/_ctypes/malloc_closure.c |
| * and has received some edits. |
| */ |
| |
| #include <ffi.h> |
| #ifdef MS_WIN32 |
| #include <windows.h> |
| #else |
| #include <sys/mman.h> |
| #include <unistd.h> |
| # if !defined(MAP_ANONYMOUS) && defined(MAP_ANON) |
| # define MAP_ANONYMOUS MAP_ANON |
| # endif |
| #endif |
| |
| /* On PaX enable kernels that have MPROTECT enable we can't use PROT_EXEC. |
| |
| This is, apparently, an undocumented change to ffi_prep_closure(): |
| depending on the Linux kernel we're running on, we must give it a |
| mmap that is either PROT_READ|PROT_WRITE|PROT_EXEC or only |
| PROT_READ|PROT_WRITE. In the latter case, just trying to obtain a |
| mmap with PROT_READ|PROT_WRITE|PROT_EXEC would kill our process(!), |
| but in that situation libffi is fine with only PROT_READ|PROT_WRITE. |
| There is nothing in the libffi API to know that, though, so we have |
| to guess by parsing /proc/self/status. "Meh." |
| */ |
| #ifdef __linux__ |
| #include <stdlib.h> |
| |
| static int emutramp_enabled = -1; |
| |
| static int |
| emutramp_enabled_check (void) |
| { |
| char *buf = NULL; |
| size_t len = 0; |
| FILE *f; |
| int ret; |
| f = fopen ("/proc/self/status", "r"); |
| if (f == NULL) |
| return 0; |
| ret = 0; |
| |
| while (getline (&buf, &len, f) != -1) |
| if (!strncmp (buf, "PaX:", 4)) |
| { |
| char emutramp; |
| if (sscanf (buf, "%*s %*c%c", &emutramp) == 1) |
| ret = (emutramp == 'E'); |
| break; |
| } |
| free (buf); |
| fclose (f); |
| return ret; |
| } |
| |
| #define is_emutramp_enabled() (emutramp_enabled >= 0 ? emutramp_enabled \ |
| : (emutramp_enabled = emutramp_enabled_check ())) |
| #else |
| #define is_emutramp_enabled() 0 |
| #endif |
| |
| |
| /* 'allocate_num_pages' is dynamically adjusted starting from one |
| page. It grows by a factor of PAGE_ALLOCATION_GROWTH_RATE. This is |
| meant to handle both the common case of not needing a lot of pages, |
| and the rare case of needing many of them. Systems in general have a |
| limit of how many mmap'd blocks can be open. |
| */ |
| |
| #define PAGE_ALLOCATION_GROWTH_RATE 1.3 |
| |
| static Py_ssize_t allocate_num_pages = 0; |
| |
| /* #define MALLOC_CLOSURE_DEBUG */ /* enable for some debugging output */ |
| |
| /******************************************************************/ |
| |
| union mmaped_block { |
| ffi_closure closure; |
| union mmaped_block *next; |
| }; |
| |
| static union mmaped_block *free_list = 0; |
| static Py_ssize_t _pagesize = 0; |
| |
| static void more_core(void) |
| { |
| union mmaped_block *item; |
| Py_ssize_t count, i; |
| |
| /* determine the pagesize */ |
| #ifdef MS_WIN32 |
| if (!_pagesize) { |
| SYSTEM_INFO systeminfo; |
| GetSystemInfo(&systeminfo); |
| _pagesize = systeminfo.dwPageSize; |
| } |
| #else |
| if (!_pagesize) { |
| #ifdef _SC_PAGESIZE |
| _pagesize = sysconf(_SC_PAGESIZE); |
| #else |
| _pagesize = getpagesize(); |
| #endif |
| } |
| #endif |
| if (_pagesize <= 0) |
| _pagesize = 4096; |
| |
| /* bump 'allocate_num_pages' */ |
| allocate_num_pages = 1 + ( |
| (Py_ssize_t)(allocate_num_pages * PAGE_ALLOCATION_GROWTH_RATE)); |
| |
| /* calculate the number of mmaped_blocks to allocate */ |
| count = (allocate_num_pages * _pagesize) / sizeof(union mmaped_block); |
| |
| /* allocate a memory block */ |
| #ifdef MS_WIN32 |
| item = (union mmaped_block *)VirtualAlloc(NULL, |
| count * sizeof(union mmaped_block), |
| MEM_COMMIT, |
| PAGE_EXECUTE_READWRITE); |
| if (item == NULL) |
| return; |
| #else |
| { |
| int prot = PROT_READ | PROT_WRITE | PROT_EXEC; |
| if (is_emutramp_enabled ()) |
| prot &= ~PROT_EXEC; |
| item = (union mmaped_block *)mmap(NULL, |
| allocate_num_pages * _pagesize, |
| prot, |
| MAP_PRIVATE | MAP_ANONYMOUS, |
| -1, |
| 0); |
| if (item == (void *)MAP_FAILED) |
| return; |
| } |
| #endif |
| |
| #ifdef MALLOC_CLOSURE_DEBUG |
| printf("block at %p allocated (%ld bytes), %ld mmaped_blocks\n", |
| item, (long)(allocate_num_pages * _pagesize), (long)count); |
| #endif |
| /* put them into the free list */ |
| for (i = 0; i < count; ++i) { |
| item->next = free_list; |
| free_list = item; |
| ++item; |
| } |
| } |
| |
| /******************************************************************/ |
| |
| /* put the item back into the free list */ |
| static void cffi_closure_free(ffi_closure *p) |
| { |
| union mmaped_block *item = (union mmaped_block *)p; |
| item->next = free_list; |
| free_list = item; |
| } |
| |
| /* return one item from the free list, allocating more if needed */ |
| static ffi_closure *cffi_closure_alloc(void) |
| { |
| union mmaped_block *item; |
| if (!free_list) |
| more_core(); |
| if (!free_list) |
| return NULL; |
| item = free_list; |
| free_list = item->next; |
| return &item->closure; |
| } |