v/vlib/builtin/prealloc.c.v

230 lines
6.5 KiB
V

@[has_globals]
module builtin
// With -prealloc, V calls libc's malloc to get chunks, each at least 16MB
// in size, as needed. Once a chunk is available, all malloc() calls within
// V code, that can fit inside the chunk, will use it instead, each bumping a
// pointer, till the chunk is filled. Once a chunk is filled, a new chunk will
// be allocated by calling libc's malloc, and the process continues.
// Each new chunk has a pointer to the old one, and at the end of the program,
// the entire linked list of chunks is freed.
// The goal of all this is to amortize the cost of calling libc's malloc,
// trading higher memory usage for a compiler (or any single threaded batch
// mode program), for a ~8-10% speed increase.
// Note: `-prealloc` is NOT safe to be used for multithreaded programs!
// size of the preallocated chunk
const prealloc_block_size = 16 * 1024 * 1024
__global g_memory_block &VMemoryBlock
@[heap]
struct VMemoryBlock {
mut:
current &u8 = 0 // 8
stop &u8 = 0 // 8
start &u8 = 0 // 8
previous &VMemoryBlock = 0 // 8
next &VMemoryBlock = 0 // 8
id int // 4
mallocs int // 4
}
fn vmemory_abort_on_nil(p voidptr, bytes isize) {
if unsafe { p == 0 } {
C.fprintf(C.stderr, c'could not allocate %td bytes\n', bytes)
exit(1)
}
}
@[unsafe]
fn vmemory_block_new(prev &VMemoryBlock, at_least isize, align isize) &VMemoryBlock {
vmem_block_size := sizeof(VMemoryBlock)
mut v := unsafe { &VMemoryBlock(C.calloc(1, vmem_block_size)) }
vmemory_abort_on_nil(v, vmem_block_size)
if unsafe { prev != 0 } {
v.id = prev.id + 1
}
v.previous = prev
if unsafe { prev != 0 } {
prev.next = v
}
base_block_size := if at_least < isize(prealloc_block_size) {
isize(prealloc_block_size)
} else {
at_least
}
block_size := if align > 0 {
if base_block_size % align == 0 {
base_block_size
} else {
base_block_size + align - (base_block_size % align)
}
} else {
base_block_size
}
$if prealloc_trace_malloc ? {
C.fprintf(C.stderr, c'vmemory_block_new id: %d, block_size: %lld, at_least: %lld, align: %lld\n',
v.id, block_size, at_least, align)
}
fixed_align := if align <= 1 { 1 } else { align }
$if windows {
v.start = unsafe { C._aligned_malloc(block_size, fixed_align) }
} $else {
if fixed_align == 1 {
v.start = unsafe { C.malloc(block_size) }
} else {
v.start = unsafe { C.aligned_alloc(fixed_align, block_size) }
}
}
vmemory_abort_on_nil(v.start, block_size)
$if prealloc_memset ? {
unsafe { C.memset(v.start, int($d('prealloc_memset_value', 0)), block_size) }
}
v.stop = unsafe { &u8(i64(v.start) + block_size) }
v.current = v.start
return v
}
@[unsafe]
fn vmemory_block_malloc(n isize, align isize) &u8 {
$if prealloc_trace_malloc ? {
C.fprintf(C.stderr, c'vmemory_block_malloc g_memory_block.id: %d, n: %lld align: %d\n',
g_memory_block.id, n, align)
}
unsafe {
remaining := i64(g_memory_block.stop) - i64(g_memory_block.current)
if _unlikely_(remaining < n) {
g_memory_block = vmemory_block_new(g_memory_block, n, align)
}
res := &u8(g_memory_block.current)
g_memory_block.current += n
$if prealloc_stats ? {
g_memory_block.mallocs++
}
return res
}
}
/////////////////////////////////////////////////
@[unsafe]
fn prealloc_vinit() {
$if prealloc_trace_vinit ? {
C.fprintf(C.stderr, c'prealloc_vinit started\n')
}
unsafe {
g_memory_block = vmemory_block_new(nil, isize(prealloc_block_size), 0)
at_exit(prealloc_vcleanup) or {}
}
}
@[unsafe]
fn prealloc_vcleanup() {
$if prealloc_trace_vcleanup ? {
C.fprintf(C.stderr, c'prealloc_vcleanup started\n')
}
$if prealloc_stats ? {
// Note: we do 2 loops here, because string interpolation
// in the first loop may still use g_memory_block
// The second loop however should *not* allocate at all.
mut nr_mallocs := i64(0)
mut total_used := i64(0)
mut mb := g_memory_block
for unsafe { mb != 0 } {
nr_mallocs += mb.mallocs
used := i64(mb.current) - i64(mb.start)
total_used += used
remaining := i64(mb.stop) - i64(mb.current)
size := i64(mb.stop) - i64(mb.start)
C.fprintf(C.stderr, c'> freeing mb: %16p, mb.id: %3d | size: %10lld | rem: %10lld | start: %16p | current: %16p | used: %10lld bytes | mallocs: %6d\n',
mb, mb.id, size, remaining, mb.start, mb.current, used, mb.mallocs)
mb = mb.previous
}
C.fprintf(C.stderr, c'> nr_mallocs: %lld, total_used: %lld bytes\n', nr_mallocs,
total_used)
}
$if prealloc_dump ? {
C.fprintf(C.stderr, c'prealloc_vcleanup dumping memory contents ...\n')
mut start := g_memory_block
unsafe {
for start.previous != 0 {
start = start.previous
}
C.fprintf(C.stderr, c'prealloc_vcleanup start: %p\n', start)
C.fprintf(C.stderr, c'prealloc_vcleanup start.id: %d\n', start.id)
C.fprintf(C.stderr, c'prealloc_vcleanup start.next: %p\n', start.next)
mut total_used := u64(0)
path := $d('memdumpfile', 'memdump.bin')
C.fprintf(C.stderr, c'prealloc_vcleanup dumping process memory to path: %s\n',
path.str)
stream := C.fopen(path.str, c'wb')
mut mb := start
for {
used := u64(mb.current) - u64(mb.start)
total_used += used
C.fprintf(C.stderr, c'prealloc_vcleanup dumping mb: %p, mb.id: %d, used: %10lld bytes\n',
mb, mb.id, used)
mut ptr := mb.start
mut remaining_bytes := isize(used)
mut x := isize(0)
for remaining_bytes > 0 {
x = isize(C.fwrite(ptr, 1, remaining_bytes, stream))
ptr += x
remaining_bytes -= x
}
if mb.next == 0 {
break
}
mb = mb.next
}
C.fclose(stream)
C.fprintf(C.stderr, c'prealloc_vcleanup total dump size in bytes: %lld\n',
total_used)
}
}
unsafe {
for g_memory_block != 0 {
$if windows {
// Warning! On windows, we always use _aligned_free to free memory.
C._aligned_free(g_memory_block.start)
} $else {
C.free(g_memory_block.start)
}
tmp := g_memory_block
g_memory_block = g_memory_block.previous
// free the link node
C.free(tmp)
}
}
}
@[unsafe]
fn prealloc_malloc(n isize) &u8 {
return unsafe { vmemory_block_malloc(n, 0) }
}
@[unsafe]
fn prealloc_realloc(old_data &u8, old_size isize, new_size isize) &u8 {
new_ptr := unsafe { vmemory_block_malloc(new_size, 0) }
min_size := if old_size < new_size { old_size } else { new_size }
unsafe { C.memcpy(new_ptr, old_data, min_size) }
return new_ptr
}
@[unsafe]
fn prealloc_calloc(n isize) &u8 {
new_ptr := unsafe { vmemory_block_malloc(n, 0) }
unsafe { C.memset(new_ptr, 0, n) }
return new_ptr
}
@[unsafe]
fn prealloc_malloc_align(n isize, align isize) &u8 {
return unsafe { vmemory_block_malloc(n, align) }
}