/** * This module contains all functions related to an object's lifetime: * allocation, resizing, deallocation, and finalization. * * Copyright: Copyright Digital Mars 2000 - 2012. * License: Distributed under the * $(LINK2 http://www.boost.org/LICENSE_1_0.txt, Boost Software License 1.0). * (See accompanying file LICENSE) * Authors: Walter Bright, Sean Kelly, Steven Schveighoffer * Source: $(DRUNTIMESRC rt/_lifetime.d) */ module rt.lifetime; import core.attribute : weak; import core.internal.array.utils : __arrayStart, __arrayClearPad; import core.memory; debug(PRINTF) import core.stdc.stdio; static import rt.tlsgc; alias BlkInfo = GC.BlkInfo; alias BlkAttr = GC.BlkAttr; private { alias bool function(Object) CollectHandler; __gshared CollectHandler collectHandler = null; extern (C) void _d_monitordelete(Object h, bool det); enum : size_t { PAGESIZE = 4096, BIGLENGTHMASK = ~(PAGESIZE - 1), SMALLPAD = 1, MEDPAD = ushort.sizeof, LARGEPREFIX = 16, // 16 bytes padding at the front of the array LARGEPAD = LARGEPREFIX + 1, MAXSMALLSIZE = 256-SMALLPAD, MAXMEDSIZE = (PAGESIZE / 2) - MEDPAD } } // Now-removed symbol, kept around for ABI // Some programs are dynamically linked, so best to err on the side of keeping symbols around for a while (especially extern(C) ones) // https://github.com/dlang/druntime/pull/3361 deprecated extern (C) void lifetime_init() { } /** Allocate memory using the garbage collector DMD uses this to allocate closures: --- void f(byte[24] x) { return () => x; // `x` is on stack, must be moved to heap to keep it alive } --- Params: sz = number of bytes to allocate Returns: pointer to `sz` bytes of free, uninitialized memory, managed by the GC. */ extern (C) void* _d_allocmemory(size_t sz) @weak { return GC.malloc(sz); } /** Create a new class instance. Allocates memory and sets fields to their initial value, but does not call a constructor. --- new Object() // _d_newclass(typeid(Object)) --- Params: ci = `TypeInfo_Class` object, to provide instance size and initial bytes to copy Returns: newly created object */ extern (C) Object _d_newclass(const ClassInfo ci) @weak { import core.stdc.stdlib; import core.exception : onOutOfMemoryError; void* p; auto init = ci.initializer; debug(PRINTF) printf("_d_newclass(ci = %p, %s)\n", ci, cast(char *)ci.name); if (ci.m_flags & TypeInfo_Class.ClassFlags.isCOMclass) { /* COM objects are not garbage collected, they are reference counted * using AddRef() and Release(). They get free'd by C's free() * function called by Release() when Release()'s reference count goes * to zero. */ p = malloc(init.length); if (!p) onOutOfMemoryError(); } else { // TODO: should this be + 1 to avoid having pointers to the next block? BlkAttr attr = BlkAttr.NONE; // extern(C++) classes don't have a classinfo pointer in their vtable so the GC can't finalize them if (ci.m_flags & TypeInfo_Class.ClassFlags.hasDtor && !(ci.m_flags & TypeInfo_Class.ClassFlags.isCPPclass)) attr |= BlkAttr.FINALIZE; if (ci.m_flags & TypeInfo_Class.ClassFlags.noPointers) attr |= BlkAttr.NO_SCAN; p = GC.malloc(init.length, attr, ci); debug(PRINTF) printf(" p = %p\n", p); } debug(PRINTF) { printf("p = %p\n", p); printf("ci = %p, ci.init.ptr = %p, len = %llu\n", ci, init.ptr, cast(ulong)init.length); printf("vptr = %p\n", *cast(void**) init); printf("vtbl[0] = %p\n", (*cast(void***) init)[0]); printf("vtbl[1] = %p\n", (*cast(void***) init)[1]); printf("init[0] = %x\n", (cast(uint*) init)[0]); printf("init[1] = %x\n", (cast(uint*) init)[1]); printf("init[2] = %x\n", (cast(uint*) init)[2]); printf("init[3] = %x\n", (cast(uint*) init)[3]); printf("init[4] = %x\n", (cast(uint*) init)[4]); } // initialize it p[0 .. init.length] = init[]; debug(PRINTF) printf("initialization done\n"); return cast(Object) p; } /** * */ extern (C) void _d_delinterface(void** p) { if (*p) { Interface* pi = **cast(Interface ***)*p; Object o = cast(Object)(*p - pi.offset); _d_delclass(&o); *p = null; } } // used for deletion private extern (D) alias void function (Object) fp_t; /** * */ extern (C) void _d_delclass(Object* p) @weak { if (*p) { debug(PRINTF) printf("_d_delclass(%p)\n", *p); ClassInfo **pc = cast(ClassInfo **)*p; if (*pc) { ClassInfo c = **pc; rt_finalize(cast(void*) *p); if (c.deallocator) { fp_t fp = cast(fp_t)c.deallocator; (*fp)(*p); // call deallocator *p = null; return; } } else { rt_finalize(cast(void*) *p); } GC.free(cast(void*) *p); *p = null; } } // strip const/immutable/shared/inout from type info inout(TypeInfo) unqualify(return scope inout(TypeInfo) cti) pure nothrow @nogc { TypeInfo ti = cast() cti; while (ti) { // avoid dynamic type casts auto tti = typeid(ti); if (tti is typeid(TypeInfo_Const)) ti = (cast(TypeInfo_Const)cast(void*)ti).base; else if (tti is typeid(TypeInfo_Invariant)) ti = (cast(TypeInfo_Invariant)cast(void*)ti).base; else if (tti is typeid(TypeInfo_Shared)) ti = (cast(TypeInfo_Shared)cast(void*)ti).base; else if (tti is typeid(TypeInfo_Inout)) ti = (cast(TypeInfo_Inout)cast(void*)ti).base; else break; } return ti; } // size used to store the TypeInfo at the end of an allocation for structs that have a destructor size_t structTypeInfoSize(const TypeInfo ti) pure nothrow @nogc { if (ti && typeid(ti) is typeid(TypeInfo_Struct)) // avoid a complete dynamic type cast { auto sti = cast(TypeInfo_Struct)cast(void*)ti; if (sti.xdtor) return size_t.sizeof; } return 0; } /** dummy class used to lock for shared array appending */ private class ArrayAllocLengthLock {} /** Set the allocated length of the array block. This is called any time an array is appended to or its length is set. The allocated block looks like this for blocks < PAGESIZE: |elem0|elem1|elem2|...|elemN-1|emptyspace|N*elemsize| The size of the allocated length at the end depends on the block size: a block of 16 to 256 bytes has an 8-bit length. a block with 512 to pagesize/2 bytes has a 16-bit length. For blocks >= pagesize, the length is a size_t and is at the beginning of the block. The reason we have to do this is because the block can extend into more pages, so we cannot trust the block length if it sits at the end of the block, because it might have just been extended. If we can prove in the future that the block is unshared, we may be able to change this, but I'm not sure it's important. In order to do put the length at the front, we have to provide 16 bytes buffer space in case the block has to be aligned properly. In x86, certain SSE instructions will only work if the data is 16-byte aligned. In addition, we need the sentinel byte to prevent accidental pointers to the next block. Because of the extra overhead, we only do this for page size and above, where the overhead is minimal compared to the block size. So for those blocks, it looks like: |N*elemsize|padding|elem0|elem1|...|elemN-1|emptyspace|sentinelbyte| where elem0 starts 16 bytes after the first byte. */ bool __setArrayAllocLength(ref BlkInfo info, size_t newlength, bool isshared, const TypeInfo tinext, size_t oldlength = ~0) pure nothrow { import core.atomic; size_t typeInfoSize = structTypeInfoSize(tinext); if (info.size <= 256) { import core.checkedint; bool overflow; auto newlength_padded = addu(newlength, addu(SMALLPAD, typeInfoSize, overflow), overflow); if (newlength_padded > info.size || overflow) // new size does not fit inside block return false; auto length = cast(ubyte *)(info.base + info.size - typeInfoSize - SMALLPAD); if (oldlength != ~0) { if (isshared) { return cas(cast(shared)length, cast(ubyte)oldlength, cast(ubyte)newlength); } else { if (*length == cast(ubyte)oldlength) *length = cast(ubyte)newlength; else return false; } } else { // setting the initial length, no cas needed *length = cast(ubyte)newlength; } if (typeInfoSize) { auto typeInfo = cast(TypeInfo*)(info.base + info.size - size_t.sizeof); *typeInfo = cast() tinext; } } else if (info.size < PAGESIZE) { if (newlength + MEDPAD + typeInfoSize > info.size) // new size does not fit inside block return false; auto length = cast(ushort *)(info.base + info.size - typeInfoSize - MEDPAD); if (oldlength != ~0) { if (isshared) { return cas(cast(shared)length, cast(ushort)oldlength, cast(ushort)newlength); } else { if (*length == oldlength) *length = cast(ushort)newlength; else return false; } } else { // setting the initial length, no cas needed *length = cast(ushort)newlength; } if (typeInfoSize) { auto typeInfo = cast(TypeInfo*)(info.base + info.size - size_t.sizeof); *typeInfo = cast() tinext; } } else { if (newlength + LARGEPAD > info.size) // new size does not fit inside block return false; auto length = cast(size_t *)(info.base); if (oldlength != ~0) { if (isshared) { return cas(cast(shared)length, cast(size_t)oldlength, cast(size_t)newlength); } else { if (*length == oldlength) *length = newlength; else return false; } } else { // setting the initial length, no cas needed *length = newlength; } if (typeInfoSize) { auto typeInfo = cast(TypeInfo*)(info.base + size_t.sizeof); *typeInfo = cast()tinext; } } return true; // resize succeeded } /** get the allocation size of the array for the given block (without padding or type info) */ private size_t __arrayAllocLength(ref BlkInfo info, const TypeInfo tinext) pure nothrow { if (info.size <= 256) return *cast(ubyte *)(info.base + info.size - structTypeInfoSize(tinext) - SMALLPAD); if (info.size < PAGESIZE) return *cast(ushort *)(info.base + info.size - structTypeInfoSize(tinext) - MEDPAD); return *cast(size_t *)(info.base); } /** get the padding required to allocate size bytes. Note that the padding is NOT included in the passed in size. Therefore, do NOT call this function with the size of an allocated block. */ private size_t __arrayPad(size_t size, const TypeInfo tinext) nothrow pure @trusted { return size > MAXMEDSIZE ? LARGEPAD : ((size > MAXSMALLSIZE ? MEDPAD : SMALLPAD) + structTypeInfoSize(tinext)); } /** allocate an array memory block by applying the proper padding and assigning block attributes if not inherited from the existing block */ private BlkInfo __arrayAlloc(size_t arrsize, const scope TypeInfo ti, const TypeInfo tinext) nothrow pure { import core.checkedint; size_t typeInfoSize = structTypeInfoSize(tinext); size_t padsize = arrsize > MAXMEDSIZE ? LARGEPAD : ((arrsize > MAXSMALLSIZE ? MEDPAD : SMALLPAD) + typeInfoSize); bool overflow; auto padded_size = addu(arrsize, padsize, overflow); if (overflow) return BlkInfo(); uint attr = (!(tinext.flags & 1) ? BlkAttr.NO_SCAN : 0) | BlkAttr.APPENDABLE; if (typeInfoSize) attr |= BlkAttr.STRUCTFINAL | BlkAttr.FINALIZE; auto bi = GC.qalloc(padded_size, attr, tinext); __arrayClearPad(bi, arrsize, padsize); return bi; } private BlkInfo __arrayAlloc(size_t arrsize, ref BlkInfo info, const scope TypeInfo ti, const TypeInfo tinext) { import core.checkedint; if (!info.base) return __arrayAlloc(arrsize, ti, tinext); immutable padsize = __arrayPad(arrsize, tinext); bool overflow; auto padded_size = addu(arrsize, padsize, overflow); if (overflow) { return BlkInfo(); } auto bi = GC.qalloc(padded_size, info.attr, tinext); __arrayClearPad(bi, arrsize, padsize); return bi; } /** cache for the lookup of the block info */ private enum N_CACHE_BLOCKS=8; // note this is TLS, so no need to sync. BlkInfo *__blkcache_storage; static if (N_CACHE_BLOCKS==1) { version=single_cache; } else { //version=simple_cache; // uncomment to test simple cache strategy //version=random_cache; // uncomment to test random cache strategy // ensure N_CACHE_BLOCKS is power of 2. static assert(!((N_CACHE_BLOCKS - 1) & N_CACHE_BLOCKS)); version (random_cache) { int __nextRndNum = 0; } int __nextBlkIdx; } @property BlkInfo *__blkcache() nothrow { if (!__blkcache_storage) { import core.stdc.stdlib; import core.stdc.string; // allocate the block cache for the first time immutable size = BlkInfo.sizeof * N_CACHE_BLOCKS; __blkcache_storage = cast(BlkInfo *)malloc(size); memset(__blkcache_storage, 0, size); } return __blkcache_storage; } // called when thread is exiting. static ~this() { // free the blkcache if (__blkcache_storage) { import core.stdc.stdlib; free(__blkcache_storage); __blkcache_storage = null; } } // we expect this to be called with the lock in place void processGCMarks(BlkInfo* cache, scope rt.tlsgc.IsMarkedDg isMarked) nothrow { // called after the mark routine to eliminate block cache data when it // might be ready to sweep debug(PRINTF) printf("processing GC Marks, %x\n", cache); if (cache) { debug(PRINTF) foreach (i; 0 .. N_CACHE_BLOCKS) { printf("cache entry %d has base ptr %x\tsize %d\tflags %x\n", i, cache[i].base, cache[i].size, cache[i].attr); } auto cache_end = cache + N_CACHE_BLOCKS; for (;cache < cache_end; ++cache) { if (cache.base != null && !isMarked(cache.base)) { debug(PRINTF) printf("clearing cache entry at %x\n", cache.base); cache.base = null; // clear that data. } } } } unittest { // Bugzilla 10701 - segfault in GC ubyte[] result; result.length = 4096; GC.free(result.ptr); GC.collect(); } /** Get the cached block info of an interior pointer. Returns null if the interior pointer's block is not cached. NOTE: The base ptr in this struct can be cleared asynchronously by the GC, so any use of the returned BlkInfo should copy it and then check the base ptr of the copy before actually using it. TODO: Change this function so the caller doesn't have to be aware of this issue. Either return by value and expect the caller to always check the base ptr as an indication of whether the struct is valid, or set the BlkInfo as a side-effect and return a bool to indicate success. */ BlkInfo *__getBlkInfo(void *interior) nothrow { BlkInfo *ptr = __blkcache; version (single_cache) { if (ptr.base && ptr.base <= interior && (interior - ptr.base) < ptr.size) return ptr; return null; // not in cache. } else version (simple_cache) { foreach (i; 0..N_CACHE_BLOCKS) { if (ptr.base && ptr.base <= interior && (interior - ptr.base) < ptr.size) return ptr; ptr++; } } else { // try to do a smart lookup, using __nextBlkIdx as the "head" auto curi = ptr + __nextBlkIdx; for (auto i = curi; i >= ptr; --i) { if (i.base && i.base <= interior && cast(size_t)(interior - i.base) < i.size) return i; } for (auto i = ptr + N_CACHE_BLOCKS - 1; i > curi; --i) { if (i.base && i.base <= interior && cast(size_t)(interior - i.base) < i.size) return i; } } return null; // not in cache. } void __insertBlkInfoCache(BlkInfo bi, BlkInfo *curpos) nothrow { version (single_cache) { *__blkcache = bi; } else { version (simple_cache) { if (curpos) *curpos = bi; else { // note, this is a super-simple algorithm that does not care about // most recently used. It simply uses a round-robin technique to // cache block info. This means that the ordering of the cache // doesn't mean anything. Certain patterns of allocation may // render the cache near-useless. __blkcache[__nextBlkIdx] = bi; __nextBlkIdx = (__nextBlkIdx+1) & (N_CACHE_BLOCKS - 1); } } else version (random_cache) { // strategy: if the block currently is in the cache, move the // current block index to the a random element and evict that // element. auto cache = __blkcache; if (!curpos) { __nextBlkIdx = (__nextRndNum = 1664525 * __nextRndNum + 1013904223) & (N_CACHE_BLOCKS - 1); curpos = cache + __nextBlkIdx; } else { __nextBlkIdx = curpos - cache; } *curpos = bi; } else { // // strategy: If the block currently is in the cache, swap it with // the head element. Otherwise, move the head element up by one, // and insert it there. // auto cache = __blkcache; if (!curpos) { __nextBlkIdx = (__nextBlkIdx+1) & (N_CACHE_BLOCKS - 1); curpos = cache + __nextBlkIdx; } else if (curpos !is cache + __nextBlkIdx) { *curpos = cache[__nextBlkIdx]; curpos = cache + __nextBlkIdx; } *curpos = bi; } } } /** Shrink the "allocated" length of an array to be the exact size of the array. It doesn't matter what the current allocated length of the array is, the user is telling the runtime that he knows what he is doing. Params: ti = `TypeInfo` of array type arr = array to shrink. Its `.length` is element length, not byte length, despite `void` type */ extern(C) void _d_arrayshrinkfit(const TypeInfo ti, void[] arr) nothrow { // note, we do not care about shared. We are setting the length no matter // what, so no lock is required. debug(PRINTF) printf("_d_arrayshrinkfit, elemsize = %d, arr.ptr = x%x arr.length = %d\n", ti.next.tsize, arr.ptr, arr.length); auto tinext = unqualify(ti.next); auto size = tinext.tsize; // array element size auto cursize = arr.length * size; auto isshared = typeid(ti) is typeid(TypeInfo_Shared); auto bic = isshared ? null : __getBlkInfo(arr.ptr); auto info = bic ? *bic : GC.query(arr.ptr); if (info.base && (info.attr & BlkAttr.APPENDABLE)) { auto newsize = (arr.ptr - __arrayStart(info)) + cursize; debug(PRINTF) printf("setting allocated size to %d\n", (arr.ptr - info.base) + cursize); // destroy structs that become unused memory when array size is shrinked if (typeid(tinext) is typeid(TypeInfo_Struct)) // avoid a complete dynamic type cast { auto sti = cast(TypeInfo_Struct)cast(void*)tinext; if (sti.xdtor) { auto oldsize = __arrayAllocLength(info, tinext); if (oldsize > cursize) { try { finalize_array(arr.ptr + cursize, oldsize - cursize, sti); } catch (Exception e) { import core.exception : onFinalizeError; onFinalizeError(sti, e); } } } } // Note: Since we "assume" the append is safe, it means it is not shared. // Since it is not shared, we also know it won't throw (no lock). if (!__setArrayAllocLength(info, newsize, false, tinext)) { import core.exception : onInvalidMemoryOperationError; onInvalidMemoryOperationError(); } // cache the block if not already done. if (!isshared && !bic) __insertBlkInfoCache(info, null); } } package bool hasPostblit(in TypeInfo ti) nothrow pure { return (&ti.postblit).funcptr !is &TypeInfo.postblit; } void __doPostblit(void *ptr, size_t len, const TypeInfo ti) { if (!hasPostblit(ti)) return; if (auto tis = cast(TypeInfo_Struct)ti) { // this is a struct, check the xpostblit member auto pblit = tis.xpostblit; if (!pblit) // postblit not specified, no point in looping. return; // optimized for struct, call xpostblit directly for each element immutable size = ti.tsize; const eptr = ptr + len; for (;ptr < eptr;ptr += size) pblit(ptr); } else { // generic case, call the typeinfo's postblit function immutable size = ti.tsize; const eptr = ptr + len; for (;ptr < eptr;ptr += size) ti.postblit(ptr); } } /** Set the array capacity. If the array capacity isn't currently large enough to hold the requested capacity (in number of elements), then the array is resized/reallocated to the appropriate size. Pass in a requested capacity of 0 to get the current capacity. Params: ti = type info of element type newcapacity = requested new capacity p = pointer to array to set. Its `length` is left unchanged. Returns: the number of elements that can actually be stored once the resizing is done */ extern(C) size_t _d_arraysetcapacity(const TypeInfo ti, size_t newcapacity, void[]* p) @weak in { assert(ti); assert(!(*p).length || (*p).ptr); } do { import core.stdc.string; import core.exception : onOutOfMemoryError; // step 1, get the block auto isshared = typeid(ti) is typeid(TypeInfo_Shared); auto bic = isshared ? null : __getBlkInfo((*p).ptr); auto info = bic ? *bic : GC.query((*p).ptr); auto tinext = unqualify(ti.next); auto size = tinext.tsize; version (D_InlineAsm_X86) { size_t reqsize = void; asm { mov EAX, newcapacity; mul EAX, size; mov reqsize, EAX; jnc Lcontinue; } } else version (D_InlineAsm_X86_64) { size_t reqsize = void; asm { mov RAX, newcapacity; mul RAX, size; mov reqsize, RAX; jnc Lcontinue; } } else { import core.checkedint : mulu; bool overflow = false; size_t reqsize = mulu(size, newcapacity, overflow); if (!overflow) goto Lcontinue; } Loverflow: onOutOfMemoryError(); assert(0); Lcontinue: // step 2, get the actual "allocated" size. If the allocated size does not // match what we expect, then we will need to reallocate anyways. // TODO: this probably isn't correct for shared arrays size_t curallocsize = void; size_t curcapacity = void; size_t offset = void; size_t arraypad = void; if (info.base && (info.attr & BlkAttr.APPENDABLE)) { if (info.size <= 256) { arraypad = SMALLPAD + structTypeInfoSize(tinext); curallocsize = *(cast(ubyte *)(info.base + info.size - arraypad)); } else if (info.size < PAGESIZE) { arraypad = MEDPAD + structTypeInfoSize(tinext); curallocsize = *(cast(ushort *)(info.base + info.size - arraypad)); } else { curallocsize = *(cast(size_t *)(info.base)); arraypad = LARGEPAD; } offset = (*p).ptr - __arrayStart(info); if (offset + (*p).length * size != curallocsize) { curcapacity = 0; } else { // figure out the current capacity of the block from the point // of view of the array. curcapacity = info.size - offset - arraypad; } } else { curallocsize = curcapacity = offset = 0; } debug(PRINTF) printf("_d_arraysetcapacity, p = x%d,%d, newcapacity=%d, info.size=%d, reqsize=%d, curallocsize=%d, curcapacity=%d, offset=%d\n", (*p).ptr, (*p).length, newcapacity, info.size, reqsize, curallocsize, curcapacity, offset); if (curcapacity >= reqsize) { // no problems, the current allocated size is large enough. return curcapacity / size; } // step 3, try to extend the array in place. if (info.size >= PAGESIZE && curcapacity != 0) { auto extendsize = reqsize + offset + LARGEPAD - info.size; auto u = GC.extend(info.base, extendsize, extendsize); if (u) { // extend worked, save the new current allocated size if (bic) bic.size = u; // update cache curcapacity = u - offset - LARGEPAD; return curcapacity / size; } } // step 4, if extending doesn't work, allocate a new array with at least the requested allocated size. auto datasize = (*p).length * size; // copy attributes from original block, or from the typeinfo if the // original block doesn't exist. info = __arrayAlloc(reqsize, info, ti, tinext); if (info.base is null) goto Loverflow; // copy the data over. // note that malloc will have initialized the data we did not request to 0. auto tgt = __arrayStart(info); memcpy(tgt, (*p).ptr, datasize); // handle postblit __doPostblit(tgt, datasize, tinext); if (!(info.attr & BlkAttr.NO_SCAN)) { // need to memset the newly requested data, except for the data that // malloc returned that we didn't request. void *endptr = tgt + reqsize; void *begptr = tgt + datasize; // sanity check assert(endptr >= begptr); memset(begptr, 0, endptr - begptr); } // set up the correct length __setArrayAllocLength(info, datasize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, bic); *p = (cast(void*)tgt)[0 .. (*p).length]; // determine the padding. This has to be done manually because __arrayPad // assumes you are not counting the pad size, and info.size does include // the pad. if (info.size <= 256) arraypad = SMALLPAD + structTypeInfoSize(tinext); else if (info.size < PAGESIZE) arraypad = MEDPAD + structTypeInfoSize(tinext); else arraypad = LARGEPAD; curcapacity = info.size - arraypad; return curcapacity / size; } /** Allocate an array with the garbage collector. Has three variants: - `_d_newarrayU` leave elements uninitialized - `_d_newarrayT` initializes to 0 (e.g `new int[]`) - `_d_newarrayiT` initializes based on initializer retrieved from TypeInfo (e.g `new float[]`) Params: ti = the type of the resulting array, (may also be the corresponding `array.ptr` type) length = `.length` of resulting array Returns: newly allocated array */ extern (C) void[] _d_newarrayU(const scope TypeInfo ti, size_t length) pure nothrow @weak { import core.exception : onOutOfMemoryError; auto tinext = unqualify(ti.next); auto size = tinext.tsize; debug(PRINTF) printf("_d_newarrayU(length = x%x, size = %d)\n", length, size); if (length == 0 || size == 0) return null; version (D_InlineAsm_X86) { asm pure nothrow @nogc { mov EAX,size ; mul EAX,length ; mov size,EAX ; jnc Lcontinue ; } } else version (D_InlineAsm_X86_64) { asm pure nothrow @nogc { mov RAX,size ; mul RAX,length ; mov size,RAX ; jnc Lcontinue ; } } else { import core.checkedint : mulu; bool overflow = false; size = mulu(size, length, overflow); if (!overflow) goto Lcontinue; } Loverflow: onOutOfMemoryError(); assert(0); Lcontinue: auto info = __arrayAlloc(size, ti, tinext); if (!info.base) goto Loverflow; debug(PRINTF) printf(" p = %p\n", info.base); // update the length of the array auto arrstart = __arrayStart(info); auto isshared = typeid(ti) is typeid(TypeInfo_Shared); __setArrayAllocLength(info, size, isshared, tinext); return arrstart[0..length]; } /// ditto extern (C) void[] _d_newarrayT(const TypeInfo ti, size_t length) pure nothrow @weak { import core.stdc.string; void[] result = _d_newarrayU(ti, length); auto tinext = unqualify(ti.next); auto size = tinext.tsize; memset(result.ptr, 0, size * length); return result; } /// ditto extern (C) void[] _d_newarrayiT(const TypeInfo ti, size_t length) pure nothrow @weak { import core.internal.traits : AliasSeq; void[] result = _d_newarrayU(ti, length); auto tinext = unqualify(ti.next); auto size = tinext.tsize; auto init = tinext.initializer(); switch (init.length) { foreach (T; AliasSeq!(ubyte, ushort, uint, ulong)) { case T.sizeof: if (tinext.talign % T.alignof == 0) { (cast(T*)result.ptr)[0 .. size * length / T.sizeof] = *cast(T*)init.ptr; return result; } goto default; } default: { import core.stdc.string; immutable sz = init.length; for (size_t u = 0; u < size * length; u += sz) memcpy(result.ptr + u, init.ptr, sz); return result; } } } /** Non-template version of $(REF _d_newitemT, core,lifetime) that does not perform initialization. Needed for $(REF allocEntry, rt,aaA). Params: _ti = `TypeInfo` of item to allocate Returns: newly allocated item */ extern (C) void* _d_newitemU(scope const TypeInfo _ti) pure nothrow @weak { auto ti = unqualify(_ti); auto flags = !(ti.flags & 1) ? BlkAttr.NO_SCAN : 0; immutable tiSize = structTypeInfoSize(ti); immutable itemSize = ti.tsize; immutable size = itemSize + tiSize; if (tiSize) flags |= BlkAttr.STRUCTFINAL | BlkAttr.FINALIZE; auto blkInf = GC.qalloc(size, flags, ti); auto p = blkInf.base; if (tiSize) { // the GC might not have cleared the padding area in the block *cast(TypeInfo*)(p + (itemSize & ~(size_t.sizeof - 1))) = null; *cast(TypeInfo*)(p + blkInf.size - tiSize) = cast() ti; } return p; } debug(PRINTF) { extern(C) void printArrayCache() { auto ptr = __blkcache; printf("CACHE: \n"); foreach (i; 0 .. N_CACHE_BLOCKS) { printf(" %d\taddr:% .8x\tsize:% .10d\tflags:% .8x\n", i, ptr[i].base, ptr[i].size, ptr[i].attr); } } } /** * */ extern (C) void _d_delmemory(void* *p) @weak { if (*p) { GC.free(*p); *p = null; } } /** * */ extern (C) void _d_callinterfacefinalizer(void *p) @weak { if (p) { Interface *pi = **cast(Interface ***)p; Object o = cast(Object)(p - pi.offset); rt_finalize(cast(void*)o); } } /** * */ extern (C) void _d_callfinalizer(void* p) @weak { rt_finalize( p ); } /** * */ extern (C) void rt_setCollectHandler(CollectHandler h) { collectHandler = h; } /** * */ extern (C) CollectHandler rt_getCollectHandler() { return collectHandler; } /** * */ extern (C) int rt_hasFinalizerInSegment(void* p, size_t size, uint attr, scope const(void)[] segment) nothrow { if (attr & BlkAttr.STRUCTFINAL) { if (attr & BlkAttr.APPENDABLE) return hasArrayFinalizerInSegment(p, size, segment); return hasStructFinalizerInSegment(p, size, segment); } // otherwise class auto ppv = cast(void**) p; if (!p || !*ppv) return false; auto c = *cast(ClassInfo*)*ppv; do { auto pf = c.destructor; if (cast(size_t)(pf - segment.ptr) < segment.length) return true; } while ((c = c.base) !is null); return false; } int hasStructFinalizerInSegment(void* p, size_t size, in void[] segment) nothrow { if (!p) return false; auto ti = *cast(TypeInfo_Struct*)(p + size - size_t.sizeof); return cast(size_t)(cast(void*)ti.xdtor - segment.ptr) < segment.length; } int hasArrayFinalizerInSegment(void* p, size_t size, in void[] segment) nothrow { if (!p) return false; TypeInfo_Struct si = void; if (size < PAGESIZE) si = *cast(TypeInfo_Struct*)(p + size - size_t.sizeof); else si = *cast(TypeInfo_Struct*)(p + size_t.sizeof); return cast(size_t)(cast(void*)si.xdtor - segment.ptr) < segment.length; } debug (VALGRIND) import etc.valgrind.valgrind; // called by the GC void finalize_array2(void* p, size_t size) nothrow { debug(PRINTF) printf("rt_finalize_array2(p = %p)\n", p); TypeInfo_Struct si = void; debug (VALGRIND) { auto block = p[0..size]; disableAddrReportingInRange(block); } if (size <= 256) { si = *cast(TypeInfo_Struct*)(p + size - size_t.sizeof); size = *cast(ubyte*)(p + size - size_t.sizeof - SMALLPAD); } else if (size < PAGESIZE) { si = *cast(TypeInfo_Struct*)(p + size - size_t.sizeof); size = *cast(ushort*)(p + size - size_t.sizeof - MEDPAD); } else { si = *cast(TypeInfo_Struct*)(p + size_t.sizeof); size = *cast(size_t*)p; p += LARGEPREFIX; } debug (VALGRIND) enableAddrReportingInRange(block); try { finalize_array(p, size, si); } catch (Exception e) { import core.exception : onFinalizeError; onFinalizeError(si, e); } } void finalize_array(void* p, size_t size, const TypeInfo_Struct si) { // Due to the fact that the delete operator calls destructors // for arrays from the last element to the first, we maintain // compatibility here by doing the same. auto tsize = si.tsize; for (auto curP = p + size - tsize; curP >= p; curP -= tsize) { // call destructor si.destroy(curP); } } // called by the GC void finalize_struct(void* p, size_t size) nothrow { debug(PRINTF) printf("finalize_struct(p = %p)\n", p); auto ti = *cast(TypeInfo_Struct*)(p + size - size_t.sizeof); try { ti.destroy(p); // call destructor } catch (Exception e) { import core.exception : onFinalizeError; onFinalizeError(ti, e); } } /** * */ extern (C) void rt_finalize2(void* p, bool det = true, bool resetMemory = true) nothrow { debug(PRINTF) printf("rt_finalize2(p = %p)\n", p); auto ppv = cast(void**) p; if (!p || !*ppv) return; auto pc = cast(ClassInfo*) *ppv; try { if (det || collectHandler is null || collectHandler(cast(Object) p)) { auto c = *pc; do { if (c.destructor) (cast(fp_t) c.destructor)(cast(Object) p); // call destructor } while ((c = c.base) !is null); } if (ppv[1]) // if monitor is not null _d_monitordelete(cast(Object) p, det); if (resetMemory) { auto w = (*pc).initializer; p[0 .. w.length] = w[]; } } catch (Exception e) { import core.exception : onFinalizeError; onFinalizeError(*pc, e); } finally { *ppv = null; // zero vptr even if `resetMemory` is false } } /// Backwards compatibility extern (C) void rt_finalize(void* p, bool det = true) nothrow { rt_finalize2(p, det, true); } extern (C) void rt_finalizeFromGC(void* p, size_t size, uint attr) nothrow { // to verify: reset memory necessary? if (!(attr & BlkAttr.STRUCTFINAL)) rt_finalize2(p, false, false); // class else if (attr & BlkAttr.APPENDABLE) finalize_array2(p, size); // array of structs else finalize_struct(p, size); // struct } /** Resize a dynamic array by setting the `.length` property Newly created elements are initialized to their default value. Has two variants: - `_d_arraysetlengthT` for arrays with elements that initialize to 0 - `_d_arraysetlengthiT` for non-zero initializers retrieved from `TypeInfo` --- void main() { int[] a = [1, 2]; a.length = 3; // gets lowered to `_d_arraysetlengthT(typeid(int[]), 3, &a)` } --- Params: ti = `TypeInfo` of array newlength = new value for the array's `.length` p = pointer to array to update the `.length` of. While it's cast to `void[]`, its `.length` is still treated as element length. Returns: `*p` after being updated */ extern (C) void[] _d_arraysetlengthT(const TypeInfo ti, size_t newlength, void[]* p) @weak in { assert(ti); assert(!(*p).length || (*p).ptr); } do { import core.stdc.string; import core.exception : onOutOfMemoryError; debug(PRINTF) { //printf("_d_arraysetlengthT(p = %p, sizeelem = %d, newlength = %d)\n", p, sizeelem, newlength); if (p) printf("\tp.ptr = %p, p.length = %d\n", (*p).ptr, (*p).length); } if (newlength <= (*p).length) { *p = (*p)[0 .. newlength]; void* newdata = (*p).ptr; return newdata[0 .. newlength]; } auto tinext = unqualify(ti.next); size_t sizeelem = tinext.tsize; /* Calculate: newsize = newlength * sizeelem */ bool overflow = false; version (D_InlineAsm_X86) { size_t newsize = void; asm pure nothrow @nogc { mov EAX, newlength; mul EAX, sizeelem; mov newsize, EAX; setc overflow; } } else version (D_InlineAsm_X86_64) { size_t newsize = void; asm pure nothrow @nogc { mov RAX, newlength; mul RAX, sizeelem; mov newsize, RAX; setc overflow; } } else { import core.checkedint : mulu; const size_t newsize = mulu(sizeelem, newlength, overflow); } if (overflow) { onOutOfMemoryError(); assert(0); } debug(PRINTF) printf("newsize = %x, newlength = %x\n", newsize, newlength); const isshared = typeid(ti) is typeid(TypeInfo_Shared); if (!(*p).ptr) { // pointer was null, need to allocate auto info = __arrayAlloc(newsize, ti, tinext); if (info.base is null) { onOutOfMemoryError(); assert(0); } __setArrayAllocLength(info, newsize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, null); void* newdata = cast(byte *)__arrayStart(info); memset(newdata, 0, newsize); *p = newdata[0 .. newlength]; return *p; } const size_t size = (*p).length * sizeelem; auto bic = isshared ? null : __getBlkInfo((*p).ptr); auto info = bic ? *bic : GC.query((*p).ptr); /* Attempt to extend past the end of the existing array. * If not possible, allocate new space for entire array and copy. */ bool allocateAndCopy = false; void* newdata = (*p).ptr; if (info.base && (info.attr & BlkAttr.APPENDABLE)) { // calculate the extent of the array given the base. const size_t offset = (*p).ptr - __arrayStart(info); if (info.size >= PAGESIZE) { // size of array is at the front of the block if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // check to see if it failed because there is not // enough space if (*(cast(size_t*)info.base) == size + offset) { // not enough space, try extending auto extendsize = newsize + offset + LARGEPAD - info.size; auto u = GC.extend(info.base, extendsize, extendsize); if (u) { // extend worked, now try setting the length // again. info.size = u; if (__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { if (!isshared) __insertBlkInfoCache(info, bic); memset(newdata + size, 0, newsize - size); *p = newdata[0 .. newlength]; return *p; } } } // couldn't do it, reallocate allocateAndCopy = true; } else if (!isshared && !bic) { // add this to the cache, it wasn't present previously. __insertBlkInfoCache(info, null); } } else if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // could not resize in place allocateAndCopy = true; } else if (!isshared && !bic) { // add this to the cache, it wasn't present previously. __insertBlkInfoCache(info, null); } } else allocateAndCopy = true; if (allocateAndCopy) { if (info.base) { if (bic) { // a chance that flags have changed since this was cached, we should fetch the most recent flags info.attr = GC.getAttr(info.base) | BlkAttr.APPENDABLE; } info = __arrayAlloc(newsize, info, ti, tinext); } else { info = __arrayAlloc(newsize, ti, tinext); } if (info.base is null) { onOutOfMemoryError(); assert(0); } __setArrayAllocLength(info, newsize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, bic); newdata = cast(byte *)__arrayStart(info); newdata[0 .. size] = (*p).ptr[0 .. size]; /* Do postblit processing, as we are making a copy and the * original array may have references. * Note that this may throw. */ __doPostblit(newdata, size, tinext); } // Zero the unused portion of the newly allocated space memset(newdata + size, 0, newsize - size); *p = newdata[0 .. newlength]; return *p; } /// ditto extern (C) void[] _d_arraysetlengthiT(const TypeInfo ti, size_t newlength, void[]* p) @weak in { assert(!(*p).length || (*p).ptr); } do { import core.stdc.string; import core.exception : onOutOfMemoryError; debug(PRINTF) { //printf("_d_arraysetlengthiT(p = %p, sizeelem = %d, newlength = %d)\n", p, sizeelem, newlength); if (p) printf("\tp.ptr = %p, p.length = %d\n", (*p).ptr, (*p).length); } if (newlength <= (*p).length) { *p = (*p)[0 .. newlength]; void* newdata = (*p).ptr; return newdata[0 .. newlength]; } auto tinext = unqualify(ti.next); size_t sizeelem = tinext.tsize; /* Calculate: newsize = newlength * sizeelem */ bool overflow = false; version (D_InlineAsm_X86) { size_t newsize = void; asm pure nothrow @nogc { mov EAX, newlength; mul EAX, sizeelem; mov newsize, EAX; setc overflow; } } else version (D_InlineAsm_X86_64) { size_t newsize = void; asm pure nothrow @nogc { mov RAX, newlength; mul RAX, sizeelem; mov newsize, RAX; setc overflow; } } else { import core.checkedint : mulu; const size_t newsize = mulu(sizeelem, newlength, overflow); } if (overflow) { onOutOfMemoryError(); assert(0); } debug(PRINTF) printf("newsize = %x, newlength = %x\n", newsize, newlength); const isshared = typeid(ti) is typeid(TypeInfo_Shared); static void doInitialize(void *start, void *end, const void[] initializer) { if (initializer.length == 1) { memset(start, *(cast(ubyte*)initializer.ptr), end - start); } else { auto q = initializer.ptr; immutable initsize = initializer.length; for (; start < end; start += initsize) { memcpy(start, q, initsize); } } } if (!(*p).ptr) { // pointer was null, need to allocate auto info = __arrayAlloc(newsize, ti, tinext); if (info.base is null) { onOutOfMemoryError(); assert(0); } __setArrayAllocLength(info, newsize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, null); void* newdata = cast(byte *)__arrayStart(info); doInitialize(newdata, newdata + newsize, tinext.initializer); *p = newdata[0 .. newlength]; return *p; } const size_t size = (*p).length * sizeelem; auto bic = isshared ? null : __getBlkInfo((*p).ptr); auto info = bic ? *bic : GC.query((*p).ptr); /* Attempt to extend past the end of the existing array. * If not possible, allocate new space for entire array and copy. */ bool allocateAndCopy = false; void* newdata = (*p).ptr; if (info.base && (info.attr & BlkAttr.APPENDABLE)) { // calculate the extent of the array given the base. const size_t offset = (*p).ptr - __arrayStart(info); if (info.size >= PAGESIZE) { // size of array is at the front of the block if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // check to see if it failed because there is not // enough space if (*(cast(size_t*)info.base) == size + offset) { // not enough space, try extending auto extendsize = newsize + offset + LARGEPAD - info.size; auto u = GC.extend(info.base, extendsize, extendsize); if (u) { // extend worked, now try setting the length // again. info.size = u; if (__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { if (!isshared) __insertBlkInfoCache(info, bic); doInitialize(newdata + size, newdata + newsize, tinext.initializer); *p = newdata[0 .. newlength]; return *p; } } } // couldn't do it, reallocate allocateAndCopy = true; } else if (!isshared && !bic) { // add this to the cache, it wasn't present previously. __insertBlkInfoCache(info, null); } } else if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // could not resize in place allocateAndCopy = true; } else if (!isshared && !bic) { // add this to the cache, it wasn't present previously. __insertBlkInfoCache(info, null); } } else allocateAndCopy = true; if (allocateAndCopy) { if (info.base) { if (bic) { // a chance that flags have changed since this was cached, we should fetch the most recent flags info.attr = GC.getAttr(info.base) | BlkAttr.APPENDABLE; } info = __arrayAlloc(newsize, info, ti, tinext); } else { info = __arrayAlloc(newsize, ti, tinext); } if (info.base is null) { onOutOfMemoryError(); assert(0); } __setArrayAllocLength(info, newsize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, bic); newdata = cast(byte *)__arrayStart(info); newdata[0 .. size] = (*p).ptr[0 .. size]; /* Do postblit processing, as we are making a copy and the * original array may have references. * Note that this may throw. */ __doPostblit(newdata, size, tinext); } // Initialize the unused portion of the newly allocated space doInitialize(newdata + size, newdata + newsize, tinext.initializer); *p = newdata[0 .. newlength]; return *p; } /** Given an array of length `size` that needs to be expanded to `newlength`, compute a new capacity. Better version by Dave Fladebo: This uses an inverse logorithmic algorithm to pre-allocate a bit more space for larger arrays. - Arrays smaller than PAGESIZE bytes are left as-is, so for the most common cases, memory allocation is 1 to 1. The small overhead added doesn't affect small array perf. (it's virtually the same as current). - Larger arrays have some space pre-allocated. - As the arrays grow, the relative pre-allocated space shrinks. - The logorithmic algorithm allocates relatively more space for mid-size arrays, making it very fast for medium arrays (for mid-to-large arrays, this turns out to be quite a bit faster than the equivalent realloc() code in C, on Linux at least. Small arrays are just as fast as GCC). - Perhaps most importantly, overall memory usage and stress on the GC is decreased significantly for demanding environments. Params: newlength = new `.length` size = old `.length` Returns: new capacity for array */ size_t newCapacity(size_t newlength, size_t size) { version (none) { size_t newcap = newlength * size; } else { size_t newcap = newlength * size; size_t newext = 0; if (newcap > PAGESIZE) { //double mult2 = 1.0 + (size / log10(pow(newcap * 2.0,2.0))); // redo above line using only integer math /*static int log2plus1(size_t c) { int i; if (c == 0) i = -1; else for (i = 1; c >>= 1; i++) { } return i; }*/ /* The following setting for mult sets how much bigger * the new size will be over what is actually needed. * 100 means the same size, more means proportionally more. * More means faster but more memory consumption. */ //long mult = 100 + (1000L * size) / (6 * log2plus1(newcap)); //long mult = 100 + (1000L * size) / log2plus1(newcap); import core.bitop; long mult = 100 + (1000L) / (bsr(newcap) + 1); // testing shows 1.02 for large arrays is about the point of diminishing return // // Commented out because the multipler will never be < 102. In order for it to be < 2, // then 1000L / (bsr(x) + 1) must be > 2. The highest bsr(x) + 1 // could be is 65 (64th bit set), and 1000L / 64 is much larger // than 2. We need 500 bit integers for 101 to be achieved :) /*if (mult < 102) mult = 102;*/ /*newext = cast(size_t)((newcap * mult) / 100); newext -= newext % size;*/ // This version rounds up to the next element, and avoids using // mod. newext = cast(size_t)((newlength * mult + 99) / 100) * size; debug(PRINTF) printf("mult: %2.2f, alloc: %2.2f\n",mult/100.0,newext / cast(double)size); } newcap = newext > newcap ? newext : newcap; debug(PRINTF) printf("newcap = %d, newlength = %d, size = %d\n", newcap, newlength, size); } return newcap; } /** Extend an array by n elements. Caller must initialize those elements. Params: ti = type info of array type (not element type) px = array to append to, cast to `byte[]` while keeping the same `.length`. Will be updated. n = number of elements to append Returns: `px` after being appended to */ extern (C) byte[] _d_arrayappendcTX(const TypeInfo ti, return scope ref byte[] px, size_t n) @weak { import core.stdc.string; // This is a cut&paste job from _d_arrayappendT(). Should be refactored. // only optimize array append where ti is not a shared type auto tinext = unqualify(ti.next); auto sizeelem = tinext.tsize; // array element size auto isshared = typeid(ti) is typeid(TypeInfo_Shared); auto bic = isshared ? null : __getBlkInfo(px.ptr); auto info = bic ? *bic : GC.query(px.ptr); auto length = px.length; auto newlength = length + n; auto newsize = newlength * sizeelem; auto size = length * sizeelem; size_t newcap = void; // for scratch space // calculate the extent of the array given the base. size_t offset = cast(void*)px.ptr - __arrayStart(info); if (info.base && (info.attr & BlkAttr.APPENDABLE)) { if (info.size >= PAGESIZE) { // size of array is at the front of the block if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // check to see if it failed because there is not // enough space newcap = newCapacity(newlength, sizeelem); if (*(cast(size_t*)info.base) == size + offset) { // not enough space, try extending auto extendoffset = offset + LARGEPAD - info.size; auto u = GC.extend(info.base, newsize + extendoffset, newcap + extendoffset); if (u) { // extend worked, now try setting the length // again. info.size = u; if (__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { if (!isshared) __insertBlkInfoCache(info, bic); goto L1; } } } // couldn't do it, reallocate goto L2; } else if (!isshared && !bic) { __insertBlkInfoCache(info, null); } } else if (!__setArrayAllocLength(info, newsize + offset, isshared, tinext, size + offset)) { // could not resize in place newcap = newCapacity(newlength, sizeelem); goto L2; } else if (!isshared && !bic) { __insertBlkInfoCache(info, null); } } else { // not appendable or is null newcap = newCapacity(newlength, sizeelem); if (info.base) { L2: if (bic) { // a chance that flags have changed since this was cached, we should fetch the most recent flags info.attr = GC.getAttr(info.base) | BlkAttr.APPENDABLE; } info = __arrayAlloc(newcap, info, ti, tinext); } else { info = __arrayAlloc(newcap, ti, tinext); } __setArrayAllocLength(info, newsize, isshared, tinext); if (!isshared) __insertBlkInfoCache(info, bic); auto newdata = cast(byte *)__arrayStart(info); memcpy(newdata, px.ptr, length * sizeelem); // do postblit processing __doPostblit(newdata, length * sizeelem, tinext); (cast(void **)(&px))[1] = newdata; } L1: *cast(size_t *)&px = newlength; return px; } /** Append `dchar` to `char[]`, converting UTF-32 to UTF-8 --- void main() { char[] s; s ~= 'α'; } --- Params: x = array to append to cast to `byte[]`. Will be modified. c = `dchar` to append Returns: updated `x` cast to `void[]` */ extern (C) void[] _d_arrayappendcd(ref byte[] x, dchar c) @weak { // c could encode into from 1 to 4 characters char[4] buf = void; char[] appendthis; // passed to appendT if (c <= 0x7F) { buf.ptr[0] = cast(char)c; appendthis = buf[0..1]; } else if (c <= 0x7FF) { buf.ptr[0] = cast(char)(0xC0 | (c >> 6)); buf.ptr[1] = cast(char)(0x80 | (c & 0x3F)); appendthis = buf[0..2]; } else if (c <= 0xFFFF) { buf.ptr[0] = cast(char)(0xE0 | (c >> 12)); buf.ptr[1] = cast(char)(0x80 | ((c >> 6) & 0x3F)); buf.ptr[2] = cast(char)(0x80 | (c & 0x3F)); appendthis = buf[0..3]; } else if (c <= 0x10FFFF) { buf.ptr[0] = cast(char)(0xF0 | (c >> 18)); buf.ptr[1] = cast(char)(0x80 | ((c >> 12) & 0x3F)); buf.ptr[2] = cast(char)(0x80 | ((c >> 6) & 0x3F)); buf.ptr[3] = cast(char)(0x80 | (c & 0x3F)); appendthis = buf[0..4]; } else { import core.exception : onUnicodeError; onUnicodeError("Invalid UTF-8 sequence", 0); // invalid utf character } // // TODO: This always assumes the array type is shared, because we do not // get a typeinfo from the compiler. Assuming shared is the safest option. // Once the compiler is fixed, the proper typeinfo should be forwarded. // // Hack because _d_arrayappendT takes `x` as a reference auto xx = cast(shared(char)[])x; object._d_arrayappendT(xx, cast(shared(char)[])appendthis); x = cast(byte[])xx; return x; } unittest { import core.exception : UnicodeException; /* Using inline try {} catch {} blocks fails to catch the UnicodeException * thrown. * https://issues.dlang.org/show_bug.cgi?id=16799 */ static void assertThrown(T : Throwable = Exception, E)(lazy E expr, string msg) { try expr; catch (T e) { assert(e.msg == msg); return; } } static void f() { string ret; int i = -1; ret ~= i; } assertThrown!UnicodeException(f(), "Invalid UTF-8 sequence"); } /** Append `dchar` to `wchar[]`, converting UTF-32 to UTF-16 --- void main() { dchar x; wchar[] s; s ~= 'α'; } --- Params: x = array to append to cast to `byte[]`. Will be modified. c = `dchar` to append Returns: updated `x` cast to `void[]` */ extern (C) void[] _d_arrayappendwd(ref byte[] x, dchar c) @weak { // c could encode into from 1 to 2 w characters wchar[2] buf = void; wchar[] appendthis; // passed to appendT if (c <= 0xFFFF) { buf.ptr[0] = cast(wchar) c; appendthis = buf[0..1]; } else { buf.ptr[0] = cast(wchar) ((((c - 0x10000) >> 10) & 0x3FF) + 0xD800); buf.ptr[1] = cast(wchar) (((c - 0x10000) & 0x3FF) + 0xDC00); appendthis = buf[0..2]; } // // TODO: This always assumes the array type is shared, because we do not // get a typeinfo from the compiler. Assuming shared is the safest option. // Once the compiler is fixed, the proper typeinfo should be forwarded. // auto xx = (cast(shared(wchar)*)x.ptr)[0 .. x.length]; object._d_arrayappendT(xx, cast(shared(wchar)[])appendthis); x = (cast(byte*)xx.ptr)[0 .. xx.length]; return x; } /** Allocate an array literal Rely on the caller to do the initialization of the array. --- int[] getArr() { return [10, 20]; // auto res = cast(int*) _d_arrayliteralTX(typeid(int[]), 2); // res[0] = 10; // res[1] = 20; // return res[0..2]; } --- Params: ti = `TypeInfo` of resulting array type length = `.length` of array literal Returns: pointer to allocated array */ extern (C) void* _d_arrayliteralTX(const TypeInfo ti, size_t length) @weak { auto tinext = unqualify(ti.next); auto sizeelem = tinext.tsize; // array element size void* result; debug(PRINTF) printf("_d_arrayliteralTX(sizeelem = %d, length = %d)\n", sizeelem, length); if (length == 0 || sizeelem == 0) result = null; else { auto allocsize = length * sizeelem; auto info = __arrayAlloc(allocsize, ti, tinext); auto isshared = typeid(ti) is typeid(TypeInfo_Shared); __setArrayAllocLength(info, allocsize, isshared, tinext); result = __arrayStart(info); } return result; } unittest { int[] a; int[] b; int i; a = new int[3]; a[0] = 1; a[1] = 2; a[2] = 3; b = a.dup; assert(b.length == 3); for (i = 0; i < 3; i++) assert(b[i] == i + 1); // test slice appending b = a[0..1]; b ~= 4; for (i = 0; i < 3; i++) assert(a[i] == i + 1); // test reserving char[] arr = new char[4093]; for (i = 0; i < arr.length; i++) arr[i] = cast(char)(i % 256); // note that these two commands used to cause corruption, which may not be // detected. arr.reserve(4094); auto arr2 = arr ~ "123"; assert(arr2[0..arr.length] == arr); assert(arr2[arr.length..$] == "123"); // test postblit on array concat, append, length, etc. static struct S { int x; int pad; this(this) { ++x; } } void testPostBlit(T)() { auto sarr = new T[1]; debug(SENTINEL) {} else assert(sarr.capacity == 1); // length extend auto sarr2 = sarr; assert(sarr[0].x == 0); sarr2.length += 1; assert(sarr2[0].x == 1); assert(sarr[0].x == 0); // append T s; sarr2 = sarr; sarr2 ~= s; assert(sarr2[0].x == 1); assert(sarr2[1].x == 1); assert(sarr[0].x == 0); assert(s.x == 0); // concat sarr2 = sarr ~ sarr; assert(sarr2[0].x == 1); assert(sarr2[1].x == 1); assert(sarr[0].x == 0); // concat multiple (calls different method) sarr2 = sarr ~ sarr ~ sarr; assert(sarr2[0].x == 1); assert(sarr2[1].x == 1); assert(sarr2[2].x == 1); assert(sarr[0].x == 0); // reserve capacity sarr2 = sarr; sarr2.reserve(2); assert(sarr2[0].x == 1); assert(sarr[0].x == 0); } testPostBlit!(S)(); testPostBlit!(const(S))(); } unittest { // Bugzilla 3454 - Inconsistent flag setting in GC.realloc() static void test(size_t multiplier) { auto p = GC.malloc(8 * multiplier, 0); assert(GC.getAttr(p) == 0); // no move, set attr p = GC.realloc(p, 8 * multiplier + 5, BlkAttr.NO_SCAN); assert(GC.getAttr(p) == BlkAttr.NO_SCAN); // shrink, copy attr p = GC.realloc(p, 2 * multiplier, 0); assert(GC.getAttr(p) == BlkAttr.NO_SCAN); // extend, copy attr p = GC.realloc(p, 8 * multiplier, 0); assert(GC.getAttr(p) == BlkAttr.NO_SCAN); } test(16); version (OnlyLowMemUnittests) {} else test(1024 * 1024); } unittest { import core.exception; try { size_t x = size_t.max; byte[] big_buf = new byte[x]; } catch (OutOfMemoryError) { } } unittest { // https://issues.dlang.org/show_bug.cgi?id=13854 auto arr = new ubyte[PAGESIZE]; // ensure page size auto info1 = GC.query(arr.ptr); assert(info1.base !is arr.ptr); // offset is required for page size or larger auto arr2 = arr[0..1]; assert(arr2.capacity == 0); // cannot append arr2 ~= 0; // add a byte assert(arr2.ptr !is arr.ptr); // reallocated auto info2 = GC.query(arr2.ptr); assert(info2.base is arr2.ptr); // no offset, the capacity is small. // do the same via setting length arr2 = arr[0..1]; assert(arr2.capacity == 0); arr2.length += 1; assert(arr2.ptr !is arr.ptr); // reallocated info2 = GC.query(arr2.ptr); assert(info2.base is arr2.ptr); // no offset, the capacity is small. // do the same for char[] since we need a type with an initializer to test certain runtime functions auto carr = new char[PAGESIZE]; info1 = GC.query(carr.ptr); assert(info1.base !is carr.ptr); // offset is required for page size or larger auto carr2 = carr[0..1]; assert(carr2.capacity == 0); // cannot append carr2 ~= 0; // add a byte assert(carr2.ptr !is carr.ptr); // reallocated info2 = GC.query(carr2.ptr); assert(info2.base is carr2.ptr); // no offset, the capacity is small. // do the same via setting length carr2 = carr[0..1]; assert(carr2.capacity == 0); carr2.length += 1; assert(carr2.ptr !is carr.ptr); // reallocated info2 = GC.query(carr2.ptr); assert(info2.base is carr2.ptr); // no offset, the capacity is small. } unittest { // https://issues.dlang.org/show_bug.cgi?id=13878 auto arr = new ubyte[1]; auto info = GC.query(arr.ptr); assert(info.attr & BlkAttr.NO_SCAN); // should be NO_SCAN arr ~= 0; // ensure array is inserted into cache debug(SENTINEL) {} else assert(arr.ptr is info.base); GC.clrAttr(arr.ptr, BlkAttr.NO_SCAN); // remove the attribute auto arr2 = arr[0..1]; assert(arr2.capacity == 0); // cannot append arr2 ~= 0; assert(arr2.ptr !is arr.ptr); info = GC.query(arr2.ptr); assert(!(info.attr & BlkAttr.NO_SCAN)); // ensure attribute sticks // do the same via setting length arr = new ubyte[1]; arr ~= 0; // ensure array is inserted into cache GC.clrAttr(arr.ptr, BlkAttr.NO_SCAN); // remove the attribute arr2 = arr[0..1]; assert(arr2.capacity == 0); arr2.length += 1; assert(arr2.ptr !is arr.ptr); // reallocated info = GC.query(arr2.ptr); assert(!(info.attr & BlkAttr.NO_SCAN)); // ensure attribute sticks // do the same for char[] since we need a type with an initializer to test certain runtime functions auto carr = new char[1]; info = GC.query(carr.ptr); assert(info.attr & BlkAttr.NO_SCAN); // should be NO_SCAN carr ~= 0; // ensure array is inserted into cache debug(SENTINEL) {} else assert(carr.ptr is info.base); GC.clrAttr(carr.ptr, BlkAttr.NO_SCAN); // remove the attribute auto carr2 = carr[0..1]; assert(carr2.capacity == 0); // cannot append carr2 ~= 0; assert(carr2.ptr !is carr.ptr); info = GC.query(carr2.ptr); assert(!(info.attr & BlkAttr.NO_SCAN)); // ensure attribute sticks // do the same via setting length carr = new char[1]; carr ~= 0; // ensure array is inserted into cache GC.clrAttr(carr.ptr, BlkAttr.NO_SCAN); // remove the attribute carr2 = carr[0..1]; assert(carr2.capacity == 0); carr2.length += 1; assert(carr2.ptr !is carr.ptr); // reallocated info = GC.query(carr2.ptr); assert(!(info.attr & BlkAttr.NO_SCAN)); // ensure attribute sticks } // test struct finalizers debug(SENTINEL) {} else deprecated unittest { __gshared int dtorCount; static struct S1 { int x; ~this() { dtorCount++; } } dtorCount = 0; S1* s2 = new S1; GC.runFinalizers((cast(char*)(typeid(S1).xdtor))[0..1]); assert(dtorCount == 1); GC.free(s2); dtorCount = 0; const(S1)* s3 = new const(S1); GC.runFinalizers((cast(char*)(typeid(S1).xdtor))[0..1]); assert(dtorCount == 1); GC.free(cast(void*)s3); dtorCount = 0; shared(S1)* s4 = new shared(S1); GC.runFinalizers((cast(char*)(typeid(S1).xdtor))[0..1]); assert(dtorCount == 1); GC.free(cast(void*)s4); dtorCount = 0; const(S1)[] carr1 = new const(S1)[5]; BlkInfo blkinf1 = GC.query(carr1.ptr); GC.runFinalizers((cast(char*)(typeid(S1).xdtor))[0..1]); assert(dtorCount == 5); GC.free(blkinf1.base); dtorCount = 0; S1[] arr2 = new S1[10]; arr2.length = 6; arr2.assumeSafeAppend; assert(dtorCount == 4); // destructors run explicitely? dtorCount = 0; BlkInfo blkinf = GC.query(arr2.ptr); GC.runFinalizers((cast(char*)(typeid(S1).xdtor))[0..1]); assert(dtorCount == 6); GC.free(blkinf.base); // associative arrays import rt.aaA : entryDtor; // throw away all existing AA entries with dtor GC.runFinalizers((cast(char*)(&entryDtor))[0..1]); S1[int] aa1; aa1[0] = S1(0); aa1[1] = S1(1); dtorCount = 0; aa1 = null; GC.runFinalizers((cast(char*)(&entryDtor))[0..1]); assert(dtorCount == 2); int[S1] aa2; aa2[S1(0)] = 0; aa2[S1(1)] = 1; aa2[S1(2)] = 2; dtorCount = 0; aa2 = null; GC.runFinalizers((cast(char*)(&entryDtor))[0..1]); assert(dtorCount == 3); S1[2][int] aa3; aa3[0] = [S1(0),S1(2)]; aa3[1] = [S1(1),S1(3)]; dtorCount = 0; aa3 = null; GC.runFinalizers((cast(char*)(&entryDtor))[0..1]); assert(dtorCount == 4); } // test struct dtor handling not causing false pointers unittest { // for 64-bit, allocate a struct of size 40 static struct S { size_t[4] data; S* ptr4; } auto p1 = new S; auto p2 = new S; p2.ptr4 = p1; // a struct with a dtor with size 32, but the dtor will cause // allocation to be larger by a pointer static struct A { size_t[3] data; S* ptr3; ~this() {} } GC.free(p2); auto a = new A; // reuse same memory if (cast(void*)a is cast(void*)p2) // reusage not guaranteed { auto ptr = cast(S**)(a + 1); assert(*ptr != p1); // still same data as p2.ptr4? } // small array static struct SArr { void*[10] data; } auto arr1 = new SArr; arr1.data[] = p1; GC.free(arr1); // allocates 2*A.sizeof + (void*).sizeof (TypeInfo) + 1 (array length) auto arr2 = new A[2]; if (cast(void*)arr1 is cast(void*)arr2.ptr) // reusage not guaranteed { auto ptr = cast(S**)(arr2.ptr + 2); assert(*ptr != p1); // still same data as p2.ptr4? } // large array static struct LArr { void*[1023] data; } auto larr1 = new LArr; larr1.data[] = p1; GC.free(larr1); auto larr2 = new S[255]; if (cast(void*)larr1 is cast(void*)larr2.ptr - LARGEPREFIX) // reusage not guaranteed { auto ptr = cast(S**)larr1; assert(ptr[0] != p1); // 16 bytes array header assert(ptr[1] != p1); version (D_LP64) {} else { assert(ptr[2] != p1); assert(ptr[3] != p1); } } } // test class finalizers exception handling unittest { bool test(E)() { import core.exception; static class C1 { E exc; this(E exc) { this.exc = exc; } ~this() { throw exc; } } bool caught = false; C1 c = new C1(new E("test onFinalizeError")); try { GC.runFinalizers((cast(uint*)&C1.__dtor)[0..1]); } catch (FinalizeError err) { caught = true; } catch (E) { } GC.free(cast(void*)c); return caught; } assert( test!Exception); import core.exception : InvalidMemoryOperationError; assert(!test!InvalidMemoryOperationError); } // test bug 14126 unittest { static struct S { S* thisptr; ~this() { assert(&this == thisptr); thisptr = null;} } S[] test14126 = new S[2048]; // make sure we allocate at least a PAGE foreach (ref s; test14126) { s.thisptr = &s; } }