mirror of
https://github.com/darlinghq/darling-objc4.git
synced 2024-11-23 04:09:46 +00:00
126 lines
3.4 KiB
Plaintext
126 lines
3.4 KiB
Plaintext
/*
|
|
* Copyright (c) 2007 Apple Inc. All Rights Reserved.
|
|
*
|
|
* @APPLE_LICENSE_HEADER_START@
|
|
*
|
|
* This file contains Original Code and/or Modifications of Original Code
|
|
* as defined in and that are subject to the Apple Public Source License
|
|
* Version 2.0 (the 'License'). You may not use this file except in
|
|
* compliance with the License. Please obtain a copy of the License at
|
|
* http://www.opensource.apple.com/apsl/ and read it before using this
|
|
* file.
|
|
*
|
|
* The Original Code and all software distributed under the License are
|
|
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
|
|
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
|
|
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
|
|
* Please see the License for the specific language governing rights and
|
|
* limitations under the License.
|
|
*
|
|
* @APPLE_LICENSE_HEADER_END@
|
|
*/
|
|
|
|
/**
|
|
* @file objc-zalloc.h
|
|
*
|
|
* "zone allocator" for objc.
|
|
*
|
|
* Provides packed allocation for data structures the runtime
|
|
* almost never frees.
|
|
*/
|
|
|
|
#include "objc-private.h"
|
|
#include "objc-zalloc.h"
|
|
|
|
namespace objc {
|
|
|
|
void *AtomicQueue::pop()
|
|
{
|
|
AtomicQueue l1, l2;
|
|
|
|
l1.pair = pair; // non atomic on purpose
|
|
|
|
do {
|
|
if (l1.head == nullptr) {
|
|
return nullptr;
|
|
}
|
|
l2.head = l1.head->next;
|
|
l2.gen = l1.gen + 1;
|
|
} while (!atomic_pair.compare_exchange_weak(l1.pair, l2.pair, relaxed, relaxed));
|
|
|
|
return reinterpret_cast<void *>(l1.head);
|
|
}
|
|
|
|
void AtomicQueue::push_list(void *_head, void *_tail)
|
|
{
|
|
Entry *head = reinterpret_cast<Entry *>(_head);
|
|
Entry *tail = reinterpret_cast<Entry *>(_tail);
|
|
AtomicQueue l1, l2;
|
|
|
|
l1.pair = pair; // non atomic load on purpose
|
|
do {
|
|
tail->next = l1.head;
|
|
l2.head = head;
|
|
l2.gen = l1.gen + 1;
|
|
} while (!atomic_pair.compare_exchange_weak(l1.pair, l2.pair, release, relaxed));
|
|
}
|
|
|
|
template<class T>
|
|
constexpr inline
|
|
T gcd(T a, T b)
|
|
{
|
|
return b == 0 ? a : gcd(b, a % b);
|
|
}
|
|
|
|
template<class T>
|
|
AtomicQueue Zone<T, false>::_freelist;
|
|
|
|
template<class T>
|
|
T *Zone<T, false>::alloc_slow()
|
|
{
|
|
// our malloc aligns to 16 bytes and this code should be used for sizes
|
|
// small enough that this should always be an actual malloc bucket.
|
|
//
|
|
// The point of this code is *NOT* speed but optimal density
|
|
constexpr size_t n_elem = MALLOC_ALIGNMENT / gcd(sizeof(T), size_t{MALLOC_ALIGNMENT});
|
|
Element *slab = reinterpret_cast<Element *>(::calloc(n_elem, sizeof(T)));
|
|
for (size_t i = 1; i < n_elem - 1; i++) {
|
|
slab[i].next = &slab[i + 1];
|
|
}
|
|
_freelist.push_list(reinterpret_cast<void *>(&slab[1]),
|
|
reinterpret_cast<void *>(&slab[n_elem - 1]));
|
|
return reinterpret_cast<T *>(&slab[0]);
|
|
}
|
|
|
|
template<class T>
|
|
T *Zone<T, false>::alloc()
|
|
{
|
|
void *e = _freelist.pop();
|
|
if (e) {
|
|
__builtin_bzero(e, sizeof(void *));
|
|
return reinterpret_cast<T *>(e);
|
|
}
|
|
return alloc_slow();
|
|
}
|
|
|
|
template<class T>
|
|
void Zone<T, false>::free(T *ptr)
|
|
{
|
|
if (ptr) {
|
|
Element *e = reinterpret_cast<Element *>(ptr);
|
|
__builtin_bzero(e->buf, sizeof(e->buf));
|
|
_freelist.push(e);
|
|
}
|
|
}
|
|
|
|
#if __OBJC2__
|
|
#define ZoneInstantiate(type) \
|
|
template class Zone<type, sizeof(type) % MALLOC_ALIGNMENT == 0>
|
|
|
|
ZoneInstantiate(class_rw_t);
|
|
ZoneInstantiate(class_rw_ext_t);
|
|
#endif
|
|
|
|
}
|