/* RetroArch - A frontend for libretro. * Copyright (C) 2010-2014 - Hans-Kristian Arntzen * Copyright (C) 2014 - Alfred Agrell * * RetroArch is free software: you can redistribute it and/or modify it under the terms * of the GNU General Public License as published by the Free Software Found- * ation, either version 3 of the License, or (at your option) any later version. * * RetroArch is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; * without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR * PURPOSE. See the GNU General Public License for more details. * * You should have received a copy of the GNU General Public License along with RetroArch. * If not, see . */ #define __STDC_LIMIT_MACROS #include "rewind.h" #include "performance.h" #include #include #include #ifndef UINT16_MAX #define UINT16_MAX 0xffff #endif #ifndef UINT32_MAX #define UINT32_MAX 0xffffffffu #endif #undef CPU_X86 #if defined(__x86_64__) || defined(__i386__) || defined(__i486__) || defined(__i686__) #define CPU_X86 #endif // Other arches SIGBUS (usually) on unaligned accesses. #ifndef CPU_X86 #define NO_UNALIGNED_MEM #endif // Format per frame: // size nextstart; // repeat { // uint16 numchanged; // everything is counted in units of uint16 // if (numchanged) { // uint16 numunchanged; // skip these before handling numchanged // uint16[numchanged] changeddata; // } // else // { // uint32 numunchanged; // if (!numunchanged) break; // } // } // size thisstart; // // The start offsets point to 'nextstart' of any given compressed frame. // Each uint16 is stored native endian; anything that claims any other endianness refers to the endianness of this specific item. // The uint32 is stored little endian. // Each size value is stored native endian if alignment is not enforced; if it is, they're little endian. // The start of the buffer contains a size pointing to the end of the buffer; the end points to its start. // Wrapping is handled by returning to the start of the buffer if the compressed data could potentially hit the edge; // if the compressed data could potentially overwrite the tail pointer, the tail retreats until it can no longer collide. // This means that on average, ~2 * maxcompsize is unused at any given moment. // These are called very few constant times per frame, keep it as simple as possible. static inline void write_size_t(void *ptr, size_t val) { memcpy(ptr, &val, sizeof(val)); } static inline size_t read_size_t(const void *ptr) { size_t ret; memcpy(&ret, ptr, sizeof(ret)); return ret; } struct state_manager { uint8_t *data; size_t capacity; /* Reading and writing is done here here. */ uint8_t *head; /* If head comes close to this, discard a frame. */ uint8_t *tail; uint8_t *thisblock; uint8_t *nextblock; /* This one is rounded up from reset::blocksize. */ size_t blocksize; /* size_t + (blocksize + 131071) / 131072 * * (blocksize + u16 + u16) + u16 + u32 + size_t * (yes, the math is a bit ugly). */ size_t maxcompsize; unsigned entries; bool thisblock_valid; }; state_manager_t *state_manager_new(size_t state_size, size_t buffer_size) { state_manager_t *state = (state_manager_t*)calloc(1, sizeof(*state)); if (!state) return NULL; size_t newblocksize = ((state_size - 1) | (sizeof(uint16_t) - 1)) + 1; state->blocksize = newblocksize; const int maxcblkcover = UINT16_MAX * sizeof(uint16_t); const int maxcblks = (state->blocksize + maxcblkcover - 1) / maxcblkcover; state->maxcompsize = state->blocksize + maxcblks * sizeof(uint16_t) * 2 + sizeof(uint16_t) + sizeof(uint32_t) + sizeof(size_t) * 2; state->data = (uint8_t*)malloc(buffer_size); state->thisblock = (uint8_t*)calloc(state->blocksize + sizeof(uint16_t) * 4 + 16, 1); state->nextblock = (uint8_t*)calloc(state->blocksize + sizeof(uint16_t) * 4 + 16, 1); if (!state->data || !state->thisblock || !state->nextblock) goto error; /* Force in a different byte at the end, so we don't need to check * bounds in the innermost loop (it's expensive). * * There is also a large amount of data that's the same, to stop the other scan * There is also some padding at the end. This is so we don't read outside the * buffer end if we're reading in large blocks; * * It doesn't make any difference to us, but sacrificing 16 bytes to get * Valgrind happy is worth it. */ *(uint16_t*)(state->thisblock + state->blocksize + sizeof(uint16_t) * 3) = 0xFFFF; *(uint16_t*)(state->nextblock + state->blocksize + sizeof(uint16_t) * 3) = 0x0000; state->capacity = buffer_size; state->head = state->data + sizeof(size_t); state->tail = state->data + sizeof(size_t); return state; error: state_manager_free(state); return NULL; } void state_manager_free(state_manager_t *state) { free(state->data); free(state->thisblock); free(state->nextblock); free(state); } bool state_manager_pop(state_manager_t *state, const void **data) { *data = NULL; if (state->thisblock_valid) { state->thisblock_valid = false; state->entries--; *data = state->thisblock; return true; } if (state->head == state->tail) return false; size_t start = read_size_t(state->head - sizeof(size_t)); state->head = state->data + start; const uint8_t *compressed = state->data + start + sizeof(size_t); uint8_t *out = state->thisblock; /* Begin decompression code * out is the last pushed (or returned) state */ const uint16_t *compressed16 = (const uint16_t*)compressed; uint16_t *out16 = (uint16_t*)out; for (;;) { uint16_t i; uint16_t numchanged = *(compressed16++); if (numchanged) { out16 += *compressed16++; /* We could do memcpy, but it seems that memcpy has a * constant-per-call overhead that actually shows up. * * Our average size in here seems to be 8 or something. * Therefore, we do something with lower overhead. */ for (i = 0; i < numchanged; i++) out16[i] = compressed16[i]; compressed16 += numchanged; out16 += numchanged; } else { uint32_t numunchanged = compressed16[0] | (compressed16[1] << 16); if (!numunchanged) break; compressed16 += 2; out16 += numunchanged; } } /* End decompression code */ state->entries--; *data = state->thisblock; return true; } void state_manager_push_where(state_manager_t *state, void **data) { /* We need to ensure we have an uncompressed copy of the last * pushed state, or we could end up applying a 'patch' to wrong * savestate, and that'd blow up rather quickly. */ if (!state->thisblock_valid) { const void *ignored; if (state_manager_pop(state, &ignored)) { state->thisblock_valid = true; state->entries++; } } *data = state->nextblock; } #if __SSE2__ #if defined(__GNUC__) static inline int compat_ctz(unsigned x) { return __builtin_ctz(x); } #else /* Only checks at nibble granularity, * because that's what we need. */ static inline int compat_ctz(unsigned x) { if (x & 0x000f) return 0; if (x & 0x00f0) return 4; if (x & 0x0f00) return 8; if (x & 0xf000) return 12; return 16; } #endif #include /* There's no equivalent in libc, you'd think so ... * std::mismatch exists, but it's not optimized at all. */ static inline size_t find_change(const uint16_t *a, const uint16_t *b) { const __m128i *a128 = (const __m128i*)a; const __m128i *b128 = (const __m128i*)b; for (;;) { __m128i v0 = _mm_loadu_si128(a128); __m128i v1 = _mm_loadu_si128(b128); __m128i c = _mm_cmpeq_epi32(v0, v1); uint32_t mask = _mm_movemask_epi8(c); if (mask != 0xffff) /* Something has changed, figure out where. */ { size_t ret = (((uint8_t*)a128 - (uint8_t*)a) | (compat_ctz(~mask))) >> 1; return ret | (a[ret] == b[ret]); } a128++; b128++; } } #else static inline size_t find_change(const uint16_t *a, const uint16_t *b) { const uint16_t *a_org = a; #ifdef NO_UNALIGNED_MEM while (((uintptr_t)a & (sizeof(size_t) - 1)) && *a == *b) { a++; b++; } if (*a == *b) #endif { const size_t *a_big = (const size_t*)a; const size_t *b_big = (const size_t*)b; while (*a_big == *b_big) { a_big++; b_big++; } a = (const uint16_t*)a_big; b = (const uint16_t*)b_big; while (*a == *b) { a++; b++; } } return a - a_org; } #endif static inline size_t find_same(const uint16_t *a, const uint16_t *b) { const uint16_t *a_org = a; #ifdef NO_UNALIGNED_MEM if (((uintptr_t)a & (sizeof(uint32_t) - 1)) && *a != *b) { a++; b++; } if (*a != *b) #endif { /* With this, it's random whether two consecutive identical * words are caught. * * Luckily, compression rate is the same for both cases, and * three is always caught. * * (We prefer to miss two-word blocks, anyways; fewer iterations * of the outer loop, as well as in the decompressor.) */ const uint32_t *a_big = (const uint32_t*)a; const uint32_t *b_big = (const uint32_t*)b; while (*a_big != *b_big) { a_big++; b_big++; } a = (const uint16_t*)a_big; b = (const uint16_t*)b_big; if (a != a_org && a[-1] == b[-1]) { a--; b--; } } return a - a_org; } void state_manager_push_do(state_manager_t *state) { if (state->thisblock_valid) { if (state->capacity < sizeof(size_t) + state->maxcompsize) return; recheckcapacity:; size_t headpos = state->head - state->data; size_t tailpos = state->tail - state->data; size_t remaining = (tailpos + state->capacity - sizeof(size_t) - headpos - 1) % state->capacity + 1; if (remaining <= state->maxcompsize) { state->tail = state->data + read_size_t(state->tail); state->entries--; goto recheckcapacity; } RARCH_PERFORMANCE_INIT(gen_deltas); RARCH_PERFORMANCE_START(gen_deltas); const uint8_t *oldb = state->thisblock; const uint8_t *newb = state->nextblock; uint8_t *compressed = state->head + sizeof(size_t); /* Begin compression code; 'compressed' will point to * the end of the compressed data (excluding the prev pointer). */ const uint16_t *old16 = (const uint16_t*)oldb; const uint16_t *new16 = (const uint16_t*)newb; uint16_t *compressed16 = (uint16_t*)compressed; size_t num16s = state->blocksize / sizeof(uint16_t); while (num16s) { size_t i; size_t skip = find_change(old16, new16); if (skip >= num16s) break; old16 += skip; new16 += skip; num16s -= skip; if (skip > UINT16_MAX) { if (skip > UINT32_MAX) { /* This will make it scan the entire thing again, * but it only hits on 8GB unchanged data anyways, * and if you're doing that, you've got bigger problems. */ skip = UINT32_MAX; } *compressed16++ = 0; *compressed16++ = skip; *compressed16++ = skip >> 16; skip = 0; continue; } size_t changed = find_same(old16, new16); if (changed > UINT16_MAX) changed = UINT16_MAX; *compressed16++ = changed; *compressed16++ = skip; for (i = 0; i < changed; i++) compressed16[i] = old16[i]; old16 += changed; new16 += changed; num16s -= changed; compressed16 += changed; } compressed16[0] = 0; compressed16[1] = 0; compressed16[2] = 0; compressed = (uint8_t*)(compressed16 + 3); /* End compression code. */ if (compressed - state->data + state->maxcompsize > state->capacity) { compressed = state->data; if (state->tail == state->data + sizeof(size_t)) state->tail = state->data + read_size_t(state->tail); } write_size_t(compressed, state->head-state->data); compressed += sizeof(size_t); write_size_t(state->head, compressed-state->data); state->head = compressed; RARCH_PERFORMANCE_STOP(gen_deltas); } else state->thisblock_valid = true; uint8_t *swap = state->thisblock; state->thisblock = state->nextblock; state->nextblock = swap; state->entries++; return; } void state_manager_capacity(state_manager_t *state, unsigned *entries, size_t *bytes, bool *full) { size_t headpos = state->head - state->data; size_t tailpos = state->tail - state->data; size_t remaining = (tailpos + state->capacity - sizeof(size_t) - headpos - 1) % state->capacity + 1; if (entries) *entries = state->entries; if (bytes) *bytes = state->capacity-remaining; if (full) *full = remaining <= state->maxcompsize * 2; }