RetroArch/state_manager.c

837 lines
23 KiB
C
Raw Normal View History

2012-04-21 21:13:50 +00:00
/* RetroArch - A frontend for libretro.
2014-01-01 00:50:59 +00:00
* Copyright (C) 2010-2014 - Hans-Kristian Arntzen
2017-01-22 12:40:32 +00:00
* Copyright (C) 2011-2017 - Daniel De Matteis
* Copyright (C) 2014-2017 - Alfred Agrell
*
2012-04-21 21:13:50 +00:00
* RetroArch is free software: you can redistribute it and/or modify it under the terms
2011-01-31 15:06:46 +00:00
* of the GNU General Public License as published by the Free Software Found-
* ation, either version 3 of the License, or (at your option) any later version.
*
2012-04-21 21:13:50 +00:00
* RetroArch is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY;
2011-01-31 15:06:46 +00:00
* without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR
* PURPOSE. See the GNU General Public License for more details.
*
2012-04-21 21:31:57 +00:00
* You should have received a copy of the GNU General Public License along with RetroArch.
2011-01-31 15:06:46 +00:00
* If not, see <http://www.gnu.org/licenses/>.
*/
2018-06-25 19:52:24 +00:00
#ifndef __STDC_LIMIT_MACROS
#define __STDC_LIMIT_MACROS
2018-06-25 19:52:24 +00:00
#endif
2017-01-25 01:57:43 +00:00
#include <stdint.h>
2011-01-31 15:06:46 +00:00
#include <stdlib.h>
#include <string.h>
2015-09-04 19:11:00 +00:00
#include <retro_inline.h>
2017-01-25 01:53:58 +00:00
#include <compat/strl.h>
2017-01-25 01:57:43 +00:00
#include <compat/intrinsics.h>
2015-09-04 19:11:00 +00:00
2016-05-09 05:14:10 +00:00
#include "state_manager.h"
#include "msg_hash.h"
#include "core.h"
#include "core_info.h"
#include "retroarch.h"
#include "verbosity.h"
#include "content.h"
#include "audio/audio_driver.h"
2011-01-31 15:06:46 +00:00
#ifdef HAVE_NETWORKING
#include "network/netplay/netplay.h"
#endif
/* This makes Valgrind throw errors if a core overflows its savestate size. */
/* Keep it off unless you're chasing a core bug, it slows things down. */
#define STRICT_BUF_SIZE 0
2015-09-15 17:02:39 +00:00
#ifndef UINT16_MAX
#define UINT16_MAX 0xffff
2011-11-01 20:28:32 +00:00
#endif
2015-07-02 16:24:55 +00:00
2015-09-15 17:02:39 +00:00
#ifndef UINT32_MAX
#define UINT32_MAX 0xffffffffu
2011-11-01 20:28:32 +00:00
#endif
2019-01-01 14:30:21 +00:00
#if defined(__x86_64__) || defined(__i386__) || defined(__i486__) || defined(__i686__) || defined(_M_IX86) || defined(_M_AMD64) || defined(_M_X64)
2017-01-25 01:57:43 +00:00
#define CPU_X86
#endif
/* Other arches SIGBUS (usually) on unaligned accesses. */
#ifndef CPU_X86
#define NO_UNALIGNED_MEM
#endif
#if __SSE2__
#include <emmintrin.h>
#endif
2020-05-29 08:23:48 +00:00
/* Format per frame (pseudocode): */
#if 0
size nextstart;
repeat {
uint16 numchanged; /* everything is counted in units of uint16 */
if (numchanged)
{
uint16 numunchanged; /* skip these before handling numchanged */
uint16[numchanged] changeddata;
}
else
{
uint32 numunchanged;
if (!numunchanged)
break;
}
}
size thisstart;
#endif
2017-01-25 01:57:43 +00:00
/* There's no equivalent in libc, you'd think so ...
* std::mismatch exists, but it's not optimized at all. */
static size_t find_change(const uint16_t *a, const uint16_t *b)
{
#if __SSE2__
const __m128i *a128 = (const __m128i*)a;
const __m128i *b128 = (const __m128i*)b;
2017-01-25 01:57:43 +00:00
for (;;)
{
__m128i v0 = _mm_loadu_si128(a128);
__m128i v1 = _mm_loadu_si128(b128);
2021-01-22 05:26:45 +00:00
__m128i c = _mm_cmpeq_epi8(v0, v1);
2017-01-25 01:57:43 +00:00
uint32_t mask = _mm_movemask_epi8(c);
if (mask != 0xffff) /* Something has changed, figure out where. */
{
2021-01-22 05:26:45 +00:00
/* calculate the real offset to the differing byte */
2017-01-25 01:57:43 +00:00
size_t ret = (((uint8_t*)a128 - (uint8_t*)a) |
2021-01-22 05:26:45 +00:00
(compat_ctz(~mask)));
/* and convert that to the uint16_t offset */
return (ret >> 1);
2017-01-25 01:57:43 +00:00
}
a128++;
b128++;
}
#else
const uint16_t *a_org = a;
#ifdef NO_UNALIGNED_MEM
while (((uintptr_t)a & (sizeof(size_t) - 1)) && *a == *b)
{
a++;
b++;
}
if (*a == *b)
#endif
{
const size_t *a_big = (const size_t*)a;
const size_t *b_big = (const size_t*)b;
2017-01-25 01:57:43 +00:00
while (*a_big == *b_big)
{
a_big++;
b_big++;
}
a = (const uint16_t*)a_big;
b = (const uint16_t*)b_big;
2017-01-25 01:57:43 +00:00
while (*a == *b)
{
a++;
b++;
}
}
return a - a_org;
#endif
}
static size_t find_same(const uint16_t *a, const uint16_t *b)
{
const uint16_t *a_org = a;
#ifdef NO_UNALIGNED_MEM
if (((uintptr_t)a & (sizeof(uint32_t) - 1)) && *a != *b)
{
a++;
b++;
}
if (*a != *b)
#endif
{
/* With this, it's random whether two consecutive identical
* words are caught.
*
* Luckily, compression rate is the same for both cases, and
2017-01-25 01:57:43 +00:00
* three is always caught.
*
* (We prefer to miss two-word blocks, anyways; fewer iterations
2017-01-25 01:57:43 +00:00
* of the outer loop, as well as in the decompressor.) */
const uint32_t *a_big = (const uint32_t*)a;
const uint32_t *b_big = (const uint32_t*)b;
2017-01-25 01:57:43 +00:00
while (*a_big != *b_big)
{
a_big++;
b_big++;
}
a = (const uint16_t*)a_big;
b = (const uint16_t*)b_big;
2017-01-25 01:57:43 +00:00
if (a != a_org && a[-1] == b[-1])
{
a--;
b--;
}
}
return a - a_org;
}
/* Returns the maximum compressed size of a savestate.
* It is very likely to compress to far less. */
static size_t state_manager_raw_maxsize(size_t uncomp)
{
2015-07-02 03:18:41 +00:00
/* bytes covered by a compressed block */
2015-09-15 17:02:39 +00:00
const int maxcblkcover = UINT16_MAX * sizeof(uint16_t);
2015-07-02 03:18:41 +00:00
/* uncompressed size, rounded to 16 bits */
size_t uncomp16 = (uncomp + sizeof(uint16_t) - 1) & -sizeof(uint16_t);
2015-07-02 03:18:41 +00:00
/* number of blocks */
size_t maxcblks = (uncomp + maxcblkcover - 1) / maxcblkcover;
return uncomp16 + maxcblks * sizeof(uint16_t) * 2 /* two u16 overhead per block */ + sizeof(uint16_t) *
3; /* three u16 to end it */
}
/*
* See state_manager_raw_compress for information about this.
* When you're done with it, send it to free().
*/
static void *state_manager_raw_alloc(size_t len, uint16_t uniq)
{
size_t len16 = (len + sizeof(uint16_t) - 1) & -sizeof(uint16_t);
uint16_t *ret = (uint16_t*)calloc(len16 + sizeof(uint16_t) * 4 + 16, 1);
if (!ret)
return NULL;
/* Force in a different byte at the end, so we don't need to check
* bounds in the innermost loop (it's expensive).
*
* There is also a large amount of data that's the same, to stop
* the other scan.
*
* There is also some padding at the end. This is so we don't
* read outside the buffer end if we're reading in large blocks;
*
* It doesn't make any difference to us, but sacrificing 16 bytes to get
* Valgrind happy is worth it. */
ret[len16/sizeof(uint16_t) + 3] = uniq;
return ret;
}
/*
* Takes two savestates and creates a patch that turns 'src' into 'dst'.
* Both 'src' and 'dst' must be returned from state_manager_raw_alloc(),
* with the same 'len', and different 'uniq'.
*
* 'patch' must be size 'state_manager_raw_maxsize(len)' or more.
* Returns the number of bytes actually written to 'patch'.
*/
static size_t state_manager_raw_compress(const void *src,
2015-07-02 16:24:55 +00:00
const void *dst, size_t len, void *patch)
{
2015-07-02 03:18:41 +00:00
const uint16_t *old16 = (const uint16_t*)src;
const uint16_t *new16 = (const uint16_t*)dst;
uint16_t *compressed16 = (uint16_t*)patch;
size_t num16s = (len + sizeof(uint16_t) - 1)
2015-07-02 16:24:55 +00:00
/ sizeof(uint16_t);
while (num16s)
{
2015-06-29 19:39:00 +00:00
size_t i, changed;
size_t skip = find_change(old16, new16);
if (skip >= num16s)
break;
2016-05-02 14:57:28 +00:00
old16 += skip;
new16 += skip;
num16s -= skip;
2015-09-15 17:02:39 +00:00
if (skip > UINT16_MAX)
{
2020-09-16 16:32:13 +00:00
/* This will make it scan the entire thing again,
* but it only hits on 8GB unchanged data anyways,
* and if you're doing that, you've got bigger problems. */
2015-09-15 17:02:39 +00:00
if (skip > UINT32_MAX)
2020-09-16 16:32:13 +00:00
skip = UINT32_MAX;
*compressed16++ = 0;
*compressed16++ = skip;
*compressed16++ = skip >> 16;
continue;
}
2020-09-16 16:32:13 +00:00
changed = find_same(old16, new16);
2015-09-15 17:02:39 +00:00
if (changed > UINT16_MAX)
changed = UINT16_MAX;
*compressed16++ = changed;
*compressed16++ = skip;
for (i = 0; i < changed; i++)
compressed16[i] = old16[i];
2020-09-16 16:32:13 +00:00
old16 += changed;
new16 += changed;
num16s -= changed;
compressed16 += changed;
}
2020-09-16 16:32:13 +00:00
compressed16[0] = 0;
compressed16[1] = 0;
compressed16[2] = 0;
2020-09-16 16:32:13 +00:00
return (uint8_t*)(compressed16 + 3) - (uint8_t*)patch;
}
/*
* Takes 'patch' from a previous call to 'state_manager_raw_compress'
* and applies it to 'data' ('src' from that call),
* yielding 'dst' in that call.
*
* If the given arguments do not match a previous call to
* state_manager_raw_compress(), anything at all can happen.
*/
static void state_manager_raw_decompress(const void *patch,
2015-07-02 03:18:41 +00:00
size_t patchlen, void *data, size_t datalen)
{
2015-07-02 03:18:41 +00:00
uint16_t *out16 = (uint16_t*)data;
const uint16_t *patch16 = (const uint16_t*)patch;
for (;;)
{
2020-09-16 16:32:13 +00:00
uint16_t numchanged = *(patch16++);
if (numchanged)
{
uint16_t i;
2020-09-16 16:32:13 +00:00
out16 += *patch16++;
/* We could do memcpy, but it seems that memcpy has a
* constant-per-call overhead that actually shows up.
*
* Our average size in here seems to be 8 or something.
* Therefore, we do something with lower overhead. */
for (i = 0; i < numchanged; i++)
2020-09-16 16:32:13 +00:00
out16[i] = patch16[i];
2020-09-16 16:32:13 +00:00
patch16 += numchanged;
out16 += numchanged;
}
else
{
uint32_t numunchanged = patch16[0] | (patch16[1] << 16);
if (!numunchanged)
break;
patch16 += 2;
2020-09-16 16:32:13 +00:00
out16 += numunchanged;
}
}
}
2014-09-09 16:31:44 +00:00
/* The start offsets point to 'nextstart' of any given compressed frame.
* Each uint16 is stored native endian; anything that claims any other
2014-09-09 16:31:44 +00:00
* endianness refers to the endianness of this specific item.
* The uint32 is stored little endian.
*
* Each size value is stored native endian if alignment is not enforced;
2014-09-09 16:31:44 +00:00
* if it is, they're little endian.
*
* The start of the buffer contains a size pointing to the end of the
2014-09-09 16:31:44 +00:00
* buffer; the end points to its start.
*
* Wrapping is handled by returning to the start of the buffer if the
2014-09-09 16:31:44 +00:00
* compressed data could potentially hit the edge;
*
* if the compressed data could potentially overwrite the tail pointer,
2014-09-09 16:31:44 +00:00
* the tail retreats until it can no longer collide.
*
* This means that on average, ~2 * maxcompsize is
2014-09-09 16:31:44 +00:00
* unused at any given moment. */
/* These are called very few constant times per frame,
2014-09-09 16:31:44 +00:00
* keep it as simple as possible. */
static INLINE void write_size_t(void *ptr, size_t val)
{
memcpy(ptr, &val, sizeof(val));
}
2014-09-09 16:31:44 +00:00
static INLINE size_t read_size_t(const void *ptr)
{
size_t ret;
2015-01-09 20:39:44 +00:00
memcpy(&ret, ptr, sizeof(ret));
return ret;
}
2016-02-01 11:46:46 +00:00
static void state_manager_free(state_manager_t *state)
{
if (!state)
return;
2017-01-09 01:49:39 +00:00
if (state->data)
free(state->data);
if (state->thisblock)
free(state->thisblock);
if (state->nextblock)
free(state->nextblock);
2016-02-01 11:46:46 +00:00
#if STRICT_BUF_SIZE
2017-01-09 01:49:39 +00:00
if (state->debugblock)
free(state->debugblock);
state->debugblock = NULL;
2016-02-01 11:46:46 +00:00
#endif
2017-01-09 01:49:39 +00:00
state->data = NULL;
state->thisblock = NULL;
state->nextblock = NULL;
2016-02-01 11:46:46 +00:00
}
2020-09-16 16:32:13 +00:00
static state_manager_t *state_manager_new(
size_t state_size, size_t buffer_size)
{
2017-01-09 01:49:39 +00:00
size_t max_comp_size, block_size;
uint8_t *next_block = NULL;
uint8_t *this_block = NULL;
uint8_t *state_data = NULL;
2011-12-24 12:46:12 +00:00
state_manager_t *state = (state_manager_t*)calloc(1, sizeof(*state));
2015-01-09 20:39:44 +00:00
2011-01-31 15:06:46 +00:00
if (!state)
return NULL;
2017-01-09 01:49:39 +00:00
block_size = (state_size + sizeof(uint16_t) - 1) & -sizeof(uint16_t);
/* the compressed data is surrounded by pointers to the other side */
2017-01-09 01:49:39 +00:00
max_comp_size = state_manager_raw_maxsize(state_size) + sizeof(size_t) * 2;
state_data = (uint8_t*)malloc(buffer_size);
2017-01-09 01:49:39 +00:00
if (!state_data)
2016-02-01 11:46:46 +00:00
goto error;
2017-01-09 01:49:39 +00:00
this_block = (uint8_t*)state_manager_raw_alloc(state_size, 0);
next_block = (uint8_t*)state_manager_raw_alloc(state_size, 1);
2016-02-01 11:46:46 +00:00
2017-01-09 01:49:39 +00:00
if (!this_block || !next_block)
2011-01-31 15:06:46 +00:00
goto error;
2017-01-09 01:49:39 +00:00
state->blocksize = block_size;
state->maxcompsize = max_comp_size;
state->data = state_data;
state->thisblock = this_block;
state->nextblock = next_block;
state->capacity = buffer_size;
2017-01-09 01:49:39 +00:00
state->head = state->data + sizeof(size_t);
state->tail = state->data + sizeof(size_t);
2011-01-31 15:06:46 +00:00
#if STRICT_BUF_SIZE
2017-01-09 01:49:39 +00:00
state->debugsize = state_size;
state->debugblock = (uint8_t*)malloc(state_size);
#endif
2011-01-31 15:06:46 +00:00
return state;
error:
2017-01-09 01:49:39 +00:00
if (state_data)
free(state_data);
state_manager_free(state);
2017-01-09 01:49:39 +00:00
free(state);
2016-02-01 11:46:46 +00:00
2011-01-31 15:06:46 +00:00
return NULL;
}
static bool state_manager_pop(state_manager_t *state, const void **data)
{
2015-01-09 20:39:44 +00:00
size_t start;
2015-06-12 15:10:16 +00:00
uint8_t *out = NULL;
const uint8_t *compressed = NULL;
2015-01-09 20:39:44 +00:00
2020-09-16 16:32:13 +00:00
*data = NULL;
if (state->thisblock_valid)
2011-02-26 00:13:52 +00:00
{
2020-09-16 16:32:13 +00:00
state->thisblock_valid = false;
state->entries--;
2020-09-16 16:32:13 +00:00
*data = state->thisblock;
2011-02-26 00:13:52 +00:00
return true;
}
2011-01-31 15:06:46 +00:00
2020-09-16 16:32:13 +00:00
*data = state->thisblock;
if (state->head == state->tail)
2011-01-31 15:06:46 +00:00
return false;
2020-09-16 16:32:13 +00:00
start = read_size_t(state->head - sizeof(size_t));
state->head = state->data + start;
compressed = state->data + start + sizeof(size_t);
out = state->thisblock;
2016-01-26 04:56:53 +00:00
state_manager_raw_decompress(compressed,
state->maxcompsize, out, state->blocksize);
2011-01-31 15:06:46 +00:00
state->entries--;
2011-01-31 15:06:46 +00:00
return true;
}
static void state_manager_push_where(state_manager_t *state, void **data)
2011-01-31 15:06:46 +00:00
{
2014-09-02 03:57:53 +00:00
/* We need to ensure we have an uncompressed copy of the last
* pushed state, or we could end up applying a 'patch' to wrong
2014-09-02 03:57:53 +00:00
* savestate, and that'd blow up rather quickly. */
if (!state->thisblock_valid)
{
const void *ignored;
if (state_manager_pop(state, &ignored))
{
state->thisblock_valid = true;
state->entries++;
}
}
*data = state->nextblock;
#if STRICT_BUF_SIZE
*data = state->debugblock;
#endif
2011-01-31 15:06:46 +00:00
}
static void state_manager_push_do(state_manager_t *state)
{
uint8_t *swap = NULL;
#if STRICT_BUF_SIZE
memcpy(state->nextblock, state->debugblock, state->debugsize);
#endif
if (state->thisblock_valid)
2011-01-31 15:06:46 +00:00
{
2015-06-26 17:55:09 +00:00
const uint8_t *oldb, *newb;
uint8_t *compressed;
size_t headpos, tailpos, remaining;
if (state->capacity < sizeof(size_t) + state->maxcompsize) {
RARCH_ERR("State capacity insufficient\n");
return;
}
recheckcapacity:;
2020-09-16 16:32:13 +00:00
headpos = state->head - state->data;
tailpos = state->tail - state->data;
2015-06-26 17:55:09 +00:00
remaining = (tailpos + state->capacity -
2014-09-02 03:57:53 +00:00
sizeof(size_t) - headpos - 1) % state->capacity + 1;
if (remaining <= state->maxcompsize)
{
state->tail = state->data + read_size_t(state->tail);
state->entries--;
goto recheckcapacity;
}
2020-09-16 16:32:13 +00:00
oldb = state->thisblock;
newb = state->nextblock;
compressed = state->head + sizeof(size_t);
2020-09-16 16:32:13 +00:00
compressed += state_manager_raw_compress(oldb, newb,
2016-01-26 04:56:53 +00:00
state->blocksize, compressed);
if (compressed - state->data + state->maxcompsize > state->capacity)
{
2020-09-16 16:32:13 +00:00
compressed = state->data;
if (state->tail == state->data + sizeof(size_t))
state->tail = state->data + read_size_t(state->tail);
}
write_size_t(compressed, state->head-state->data);
2020-09-16 16:32:13 +00:00
compressed += sizeof(size_t);
write_size_t(state->head, compressed-state->data);
2020-09-16 16:32:13 +00:00
state->head = compressed;
2011-01-31 15:06:46 +00:00
}
else
state->thisblock_valid = true;
2011-01-31 15:06:46 +00:00
2020-09-16 16:32:13 +00:00
swap = state->thisblock;
state->thisblock = state->nextblock;
state->nextblock = swap;
state->entries++;
2011-01-31 15:06:46 +00:00
}
#if 0
static void state_manager_capacity(state_manager_t *state,
2014-09-02 03:57:53 +00:00
unsigned *entries, size_t *bytes, bool *full)
2011-01-31 15:06:46 +00:00
{
2015-01-09 20:39:44 +00:00
size_t headpos = state->head - state->data;
size_t tailpos = state->tail - state->data;
2014-09-02 03:57:53 +00:00
size_t remaining = (tailpos + state->capacity -
sizeof(size_t) - headpos - 1) % state->capacity + 1;
if (entries)
2020-09-16 16:32:13 +00:00
*entries = state->entries;
if (bytes)
2020-09-16 16:32:13 +00:00
*bytes = state->capacity-remaining;
if (full)
2020-09-16 16:32:13 +00:00
*full = remaining <= state->maxcompsize * 2;
2011-01-31 15:06:46 +00:00
}
#endif
2015-04-11 11:36:33 +00:00
void state_manager_event_init(
struct state_manager_rewind_state *rewind_st,
unsigned rewind_buffer_size)
2015-04-11 11:36:33 +00:00
{
core_info_t *core_info = NULL;
void *state = NULL;
2015-04-11 11:36:33 +00:00
if ( !rewind_st
|| (rewind_st->flags & STATE_MGR_REWIND_ST_FLAG_INIT_ATTEMPTED)
|| rewind_st->state)
2015-04-11 11:36:33 +00:00
return;
rewind_st->size = 0;
rewind_st->flags &= ~(
STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED
| STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_CHECKED
| STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED
);
rewind_st->flags |= STATE_MGR_REWIND_ST_FLAG_INIT_ATTEMPTED;
/* We cannot initialise the rewind buffer
* unless the core info struct for the current
* core has been initialised (i.e. without this,
* the savestate support level for the current
* core is unknown) */
if (!core_info_get_current_core(&core_info) || !core_info)
return;
if (!core_info_current_supports_rewind())
{
RARCH_ERR("%s\n", msg_hash_to_str(MSG_REWIND_UNSUPPORTED));
return;
}
2016-05-08 15:03:51 +00:00
if (audio_driver_has_callback())
2015-04-11 11:36:33 +00:00
{
RARCH_ERR("%s.\n", msg_hash_to_str(MSG_REWIND_INIT_FAILED_THREADED_AUDIO));
2015-04-11 11:36:33 +00:00
return;
}
rewind_st->size = content_get_serialized_size_rewind();
2015-04-11 11:36:33 +00:00
if (!rewind_st->size)
2015-04-11 11:36:33 +00:00
{
2015-07-02 16:23:18 +00:00
RARCH_ERR("%s.\n",
msg_hash_to_str(MSG_REWIND_INIT_FAILED));
2015-04-11 11:36:33 +00:00
return;
}
2015-07-02 16:23:18 +00:00
RARCH_LOG("%s: %u MB\n",
msg_hash_to_str(MSG_REWIND_INIT),
(unsigned)(rewind_buffer_size / 1000000));
2015-04-11 11:36:33 +00:00
rewind_st->state = state_manager_new(rewind_st->size,
rewind_buffer_size);
2015-04-11 11:36:33 +00:00
if (!rewind_st->state)
2015-07-02 16:23:18 +00:00
RARCH_WARN("%s.\n", msg_hash_to_str(MSG_REWIND_INIT_FAILED));
2015-04-11 11:36:33 +00:00
state_manager_push_where(rewind_st->state, &state);
2016-01-27 06:23:50 +00:00
content_serialize_state_rewind(state, rewind_st->size);
2016-01-27 06:23:50 +00:00
state_manager_push_do(rewind_st->state);
2015-04-11 11:36:33 +00:00
}
void state_manager_event_deinit(
struct state_manager_rewind_state *rewind_st,
struct retro_core_t *current_core)
{
bool restore_callbacks = false;
if (!rewind_st)
return;
restore_callbacks =
(rewind_st->flags & STATE_MGR_REWIND_ST_FLAG_INIT_ATTEMPTED)
&& (rewind_st->state)
&& (current_core);
if (rewind_st->state)
2017-01-09 01:49:39 +00:00
{
state_manager_free(rewind_st->state);
free(rewind_st->state);
2017-01-09 01:49:39 +00:00
}
rewind_st->state = NULL;
rewind_st->size = 0;
rewind_st->flags &= ~(
STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED
| STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_CHECKED
| STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED
| STATE_MGR_REWIND_ST_FLAG_INIT_ATTEMPTED
);
/* Restore regular (non-rewind) core audio
* callbacks if required */
if (restore_callbacks)
{
if (current_core->retro_set_audio_sample)
current_core->retro_set_audio_sample(audio_driver_sample);
if (current_core->retro_set_audio_sample_batch)
current_core->retro_set_audio_sample_batch(audio_driver_sample_batch);
}
2015-11-30 21:25:13 +00:00
}
/**
* check_rewind:
* @pressed : was rewind key pressed or held?
*
* Checks if rewind toggle/hold was being pressed and/or held.
**/
bool state_manager_check_rewind(
struct state_manager_rewind_state *rewind_st,
struct retro_core_t *current_core,
bool pressed,
2017-01-25 01:53:58 +00:00
unsigned rewind_granularity, bool is_paused,
char *s, size_t len, unsigned *time)
2015-11-30 21:25:13 +00:00
{
bool ret = false;
2017-04-28 13:31:33 +00:00
#ifdef HAVE_NETWORKING
bool was_reversed = false;
2017-04-28 13:31:33 +00:00
#endif
2015-11-30 21:25:13 +00:00
if ( !rewind_st
|| (!(rewind_st->flags & STATE_MGR_REWIND_ST_FLAG_INIT_ATTEMPTED)))
return false;
if (!(rewind_st->flags & STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_CHECKED))
2015-11-30 21:25:13 +00:00
{
rewind_st->flags |= STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_CHECKED;
2017-01-25 01:53:58 +00:00
return false;
2015-11-30 21:25:13 +00:00
}
if (!rewind_st->state)
{
if ((pressed
&& (!(rewind_st->flags
& STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED)))
&& !core_info_current_supports_rewind())
runloop_msg_queue_push(msg_hash_to_str(MSG_REWIND_UNSUPPORTED),
1, 100, false, NULL,
MESSAGE_QUEUE_ICON_DEFAULT, MESSAGE_QUEUE_CATEGORY_INFO);
if (pressed)
rewind_st->flags |= STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED;
else
rewind_st->flags &= ~STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED;
2017-01-25 01:53:58 +00:00
return false;
}
2015-11-30 21:25:13 +00:00
if (rewind_st->flags & STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED)
{
#ifdef HAVE_NETWORKING
was_reversed = true;
#endif
audio_driver_frame_is_reverse();
rewind_st->flags &= ~STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED;
}
2015-11-30 21:25:13 +00:00
if (pressed)
{
const void *buf = NULL;
if (state_manager_pop(rewind_st->state, &buf))
2015-11-30 21:25:13 +00:00
{
#ifdef HAVE_NETWORKING
2017-04-28 13:31:33 +00:00
/* Make sure netplay isn't confused */
if (!was_reversed)
netplay_driver_ctl(RARCH_NETPLAY_CTL_DESYNC_PUSH, NULL);
#endif
rewind_st->flags |= STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED;
2016-05-08 14:28:24 +00:00
audio_driver_setup_rewind();
2015-11-30 21:25:13 +00:00
2017-01-25 01:53:58 +00:00
strlcpy(s, msg_hash_to_str(MSG_REWINDING), len);
*time = is_paused ? 1 : 30;
ret = true;
2016-01-27 06:23:50 +00:00
content_deserialize_state(buf, rewind_st->size);
#ifdef HAVE_BSV_MOVIE
bsv_movie_frame_rewind();
#endif
2015-11-30 21:25:13 +00:00
}
else
2017-01-18 17:56:52 +00:00
{
content_deserialize_state(buf, rewind_st->size);
2017-01-25 01:53:58 +00:00
#ifdef HAVE_NETWORKING
2017-04-28 13:31:33 +00:00
/* Tell netplay we're done */
if (was_reversed)
netplay_driver_ctl(RARCH_NETPLAY_CTL_DESYNC_POP, NULL);
#endif
strlcpy(s,
msg_hash_to_str(MSG_REWIND_REACHED_END),
2017-01-25 01:53:58 +00:00
len);
2017-01-25 01:53:58 +00:00
*time = 30;
ret = true;
2017-01-18 17:56:52 +00:00
}
2015-11-30 21:25:13 +00:00
}
else
{
static unsigned cnt = 0;
#ifdef HAVE_NETWORKING
2017-04-28 13:31:33 +00:00
/* Tell netplay we're done */
if (was_reversed)
netplay_driver_ctl(RARCH_NETPLAY_CTL_DESYNC_POP, NULL);
#endif
cnt = (cnt + 1) % (rewind_granularity ?
rewind_granularity : 1); /* Avoid possible SIGFPE. */
2015-11-30 21:25:13 +00:00
if ( !is_paused
&& ((cnt == 0) || retroarch_ctl(RARCH_CTL_BSV_MOVIE_IS_INITED, NULL)))
2015-11-30 21:25:13 +00:00
{
void *state = NULL;
state_manager_push_where(rewind_st->state, &state);
2015-11-30 21:25:13 +00:00
content_serialize_state_rewind(state, rewind_st->size);
2016-01-27 06:23:50 +00:00
state_manager_push_do(rewind_st->state);
2015-11-30 21:25:13 +00:00
}
}
/* Update core audio callbacks */
if (current_core)
{
if (current_core->retro_set_audio_sample)
current_core->retro_set_audio_sample(
(rewind_st->flags
& STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED)
? audio_driver_sample_rewind
: audio_driver_sample);
if (current_core->retro_set_audio_sample_batch)
current_core->retro_set_audio_sample_batch(
( rewind_st->flags
& STATE_MGR_REWIND_ST_FLAG_FRAME_IS_REVERSED)
? audio_driver_sample_batch_rewind
: audio_driver_sample_batch);
}
2017-01-25 01:53:58 +00:00
if (pressed)
rewind_st->flags |= STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED;
else
rewind_st->flags &= ~STATE_MGR_REWIND_ST_FLAG_HOTKEY_WAS_PRESSED;
2017-01-25 01:53:58 +00:00
return ret;
2015-11-30 21:25:13 +00:00
}