mirror of
https://github.com/xemu-project/xemu.git
synced 2024-12-19 09:48:16 +00:00
d73415a315
clang's C11 atomic_fetch_*() functions only take a C11 atomic type pointer argument. QEMU uses direct types (int, etc) and this causes a compiler error when a QEMU code calls these functions in a source file that also included <stdatomic.h> via a system header file: $ CC=clang CXX=clang++ ./configure ... && make ../util/async.c:79:17: error: address argument to atomic operation must be a pointer to _Atomic type ('unsigned int *' invalid) Avoid using atomic_*() names in QEMU's atomic.h since that namespace is used by <stdatomic.h>. Prefix QEMU's APIs with 'q' so that atomic.h and <stdatomic.h> can co-exist. I checked /usr/include on my machine and searched GitHub for existing "qatomic_" users but there seem to be none. This patch was generated using: $ git grep -h -o '\<atomic\(64\)\?_[a-z0-9_]\+' include/qemu/atomic.h | \ sort -u >/tmp/changed_identifiers $ for identifier in $(</tmp/changed_identifiers); do sed -i "s%\<$identifier\>%q$identifier%g" \ $(git grep -I -l "\<$identifier\>") done I manually fixed line-wrap issues and misaligned rST tables. Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com> Acked-by: Paolo Bonzini <pbonzini@redhat.com> Message-Id: <20200923105646.47864-1-stefanha@redhat.com>
814 lines
24 KiB
C
814 lines
24 KiB
C
/*
|
|
* qsp.c - QEMU Synchronization Profiler
|
|
*
|
|
* Copyright (C) 2018, Emilio G. Cota <cota@braap.org>
|
|
*
|
|
* License: GNU GPL, version 2 or later.
|
|
* See the COPYING file in the top-level directory.
|
|
*
|
|
* QSP profiles the time spent in synchronization primitives, which can
|
|
* help diagnose performance problems, e.g. scalability issues when
|
|
* contention is high.
|
|
*
|
|
* The primitives currently supported are mutexes, recursive mutexes and
|
|
* condition variables. Note that not all related functions are intercepted;
|
|
* instead we profile only those functions that can have a performance impact,
|
|
* either due to blocking (e.g. cond_wait, mutex_lock) or cache line
|
|
* contention (e.g. mutex_lock, mutex_trylock).
|
|
*
|
|
* QSP's design focuses on speed and scalability. This is achieved
|
|
* by having threads do their profiling entirely on thread-local data.
|
|
* The appropriate thread-local data is found via a QHT, i.e. a concurrent hash
|
|
* table. To aggregate data in order to generate a report, we iterate over
|
|
* all entries in the hash table. Depending on the number of threads and
|
|
* synchronization objects this might be expensive, but note that it is
|
|
* very rarely called -- reports are generated only when requested by users.
|
|
*
|
|
* Reports are generated as a table where each row represents a call site. A
|
|
* call site is the triplet formed by the __file__ and __LINE__ of the caller
|
|
* as well as the address of the "object" (i.e. mutex, rec. mutex or condvar)
|
|
* being operated on. Optionally, call sites that operate on different objects
|
|
* of the same type can be coalesced, which can be particularly useful when
|
|
* profiling dynamically-allocated objects.
|
|
*
|
|
* Alternative designs considered:
|
|
*
|
|
* - Use an off-the-shelf profiler such as mutrace. This is not a viable option
|
|
* for us because QEMU has __malloc_hook set (by one of the libraries it
|
|
* uses); leaving this hook unset is required to avoid deadlock in mutrace.
|
|
*
|
|
* - Use a glib HT for each thread, protecting each HT with its own lock.
|
|
* This isn't simpler than the current design, and is 10% slower in the
|
|
* atomic_add-bench microbenchmark (-m option).
|
|
*
|
|
* - For reports, just use a binary tree as we aggregate data, instead of having
|
|
* an intermediate hash table. This would simplify the code only slightly, but
|
|
* would perform badly if there were many threads and objects to track.
|
|
*
|
|
* - Wrap operations on qsp entries with RCU read-side critical sections, so
|
|
* that qsp_reset() can delete entries. Unfortunately, the overhead of calling
|
|
* rcu_read_lock/unlock slows down atomic_add-bench -m by 24%. Having
|
|
* a snapshot that is updated on qsp_reset() avoids this overhead.
|
|
*
|
|
* Related Work:
|
|
* - Lennart Poettering's mutrace: http://0pointer.de/blog/projects/mutrace.html
|
|
* - Lozi, David, Thomas, Lawall and Muller. "Remote Core Locking: Migrating
|
|
* Critical-Section Execution to Improve the Performance of Multithreaded
|
|
* Applications", USENIX ATC'12.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/qemu-print.h"
|
|
#include "qemu/thread.h"
|
|
#include "qemu/timer.h"
|
|
#include "qemu/qht.h"
|
|
#include "qemu/rcu.h"
|
|
#include "qemu/xxhash.h"
|
|
|
|
enum QSPType {
|
|
QSP_MUTEX,
|
|
QSP_BQL_MUTEX,
|
|
QSP_REC_MUTEX,
|
|
QSP_CONDVAR,
|
|
};
|
|
|
|
struct QSPCallSite {
|
|
const void *obj;
|
|
const char *file; /* i.e. __FILE__; shortened later */
|
|
int line;
|
|
enum QSPType type;
|
|
};
|
|
typedef struct QSPCallSite QSPCallSite;
|
|
|
|
struct QSPEntry {
|
|
void *thread_ptr;
|
|
const QSPCallSite *callsite;
|
|
uint64_t n_acqs;
|
|
uint64_t ns;
|
|
unsigned int n_objs; /* count of coalesced objs; only used for reporting */
|
|
};
|
|
typedef struct QSPEntry QSPEntry;
|
|
|
|
struct QSPSnapshot {
|
|
struct rcu_head rcu;
|
|
struct qht ht;
|
|
};
|
|
typedef struct QSPSnapshot QSPSnapshot;
|
|
|
|
/* initial sizing for hash tables */
|
|
#define QSP_INITIAL_SIZE 64
|
|
|
|
/* If this file is moved, QSP_REL_PATH should be updated accordingly */
|
|
#define QSP_REL_PATH "util/qsp.c"
|
|
|
|
/* this file's full path. Used to present all call sites with relative paths */
|
|
static size_t qsp_qemu_path_len;
|
|
|
|
/* the address of qsp_thread gives us a unique 'thread ID' */
|
|
static __thread int qsp_thread;
|
|
|
|
/*
|
|
* Call sites are the same for all threads, so we track them in a separate hash
|
|
* table to save memory.
|
|
*/
|
|
static struct qht qsp_callsite_ht;
|
|
|
|
static struct qht qsp_ht;
|
|
static QSPSnapshot *qsp_snapshot;
|
|
static bool qsp_initialized, qsp_initializing;
|
|
|
|
static const char * const qsp_typenames[] = {
|
|
[QSP_MUTEX] = "mutex",
|
|
[QSP_BQL_MUTEX] = "BQL mutex",
|
|
[QSP_REC_MUTEX] = "rec_mutex",
|
|
[QSP_CONDVAR] = "condvar",
|
|
};
|
|
|
|
QemuMutexLockFunc qemu_bql_mutex_lock_func = qemu_mutex_lock_impl;
|
|
QemuMutexLockFunc qemu_mutex_lock_func = qemu_mutex_lock_impl;
|
|
QemuMutexTrylockFunc qemu_mutex_trylock_func = qemu_mutex_trylock_impl;
|
|
QemuRecMutexLockFunc qemu_rec_mutex_lock_func = qemu_rec_mutex_lock_impl;
|
|
QemuRecMutexTrylockFunc qemu_rec_mutex_trylock_func =
|
|
qemu_rec_mutex_trylock_impl;
|
|
QemuCondWaitFunc qemu_cond_wait_func = qemu_cond_wait_impl;
|
|
QemuCondTimedWaitFunc qemu_cond_timedwait_func = qemu_cond_timedwait_impl;
|
|
|
|
/*
|
|
* It pays off to _not_ hash callsite->file; hashing a string is slow, and
|
|
* without it we still get a pretty unique hash.
|
|
*/
|
|
static inline
|
|
uint32_t do_qsp_callsite_hash(const QSPCallSite *callsite, uint64_t ab)
|
|
{
|
|
uint64_t cd = (uint64_t)(uintptr_t)callsite->obj;
|
|
uint32_t e = callsite->line;
|
|
uint32_t f = callsite->type;
|
|
|
|
return qemu_xxhash6(ab, cd, e, f);
|
|
}
|
|
|
|
static inline
|
|
uint32_t qsp_callsite_hash(const QSPCallSite *callsite)
|
|
{
|
|
return do_qsp_callsite_hash(callsite, 0);
|
|
}
|
|
|
|
static inline uint32_t do_qsp_entry_hash(const QSPEntry *entry, uint64_t a)
|
|
{
|
|
return do_qsp_callsite_hash(entry->callsite, a);
|
|
}
|
|
|
|
static uint32_t qsp_entry_hash(const QSPEntry *entry)
|
|
{
|
|
return do_qsp_entry_hash(entry, (uint64_t)(uintptr_t)entry->thread_ptr);
|
|
}
|
|
|
|
static uint32_t qsp_entry_no_thread_hash(const QSPEntry *entry)
|
|
{
|
|
return do_qsp_entry_hash(entry, 0);
|
|
}
|
|
|
|
/* without the objects we need to hash the file name to get a decent hash */
|
|
static uint32_t qsp_entry_no_thread_obj_hash(const QSPEntry *entry)
|
|
{
|
|
const QSPCallSite *callsite = entry->callsite;
|
|
uint64_t ab = g_str_hash(callsite->file);
|
|
uint64_t cd = callsite->line;
|
|
uint32_t e = callsite->type;
|
|
|
|
return qemu_xxhash5(ab, cd, e);
|
|
}
|
|
|
|
static bool qsp_callsite_cmp(const void *ap, const void *bp)
|
|
{
|
|
const QSPCallSite *a = ap;
|
|
const QSPCallSite *b = bp;
|
|
|
|
return a == b ||
|
|
(a->obj == b->obj &&
|
|
a->line == b->line &&
|
|
a->type == b->type &&
|
|
(a->file == b->file || !strcmp(a->file, b->file)));
|
|
}
|
|
|
|
static bool qsp_callsite_no_obj_cmp(const void *ap, const void *bp)
|
|
{
|
|
const QSPCallSite *a = ap;
|
|
const QSPCallSite *b = bp;
|
|
|
|
return a == b ||
|
|
(a->line == b->line &&
|
|
a->type == b->type &&
|
|
(a->file == b->file || !strcmp(a->file, b->file)));
|
|
}
|
|
|
|
static bool qsp_entry_no_thread_cmp(const void *ap, const void *bp)
|
|
{
|
|
const QSPEntry *a = ap;
|
|
const QSPEntry *b = bp;
|
|
|
|
return qsp_callsite_cmp(a->callsite, b->callsite);
|
|
}
|
|
|
|
static bool qsp_entry_no_thread_obj_cmp(const void *ap, const void *bp)
|
|
{
|
|
const QSPEntry *a = ap;
|
|
const QSPEntry *b = bp;
|
|
|
|
return qsp_callsite_no_obj_cmp(a->callsite, b->callsite);
|
|
}
|
|
|
|
static bool qsp_entry_cmp(const void *ap, const void *bp)
|
|
{
|
|
const QSPEntry *a = ap;
|
|
const QSPEntry *b = bp;
|
|
|
|
return a->thread_ptr == b->thread_ptr &&
|
|
qsp_callsite_cmp(a->callsite, b->callsite);
|
|
}
|
|
|
|
/*
|
|
* Normally we'd call this from a constructor function, but we want it to work
|
|
* via libutil as well.
|
|
*/
|
|
static void qsp_do_init(void)
|
|
{
|
|
/* make sure this file's path in the tree is up to date with QSP_REL_PATH */
|
|
g_assert(strstr(__FILE__, QSP_REL_PATH));
|
|
qsp_qemu_path_len = strlen(__FILE__) - strlen(QSP_REL_PATH);
|
|
|
|
qht_init(&qsp_ht, qsp_entry_cmp, QSP_INITIAL_SIZE,
|
|
QHT_MODE_AUTO_RESIZE | QHT_MODE_RAW_MUTEXES);
|
|
qht_init(&qsp_callsite_ht, qsp_callsite_cmp, QSP_INITIAL_SIZE,
|
|
QHT_MODE_AUTO_RESIZE | QHT_MODE_RAW_MUTEXES);
|
|
}
|
|
|
|
static __attribute__((noinline)) void qsp_init__slowpath(void)
|
|
{
|
|
if (qatomic_cmpxchg(&qsp_initializing, false, true) == false) {
|
|
qsp_do_init();
|
|
qatomic_set(&qsp_initialized, true);
|
|
} else {
|
|
while (!qatomic_read(&qsp_initialized)) {
|
|
cpu_relax();
|
|
}
|
|
}
|
|
}
|
|
|
|
/* qsp_init() must be called from _all_ exported functions */
|
|
static inline void qsp_init(void)
|
|
{
|
|
if (likely(qatomic_read(&qsp_initialized))) {
|
|
return;
|
|
}
|
|
qsp_init__slowpath();
|
|
}
|
|
|
|
static QSPCallSite *qsp_callsite_find(const QSPCallSite *orig)
|
|
{
|
|
QSPCallSite *callsite;
|
|
uint32_t hash;
|
|
|
|
hash = qsp_callsite_hash(orig);
|
|
callsite = qht_lookup(&qsp_callsite_ht, orig, hash);
|
|
if (callsite == NULL) {
|
|
void *existing = NULL;
|
|
|
|
callsite = g_new(QSPCallSite, 1);
|
|
memcpy(callsite, orig, sizeof(*callsite));
|
|
qht_insert(&qsp_callsite_ht, callsite, hash, &existing);
|
|
if (unlikely(existing)) {
|
|
g_free(callsite);
|
|
callsite = existing;
|
|
}
|
|
}
|
|
return callsite;
|
|
}
|
|
|
|
static QSPEntry *
|
|
qsp_entry_create(struct qht *ht, const QSPEntry *entry, uint32_t hash)
|
|
{
|
|
QSPEntry *e;
|
|
void *existing = NULL;
|
|
|
|
e = g_new0(QSPEntry, 1);
|
|
e->thread_ptr = entry->thread_ptr;
|
|
e->callsite = qsp_callsite_find(entry->callsite);
|
|
|
|
qht_insert(ht, e, hash, &existing);
|
|
if (unlikely(existing)) {
|
|
g_free(e);
|
|
e = existing;
|
|
}
|
|
return e;
|
|
}
|
|
|
|
static QSPEntry *
|
|
qsp_entry_find(struct qht *ht, const QSPEntry *entry, uint32_t hash)
|
|
{
|
|
QSPEntry *e;
|
|
|
|
e = qht_lookup(ht, entry, hash);
|
|
if (e == NULL) {
|
|
e = qsp_entry_create(ht, entry, hash);
|
|
}
|
|
return e;
|
|
}
|
|
|
|
/*
|
|
* Note: Entries are never removed, so callers do not have to be in an RCU
|
|
* read-side critical section.
|
|
*/
|
|
static QSPEntry *qsp_entry_get(const void *obj, const char *file, int line,
|
|
enum QSPType type)
|
|
{
|
|
QSPCallSite callsite = {
|
|
.obj = obj,
|
|
.file = file,
|
|
.line = line,
|
|
.type = type,
|
|
};
|
|
QSPEntry orig;
|
|
uint32_t hash;
|
|
|
|
qsp_init();
|
|
|
|
orig.thread_ptr = &qsp_thread;
|
|
orig.callsite = &callsite;
|
|
|
|
hash = qsp_entry_hash(&orig);
|
|
return qsp_entry_find(&qsp_ht, &orig, hash);
|
|
}
|
|
|
|
/*
|
|
* @e is in the global hash table; it is only written to by the current thread,
|
|
* so we write to it atomically (as in "write once") to prevent torn reads.
|
|
*/
|
|
static inline void do_qsp_entry_record(QSPEntry *e, int64_t delta, bool acq)
|
|
{
|
|
qatomic_set_u64(&e->ns, e->ns + delta);
|
|
if (acq) {
|
|
qatomic_set_u64(&e->n_acqs, e->n_acqs + 1);
|
|
}
|
|
}
|
|
|
|
static inline void qsp_entry_record(QSPEntry *e, int64_t delta)
|
|
{
|
|
do_qsp_entry_record(e, delta, true);
|
|
}
|
|
|
|
#define QSP_GEN_VOID(type_, qsp_t_, func_, impl_) \
|
|
static void func_(type_ *obj, const char *file, int line) \
|
|
{ \
|
|
QSPEntry *e; \
|
|
int64_t t0, t1; \
|
|
\
|
|
t0 = get_clock(); \
|
|
impl_(obj, file, line); \
|
|
t1 = get_clock(); \
|
|
\
|
|
e = qsp_entry_get(obj, file, line, qsp_t_); \
|
|
qsp_entry_record(e, t1 - t0); \
|
|
}
|
|
|
|
#define QSP_GEN_RET1(type_, qsp_t_, func_, impl_) \
|
|
static int func_(type_ *obj, const char *file, int line) \
|
|
{ \
|
|
QSPEntry *e; \
|
|
int64_t t0, t1; \
|
|
int err; \
|
|
\
|
|
t0 = get_clock(); \
|
|
err = impl_(obj, file, line); \
|
|
t1 = get_clock(); \
|
|
\
|
|
e = qsp_entry_get(obj, file, line, qsp_t_); \
|
|
do_qsp_entry_record(e, t1 - t0, !err); \
|
|
return err; \
|
|
}
|
|
|
|
QSP_GEN_VOID(QemuMutex, QSP_BQL_MUTEX, qsp_bql_mutex_lock, qemu_mutex_lock_impl)
|
|
QSP_GEN_VOID(QemuMutex, QSP_MUTEX, qsp_mutex_lock, qemu_mutex_lock_impl)
|
|
QSP_GEN_RET1(QemuMutex, QSP_MUTEX, qsp_mutex_trylock, qemu_mutex_trylock_impl)
|
|
|
|
QSP_GEN_VOID(QemuRecMutex, QSP_REC_MUTEX, qsp_rec_mutex_lock,
|
|
qemu_rec_mutex_lock_impl)
|
|
QSP_GEN_RET1(QemuRecMutex, QSP_REC_MUTEX, qsp_rec_mutex_trylock,
|
|
qemu_rec_mutex_trylock_impl)
|
|
|
|
#undef QSP_GEN_RET1
|
|
#undef QSP_GEN_VOID
|
|
|
|
static void
|
|
qsp_cond_wait(QemuCond *cond, QemuMutex *mutex, const char *file, int line)
|
|
{
|
|
QSPEntry *e;
|
|
int64_t t0, t1;
|
|
|
|
t0 = get_clock();
|
|
qemu_cond_wait_impl(cond, mutex, file, line);
|
|
t1 = get_clock();
|
|
|
|
e = qsp_entry_get(cond, file, line, QSP_CONDVAR);
|
|
qsp_entry_record(e, t1 - t0);
|
|
}
|
|
|
|
static bool
|
|
qsp_cond_timedwait(QemuCond *cond, QemuMutex *mutex, int ms,
|
|
const char *file, int line)
|
|
{
|
|
QSPEntry *e;
|
|
int64_t t0, t1;
|
|
bool ret;
|
|
|
|
t0 = get_clock();
|
|
ret = qemu_cond_timedwait_impl(cond, mutex, ms, file, line);
|
|
t1 = get_clock();
|
|
|
|
e = qsp_entry_get(cond, file, line, QSP_CONDVAR);
|
|
qsp_entry_record(e, t1 - t0);
|
|
return ret;
|
|
}
|
|
|
|
bool qsp_is_enabled(void)
|
|
{
|
|
return qatomic_read(&qemu_mutex_lock_func) == qsp_mutex_lock;
|
|
}
|
|
|
|
void qsp_enable(void)
|
|
{
|
|
qatomic_set(&qemu_mutex_lock_func, qsp_mutex_lock);
|
|
qatomic_set(&qemu_mutex_trylock_func, qsp_mutex_trylock);
|
|
qatomic_set(&qemu_bql_mutex_lock_func, qsp_bql_mutex_lock);
|
|
qatomic_set(&qemu_rec_mutex_lock_func, qsp_rec_mutex_lock);
|
|
qatomic_set(&qemu_rec_mutex_trylock_func, qsp_rec_mutex_trylock);
|
|
qatomic_set(&qemu_cond_wait_func, qsp_cond_wait);
|
|
qatomic_set(&qemu_cond_timedwait_func, qsp_cond_timedwait);
|
|
}
|
|
|
|
void qsp_disable(void)
|
|
{
|
|
qatomic_set(&qemu_mutex_lock_func, qemu_mutex_lock_impl);
|
|
qatomic_set(&qemu_mutex_trylock_func, qemu_mutex_trylock_impl);
|
|
qatomic_set(&qemu_bql_mutex_lock_func, qemu_mutex_lock_impl);
|
|
qatomic_set(&qemu_rec_mutex_lock_func, qemu_rec_mutex_lock_impl);
|
|
qatomic_set(&qemu_rec_mutex_trylock_func, qemu_rec_mutex_trylock_impl);
|
|
qatomic_set(&qemu_cond_wait_func, qemu_cond_wait_impl);
|
|
qatomic_set(&qemu_cond_timedwait_func, qemu_cond_timedwait_impl);
|
|
}
|
|
|
|
static gint qsp_tree_cmp(gconstpointer ap, gconstpointer bp, gpointer up)
|
|
{
|
|
const QSPEntry *a = ap;
|
|
const QSPEntry *b = bp;
|
|
enum QSPSortBy sort_by = *(enum QSPSortBy *)up;
|
|
const QSPCallSite *ca;
|
|
const QSPCallSite *cb;
|
|
|
|
switch (sort_by) {
|
|
case QSP_SORT_BY_TOTAL_WAIT_TIME:
|
|
if (a->ns > b->ns) {
|
|
return -1;
|
|
} else if (a->ns < b->ns) {
|
|
return 1;
|
|
}
|
|
break;
|
|
case QSP_SORT_BY_AVG_WAIT_TIME:
|
|
{
|
|
double avg_a = a->n_acqs ? a->ns / a->n_acqs : 0;
|
|
double avg_b = b->n_acqs ? b->ns / b->n_acqs : 0;
|
|
|
|
if (avg_a > avg_b) {
|
|
return -1;
|
|
} else if (avg_a < avg_b) {
|
|
return 1;
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
|
|
ca = a->callsite;
|
|
cb = b->callsite;
|
|
/* Break the tie with the object's address */
|
|
if (ca->obj < cb->obj) {
|
|
return -1;
|
|
} else if (ca->obj > cb->obj) {
|
|
return 1;
|
|
} else {
|
|
int cmp;
|
|
|
|
/* same obj. Break the tie with the callsite's file */
|
|
cmp = strcmp(ca->file, cb->file);
|
|
if (cmp) {
|
|
return cmp;
|
|
}
|
|
/* same callsite file. Break the tie with the callsite's line */
|
|
g_assert(ca->line != cb->line);
|
|
if (ca->line < cb->line) {
|
|
return -1;
|
|
} else if (ca->line > cb->line) {
|
|
return 1;
|
|
} else {
|
|
/* break the tie with the callsite's type */
|
|
return cb->type - ca->type;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void qsp_sort(void *p, uint32_t h, void *userp)
|
|
{
|
|
QSPEntry *e = p;
|
|
GTree *tree = userp;
|
|
|
|
g_tree_insert(tree, e, NULL);
|
|
}
|
|
|
|
static void qsp_aggregate(void *p, uint32_t h, void *up)
|
|
{
|
|
struct qht *ht = up;
|
|
const QSPEntry *e = p;
|
|
QSPEntry *agg;
|
|
uint32_t hash;
|
|
|
|
hash = qsp_entry_no_thread_hash(e);
|
|
agg = qsp_entry_find(ht, e, hash);
|
|
/*
|
|
* The entry is in the global hash table; read from it atomically (as in
|
|
* "read once").
|
|
*/
|
|
agg->ns += qatomic_read_u64(&e->ns);
|
|
agg->n_acqs += qatomic_read_u64(&e->n_acqs);
|
|
}
|
|
|
|
static void qsp_iter_diff(void *p, uint32_t hash, void *htp)
|
|
{
|
|
struct qht *ht = htp;
|
|
QSPEntry *old = p;
|
|
QSPEntry *new;
|
|
|
|
new = qht_lookup(ht, old, hash);
|
|
/* entries are never deleted, so we must have this one */
|
|
g_assert(new != NULL);
|
|
/* our reading of the stats happened after the snapshot was taken */
|
|
g_assert(new->n_acqs >= old->n_acqs);
|
|
g_assert(new->ns >= old->ns);
|
|
|
|
new->n_acqs -= old->n_acqs;
|
|
new->ns -= old->ns;
|
|
|
|
/* No point in reporting an empty entry */
|
|
if (new->n_acqs == 0 && new->ns == 0) {
|
|
bool removed = qht_remove(ht, new, hash);
|
|
|
|
g_assert(removed);
|
|
g_free(new);
|
|
}
|
|
}
|
|
|
|
static void qsp_diff(struct qht *orig, struct qht *new)
|
|
{
|
|
qht_iter(orig, qsp_iter_diff, new);
|
|
}
|
|
|
|
static void qsp_iter_callsite_coalesce(void *p, uint32_t h, void *htp)
|
|
{
|
|
struct qht *ht = htp;
|
|
QSPEntry *old = p;
|
|
QSPEntry *e;
|
|
uint32_t hash;
|
|
|
|
hash = qsp_entry_no_thread_obj_hash(old);
|
|
e = qht_lookup(ht, old, hash);
|
|
if (e == NULL) {
|
|
e = qsp_entry_create(ht, old, hash);
|
|
e->n_objs = 1;
|
|
} else if (e->callsite->obj != old->callsite->obj) {
|
|
e->n_objs++;
|
|
}
|
|
e->ns += old->ns;
|
|
e->n_acqs += old->n_acqs;
|
|
}
|
|
|
|
static void qsp_ht_delete(void *p, uint32_t h, void *htp)
|
|
{
|
|
g_free(p);
|
|
}
|
|
|
|
static void qsp_mktree(GTree *tree, bool callsite_coalesce)
|
|
{
|
|
struct qht ht, coalesce_ht;
|
|
struct qht *htp;
|
|
|
|
/*
|
|
* First, see if there's a prior snapshot, so that we read the global hash
|
|
* table _after_ the snapshot has been created, which guarantees that
|
|
* the entries we'll read will be a superset of the snapshot's entries.
|
|
*
|
|
* We must remain in an RCU read-side critical section until we're done
|
|
* with the snapshot.
|
|
*/
|
|
WITH_RCU_READ_LOCK_GUARD() {
|
|
QSPSnapshot *snap = qatomic_rcu_read(&qsp_snapshot);
|
|
|
|
/* Aggregate all results from the global hash table into a local one */
|
|
qht_init(&ht, qsp_entry_no_thread_cmp, QSP_INITIAL_SIZE,
|
|
QHT_MODE_AUTO_RESIZE | QHT_MODE_RAW_MUTEXES);
|
|
qht_iter(&qsp_ht, qsp_aggregate, &ht);
|
|
|
|
/* compute the difference wrt the snapshot, if any */
|
|
if (snap) {
|
|
qsp_diff(&snap->ht, &ht);
|
|
}
|
|
}
|
|
|
|
htp = &ht;
|
|
if (callsite_coalesce) {
|
|
qht_init(&coalesce_ht, qsp_entry_no_thread_obj_cmp, QSP_INITIAL_SIZE,
|
|
QHT_MODE_AUTO_RESIZE | QHT_MODE_RAW_MUTEXES);
|
|
qht_iter(&ht, qsp_iter_callsite_coalesce, &coalesce_ht);
|
|
|
|
/* free the previous hash table, and point htp to coalesce_ht */
|
|
qht_iter(&ht, qsp_ht_delete, NULL);
|
|
qht_destroy(&ht);
|
|
htp = &coalesce_ht;
|
|
}
|
|
|
|
/* sort the hash table elements by using a tree */
|
|
qht_iter(htp, qsp_sort, tree);
|
|
|
|
/* free the hash table, but keep the elements (those are in the tree now) */
|
|
qht_destroy(htp);
|
|
}
|
|
|
|
/* free string with g_free */
|
|
static char *qsp_at(const QSPCallSite *callsite)
|
|
{
|
|
GString *s = g_string_new(NULL);
|
|
const char *shortened;
|
|
|
|
/* remove the absolute path to qemu */
|
|
if (unlikely(strlen(callsite->file) < qsp_qemu_path_len)) {
|
|
shortened = callsite->file;
|
|
} else {
|
|
shortened = callsite->file + qsp_qemu_path_len;
|
|
}
|
|
g_string_append_printf(s, "%s:%u", shortened, callsite->line);
|
|
return g_string_free(s, FALSE);
|
|
}
|
|
|
|
struct QSPReportEntry {
|
|
const void *obj;
|
|
char *callsite_at;
|
|
const char *typename;
|
|
double time_s;
|
|
double ns_avg;
|
|
uint64_t n_acqs;
|
|
unsigned int n_objs;
|
|
};
|
|
typedef struct QSPReportEntry QSPReportEntry;
|
|
|
|
struct QSPReport {
|
|
QSPReportEntry *entries;
|
|
size_t n_entries;
|
|
size_t max_n_entries;
|
|
};
|
|
typedef struct QSPReport QSPReport;
|
|
|
|
static gboolean qsp_tree_report(gpointer key, gpointer value, gpointer udata)
|
|
{
|
|
const QSPEntry *e = key;
|
|
QSPReport *report = udata;
|
|
QSPReportEntry *entry;
|
|
|
|
if (report->n_entries == report->max_n_entries) {
|
|
return TRUE;
|
|
}
|
|
entry = &report->entries[report->n_entries];
|
|
report->n_entries++;
|
|
|
|
entry->obj = e->callsite->obj;
|
|
entry->n_objs = e->n_objs;
|
|
entry->callsite_at = qsp_at(e->callsite);
|
|
entry->typename = qsp_typenames[e->callsite->type];
|
|
entry->time_s = e->ns * 1e-9;
|
|
entry->n_acqs = e->n_acqs;
|
|
entry->ns_avg = e->n_acqs ? e->ns / e->n_acqs : 0;
|
|
return FALSE;
|
|
}
|
|
|
|
static void pr_report(const QSPReport *rep)
|
|
{
|
|
char *dashes;
|
|
size_t max_len = 0;
|
|
int callsite_len = 0;
|
|
int callsite_rspace;
|
|
int n_dashes;
|
|
size_t i;
|
|
|
|
/* find out the maximum length of all 'callsite' fields */
|
|
for (i = 0; i < rep->n_entries; i++) {
|
|
const QSPReportEntry *e = &rep->entries[i];
|
|
size_t len = strlen(e->callsite_at);
|
|
|
|
if (len > max_len) {
|
|
max_len = len;
|
|
}
|
|
}
|
|
|
|
callsite_len = MAX(max_len, strlen("Call site"));
|
|
/* white space to leave to the right of "Call site" */
|
|
callsite_rspace = callsite_len - strlen("Call site");
|
|
|
|
qemu_printf("Type Object Call site%*s Wait Time (s) "
|
|
" Count Average (us)\n", callsite_rspace, "");
|
|
|
|
/* build a horizontal rule with dashes */
|
|
n_dashes = 79 + callsite_rspace;
|
|
dashes = g_malloc(n_dashes + 1);
|
|
memset(dashes, '-', n_dashes);
|
|
dashes[n_dashes] = '\0';
|
|
qemu_printf("%s\n", dashes);
|
|
|
|
for (i = 0; i < rep->n_entries; i++) {
|
|
const QSPReportEntry *e = &rep->entries[i];
|
|
GString *s = g_string_new(NULL);
|
|
|
|
g_string_append_printf(s, "%-9s ", e->typename);
|
|
if (e->n_objs > 1) {
|
|
g_string_append_printf(s, "[%12u]", e->n_objs);
|
|
} else {
|
|
g_string_append_printf(s, "%14p", e->obj);
|
|
}
|
|
g_string_append_printf(s, " %s%*s %13.5f %12" PRIu64 " %12.2f\n",
|
|
e->callsite_at,
|
|
callsite_len - (int)strlen(e->callsite_at), "",
|
|
e->time_s, e->n_acqs, e->ns_avg * 1e-3);
|
|
qemu_printf("%s", s->str);
|
|
g_string_free(s, TRUE);
|
|
}
|
|
|
|
qemu_printf("%s\n", dashes);
|
|
g_free(dashes);
|
|
}
|
|
|
|
static void report_destroy(QSPReport *rep)
|
|
{
|
|
size_t i;
|
|
|
|
for (i = 0; i < rep->n_entries; i++) {
|
|
QSPReportEntry *e = &rep->entries[i];
|
|
|
|
g_free(e->callsite_at);
|
|
}
|
|
g_free(rep->entries);
|
|
}
|
|
|
|
void qsp_report(size_t max, enum QSPSortBy sort_by,
|
|
bool callsite_coalesce)
|
|
{
|
|
GTree *tree = g_tree_new_full(qsp_tree_cmp, &sort_by, g_free, NULL);
|
|
QSPReport rep;
|
|
|
|
qsp_init();
|
|
|
|
rep.entries = g_new0(QSPReportEntry, max);
|
|
rep.n_entries = 0;
|
|
rep.max_n_entries = max;
|
|
|
|
qsp_mktree(tree, callsite_coalesce);
|
|
g_tree_foreach(tree, qsp_tree_report, &rep);
|
|
g_tree_destroy(tree);
|
|
|
|
pr_report(&rep);
|
|
report_destroy(&rep);
|
|
}
|
|
|
|
static void qsp_snapshot_destroy(QSPSnapshot *snap)
|
|
{
|
|
qht_iter(&snap->ht, qsp_ht_delete, NULL);
|
|
qht_destroy(&snap->ht);
|
|
g_free(snap);
|
|
}
|
|
|
|
void qsp_reset(void)
|
|
{
|
|
QSPSnapshot *new = g_new(QSPSnapshot, 1);
|
|
QSPSnapshot *old;
|
|
|
|
qsp_init();
|
|
|
|
qht_init(&new->ht, qsp_entry_cmp, QSP_INITIAL_SIZE,
|
|
QHT_MODE_AUTO_RESIZE | QHT_MODE_RAW_MUTEXES);
|
|
|
|
/* take a snapshot of the current state */
|
|
qht_iter(&qsp_ht, qsp_aggregate, &new->ht);
|
|
|
|
/* replace the previous snapshot, if any */
|
|
old = qatomic_xchg(&qsp_snapshot, new);
|
|
if (old) {
|
|
call_rcu(old, qsp_snapshot_destroy, rcu);
|
|
}
|
|
}
|