mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-24 05:11:16 +00:00
bea11a62d5
nsIGfxInfo::FEATURE_HARDWARE_VIDEO_DECODING is used on all platforms so let's use it on Linux too and don't add new feature for Linux only. Differential Revision: https://phabricator.services.mozilla.com/D149765
999 lines
32 KiB
C++
999 lines
32 KiB
C++
/* -*- Mode: C++; tab-width: 20; indent-tabs-mode: nil; c-basic-offset: 2 -*-
|
|
* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#define PANGO_ENABLE_BACKEND
|
|
#define PANGO_ENABLE_ENGINE
|
|
|
|
#include "gfxPlatformGtk.h"
|
|
|
|
#include <gtk/gtk.h>
|
|
#include <fontconfig/fontconfig.h>
|
|
|
|
#include "base/task.h"
|
|
#include "base/thread.h"
|
|
#include "base/message_loop.h"
|
|
#include "cairo.h"
|
|
#include "gfx2DGlue.h"
|
|
#include "gfxFcPlatformFontList.h"
|
|
#include "gfxConfig.h"
|
|
#include "gfxContext.h"
|
|
#include "gfxImageSurface.h"
|
|
#include "gfxUserFontSet.h"
|
|
#include "gfxUtils.h"
|
|
#include "gfxFT2FontBase.h"
|
|
#include "gfxTextRun.h"
|
|
#include "GLContextProvider.h"
|
|
#include "mozilla/Atomics.h"
|
|
#include "mozilla/Components.h"
|
|
#include "mozilla/dom/ContentChild.h"
|
|
#include "mozilla/FontPropertyTypes.h"
|
|
#include "mozilla/gfx/2D.h"
|
|
#include "mozilla/gfx/Logging.h"
|
|
#include "mozilla/Monitor.h"
|
|
#include "mozilla/Preferences.h"
|
|
#include "mozilla/StaticPrefs_gfx.h"
|
|
#include "mozilla/StaticPrefs_layers.h"
|
|
#include "mozilla/StaticPrefs_media.h"
|
|
#include "nsAppRunner.h"
|
|
#include "nsIGfxInfo.h"
|
|
#include "nsMathUtils.h"
|
|
#include "nsUnicharUtils.h"
|
|
#include "nsUnicodeProperties.h"
|
|
#include "prenv.h"
|
|
#include "VsyncSource.h"
|
|
#include "mozilla/WidgetUtilsGtk.h"
|
|
|
|
#ifdef MOZ_X11
|
|
# include "mozilla/gfx/XlibDisplay.h"
|
|
# include <gdk/gdkx.h>
|
|
# include <X11/extensions/Xrandr.h>
|
|
# include "cairo-xlib.h"
|
|
# include "gfxXlibSurface.h"
|
|
# include "GLContextGLX.h"
|
|
# include "GLXLibrary.h"
|
|
# include "mozilla/X11Util.h"
|
|
# include "SoftwareVsyncSource.h"
|
|
|
|
/* Undefine the Status from Xlib since it will conflict with system headers on
|
|
* OSX */
|
|
# if defined(__APPLE__) && defined(Status)
|
|
# undef Status
|
|
# endif
|
|
#endif /* MOZ_X11 */
|
|
|
|
#ifdef MOZ_WAYLAND
|
|
# include <gdk/gdkwayland.h>
|
|
# include "mozilla/widget/nsWaylandDisplay.h"
|
|
# include "mozilla/widget/DMABufLibWrapper.h"
|
|
# include "mozilla/StaticPrefs_widget.h"
|
|
#endif
|
|
|
|
#define GDK_PIXMAP_SIZE_MAX 32767
|
|
|
|
#define GFX_PREF_MAX_GENERIC_SUBSTITUTIONS \
|
|
"gfx.font_rendering.fontconfig.max_generic_substitutions"
|
|
|
|
using namespace mozilla;
|
|
using namespace mozilla::gfx;
|
|
using namespace mozilla::unicode;
|
|
using namespace mozilla::widget;
|
|
|
|
static FT_Library gPlatformFTLibrary = nullptr;
|
|
static int32_t sDPI;
|
|
|
|
static void screen_resolution_changed(GdkScreen* aScreen, GParamSpec* aPspec,
|
|
gpointer aClosure) {
|
|
sDPI = 0;
|
|
}
|
|
|
|
#if defined(MOZ_X11)
|
|
// TODO(aosmond): The envvar is deprecated. We should remove it once EGL is the
|
|
// default in release.
|
|
static bool IsX11EGLEnvvarEnabled() {
|
|
const char* eglPref = PR_GetEnv("MOZ_X11_EGL");
|
|
return (eglPref && *eglPref);
|
|
}
|
|
#endif
|
|
|
|
gfxPlatformGtk::gfxPlatformGtk() {
|
|
if (!gfxPlatform::IsHeadless()) {
|
|
gtk_init(nullptr, nullptr);
|
|
}
|
|
|
|
mMaxGenericSubstitutions = UNINITIALIZED_VALUE;
|
|
mIsX11Display = gfxPlatform::IsHeadless() ? false : GdkIsX11Display();
|
|
if (XRE_IsParentProcess()) {
|
|
InitX11EGLConfig();
|
|
if (IsWaylandDisplay() || gfxConfig::IsEnabled(Feature::X11_EGL)) {
|
|
gfxVars::SetUseEGL(true);
|
|
}
|
|
InitDmabufConfig();
|
|
if (gfxConfig::IsEnabled(Feature::DMABUF)) {
|
|
gfxVars::SetUseDMABuf(true);
|
|
}
|
|
}
|
|
|
|
InitBackendPrefs(GetBackendPrefs());
|
|
|
|
gPlatformFTLibrary = Factory::NewFTLibrary();
|
|
MOZ_RELEASE_ASSERT(gPlatformFTLibrary);
|
|
Factory::SetFTLibrary(gPlatformFTLibrary);
|
|
|
|
GdkScreen* gdkScreen = gdk_screen_get_default();
|
|
if (gdkScreen) {
|
|
g_signal_connect(gdkScreen, "notify::resolution",
|
|
G_CALLBACK(screen_resolution_changed), nullptr);
|
|
}
|
|
|
|
// Bug 1714483: Force disable FXAA Antialiasing on NV drivers. This is a
|
|
// temporary workaround for a driver bug.
|
|
PR_SetEnv("__GL_ALLOW_FXAA_USAGE=0");
|
|
}
|
|
|
|
gfxPlatformGtk::~gfxPlatformGtk() {
|
|
Factory::ReleaseFTLibrary(gPlatformFTLibrary);
|
|
gPlatformFTLibrary = nullptr;
|
|
}
|
|
|
|
void gfxPlatformGtk::InitX11EGLConfig() {
|
|
FeatureState& feature = gfxConfig::GetFeature(Feature::X11_EGL);
|
|
#ifdef MOZ_X11
|
|
feature.EnableByDefault();
|
|
|
|
if (StaticPrefs::gfx_x11_egl_force_enabled_AtStartup()) {
|
|
feature.UserForceEnable("Force enabled by pref");
|
|
} else if (IsX11EGLEnvvarEnabled()) {
|
|
feature.UserForceEnable("Force enabled by envvar");
|
|
} else if (StaticPrefs::gfx_x11_egl_force_disabled_AtStartup()) {
|
|
feature.UserDisable("Force disabled by pref",
|
|
"FEATURE_FAILURE_USER_FORCE_DISABLED"_ns);
|
|
}
|
|
|
|
nsCString failureId;
|
|
int32_t status;
|
|
nsCOMPtr<nsIGfxInfo> gfxInfo = components::GfxInfo::Service();
|
|
if (NS_FAILED(gfxInfo->GetFeatureStatus(nsIGfxInfo::FEATURE_X11_EGL,
|
|
failureId, &status))) {
|
|
feature.Disable(FeatureStatus::BlockedNoGfxInfo, "gfxInfo is broken",
|
|
"FEATURE_FAILURE_NO_GFX_INFO"_ns);
|
|
} else if (status != nsIGfxInfo::FEATURE_STATUS_OK) {
|
|
feature.Disable(FeatureStatus::Blocklisted, "Blocklisted by gfxInfo",
|
|
failureId);
|
|
}
|
|
|
|
nsAutoString testType;
|
|
gfxInfo->GetTestType(testType);
|
|
// We can only use X11/EGL if we actually found the EGL library and
|
|
// successfully use it to determine system information in glxtest.
|
|
if (testType != u"EGL") {
|
|
feature.ForceDisable(FeatureStatus::Broken, "glxtest could not use EGL",
|
|
"FEATURE_FAILURE_GLXTEST_NO_EGL"_ns);
|
|
}
|
|
|
|
if (feature.IsEnabled() && IsX11Display()) {
|
|
// Enabling glthread crashes on X11/EGL, see bug 1670545
|
|
PR_SetEnv("mesa_glthread=false");
|
|
}
|
|
#else
|
|
feature.DisableByDefault(FeatureStatus::Unavailable, "X11 support missing",
|
|
"FEATURE_FAILURE_NO_X11"_ns);
|
|
#endif
|
|
}
|
|
|
|
void gfxPlatformGtk::InitDmabufConfig() {
|
|
FeatureState& feature = gfxConfig::GetFeature(Feature::DMABUF);
|
|
#ifdef MOZ_WAYLAND
|
|
feature.EnableByDefault();
|
|
|
|
if (StaticPrefs::widget_dmabuf_force_enabled_AtStartup()) {
|
|
feature.UserForceEnable("Force enabled by pref");
|
|
}
|
|
|
|
nsCString failureId;
|
|
int32_t status;
|
|
nsCOMPtr<nsIGfxInfo> gfxInfo = components::GfxInfo::Service();
|
|
if (NS_FAILED(gfxInfo->GetFeatureStatus(nsIGfxInfo::FEATURE_DMABUF, failureId,
|
|
&status))) {
|
|
feature.Disable(FeatureStatus::BlockedNoGfxInfo, "gfxInfo is broken",
|
|
"FEATURE_FAILURE_NO_GFX_INFO"_ns);
|
|
} else if (status != nsIGfxInfo::FEATURE_STATUS_OK) {
|
|
feature.Disable(FeatureStatus::Blocklisted, "Blocklisted by gfxInfo",
|
|
failureId);
|
|
}
|
|
|
|
if (!gfxVars::UseEGL()) {
|
|
feature.ForceDisable(FeatureStatus::Unavailable, "Requires EGL",
|
|
"FEATURE_FAILURE_REQUIRES_EGL"_ns);
|
|
}
|
|
|
|
if (feature.IsEnabled()) {
|
|
nsAutoCString drmRenderDevice;
|
|
gfxInfo->GetDrmRenderDevice(drmRenderDevice);
|
|
gfxVars::SetDrmRenderDevice(drmRenderDevice);
|
|
|
|
if (!GetDMABufDevice()->Configure(failureId)) {
|
|
feature.ForceDisable(FeatureStatus::Failed, "Failed to configure",
|
|
failureId);
|
|
}
|
|
}
|
|
#else
|
|
feature.DisableByDefault(FeatureStatus::Unavailable,
|
|
"Wayland support missing",
|
|
"FEATURE_FAILURE_NO_WAYLAND"_ns);
|
|
#endif
|
|
}
|
|
|
|
bool gfxPlatformGtk::InitVAAPIConfig(bool aEnabledByPlatform) {
|
|
FeatureState& feature =
|
|
gfxConfig::GetFeature(Feature::HARDWARE_VIDEO_DECODING);
|
|
#ifdef MOZ_WAYLAND
|
|
feature.EnableByDefault();
|
|
|
|
nsCString failureId;
|
|
int32_t status = nsIGfxInfo::FEATURE_STATUS_UNKNOWN;
|
|
nsCOMPtr<nsIGfxInfo> gfxInfo = components::GfxInfo::Service();
|
|
if (NS_FAILED(gfxInfo->GetFeatureStatus(
|
|
nsIGfxInfo::FEATURE_HARDWARE_VIDEO_DECODING, failureId, &status))) {
|
|
feature.Disable(FeatureStatus::BlockedNoGfxInfo, "gfxInfo is broken",
|
|
"FEATURE_FAILURE_NO_GFX_INFO"_ns);
|
|
} else if (status != nsIGfxInfo::FEATURE_STATUS_OK) {
|
|
feature.Disable(FeatureStatus::Blocklisted, "Blocklisted by gfxInfo",
|
|
failureId);
|
|
}
|
|
if (status != nsIGfxInfo::FEATURE_STATUS_OK && aEnabledByPlatform) {
|
|
feature.UserForceEnable("Force enabled by pref");
|
|
}
|
|
if (!aEnabledByPlatform) {
|
|
feature.Disable(FeatureStatus::Blocked, "Blocked by platform", failureId);
|
|
}
|
|
if (!gfxVars::UseEGL()) {
|
|
feature.ForceDisable(FeatureStatus::Unavailable, "Requires EGL",
|
|
"FEATURE_FAILURE_REQUIRES_EGL"_ns);
|
|
}
|
|
#else
|
|
feature.DisableByDefault(FeatureStatus::Unavailable,
|
|
"Wayland support missing",
|
|
"FEATURE_FAILURE_NO_WAYLAND"_ns);
|
|
#endif
|
|
return feature.IsEnabled();
|
|
}
|
|
|
|
void gfxPlatformGtk::InitWebRenderConfig() {
|
|
gfxPlatform::InitWebRenderConfig();
|
|
|
|
if (!XRE_IsParentProcess()) {
|
|
return;
|
|
}
|
|
|
|
FeatureState& feature = gfxConfig::GetFeature(Feature::WEBRENDER_COMPOSITOR);
|
|
#ifdef RELEASE_OR_BETA
|
|
feature.ForceDisable(FeatureStatus::Blocked,
|
|
"Cannot be enabled in release or beta",
|
|
"FEATURE_FAILURE_DISABLE_RELEASE_OR_BETA"_ns);
|
|
#else
|
|
if (feature.IsEnabled()) {
|
|
if (!(gfxConfig::IsEnabled(Feature::WEBRENDER) ||
|
|
gfxConfig::IsEnabled(Feature::WEBRENDER_SOFTWARE))) {
|
|
feature.ForceDisable(FeatureStatus::Unavailable, "WebRender disabled",
|
|
"FEATURE_FAILURE_WR_DISABLED"_ns);
|
|
} else if (!IsWaylandDisplay()) {
|
|
feature.ForceDisable(FeatureStatus::Unavailable,
|
|
"Wayland support missing",
|
|
"FEATURE_FAILURE_NO_WAYLAND"_ns);
|
|
}
|
|
# ifdef MOZ_WAYLAND
|
|
else if (gfxConfig::IsEnabled(Feature::WEBRENDER) &&
|
|
!gfxConfig::IsEnabled(Feature::DMABUF)) {
|
|
// We use zwp_linux_dmabuf_v1 and GBM directly to manage FBOs. In theory
|
|
// this is also possible vie EGLstreams, but we don't bother to implement
|
|
// it as recent NVidia drivers support GBM and DMABuf as well.
|
|
feature.ForceDisable(FeatureStatus::Unavailable,
|
|
"Hardware Webrender requires DMAbuf support",
|
|
"FEATURE_FAILURE_NO_DMABUF"_ns);
|
|
} else if (!widget::WaylandDisplayGet()->GetViewporter()) {
|
|
feature.ForceDisable(FeatureStatus::Unavailable,
|
|
"Requires wp_viewporter protocol support",
|
|
"FEATURE_FAILURE_REQUIRES_WPVIEWPORTER"_ns);
|
|
}
|
|
# endif // MOZ_WAYLAND
|
|
}
|
|
#endif // RELEASE_OR_BETA
|
|
|
|
gfxVars::SetUseWebRenderCompositor(feature.IsEnabled());
|
|
}
|
|
|
|
void gfxPlatformGtk::InitPlatformGPUProcessPrefs() {
|
|
#ifdef MOZ_WAYLAND
|
|
if (IsWaylandDisplay()) {
|
|
FeatureState& gpuProc = gfxConfig::GetFeature(Feature::GPU_PROCESS);
|
|
gpuProc.ForceDisable(FeatureStatus::Blocked,
|
|
"Wayland does not work in the GPU process",
|
|
"FEATURE_FAILURE_WAYLAND"_ns);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
already_AddRefed<gfxASurface> gfxPlatformGtk::CreateOffscreenSurface(
|
|
const IntSize& aSize, gfxImageFormat aFormat) {
|
|
if (!Factory::AllowedSurfaceSize(aSize)) {
|
|
return nullptr;
|
|
}
|
|
|
|
RefPtr<gfxASurface> newSurface;
|
|
bool needsClear = true;
|
|
// XXX we really need a different interface here, something that passes
|
|
// in more context, including the display and/or target surface type that
|
|
// we should try to match
|
|
GdkScreen* gdkScreen = gdk_screen_get_default();
|
|
if (gdkScreen) {
|
|
newSurface = new gfxImageSurface(aSize, aFormat);
|
|
// The gfxImageSurface ctor zeroes this for us, no need to
|
|
// waste time clearing again
|
|
needsClear = false;
|
|
}
|
|
|
|
if (!newSurface) {
|
|
// We couldn't create a native surface for whatever reason;
|
|
// e.g., no display, no RENDER, bad size, etc.
|
|
// Fall back to image surface for the data.
|
|
newSurface = new gfxImageSurface(aSize, aFormat);
|
|
}
|
|
|
|
if (newSurface->CairoStatus()) {
|
|
newSurface = nullptr; // surface isn't valid for some reason
|
|
}
|
|
|
|
if (newSurface && needsClear) {
|
|
gfxUtils::ClearThebesSurface(newSurface);
|
|
}
|
|
|
|
return newSurface.forget();
|
|
}
|
|
|
|
nsresult gfxPlatformGtk::GetFontList(nsAtom* aLangGroup,
|
|
const nsACString& aGenericFamily,
|
|
nsTArray<nsString>& aListOfFonts) {
|
|
gfxPlatformFontList::PlatformFontList()->GetFontList(
|
|
aLangGroup, aGenericFamily, aListOfFonts);
|
|
return NS_OK;
|
|
}
|
|
|
|
// xxx - this is ubuntu centric, need to go through other distros and flesh
|
|
// out a more general list
|
|
static const char kFontDejaVuSans[] = "DejaVu Sans";
|
|
static const char kFontDejaVuSerif[] = "DejaVu Serif";
|
|
static const char kFontFreeSans[] = "FreeSans";
|
|
static const char kFontFreeSerif[] = "FreeSerif";
|
|
static const char kFontTakaoPGothic[] = "TakaoPGothic";
|
|
static const char kFontTwemojiMozilla[] = "Twemoji Mozilla";
|
|
static const char kFontDroidSansFallback[] = "Droid Sans Fallback";
|
|
static const char kFontWenQuanYiMicroHei[] = "WenQuanYi Micro Hei";
|
|
static const char kFontNanumGothic[] = "NanumGothic";
|
|
static const char kFontSymbola[] = "Symbola";
|
|
static const char kFontNotoSansSymbols[] = "Noto Sans Symbols";
|
|
static const char kFontNotoSansSymbols2[] = "Noto Sans Symbols2";
|
|
|
|
void gfxPlatformGtk::GetCommonFallbackFonts(uint32_t aCh, Script aRunScript,
|
|
eFontPresentation aPresentation,
|
|
nsTArray<const char*>& aFontList) {
|
|
if (PrefersColor(aPresentation)) {
|
|
aFontList.AppendElement(kFontTwemojiMozilla);
|
|
}
|
|
|
|
aFontList.AppendElement(kFontDejaVuSerif);
|
|
aFontList.AppendElement(kFontFreeSerif);
|
|
aFontList.AppendElement(kFontDejaVuSans);
|
|
aFontList.AppendElement(kFontFreeSans);
|
|
aFontList.AppendElement(kFontSymbola);
|
|
aFontList.AppendElement(kFontNotoSansSymbols);
|
|
aFontList.AppendElement(kFontNotoSansSymbols2);
|
|
|
|
// add fonts for CJK ranges
|
|
// xxx - this isn't really correct, should use the same CJK font ordering
|
|
// as the pref font code
|
|
if (aCh >= 0x3000 && ((aCh < 0xe000) || (aCh >= 0xf900 && aCh < 0xfff0) ||
|
|
((aCh >> 16) == 2))) {
|
|
aFontList.AppendElement(kFontTakaoPGothic);
|
|
aFontList.AppendElement(kFontDroidSansFallback);
|
|
aFontList.AppendElement(kFontWenQuanYiMicroHei);
|
|
aFontList.AppendElement(kFontNanumGothic);
|
|
}
|
|
}
|
|
|
|
void gfxPlatformGtk::ReadSystemFontList(
|
|
mozilla::dom::SystemFontList* retValue) {
|
|
gfxFcPlatformFontList::PlatformFontList()->ReadSystemFontList(retValue);
|
|
}
|
|
|
|
bool gfxPlatformGtk::CreatePlatformFontList() {
|
|
return gfxPlatformFontList::Initialize(new gfxFcPlatformFontList);
|
|
}
|
|
|
|
int32_t gfxPlatformGtk::GetFontScaleDPI() {
|
|
MOZ_ASSERT(XRE_IsParentProcess(),
|
|
"You can access this via LookAndFeel if you need it in child "
|
|
"processes");
|
|
if (MOZ_LIKELY(sDPI != 0)) {
|
|
return sDPI;
|
|
}
|
|
GdkScreen* screen = gdk_screen_get_default();
|
|
// Ensure settings in config files are processed.
|
|
gtk_settings_get_for_screen(screen);
|
|
int32_t dpi = int32_t(round(gdk_screen_get_resolution(screen)));
|
|
if (dpi <= 0) {
|
|
// Fall back to something sane
|
|
dpi = 96;
|
|
}
|
|
sDPI = dpi;
|
|
return dpi;
|
|
}
|
|
|
|
double gfxPlatformGtk::GetFontScaleFactor() {
|
|
// Integer scale factors work well with GTK window scaling, image scaling, and
|
|
// pixel alignment, but there is a range where 1 is too small and 2 is too
|
|
// big.
|
|
//
|
|
// An additional step of 1.5 is added because this is common scale on WINNT
|
|
// and at this ratio the advantages of larger rendering outweigh the
|
|
// disadvantages from scaling and pixel mis-alignment.
|
|
//
|
|
// A similar step for 1.25 is added as well, because this is the scale that
|
|
// "Large text" settings use in gnome, and it seems worth to allow, especially
|
|
// on already-hidpi environments.
|
|
int32_t dpi = GetFontScaleDPI();
|
|
if (dpi < 120) {
|
|
return 1.0;
|
|
}
|
|
if (dpi < 132) {
|
|
return 1.25;
|
|
}
|
|
if (dpi < 168) {
|
|
return 1.5;
|
|
}
|
|
return round(dpi / 96.0);
|
|
}
|
|
|
|
gfxImageFormat gfxPlatformGtk::GetOffscreenFormat() {
|
|
// Make sure there is a screen
|
|
GdkScreen* screen = gdk_screen_get_default();
|
|
if (screen && gdk_visual_get_depth(gdk_visual_get_system()) == 16) {
|
|
return SurfaceFormat::R5G6B5_UINT16;
|
|
}
|
|
|
|
return SurfaceFormat::X8R8G8B8_UINT32;
|
|
}
|
|
|
|
void gfxPlatformGtk::FontsPrefsChanged(const char* aPref) {
|
|
// only checking for generic substitions, pass other changes up
|
|
if (strcmp(GFX_PREF_MAX_GENERIC_SUBSTITUTIONS, aPref) != 0) {
|
|
gfxPlatform::FontsPrefsChanged(aPref);
|
|
return;
|
|
}
|
|
|
|
mMaxGenericSubstitutions = UNINITIALIZED_VALUE;
|
|
gfxFcPlatformFontList* pfl = gfxFcPlatformFontList::PlatformFontList();
|
|
pfl->ClearGenericMappings();
|
|
FlushFontAndWordCaches();
|
|
}
|
|
|
|
uint32_t gfxPlatformGtk::MaxGenericSubstitions() {
|
|
if (mMaxGenericSubstitutions == UNINITIALIZED_VALUE) {
|
|
mMaxGenericSubstitutions =
|
|
Preferences::GetInt(GFX_PREF_MAX_GENERIC_SUBSTITUTIONS, 3);
|
|
if (mMaxGenericSubstitutions < 0) {
|
|
mMaxGenericSubstitutions = 3;
|
|
}
|
|
}
|
|
|
|
return uint32_t(mMaxGenericSubstitutions);
|
|
}
|
|
|
|
bool gfxPlatformGtk::AccelerateLayersByDefault() { return true; }
|
|
|
|
#if defined(MOZ_X11)
|
|
|
|
static nsTArray<uint8_t> GetDisplayICCProfile(Display* dpy, Window& root) {
|
|
const char kIccProfileAtomName[] = "_ICC_PROFILE";
|
|
Atom iccAtom = XInternAtom(dpy, kIccProfileAtomName, TRUE);
|
|
if (!iccAtom) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
Atom retAtom;
|
|
int retFormat;
|
|
unsigned long retLength, retAfter;
|
|
unsigned char* retProperty;
|
|
|
|
if (XGetWindowProperty(dpy, root, iccAtom, 0, INT_MAX /* length */, X11False,
|
|
AnyPropertyType, &retAtom, &retFormat, &retLength,
|
|
&retAfter, &retProperty) != Success) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
nsTArray<uint8_t> result;
|
|
|
|
if (retLength > 0) {
|
|
result.AppendElements(static_cast<uint8_t*>(retProperty), retLength);
|
|
}
|
|
|
|
XFree(retProperty);
|
|
|
|
return result;
|
|
}
|
|
|
|
nsTArray<uint8_t> gfxPlatformGtk::GetPlatformCMSOutputProfileData() {
|
|
nsTArray<uint8_t> prefProfileData = GetPrefCMSOutputProfileData();
|
|
if (!prefProfileData.IsEmpty()) {
|
|
return prefProfileData;
|
|
}
|
|
|
|
if (XRE_IsContentProcess()) {
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
// This will be passed in during InitChild so we can avoid sending a
|
|
// sync message back to the parent during init.
|
|
const mozilla::gfx::ContentDeviceData* contentDeviceData =
|
|
GetInitContentDeviceData();
|
|
if (contentDeviceData) {
|
|
// On Windows, we assert that the profile isn't empty, but on
|
|
// Linux it can legitimately be empty if the display isn't
|
|
// calibrated. Thus, no assertion here.
|
|
return contentDeviceData->cmsOutputProfileData().Clone();
|
|
}
|
|
|
|
// Otherwise we need to ask the parent for the updated color profile
|
|
mozilla::dom::ContentChild* cc = mozilla::dom::ContentChild::GetSingleton();
|
|
nsTArray<uint8_t> result;
|
|
Unused << cc->SendGetOutputColorProfileData(&result);
|
|
return result;
|
|
}
|
|
|
|
if (!mIsX11Display) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
GdkDisplay* display = gdk_display_get_default();
|
|
Display* dpy = GDK_DISPLAY_XDISPLAY(display);
|
|
// In xpcshell tests, we never initialize X and hence don't have a Display.
|
|
// In this case, there's no output colour management to be done, so we just
|
|
// return with nullptr.
|
|
if (!dpy) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
Window root = gdk_x11_get_default_root_xwindow();
|
|
|
|
// First try ICC Profile
|
|
nsTArray<uint8_t> iccResult = GetDisplayICCProfile(dpy, root);
|
|
if (!iccResult.IsEmpty()) {
|
|
return iccResult;
|
|
}
|
|
|
|
// If ICC doesn't work, then try EDID
|
|
const char kEdid1AtomName[] = "XFree86_DDC_EDID1_RAWDATA";
|
|
Atom edidAtom = XInternAtom(dpy, kEdid1AtomName, TRUE);
|
|
if (!edidAtom) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
Atom retAtom;
|
|
int retFormat;
|
|
unsigned long retLength, retAfter;
|
|
unsigned char* retProperty;
|
|
|
|
if (XGetWindowProperty(dpy, root, edidAtom, 0, 32, X11False, AnyPropertyType,
|
|
&retAtom, &retFormat, &retLength, &retAfter,
|
|
&retProperty) != Success) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
if (retLength != 128) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
// Format documented in "VESA E-EDID Implementation Guide"
|
|
float gamma = (100 + (float)retProperty[0x17]) / 100.0f;
|
|
|
|
qcms_CIE_xyY whitePoint;
|
|
whitePoint.x =
|
|
((retProperty[0x21] << 2) | (retProperty[0x1a] >> 2 & 3)) / 1024.0;
|
|
whitePoint.y =
|
|
((retProperty[0x22] << 2) | (retProperty[0x1a] >> 0 & 3)) / 1024.0;
|
|
whitePoint.Y = 1.0;
|
|
|
|
qcms_CIE_xyYTRIPLE primaries;
|
|
primaries.red.x =
|
|
((retProperty[0x1b] << 2) | (retProperty[0x19] >> 6 & 3)) / 1024.0;
|
|
primaries.red.y =
|
|
((retProperty[0x1c] << 2) | (retProperty[0x19] >> 4 & 3)) / 1024.0;
|
|
primaries.red.Y = 1.0;
|
|
|
|
primaries.green.x =
|
|
((retProperty[0x1d] << 2) | (retProperty[0x19] >> 2 & 3)) / 1024.0;
|
|
primaries.green.y =
|
|
((retProperty[0x1e] << 2) | (retProperty[0x19] >> 0 & 3)) / 1024.0;
|
|
primaries.green.Y = 1.0;
|
|
|
|
primaries.blue.x =
|
|
((retProperty[0x1f] << 2) | (retProperty[0x1a] >> 6 & 3)) / 1024.0;
|
|
primaries.blue.y =
|
|
((retProperty[0x20] << 2) | (retProperty[0x1a] >> 4 & 3)) / 1024.0;
|
|
primaries.blue.Y = 1.0;
|
|
|
|
XFree(retProperty);
|
|
|
|
void* mem = nullptr;
|
|
size_t size = 0;
|
|
qcms_data_create_rgb_with_gamma(whitePoint, primaries, gamma, &mem, &size);
|
|
if (!mem) {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
nsTArray<uint8_t> result;
|
|
result.AppendElements(static_cast<uint8_t*>(mem), size);
|
|
free(mem);
|
|
|
|
// XXX: It seems like we get wrong colors when using this constructed profile:
|
|
// See bug 1696819. For now just forget that we made it.
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
#else // defined(MOZ_X11)
|
|
|
|
nsTArray<uint8_t> gfxPlatformGtk::GetPlatformCMSOutputProfileData() {
|
|
return nsTArray<uint8_t>();
|
|
}
|
|
|
|
#endif
|
|
|
|
bool gfxPlatformGtk::CheckVariationFontSupport() {
|
|
// Although there was some variation/multiple-master support in FreeType
|
|
// in older versions, it seems too incomplete/unstable for us to use
|
|
// until at least 2.7.1.
|
|
FT_Int major, minor, patch;
|
|
FT_Library_Version(Factory::GetFTLibrary(), &major, &minor, &patch);
|
|
return major * 1000000 + minor * 1000 + patch >= 2007001;
|
|
}
|
|
|
|
#ifdef MOZ_X11
|
|
|
|
class GtkVsyncSource final : public VsyncSource {
|
|
public:
|
|
GtkVsyncSource()
|
|
: mGLContext(nullptr),
|
|
mXDisplay(nullptr),
|
|
mSetupLock("GLXVsyncSetupLock"),
|
|
mVsyncThread("GLXVsyncThread"),
|
|
mVsyncTask(nullptr),
|
|
mVsyncEnabledLock("GLXVsyncEnabledLock"),
|
|
mVsyncEnabled(false) {
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
}
|
|
|
|
virtual ~GtkVsyncSource() { MOZ_ASSERT(NS_IsMainThread()); }
|
|
|
|
// Sets up the display's GL context on a worker thread.
|
|
// Required as GLContexts may only be used by the creating thread.
|
|
// Returns true if setup was a success.
|
|
bool Setup() {
|
|
MonitorAutoLock lock(mSetupLock);
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
if (!mVsyncThread.Start()) return false;
|
|
|
|
RefPtr<Runnable> vsyncSetup =
|
|
NewRunnableMethod("GtkVsyncSource::SetupGLContext", this,
|
|
&GtkVsyncSource::SetupGLContext);
|
|
mVsyncThread.message_loop()->PostTask(vsyncSetup.forget());
|
|
// Wait until the setup has completed.
|
|
lock.Wait();
|
|
return mGLContext != nullptr;
|
|
}
|
|
|
|
// Called on the Vsync thread to setup the GL context.
|
|
void SetupGLContext() {
|
|
MonitorAutoLock lock(mSetupLock);
|
|
MOZ_ASSERT(!NS_IsMainThread());
|
|
MOZ_ASSERT(!mGLContext, "GLContext already setup!");
|
|
|
|
// Create video sync timer on a separate Display to prevent locking the
|
|
// main thread X display.
|
|
mXDisplay = XOpenDisplay(nullptr);
|
|
if (!mXDisplay) {
|
|
lock.NotifyAll();
|
|
return;
|
|
}
|
|
|
|
// Most compositors wait for vsync events on the root window.
|
|
Window root = DefaultRootWindow(mXDisplay);
|
|
int screen = DefaultScreen(mXDisplay);
|
|
|
|
ScopedXFree<GLXFBConfig> cfgs;
|
|
GLXFBConfig config;
|
|
int visid;
|
|
bool forWebRender = false;
|
|
if (!gl::GLContextGLX::FindFBConfigForWindow(
|
|
mXDisplay, screen, root, &cfgs, &config, &visid, forWebRender)) {
|
|
lock.NotifyAll();
|
|
return;
|
|
}
|
|
|
|
mGLContext = gl::GLContextGLX::CreateGLContext(
|
|
{}, gfx::XlibDisplay::Borrow(mXDisplay), root, config, false, nullptr);
|
|
|
|
if (!mGLContext) {
|
|
lock.NotifyAll();
|
|
return;
|
|
}
|
|
|
|
mGLContext->MakeCurrent();
|
|
|
|
// Test that SGI_video_sync lets us get the counter.
|
|
unsigned int syncCounter = 0;
|
|
if (gl::sGLXLibrary.fGetVideoSync(&syncCounter) != 0) {
|
|
mGLContext = nullptr;
|
|
}
|
|
|
|
lock.NotifyAll();
|
|
}
|
|
|
|
virtual void EnableVsync() override {
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
MOZ_ASSERT(mGLContext, "GLContext not setup!");
|
|
|
|
MonitorAutoLock lock(mVsyncEnabledLock);
|
|
if (mVsyncEnabled) {
|
|
return;
|
|
}
|
|
mVsyncEnabled = true;
|
|
|
|
// If the task has not nulled itself out, it hasn't yet realized
|
|
// that vsync was disabled earlier, so continue its execution.
|
|
if (!mVsyncTask) {
|
|
mVsyncTask = NewRunnableMethod("GtkVsyncSource::RunVsync", this,
|
|
&GtkVsyncSource::RunVsync);
|
|
RefPtr<Runnable> addrefedTask = mVsyncTask;
|
|
mVsyncThread.message_loop()->PostTask(addrefedTask.forget());
|
|
}
|
|
}
|
|
|
|
virtual void DisableVsync() override {
|
|
MonitorAutoLock lock(mVsyncEnabledLock);
|
|
mVsyncEnabled = false;
|
|
}
|
|
|
|
virtual bool IsVsyncEnabled() override {
|
|
MonitorAutoLock lock(mVsyncEnabledLock);
|
|
return mVsyncEnabled;
|
|
}
|
|
|
|
virtual void Shutdown() override {
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
DisableVsync();
|
|
|
|
// Cleanup thread-specific resources before shutting down.
|
|
RefPtr<Runnable> shutdownTask = NewRunnableMethod(
|
|
"GtkVsyncSource::Cleanup", this, &GtkVsyncSource::Cleanup);
|
|
mVsyncThread.message_loop()->PostTask(shutdownTask.forget());
|
|
|
|
// Stop, waiting for the cleanup task to finish execution.
|
|
mVsyncThread.Stop();
|
|
}
|
|
|
|
private:
|
|
void RunVsync() {
|
|
MOZ_ASSERT(!NS_IsMainThread());
|
|
|
|
mGLContext->MakeCurrent();
|
|
|
|
unsigned int syncCounter = 0;
|
|
gl::sGLXLibrary.fGetVideoSync(&syncCounter);
|
|
for (;;) {
|
|
{
|
|
MonitorAutoLock lock(mVsyncEnabledLock);
|
|
if (!mVsyncEnabled) {
|
|
mVsyncTask = nullptr;
|
|
return;
|
|
}
|
|
}
|
|
|
|
TimeStamp lastVsync = TimeStamp::Now();
|
|
bool useSoftware = false;
|
|
|
|
// Wait until the video sync counter reaches the next value by waiting
|
|
// until the parity of the counter value changes.
|
|
unsigned int nextSync = syncCounter + 1;
|
|
int status;
|
|
if ((status = gl::sGLXLibrary.fWaitVideoSync(2, (int)nextSync % 2,
|
|
&syncCounter)) != 0) {
|
|
gfxWarningOnce() << "glXWaitVideoSync returned " << status;
|
|
useSoftware = true;
|
|
}
|
|
|
|
if (syncCounter == (nextSync - 1)) {
|
|
gfxWarningOnce()
|
|
<< "glXWaitVideoSync failed to increment the sync counter.";
|
|
useSoftware = true;
|
|
}
|
|
|
|
if (useSoftware) {
|
|
double remaining =
|
|
(1000.f / 60.f) - (TimeStamp::Now() - lastVsync).ToMilliseconds();
|
|
if (remaining > 0) {
|
|
AUTO_PROFILER_THREAD_SLEEP;
|
|
PlatformThread::Sleep((int)remaining);
|
|
}
|
|
}
|
|
|
|
lastVsync = TimeStamp::Now();
|
|
TimeStamp outputTime = lastVsync + GetVsyncRate();
|
|
NotifyVsync(lastVsync, outputTime);
|
|
}
|
|
}
|
|
|
|
void Cleanup() {
|
|
MOZ_ASSERT(!NS_IsMainThread());
|
|
|
|
mGLContext = nullptr;
|
|
if (mXDisplay) XCloseDisplay(mXDisplay);
|
|
}
|
|
|
|
// Owned by the vsync thread.
|
|
RefPtr<gl::GLContextGLX> mGLContext;
|
|
_XDisplay* mXDisplay;
|
|
Monitor mSetupLock MOZ_UNANNOTATED;
|
|
base::Thread mVsyncThread;
|
|
RefPtr<Runnable> mVsyncTask;
|
|
Monitor mVsyncEnabledLock MOZ_UNANNOTATED;
|
|
bool mVsyncEnabled;
|
|
};
|
|
|
|
class XrandrSoftwareVsyncSource final
|
|
: public mozilla::gfx::SoftwareVsyncSource {
|
|
public:
|
|
XrandrSoftwareVsyncSource() : SoftwareVsyncSource(ComputeVsyncRate()) {
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
|
|
GdkScreen* defaultScreen = gdk_screen_get_default();
|
|
g_signal_connect(defaultScreen, "monitors-changed",
|
|
G_CALLBACK(monitors_changed), this);
|
|
}
|
|
|
|
private:
|
|
// Request the current refresh rate via xrandr. It is hard to find the
|
|
// "correct" one, thus choose the highest one, assuming this will usually
|
|
// give the best user experience.
|
|
static mozilla::TimeDuration ComputeVsyncRate() {
|
|
struct _XDisplay* dpy = gdk_x11_get_default_xdisplay();
|
|
|
|
// Use the default software refresh rate as lower bound. Allowing lower
|
|
// rates makes a bunch of tests start to fail on CI. The main goal of this
|
|
// VsyncSource is to support refresh rates greater than the default one.
|
|
double highestRefreshRate = gfxPlatform::GetSoftwareVsyncRate();
|
|
|
|
// When running on remote X11 the xrandr version may be stuck on an
|
|
// ancient version. There are still setups using remote X11 out there, so
|
|
// make sure we don't crash.
|
|
int eventBase, errorBase, major, minor;
|
|
if (XRRQueryExtension(dpy, &eventBase, &errorBase) &&
|
|
XRRQueryVersion(dpy, &major, &minor) &&
|
|
(major > 1 || (major == 1 && minor >= 3))) {
|
|
Window root = gdk_x11_get_default_root_xwindow();
|
|
XRRScreenResources* res = XRRGetScreenResourcesCurrent(dpy, root);
|
|
|
|
// We can't use refresh rates far below the default one (60Hz) because
|
|
// otherwise random CI tests start to fail. However, many users have
|
|
// screens just below the default rate, e.g. 59.95Hz. So slightly
|
|
// decrease the lower bound.
|
|
highestRefreshRate -= 1.0;
|
|
|
|
for (int i = 0; i < res->noutput; i++) {
|
|
XRROutputInfo* outputInfo = XRRGetOutputInfo(dpy, res, res->outputs[i]);
|
|
if (!outputInfo->crtc) {
|
|
XRRFreeOutputInfo(outputInfo);
|
|
continue;
|
|
}
|
|
|
|
XRRCrtcInfo* crtcInfo = XRRGetCrtcInfo(dpy, res, outputInfo->crtc);
|
|
for (int j = 0; j < res->nmode; j++) {
|
|
if (res->modes[j].id == crtcInfo->mode) {
|
|
double refreshRate = mode_refresh(&res->modes[j]);
|
|
if (refreshRate > highestRefreshRate) {
|
|
highestRefreshRate = refreshRate;
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
XRRFreeCrtcInfo(crtcInfo);
|
|
XRRFreeOutputInfo(outputInfo);
|
|
}
|
|
XRRFreeScreenResources(res);
|
|
}
|
|
|
|
const double rate = 1000.0 / highestRefreshRate;
|
|
return mozilla::TimeDuration::FromMilliseconds(rate);
|
|
}
|
|
|
|
static void monitors_changed(GdkScreen* aScreen, gpointer aClosure) {
|
|
XrandrSoftwareVsyncSource* self =
|
|
static_cast<XrandrSoftwareVsyncSource*>(aClosure);
|
|
self->SetVsyncRate(ComputeVsyncRate());
|
|
}
|
|
|
|
// from xrandr.c
|
|
static double mode_refresh(const XRRModeInfo* mode_info) {
|
|
double rate;
|
|
double vTotal = mode_info->vTotal;
|
|
|
|
if (mode_info->modeFlags & RR_DoubleScan) {
|
|
/* doublescan doubles the number of lines */
|
|
vTotal *= 2;
|
|
}
|
|
|
|
if (mode_info->modeFlags & RR_Interlace) {
|
|
/* interlace splits the frame into two fields */
|
|
/* the field rate is what is typically reported by monitors */
|
|
vTotal /= 2;
|
|
}
|
|
|
|
if (mode_info->hTotal && vTotal) {
|
|
rate = ((double)mode_info->dotClock /
|
|
((double)mode_info->hTotal * (double)vTotal));
|
|
} else {
|
|
rate = 0;
|
|
}
|
|
return rate;
|
|
}
|
|
};
|
|
#endif
|
|
|
|
already_AddRefed<gfx::VsyncSource>
|
|
gfxPlatformGtk::CreateGlobalHardwareVsyncSource() {
|
|
#ifdef MOZ_X11
|
|
if (IsHeadless() || IsWaylandDisplay()) {
|
|
// On Wayland we can not create a global hardware based vsync source, thus
|
|
// use a software based one here. We create window specific ones later.
|
|
return GetSoftwareVsyncSource();
|
|
}
|
|
|
|
nsCOMPtr<nsIGfxInfo> gfxInfo = components::GfxInfo::Service();
|
|
nsString windowProtocol;
|
|
gfxInfo->GetWindowProtocol(windowProtocol);
|
|
bool isXwayland = windowProtocol.Find("xwayland") != -1;
|
|
nsString adapterDriverVendor;
|
|
gfxInfo->GetAdapterDriverVendor(adapterDriverVendor);
|
|
bool isMesa = adapterDriverVendor.Find("mesa") != -1;
|
|
|
|
// Only use GLX vsync when the OpenGL compositor / WebRender is being used.
|
|
// The extra cost of initializing a GLX context while blocking the main thread
|
|
// is not worth it when using basic composition. Do not use it on Xwayland, as
|
|
// Xwayland will give us a software timer as we are listening for the root
|
|
// window, which does not have a Wayland equivalent. Don't call
|
|
// gl::sGLXLibrary.SupportsVideoSync() when EGL is used as NVIDIA drivers
|
|
// refuse to use EGL GL context when GLX was initialized first and fail
|
|
// silently.
|
|
if (gfxConfig::IsEnabled(Feature::HW_COMPOSITING) && !isXwayland &&
|
|
(!gfxVars::UseEGL() || isMesa) &&
|
|
gl::sGLXLibrary.SupportsVideoSync(DefaultXDisplay())) {
|
|
RefPtr<GtkVsyncSource> vsyncSource = new GtkVsyncSource();
|
|
if (!vsyncSource->Setup()) {
|
|
NS_WARNING("Failed to setup GLContext, falling back to software vsync.");
|
|
return GetSoftwareVsyncSource();
|
|
}
|
|
return vsyncSource.forget();
|
|
}
|
|
|
|
RefPtr<VsyncSource> softwareVsync = new XrandrSoftwareVsyncSource();
|
|
return softwareVsync.forget();
|
|
#else
|
|
return CreateSoftwareVsyncSource();
|
|
#endif
|
|
}
|
|
|
|
void gfxPlatformGtk::BuildContentDeviceData(ContentDeviceData* aOut) {
|
|
gfxPlatform::BuildContentDeviceData(aOut);
|
|
|
|
aOut->cmsOutputProfileData() = GetPlatformCMSOutputProfileData();
|
|
}
|