mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-26 06:11:37 +00:00
34f99e4b4c
PR_LOGGING is now always defined, we can remove #ifdefs checking for it.
425 lines
12 KiB
C++
425 lines
12 KiB
C++
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
* You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#include "nsIPrefService.h"
|
|
#include "nsIPrefBranch.h"
|
|
|
|
#include "CSFLog.h"
|
|
#include "prenv.h"
|
|
|
|
static PRLogModuleInfo*
|
|
GetUserMediaLog()
|
|
{
|
|
static PRLogModuleInfo *sLog;
|
|
if (!sLog)
|
|
sLog = PR_NewLogModule("GetUserMedia");
|
|
return sLog;
|
|
}
|
|
|
|
#include "MediaEngineWebRTC.h"
|
|
#include "ImageContainer.h"
|
|
#include "nsIComponentRegistrar.h"
|
|
#include "MediaEngineTabVideoSource.h"
|
|
#include "nsITabSource.h"
|
|
#include "MediaTrackConstraints.h"
|
|
|
|
#ifdef MOZ_WIDGET_ANDROID
|
|
#include "AndroidJNIWrapper.h"
|
|
#include "AndroidBridge.h"
|
|
#endif
|
|
|
|
#if defined(MOZ_B2G_CAMERA) && defined(MOZ_WIDGET_GONK)
|
|
#include "ICameraControl.h"
|
|
#include "MediaEngineGonkVideoSource.h"
|
|
#endif
|
|
|
|
#undef LOG
|
|
#define LOG(args) PR_LOG(GetUserMediaLog(), PR_LOG_DEBUG, args)
|
|
|
|
namespace mozilla {
|
|
|
|
MediaEngineWebRTC::MediaEngineWebRTC(MediaEnginePrefs &aPrefs)
|
|
: mMutex("mozilla::MediaEngineWebRTC")
|
|
, mScreenEngine(nullptr)
|
|
, mBrowserEngine(nullptr)
|
|
, mWinEngine(nullptr)
|
|
, mAppEngine(nullptr)
|
|
, mVideoEngine(nullptr)
|
|
, mVoiceEngine(nullptr)
|
|
, mVideoEngineInit(false)
|
|
, mAudioEngineInit(false)
|
|
, mScreenEngineInit(false)
|
|
, mBrowserEngineInit(false)
|
|
, mAppEngineInit(false)
|
|
{
|
|
#ifndef MOZ_B2G_CAMERA
|
|
nsCOMPtr<nsIComponentRegistrar> compMgr;
|
|
NS_GetComponentRegistrar(getter_AddRefs(compMgr));
|
|
if (compMgr) {
|
|
compMgr->IsContractIDRegistered(NS_TABSOURCESERVICE_CONTRACTID, &mHasTabVideoSource);
|
|
}
|
|
#else
|
|
#ifdef MOZ_WIDGET_GONK
|
|
AsyncLatencyLogger::Get()->AddRef();
|
|
#endif
|
|
#endif
|
|
// XXX
|
|
gFarendObserver = new AudioOutputObserver();
|
|
|
|
NS_NewNamedThread("AudioGUM", getter_AddRefs(mThread));
|
|
MOZ_ASSERT(mThread);
|
|
}
|
|
|
|
void
|
|
MediaEngineWebRTC::EnumerateVideoDevices(dom::MediaSourceEnum aMediaSource,
|
|
nsTArray<nsRefPtr<MediaEngineVideoSource> >* aVSources)
|
|
{
|
|
// We spawn threads to handle gUM runnables, so we must protect the member vars
|
|
MutexAutoLock lock(mMutex);
|
|
|
|
#if defined(MOZ_B2G_CAMERA) && defined(MOZ_WIDGET_GONK)
|
|
if (aMediaSource != dom::MediaSourceEnum::Camera) {
|
|
// only supports camera sources
|
|
return;
|
|
}
|
|
|
|
/**
|
|
* We still enumerate every time, in case a new device was plugged in since
|
|
* the last call. TODO: Verify that WebRTC actually does deal with hotplugging
|
|
* new devices (with or without new engine creation) and accordingly adjust.
|
|
* Enumeration is not neccessary if GIPS reports the same set of devices
|
|
* for a given instance of the engine. Likewise, if a device was plugged out,
|
|
* mVideoSources must be updated.
|
|
*/
|
|
int num = 0;
|
|
nsresult result;
|
|
result = ICameraControl::GetNumberOfCameras(num);
|
|
if (num <= 0 || result != NS_OK) {
|
|
return;
|
|
}
|
|
|
|
for (int i = 0; i < num; i++) {
|
|
nsCString cameraName;
|
|
result = ICameraControl::GetCameraName(i, cameraName);
|
|
if (result != NS_OK) {
|
|
continue;
|
|
}
|
|
|
|
nsRefPtr<MediaEngineVideoSource> vSource;
|
|
NS_ConvertUTF8toUTF16 uuid(cameraName);
|
|
if (mVideoSources.Get(uuid, getter_AddRefs(vSource))) {
|
|
// We've already seen this device, just append.
|
|
aVSources->AppendElement(vSource.get());
|
|
} else {
|
|
vSource = new MediaEngineGonkVideoSource(i);
|
|
mVideoSources.Put(uuid, vSource); // Hashtable takes ownership.
|
|
aVSources->AppendElement(vSource);
|
|
}
|
|
}
|
|
|
|
return;
|
|
#else
|
|
ScopedCustomReleasePtr<webrtc::ViEBase> ptrViEBase;
|
|
ScopedCustomReleasePtr<webrtc::ViECapture> ptrViECapture;
|
|
webrtc::Config configSet;
|
|
webrtc::VideoEngine *videoEngine = nullptr;
|
|
bool *videoEngineInit = nullptr;
|
|
|
|
#ifdef MOZ_WIDGET_ANDROID
|
|
// get the JVM
|
|
JavaVM *jvm = mozilla::AndroidBridge::Bridge()->GetVM();
|
|
|
|
if (webrtc::VideoEngine::SetAndroidObjects(jvm) != 0) {
|
|
LOG(("VieCapture:SetAndroidObjects Failed"));
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
switch (aMediaSource) {
|
|
case dom::MediaSourceEnum::Window:
|
|
mWinEngineConfig.Set<webrtc::CaptureDeviceInfo>(
|
|
new webrtc::CaptureDeviceInfo(webrtc::CaptureDeviceType::Window));
|
|
if (!mWinEngine) {
|
|
if (!(mWinEngine = webrtc::VideoEngine::Create(mWinEngineConfig))) {
|
|
return;
|
|
}
|
|
}
|
|
videoEngine = mWinEngine;
|
|
videoEngineInit = &mWinEngineInit;
|
|
break;
|
|
case dom::MediaSourceEnum::Application:
|
|
mAppEngineConfig.Set<webrtc::CaptureDeviceInfo>(
|
|
new webrtc::CaptureDeviceInfo(webrtc::CaptureDeviceType::Application));
|
|
if (!mAppEngine) {
|
|
if (!(mAppEngine = webrtc::VideoEngine::Create(mAppEngineConfig))) {
|
|
return;
|
|
}
|
|
}
|
|
videoEngine = mAppEngine;
|
|
videoEngineInit = &mAppEngineInit;
|
|
break;
|
|
case dom::MediaSourceEnum::Screen:
|
|
mScreenEngineConfig.Set<webrtc::CaptureDeviceInfo>(
|
|
new webrtc::CaptureDeviceInfo(webrtc::CaptureDeviceType::Screen));
|
|
if (!mScreenEngine) {
|
|
if (!(mScreenEngine = webrtc::VideoEngine::Create(mScreenEngineConfig))) {
|
|
return;
|
|
}
|
|
}
|
|
videoEngine = mScreenEngine;
|
|
videoEngineInit = &mScreenEngineInit;
|
|
break;
|
|
case dom::MediaSourceEnum::Browser:
|
|
mBrowserEngineConfig.Set<webrtc::CaptureDeviceInfo>(
|
|
new webrtc::CaptureDeviceInfo(webrtc::CaptureDeviceType::Browser));
|
|
if (!mBrowserEngine) {
|
|
if (!(mBrowserEngine = webrtc::VideoEngine::Create(mBrowserEngineConfig))) {
|
|
return;
|
|
}
|
|
}
|
|
videoEngine = mBrowserEngine;
|
|
videoEngineInit = &mBrowserEngineInit;
|
|
break;
|
|
case dom::MediaSourceEnum::Camera:
|
|
// fall through
|
|
default:
|
|
if (!mVideoEngine) {
|
|
if (!(mVideoEngine = webrtc::VideoEngine::Create())) {
|
|
return;
|
|
}
|
|
}
|
|
videoEngine = mVideoEngine;
|
|
videoEngineInit = &mVideoEngineInit;
|
|
break;
|
|
}
|
|
|
|
ptrViEBase = webrtc::ViEBase::GetInterface(videoEngine);
|
|
if (!ptrViEBase) {
|
|
return;
|
|
}
|
|
if (ptrViEBase->Init() < 0) {
|
|
return;
|
|
}
|
|
*videoEngineInit = true;
|
|
|
|
ptrViECapture = webrtc::ViECapture::GetInterface(videoEngine);
|
|
if (!ptrViECapture) {
|
|
return;
|
|
}
|
|
|
|
/**
|
|
* We still enumerate every time, in case a new device was plugged in since
|
|
* the last call. TODO: Verify that WebRTC actually does deal with hotplugging
|
|
* new devices (with or without new engine creation) and accordingly adjust.
|
|
* Enumeration is not neccessary if GIPS reports the same set of devices
|
|
* for a given instance of the engine. Likewise, if a device was plugged out,
|
|
* mVideoSources must be updated.
|
|
*/
|
|
int num = ptrViECapture->NumberOfCaptureDevices();
|
|
if (num <= 0) {
|
|
return;
|
|
}
|
|
|
|
for (int i = 0; i < num; i++) {
|
|
char deviceName[MediaEngineSource::kMaxDeviceNameLength];
|
|
char uniqueId[MediaEngineSource::kMaxUniqueIdLength];
|
|
|
|
// paranoia
|
|
deviceName[0] = '\0';
|
|
uniqueId[0] = '\0';
|
|
int error = ptrViECapture->GetCaptureDevice(i, deviceName,
|
|
sizeof(deviceName), uniqueId,
|
|
sizeof(uniqueId));
|
|
|
|
if (error) {
|
|
LOG((" VieCapture:GetCaptureDevice: Failed %d",
|
|
ptrViEBase->LastError() ));
|
|
continue;
|
|
}
|
|
#ifdef DEBUG
|
|
LOG((" Capture Device Index %d, Name %s", i, deviceName));
|
|
|
|
webrtc::CaptureCapability cap;
|
|
int numCaps = ptrViECapture->NumberOfCapabilities(uniqueId,
|
|
MediaEngineSource::kMaxUniqueIdLength);
|
|
LOG(("Number of Capabilities %d", numCaps));
|
|
for (int j = 0; j < numCaps; j++) {
|
|
if (ptrViECapture->GetCaptureCapability(uniqueId,
|
|
MediaEngineSource::kMaxUniqueIdLength,
|
|
j, cap ) != 0 ) {
|
|
break;
|
|
}
|
|
LOG(("type=%d width=%d height=%d maxFPS=%d",
|
|
cap.rawType, cap.width, cap.height, cap.maxFPS ));
|
|
}
|
|
#endif
|
|
|
|
if (uniqueId[0] == '\0') {
|
|
// In case a device doesn't set uniqueId!
|
|
strncpy(uniqueId, deviceName, sizeof(uniqueId));
|
|
uniqueId[sizeof(uniqueId)-1] = '\0'; // strncpy isn't safe
|
|
}
|
|
|
|
nsRefPtr<MediaEngineVideoSource> vSource;
|
|
NS_ConvertUTF8toUTF16 uuid(uniqueId);
|
|
if (mVideoSources.Get(uuid, getter_AddRefs(vSource))) {
|
|
// We've already seen this device, just refresh and append.
|
|
static_cast<MediaEngineWebRTCVideoSource*>(vSource.get())->Refresh(i);
|
|
aVSources->AppendElement(vSource.get());
|
|
} else {
|
|
vSource = new MediaEngineWebRTCVideoSource(videoEngine, i, aMediaSource);
|
|
mVideoSources.Put(uuid, vSource); // Hashtable takes ownership.
|
|
aVSources->AppendElement(vSource);
|
|
}
|
|
}
|
|
|
|
if (mHasTabVideoSource || dom::MediaSourceEnum::Browser == aMediaSource)
|
|
aVSources->AppendElement(new MediaEngineTabVideoSource());
|
|
|
|
return;
|
|
#endif
|
|
}
|
|
|
|
void
|
|
MediaEngineWebRTC::EnumerateAudioDevices(dom::MediaSourceEnum aMediaSource,
|
|
nsTArray<nsRefPtr<MediaEngineAudioSource> >* aASources)
|
|
{
|
|
ScopedCustomReleasePtr<webrtc::VoEBase> ptrVoEBase;
|
|
ScopedCustomReleasePtr<webrtc::VoEHardware> ptrVoEHw;
|
|
// We spawn threads to handle gUM runnables, so we must protect the member vars
|
|
MutexAutoLock lock(mMutex);
|
|
|
|
#ifdef MOZ_WIDGET_ANDROID
|
|
jobject context = mozilla::AndroidBridge::Bridge()->GetGlobalContextRef();
|
|
|
|
// get the JVM
|
|
JavaVM *jvm = mozilla::AndroidBridge::Bridge()->GetVM();
|
|
JNIEnv *env = GetJNIForThread();
|
|
|
|
if (webrtc::VoiceEngine::SetAndroidObjects(jvm, env, (void*)context) != 0) {
|
|
LOG(("VoiceEngine:SetAndroidObjects Failed"));
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (!mVoiceEngine) {
|
|
mVoiceEngine = webrtc::VoiceEngine::Create();
|
|
if (!mVoiceEngine) {
|
|
return;
|
|
}
|
|
}
|
|
|
|
ptrVoEBase = webrtc::VoEBase::GetInterface(mVoiceEngine);
|
|
if (!ptrVoEBase) {
|
|
return;
|
|
}
|
|
|
|
if (!mAudioEngineInit) {
|
|
if (ptrVoEBase->Init() < 0) {
|
|
return;
|
|
}
|
|
mAudioEngineInit = true;
|
|
}
|
|
|
|
ptrVoEHw = webrtc::VoEHardware::GetInterface(mVoiceEngine);
|
|
if (!ptrVoEHw) {
|
|
return;
|
|
}
|
|
|
|
int nDevices = 0;
|
|
ptrVoEHw->GetNumOfRecordingDevices(nDevices);
|
|
int i;
|
|
#if defined(MOZ_WIDGET_ANDROID) || defined(MOZ_WIDGET_GONK)
|
|
i = 0; // Bug 1037025 - let the OS handle defaulting for now on android/b2g
|
|
#else
|
|
// -1 is "default communications device" depending on OS in webrtc.org code
|
|
i = -1;
|
|
#endif
|
|
for (; i < nDevices; i++) {
|
|
// We use constants here because GetRecordingDeviceName takes char[128].
|
|
char deviceName[128];
|
|
char uniqueId[128];
|
|
// paranoia; jingle doesn't bother with this
|
|
deviceName[0] = '\0';
|
|
uniqueId[0] = '\0';
|
|
|
|
int error = ptrVoEHw->GetRecordingDeviceName(i, deviceName, uniqueId);
|
|
if (error) {
|
|
LOG((" VoEHardware:GetRecordingDeviceName: Failed %d",
|
|
ptrVoEBase->LastError() ));
|
|
continue;
|
|
}
|
|
|
|
if (uniqueId[0] == '\0') {
|
|
// Mac and Linux don't set uniqueId!
|
|
MOZ_ASSERT(sizeof(deviceName) == sizeof(uniqueId)); // total paranoia
|
|
strcpy(uniqueId,deviceName); // safe given assert and initialization/error-check
|
|
}
|
|
|
|
nsRefPtr<MediaEngineWebRTCAudioSource> aSource;
|
|
NS_ConvertUTF8toUTF16 uuid(uniqueId);
|
|
if (mAudioSources.Get(uuid, getter_AddRefs(aSource))) {
|
|
// We've already seen this device, just append.
|
|
aASources->AppendElement(aSource.get());
|
|
} else {
|
|
aSource = new MediaEngineWebRTCAudioSource(
|
|
mThread, mVoiceEngine, i, deviceName, uniqueId
|
|
);
|
|
mAudioSources.Put(uuid, aSource); // Hashtable takes ownership.
|
|
aASources->AppendElement(aSource);
|
|
}
|
|
}
|
|
}
|
|
|
|
void
|
|
MediaEngineWebRTC::Shutdown()
|
|
{
|
|
// This is likely paranoia
|
|
MutexAutoLock lock(mMutex);
|
|
|
|
// Clear callbacks before we go away since the engines may outlive us
|
|
mVideoSources.Clear();
|
|
mAudioSources.Clear();
|
|
if (mVideoEngine) {
|
|
mVideoEngine->SetTraceCallback(nullptr);
|
|
webrtc::VideoEngine::Delete(mVideoEngine);
|
|
}
|
|
|
|
if (mScreenEngine) {
|
|
mScreenEngine->SetTraceCallback(nullptr);
|
|
webrtc::VideoEngine::Delete(mScreenEngine);
|
|
}
|
|
if (mWinEngine) {
|
|
mWinEngine->SetTraceCallback(nullptr);
|
|
webrtc::VideoEngine::Delete(mWinEngine);
|
|
}
|
|
if (mBrowserEngine) {
|
|
mBrowserEngine->SetTraceCallback(nullptr);
|
|
webrtc::VideoEngine::Delete(mBrowserEngine);
|
|
}
|
|
if (mAppEngine) {
|
|
mAppEngine->SetTraceCallback(nullptr);
|
|
webrtc::VideoEngine::Delete(mAppEngine);
|
|
}
|
|
|
|
if (mVoiceEngine) {
|
|
mVoiceEngine->SetTraceCallback(nullptr);
|
|
webrtc::VoiceEngine::Delete(mVoiceEngine);
|
|
}
|
|
|
|
mVideoEngine = nullptr;
|
|
mVoiceEngine = nullptr;
|
|
mScreenEngine = nullptr;
|
|
mWinEngine = nullptr;
|
|
mBrowserEngine = nullptr;
|
|
mAppEngine = nullptr;
|
|
|
|
if (mThread) {
|
|
mThread->Shutdown();
|
|
mThread = nullptr;
|
|
}
|
|
}
|
|
|
|
}
|