/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ /* vim:set ts=2 sw=2 sts=2 et cindent: */ /* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this file, * You can obtain one at http://mozilla.org/MPL/2.0/. */ #include #include #include "base/basictypes.h" #include #include #include #include #include #include #include #include #if MOZ_WIDGET_GONK && ANDROID_VERSION >= 17 #include #endif #include "mozilla/layers/GrallocTextureClient.h" #include "mozilla/layers/TextureClient.h" #include "mozilla/Preferences.h" #include "mozilla/Types.h" #include "mozilla/Monitor.h" #include "nsMimeTypes.h" #include "MPAPI.h" #include "prlog.h" #include "GonkNativeWindow.h" #include "GonkNativeWindowClient.h" #include "OMXCodecProxy.h" #include "OmxDecoder.h" #include #define OD_LOG(...) __android_log_print(ANDROID_LOG_DEBUG, "OmxDecoder", __VA_ARGS__) #undef LOG #ifdef PR_LOGGING PRLogModuleInfo *gOmxDecoderLog; #define LOG(type, msg...) PR_LOG(gOmxDecoderLog, type, (msg)) #else #define LOG(x...) #endif using namespace MPAPI; using namespace mozilla; using namespace mozilla::gfx; using namespace mozilla::layers; using namespace android; OmxDecoder::OmxDecoder(MediaResource *aResource, AbstractMediaDecoder *aDecoder) : mDecoder(aDecoder), mResource(aResource), mDisplayWidth(0), mDisplayHeight(0), mVideoWidth(0), mVideoHeight(0), mVideoColorFormat(0), mVideoStride(0), mVideoSliceHeight(0), mVideoRotation(0), mAudioChannels(-1), mAudioSampleRate(-1), mDurationUs(-1), mVideoLastFrameTime(-1), mVideoBuffer(nullptr), mAudioBuffer(nullptr), mIsVideoSeeking(false), mAudioMetadataRead(false), mAudioPaused(false), mVideoPaused(false) { mLooper = new ALooper; mLooper->setName("OmxDecoder"); mReflector = new AHandlerReflector(this); // Register AMessage handler to ALooper. mLooper->registerHandler(mReflector); // Start ALooper thread. mLooper->start(); } OmxDecoder::~OmxDecoder() { MOZ_ASSERT(NS_IsMainThread()); ReleaseMediaResources(); // unregister AMessage handler from ALooper. mLooper->unregisterHandler(mReflector->id()); // Stop ALooper thread. mLooper->stop(); } void OmxDecoder::statusChanged() { sp notify = new AMessage(kNotifyStatusChanged, mReflector->id()); // post AMessage to OmxDecoder via ALooper. notify->post(); } static sp sOMX = nullptr; static sp GetOMX() { if(sOMX.get() == nullptr) { sOMX = new OMX; } return sOMX; } bool OmxDecoder::Init(sp& extractor) { #ifdef PR_LOGGING if (!gOmxDecoderLog) { gOmxDecoderLog = PR_NewLogModule("OmxDecoder"); } #endif sp meta = extractor->getMetaData(); ssize_t audioTrackIndex = -1; ssize_t videoTrackIndex = -1; for (size_t i = 0; i < extractor->countTracks(); ++i) { sp meta = extractor->getTrackMetaData(i); int32_t bitRate; if (!meta->findInt32(kKeyBitRate, &bitRate)) bitRate = 0; const char *mime; if (!meta->findCString(kKeyMIMEType, &mime)) { continue; } if (videoTrackIndex == -1 && !strncasecmp(mime, "video/", 6)) { videoTrackIndex = i; } else if (audioTrackIndex == -1 && !strncasecmp(mime, "audio/", 6)) { audioTrackIndex = i; } } if (videoTrackIndex == -1 && audioTrackIndex == -1) { NS_WARNING("OMX decoder could not find video or audio tracks"); return false; } mResource->SetReadMode(MediaCacheStream::MODE_PLAYBACK); if (videoTrackIndex != -1) { mVideoTrack = extractor->getTrack(videoTrackIndex); } if (audioTrackIndex != -1) { mAudioTrack = extractor->getTrack(audioTrackIndex); #ifdef MOZ_AUDIO_OFFLOAD // mAudioTrack is be used by OMXCodec. For offloaded audio track, using same // object gives undetermined behavior. So get a new track mAudioOffloadTrack = extractor->getTrack(audioTrackIndex); #endif } return true; } bool OmxDecoder::EnsureMetadata() { // calculate duration int64_t totalDurationUs = 0; int64_t durationUs = 0; if (mVideoTrack.get() && mVideoTrack->getFormat()->findInt64(kKeyDuration, &durationUs)) { if (durationUs > totalDurationUs) totalDurationUs = durationUs; } if (mAudioTrack.get()) { durationUs = -1; sp meta = mAudioTrack->getFormat(); if ((durationUs == -1) && meta->findInt64(kKeyDuration, &durationUs)) { if (durationUs > totalDurationUs) { totalDurationUs = durationUs; } } } mDurationUs = totalDurationUs; // read video metadata if (mVideoSource.get() && !SetVideoFormat()) { NS_WARNING("Couldn't set OMX video format"); return false; } // read audio metadata if (mAudioSource.get()) { // To reliably get the channel and sample rate data we need to read from the // audio source until we get a INFO_FORMAT_CHANGE status status_t err = mAudioSource->read(&mAudioBuffer); if (err != INFO_FORMAT_CHANGED) { if (err != OK) { NS_WARNING("Couldn't read audio buffer from OMX decoder"); return false; } sp meta = mAudioSource->getFormat(); if (!meta->findInt32(kKeyChannelCount, &mAudioChannels) || !meta->findInt32(kKeySampleRate, &mAudioSampleRate)) { NS_WARNING("Couldn't get audio metadata from OMX decoder"); return false; } mAudioMetadataRead = true; } else if (!SetAudioFormat()) { NS_WARNING("Couldn't set audio format"); return false; } } return true; } bool OmxDecoder::IsDormantNeeded() { if (mVideoTrack.get()) { return true; } return false; } bool OmxDecoder::IsWaitingMediaResources() { if (mVideoSource.get()) { return mVideoSource->IsWaitingResources(); } return false; } static bool isInEmulator() { char propQemu[PROPERTY_VALUE_MAX]; property_get("ro.kernel.qemu", propQemu, ""); return !strncmp(propQemu, "1", 1); } bool OmxDecoder::AllocateMediaResources() { if ((mVideoTrack != nullptr) && (mVideoSource == nullptr)) { // OMXClient::connect() always returns OK and abort's fatally if // it can't connect. OMXClient client; DebugOnly err = client.connect(); NS_ASSERTION(err == OK, "Failed to connect to OMX in mediaserver."); sp omx = client.interface(); #if defined(MOZ_WIDGET_GONK) && ANDROID_VERSION >= 21 sp producer; sp consumer; GonkBufferQueue::createBufferQueue(&producer, &consumer); mNativeWindow = new GonkNativeWindow(consumer); #else mNativeWindow = new GonkNativeWindow(); #endif #if defined(MOZ_WIDGET_GONK) && ANDROID_VERSION >= 21 mNativeWindowClient = new GonkNativeWindowClient(producer); #elif defined(MOZ_WIDGET_GONK) && ANDROID_VERSION >= 17 mNativeWindowClient = new GonkNativeWindowClient(mNativeWindow->getBufferQueue()); #else mNativeWindowClient = new GonkNativeWindowClient(mNativeWindow); #endif // Experience with OMX codecs is that only the HW decoders are // worth bothering with, at least on the platforms where this code // is currently used, and for formats this code is currently used // for (h.264). So if we don't get a hardware decoder, just give // up. #ifdef MOZ_OMX_WEBM_DECODER int flags = 0;//fallback to omx sw decoder if there is no hw decoder #else int flags = kHardwareCodecsOnly; #endif//MOZ_OMX_WEBM_DECODER if (isInEmulator()) { // If we are in emulator, allow to fall back to software. flags = 0; } mVideoSource = OMXCodecProxy::Create(omx, mVideoTrack->getFormat(), false, // decoder mVideoTrack, nullptr, flags, mNativeWindowClient); if (mVideoSource == nullptr) { NS_WARNING("Couldn't create OMX video source"); return false; } else { sp listener = this; mVideoSource->setEventListener(listener); mVideoSource->requestResource(); } } if ((mAudioTrack != nullptr) && (mAudioSource == nullptr)) { // OMXClient::connect() always returns OK and abort's fatally if // it can't connect. OMXClient client; DebugOnly err = client.connect(); NS_ASSERTION(err == OK, "Failed to connect to OMX in mediaserver."); sp omx = client.interface(); const char *audioMime = nullptr; sp meta = mAudioTrack->getFormat(); if (!meta->findCString(kKeyMIMEType, &audioMime)) { return false; } if (!strcasecmp(audioMime, "audio/raw")) { mAudioSource = mAudioTrack; } else { // try to load hardware codec in mediaserver process. int flags = kHardwareCodecsOnly; mAudioSource = OMXCodec::Create(omx, mAudioTrack->getFormat(), false, // decoder mAudioTrack, nullptr, flags); } if (mAudioSource == nullptr) { // try to load software codec in this process. int flags = kSoftwareCodecsOnly; mAudioSource = OMXCodec::Create(GetOMX(), mAudioTrack->getFormat(), false, // decoder mAudioTrack, nullptr, flags); if (mAudioSource == nullptr) { NS_WARNING("Couldn't create OMX audio source"); return false; } } if (mAudioSource->start() != OK) { NS_WARNING("Couldn't start OMX audio source"); mAudioSource.clear(); return false; } } return true; } void OmxDecoder::ReleaseMediaResources() { ReleaseVideoBuffer(); ReleaseAudioBuffer(); { Mutex::Autolock autoLock(mPendingVideoBuffersLock); MOZ_ASSERT(mPendingRecycleTexutreClients.empty()); // Release all pending recycle TextureClients, if they are not recycled yet. // This should not happen. See Bug 1042308. if (!mPendingRecycleTexutreClients.empty()) { printf_stderr("OmxDecoder::ReleaseMediaResources -- TextureClients are not recycled yet\n"); for (std::set::iterator it=mPendingRecycleTexutreClients.begin(); it!=mPendingRecycleTexutreClients.end(); it++) { GrallocTextureClientOGL* client = static_cast(*it); client->ClearRecycleCallback(); if (client->GetMediaBuffer()) { mPendingVideoBuffers.push(BufferItem(client->GetMediaBuffer(), client->GetReleaseFenceHandle())); } } mPendingRecycleTexutreClients.clear(); } } { // Free all pending video buffers. Mutex::Autolock autoLock(mSeekLock); ReleaseAllPendingVideoBuffersLocked(); } if (mVideoSource.get()) { mVideoSource->stop(); mVideoSource.clear(); } if (mAudioSource.get()) { mAudioSource->stop(); mAudioSource.clear(); } mNativeWindowClient.clear(); mNativeWindow.clear(); } bool OmxDecoder::SetVideoFormat() { const char *componentName; if (!mVideoSource->getFormat()->findInt32(kKeyWidth, &mVideoWidth) || !mVideoSource->getFormat()->findInt32(kKeyHeight, &mVideoHeight) || !mVideoSource->getFormat()->findCString(kKeyDecoderComponent, &componentName) || !mVideoSource->getFormat()->findInt32(kKeyColorFormat, &mVideoColorFormat) ) { return false; } if (!mVideoTrack.get() || !mVideoTrack->getFormat()->findInt32(kKeyDisplayWidth, &mDisplayWidth)) { mDisplayWidth = mVideoWidth; NS_WARNING("display width not available, assuming width"); } if (!mVideoTrack.get() || !mVideoTrack->getFormat()->findInt32(kKeyDisplayHeight, &mDisplayHeight)) { mDisplayHeight = mVideoHeight; NS_WARNING("display height not available, assuming height"); } if (!mVideoSource->getFormat()->findInt32(kKeyStride, &mVideoStride)) { mVideoStride = mVideoWidth; NS_WARNING("stride not available, assuming width"); } if (!mVideoSource->getFormat()->findInt32(kKeySliceHeight, &mVideoSliceHeight)) { mVideoSliceHeight = mVideoHeight; NS_WARNING("slice height not available, assuming height"); } // Since ICS, valid video side is caluculated from kKeyCropRect. // kKeyWidth means decoded video buffer width. // kKeyHeight means decoded video buffer height. // On some hardwares, decoded video buffer and valid video size are different. int32_t crop_left, crop_top, crop_right, crop_bottom; if (mVideoSource->getFormat()->findRect(kKeyCropRect, &crop_left, &crop_top, &crop_right, &crop_bottom)) { mVideoWidth = crop_right - crop_left + 1; mVideoHeight = crop_bottom - crop_top + 1; } if (!mVideoSource->getFormat()->findInt32(kKeyRotation, &mVideoRotation)) { mVideoRotation = 0; NS_WARNING("rotation not available, assuming 0"); } LOG(PR_LOG_DEBUG, "display width: %d display height %d width: %d height: %d component: %s format: %d stride: %d sliceHeight: %d rotation: %d", mDisplayWidth, mDisplayHeight, mVideoWidth, mVideoHeight, componentName, mVideoColorFormat, mVideoStride, mVideoSliceHeight, mVideoRotation); return true; } bool OmxDecoder::SetAudioFormat() { // If the format changed, update our cached info. if (!mAudioSource->getFormat()->findInt32(kKeyChannelCount, &mAudioChannels) || !mAudioSource->getFormat()->findInt32(kKeySampleRate, &mAudioSampleRate)) { return false; } LOG(PR_LOG_DEBUG, "channelCount: %d sampleRate: %d", mAudioChannels, mAudioSampleRate); return true; } void OmxDecoder::ReleaseDecoder() { mDecoder = nullptr; } void OmxDecoder::ReleaseVideoBuffer() { if (mVideoBuffer) { mVideoBuffer->release(); mVideoBuffer = nullptr; } } void OmxDecoder::ReleaseAudioBuffer() { if (mAudioBuffer) { mAudioBuffer->release(); mAudioBuffer = nullptr; } } void OmxDecoder::PlanarYUV420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame) { void *y = aData; void *u = static_cast(y) + mVideoStride * mVideoSliceHeight; void *v = static_cast(u) + mVideoStride/2 * mVideoSliceHeight/2; aFrame->Set(aTimeUs, aKeyFrame, aData, aSize, mVideoStride, mVideoSliceHeight, mVideoRotation, y, mVideoStride, mVideoWidth, mVideoHeight, 0, 0, u, mVideoStride/2, mVideoWidth/2, mVideoHeight/2, 0, 0, v, mVideoStride/2, mVideoWidth/2, mVideoHeight/2, 0, 0); } void OmxDecoder::CbYCrYFrame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame) { aFrame->Set(aTimeUs, aKeyFrame, aData, aSize, mVideoStride, mVideoSliceHeight, mVideoRotation, aData, mVideoStride, mVideoWidth, mVideoHeight, 1, 1, aData, mVideoStride, mVideoWidth/2, mVideoHeight/2, 0, 3, aData, mVideoStride, mVideoWidth/2, mVideoHeight/2, 2, 3); } void OmxDecoder::SemiPlanarYUV420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame) { void *y = aData; void *uv = static_cast(y) + (mVideoStride * mVideoSliceHeight); aFrame->Set(aTimeUs, aKeyFrame, aData, aSize, mVideoStride, mVideoSliceHeight, mVideoRotation, y, mVideoStride, mVideoWidth, mVideoHeight, 0, 0, uv, mVideoStride, mVideoWidth/2, mVideoHeight/2, 0, 1, uv, mVideoStride, mVideoWidth/2, mVideoHeight/2, 1, 1); } void OmxDecoder::SemiPlanarYVU420Frame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame) { SemiPlanarYUV420Frame(aFrame, aTimeUs, aData, aSize, aKeyFrame); aFrame->Cb.mOffset = 1; aFrame->Cr.mOffset = 0; } bool OmxDecoder::ToVideoFrame(VideoFrame *aFrame, int64_t aTimeUs, void *aData, size_t aSize, bool aKeyFrame) { const int OMX_QCOM_COLOR_FormatYVU420SemiPlanar = 0x7FA30C00; aFrame->mGraphicBuffer = nullptr; switch (mVideoColorFormat) { case OMX_COLOR_FormatYUV420Planar: PlanarYUV420Frame(aFrame, aTimeUs, aData, aSize, aKeyFrame); break; case OMX_COLOR_FormatCbYCrY: CbYCrYFrame(aFrame, aTimeUs, aData, aSize, aKeyFrame); break; case OMX_COLOR_FormatYUV420SemiPlanar: SemiPlanarYUV420Frame(aFrame, aTimeUs, aData, aSize, aKeyFrame); break; case OMX_QCOM_COLOR_FormatYVU420SemiPlanar: SemiPlanarYVU420Frame(aFrame, aTimeUs, aData, aSize, aKeyFrame); break; default: LOG(PR_LOG_DEBUG, "Unknown video color format %08x", mVideoColorFormat); return false; } return true; } bool OmxDecoder::ToAudioFrame(AudioFrame *aFrame, int64_t aTimeUs, void *aData, size_t aDataOffset, size_t aSize, int32_t aAudioChannels, int32_t aAudioSampleRate) { aFrame->Set(aTimeUs, static_cast(aData) + aDataOffset, aSize, aAudioChannels, aAudioSampleRate); return true; } bool OmxDecoder::ReadVideo(VideoFrame *aFrame, int64_t aTimeUs, bool aKeyframeSkip, bool aDoSeek) { if (!mVideoSource.get()) return false; ReleaseVideoBuffer(); status_t err; if (aDoSeek) { { Mutex::Autolock autoLock(mSeekLock); ReleaseAllPendingVideoBuffersLocked(); mIsVideoSeeking = true; } MediaSource::ReadOptions options; MediaSource::ReadOptions::SeekMode seekMode; // If the last timestamp of decoded frame is smaller than seekTime, // seek to next key frame. Otherwise seek to the previos one. if (mVideoLastFrameTime > aTimeUs || mVideoLastFrameTime == -1) { seekMode = MediaSource::ReadOptions::SEEK_PREVIOUS_SYNC; } else { seekMode = MediaSource::ReadOptions::SEEK_NEXT_SYNC; } bool findNextBuffer = true; while (findNextBuffer) { options.setSeekTo(aTimeUs, seekMode); findNextBuffer = false; if (mIsVideoSeeking) { err = mVideoSource->read(&mVideoBuffer, &options); Mutex::Autolock autoLock(mSeekLock); mIsVideoSeeking = false; PostReleaseVideoBuffer(nullptr, FenceHandle()); } else { err = mVideoSource->read(&mVideoBuffer); } // If there is no next Keyframe, jump to the previous key frame. if (err == ERROR_END_OF_STREAM && seekMode == MediaSource::ReadOptions::SEEK_NEXT_SYNC) { seekMode = MediaSource::ReadOptions::SEEK_PREVIOUS_SYNC; findNextBuffer = true; { Mutex::Autolock autoLock(mSeekLock); mIsVideoSeeking = true; } continue; } else if (err != OK) { OD_LOG("Unexpected error when seeking to %lld", aTimeUs); break; } // For some codecs, the length of first decoded frame after seek is 0. // Need to ignore it and continue to find the next one if (mVideoBuffer->range_length() == 0) { ReleaseVideoBuffer(); findNextBuffer = true; } } aDoSeek = false; } else { err = mVideoSource->read(&mVideoBuffer); } aFrame->mSize = 0; if (err == OK) { int64_t timeUs; int32_t unreadable; int32_t keyFrame; size_t length = mVideoBuffer->range_length(); if (!mVideoBuffer->meta_data()->findInt64(kKeyTime, &timeUs) ) { NS_WARNING("OMX decoder did not return frame time"); return false; } if (!mVideoBuffer->meta_data()->findInt32(kKeyIsSyncFrame, &keyFrame)) { keyFrame = 0; } if (!mVideoBuffer->meta_data()->findInt32(kKeyIsUnreadable, &unreadable)) { unreadable = 0; } RefPtr textureClient; if ((mVideoBuffer->graphicBuffer().get())) { textureClient = mNativeWindow->getTextureClientFromBuffer(mVideoBuffer->graphicBuffer().get()); } if (textureClient) { // Manually increment reference count to keep MediaBuffer alive // during TextureClient is in use. mVideoBuffer->add_ref(); GrallocTextureClientOGL* grallocClient = static_cast(textureClient.get()); grallocClient->SetMediaBuffer(mVideoBuffer); // Set recycle callback for TextureClient textureClient->SetRecycleCallback(OmxDecoder::RecycleCallback, this); { Mutex::Autolock autoLock(mPendingVideoBuffersLock); // Store pending recycle TextureClient. MOZ_ASSERT(mPendingRecycleTexutreClients.find(textureClient) == mPendingRecycleTexutreClients.end()); mPendingRecycleTexutreClients.insert(textureClient); } aFrame->mGraphicBuffer = textureClient; aFrame->mRotation = mVideoRotation; aFrame->mTimeUs = timeUs; aFrame->mKeyFrame = keyFrame; aFrame->Y.mWidth = mVideoWidth; aFrame->Y.mHeight = mVideoHeight; // Release to hold video buffer in OmxDecoder more. // MediaBuffer's ref count is changed from 2 to 1. ReleaseVideoBuffer(); } else if (length > 0) { char *data = static_cast(mVideoBuffer->data()) + mVideoBuffer->range_offset(); if (unreadable) { LOG(PR_LOG_DEBUG, "video frame is unreadable"); } if (!ToVideoFrame(aFrame, timeUs, data, length, keyFrame)) { return false; } } // Check if this frame is valid or not. If not, skip it. if ((aKeyframeSkip && timeUs < aTimeUs) || length == 0) { aFrame->mShouldSkip = true; } mVideoLastFrameTime = timeUs; } else if (err == INFO_FORMAT_CHANGED) { // If the format changed, update our cached info. if (!SetVideoFormat()) { return false; } else { return ReadVideo(aFrame, aTimeUs, aKeyframeSkip, aDoSeek); } } else if (err == ERROR_END_OF_STREAM) { return false; } else if (err == -ETIMEDOUT) { LOG(PR_LOG_DEBUG, "OmxDecoder::ReadVideo timed out, will retry"); return true; } else { // UNKNOWN_ERROR is sometimes is used to mean "out of memory", but // regardless, don't keep trying to decode if the decoder doesn't want to. LOG(PR_LOG_DEBUG, "OmxDecoder::ReadVideo failed, err=%d", err); return false; } return true; } bool OmxDecoder::ReadAudio(AudioFrame *aFrame, int64_t aSeekTimeUs) { status_t err; if (mAudioMetadataRead && aSeekTimeUs == -1) { // Use the data read into the buffer during metadata time err = OK; } else { ReleaseAudioBuffer(); if (aSeekTimeUs != -1) { MediaSource::ReadOptions options; options.setSeekTo(aSeekTimeUs); err = mAudioSource->read(&mAudioBuffer, &options); } else { err = mAudioSource->read(&mAudioBuffer); } } mAudioMetadataRead = false; aSeekTimeUs = -1; aFrame->mSize = 0; if (err == OK && mAudioBuffer && mAudioBuffer->range_length() != 0) { int64_t timeUs; if (!mAudioBuffer->meta_data()->findInt64(kKeyTime, &timeUs)) return false; return ToAudioFrame(aFrame, timeUs, mAudioBuffer->data(), mAudioBuffer->range_offset(), mAudioBuffer->range_length(), mAudioChannels, mAudioSampleRate); } else if (err == INFO_FORMAT_CHANGED) { // If the format changed, update our cached info. if (!SetAudioFormat()) { return false; } else { return ReadAudio(aFrame, aSeekTimeUs); } } else if (err == ERROR_END_OF_STREAM) { if (aFrame->mSize == 0) { return false; } } else if (err == -ETIMEDOUT) { LOG(PR_LOG_DEBUG, "OmxDecoder::ReadAudio timed out, will retry"); return true; } else if (err != OK) { LOG(PR_LOG_DEBUG, "OmxDecoder::ReadAudio failed, err=%d", err); return false; } return true; } nsresult OmxDecoder::Play() { if (!mVideoPaused && !mAudioPaused) { return NS_OK; } if (mVideoPaused && mVideoSource.get() && mVideoSource->start() != OK) { return NS_ERROR_UNEXPECTED; } mVideoPaused = false; if (mAudioPaused && mAudioSource.get() && mAudioSource->start() != OK) { return NS_ERROR_UNEXPECTED; } mAudioPaused = false; return NS_OK; } // AOSP didn't give implementation on OMXCodec::Pause() and not define // OMXCodec::Start() should be called for resuming the decoding. Currently // it is customized by a specific open source repository only. // ToDo The one not supported OMXCodec::Pause() should return error code here, // so OMXCodec::Start() doesn't be called again for resuming. But if someone // implement the OMXCodec::Pause() and need a following OMXCodec::Read() with // seek option (define in MediaSource.h) then it is still not supported here. // We need to fix it until it is really happened. void OmxDecoder::Pause() { /* The implementation of OMXCodec::pause is flawed. * OMXCodec::start will not restore from the paused state and result in * buffer timeout which causes timeouts in mochitests. * Since there is not power consumption problem in emulator, we will just * return when running in emulator to fix timeouts in mochitests. */ if (isInEmulator()) { return; } if (mVideoPaused || mAudioPaused) { return; } if (mVideoSource.get() && mVideoSource->pause() == OK) { mVideoPaused = true; } if (mAudioSource.get() && mAudioSource->pause() == OK) { mAudioPaused = true; } } // Called on ALooper thread. void OmxDecoder::onMessageReceived(const sp &msg) { switch (msg->what()) { case kNotifyPostReleaseVideoBuffer: { Mutex::Autolock autoLock(mSeekLock); // Free pending video buffers when OmxDecoder is not seeking video. // If OmxDecoder is seeking video, the buffers are freed on seek exit. if (!mIsVideoSeeking) { ReleaseAllPendingVideoBuffersLocked(); } break; } case kNotifyStatusChanged: { // Our decode may have acquired the hardware resource that it needs // to start. Notify the state machine to resume loading metadata. mDecoder->NotifyWaitingForResourcesStatusChanged(); break; } default: TRESPASS(); break; } } void OmxDecoder::PostReleaseVideoBuffer(MediaBuffer *aBuffer, const FenceHandle& aReleaseFenceHandle) { { Mutex::Autolock autoLock(mPendingVideoBuffersLock); if (aBuffer) { mPendingVideoBuffers.push(BufferItem(aBuffer, aReleaseFenceHandle)); } } sp notify = new AMessage(kNotifyPostReleaseVideoBuffer, mReflector->id()); // post AMessage to OmxDecoder via ALooper. notify->post(); } void OmxDecoder::ReleaseAllPendingVideoBuffersLocked() { Vector releasingVideoBuffers; { Mutex::Autolock autoLock(mPendingVideoBuffersLock); int size = mPendingVideoBuffers.size(); for (int i = 0; i < size; i++) { releasingVideoBuffers.push(mPendingVideoBuffers[i]); } mPendingVideoBuffers.clear(); } // Free all pending video buffers without holding mPendingVideoBuffersLock. int size = releasingVideoBuffers.size(); for (int i = 0; i < size; i++) { MediaBuffer *buffer; buffer = releasingVideoBuffers[i].mMediaBuffer; #if defined(MOZ_WIDGET_GONK) && ANDROID_VERSION >= 17 android::sp fence; int fenceFd = -1; fence = releasingVideoBuffers[i].mReleaseFenceHandle.mFence; if (fence.get() && fence->isValid()) { fenceFd = fence->dup(); } MOZ_ASSERT(buffer->refcount() == 1); // This code expect MediaBuffer's ref count is 1. // Return gralloc buffer to ANativeWindow ANativeWindow* window = static_cast(mNativeWindowClient.get()); window->cancelBuffer(window, buffer->graphicBuffer().get(), fenceFd); // Mark MediaBuffer as rendered. // When gralloc buffer is directly returned to ANativeWindow, // this mark is necesary. sp metaData = buffer->meta_data(); metaData->setInt32(kKeyRendered, 1); #endif // Return MediaBuffer to OMXCodec. buffer->release(); } releasingVideoBuffers.clear(); } void OmxDecoder::RecycleCallbackImp(TextureClient* aClient) { aClient->ClearRecycleCallback(); { Mutex::Autolock autoLock(mPendingVideoBuffersLock); if (mPendingRecycleTexutreClients.find(aClient) == mPendingRecycleTexutreClients.end()) { printf_stderr("OmxDecoder::RecycleCallbackImp -- TextureClient is not pending recycle\n"); return; } mPendingRecycleTexutreClients.erase(aClient); GrallocTextureClientOGL* client = static_cast(aClient); if (client->GetMediaBuffer()) { mPendingVideoBuffers.push(BufferItem(client->GetMediaBuffer(), client->GetReleaseFenceHandle())); } } sp notify = new AMessage(kNotifyPostReleaseVideoBuffer, mReflector->id()); // post AMessage to OmxDecoder via ALooper. notify->post(); } /* static */ void OmxDecoder::RecycleCallback(TextureClient* aClient, void* aClosure) { OmxDecoder* decoder = static_cast(aClosure); decoder->RecycleCallbackImp(aClient); }