mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-07 20:17:37 +00:00
92bcfe6282
Implements a new class AudioOffloadPlayer to make use of Android::AudioTrack class's offloading capability
670 lines
18 KiB
C++
Executable File
670 lines
18 KiB
C++
Executable File
/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
|
|
/* vim:set ts=2 sw=2 sts=2 et cindent: */
|
|
/*
|
|
* Copyright (c) 2014 The Linux Foundation. All rights reserved.
|
|
* Copyright (C) 2009 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include "AudioOffloadPlayer.h"
|
|
#include "nsComponentManagerUtils.h"
|
|
#include "nsITimer.h"
|
|
#include "mozilla/dom/HTMLMediaElement.h"
|
|
|
|
#include <binder/IPCThreadState.h>
|
|
#include <stagefright/foundation/ADebug.h>
|
|
#include <stagefright/foundation/ALooper.h>
|
|
#include <stagefright/MediaDefs.h>
|
|
#include <stagefright/MediaErrors.h>
|
|
#include <stagefright/MediaSource.h>
|
|
#include <stagefright/MetaData.h>
|
|
#include <stagefright/Utils.h>
|
|
#include <AudioTrack.h>
|
|
#include <AudioSystem.h>
|
|
#include <AudioParameter.h>
|
|
#include <hardware/audio.h>
|
|
|
|
using namespace android;
|
|
|
|
namespace mozilla {
|
|
|
|
#ifdef PR_LOGGING
|
|
PRLogModuleInfo* gAudioOffloadPlayerLog;
|
|
#define AUDIO_OFFLOAD_LOG(type, msg) \
|
|
PR_LOG(gAudioOffloadPlayerLog, type, msg)
|
|
#else
|
|
#define AUDIO_OFFLOAD_LOG(type, msg)
|
|
#endif
|
|
|
|
AudioOffloadPlayer::AudioOffloadPlayer(MediaOmxDecoder* aObserver) :
|
|
mObserver(aObserver),
|
|
mInputBuffer(nullptr),
|
|
mSampleRate(0),
|
|
mSeeking(false),
|
|
mSeekDuringPause(false),
|
|
mReachedEOS(false),
|
|
mSeekTimeUs(0),
|
|
mStartPosUs(0),
|
|
mPositionTimeMediaUs(-1),
|
|
mStarted(false),
|
|
mPlaying(false),
|
|
mIsElementVisible(true)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
|
|
#ifdef PR_LOGGING
|
|
if (!gAudioOffloadPlayerLog) {
|
|
gAudioOffloadPlayerLog = PR_NewLogModule("AudioOffloadPlayer");
|
|
}
|
|
#endif
|
|
|
|
CHECK(aObserver);
|
|
mSessionId = AudioSystem::newAudioSessionId();
|
|
AudioSystem::acquireAudioSessionId(mSessionId);
|
|
mAudioSink = new AudioOutput(mSessionId,
|
|
IPCThreadState::self()->getCallingUid());
|
|
}
|
|
|
|
AudioOffloadPlayer::~AudioOffloadPlayer()
|
|
{
|
|
if (mStarted) {
|
|
Reset();
|
|
}
|
|
AudioSystem::releaseAudioSessionId(mSessionId);
|
|
}
|
|
|
|
void AudioOffloadPlayer::SetSource(const sp<MediaSource> &aSource)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(!mSource.get());
|
|
|
|
mSource = aSource;
|
|
}
|
|
|
|
status_t AudioOffloadPlayer::Start(bool aSourceAlreadyStarted)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(!mStarted);
|
|
CHECK(mSource.get());
|
|
|
|
status_t err;
|
|
CHECK(mAudioSink.get());
|
|
|
|
if (!aSourceAlreadyStarted) {
|
|
err = mSource->start();
|
|
|
|
if (err != OK) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
MediaSource::ReadOptions options;
|
|
if (mSeeking) {
|
|
options.setSeekTo(mSeekTimeUs);
|
|
mSeeking = false;
|
|
}
|
|
|
|
sp<MetaData> format = mSource->getFormat();
|
|
const char* mime;
|
|
int avgBitRate = -1;
|
|
int32_t channelMask;
|
|
int32_t numChannels;
|
|
int64_t durationUs = -1;
|
|
audio_format_t audioFormat = AUDIO_FORMAT_PCM_16_BIT;
|
|
uint32_t flags = AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD;
|
|
audio_offload_info_t offloadInfo = AUDIO_INFO_INITIALIZER;
|
|
|
|
CHECK(format->findCString(kKeyMIMEType, &mime));
|
|
CHECK(format->findInt32(kKeySampleRate, &mSampleRate));
|
|
CHECK(format->findInt32(kKeyChannelCount, &numChannels));
|
|
format->findInt32(kKeyBitRate, &avgBitRate);
|
|
format->findInt64(kKeyDuration, &durationUs);
|
|
|
|
if(!format->findInt32(kKeyChannelMask, &channelMask)) {
|
|
channelMask = CHANNEL_MASK_USE_CHANNEL_ORDER;
|
|
}
|
|
|
|
if (mapMimeToAudioFormat(audioFormat, mime) != OK) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_ERROR, ("Couldn't map mime type \"%s\" to a valid "
|
|
"AudioSystem::audio_format", mime));
|
|
audioFormat = AUDIO_FORMAT_INVALID;
|
|
}
|
|
|
|
offloadInfo.duration_us = durationUs;
|
|
offloadInfo.sample_rate = mSampleRate;
|
|
offloadInfo.channel_mask = channelMask;
|
|
offloadInfo.format = audioFormat;
|
|
offloadInfo.stream_type = AUDIO_STREAM_MUSIC;
|
|
offloadInfo.bit_rate = avgBitRate;
|
|
offloadInfo.has_video = false;
|
|
offloadInfo.is_streaming = false;
|
|
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("isOffloadSupported: SR=%u, CM=0x%x, "
|
|
"Format=0x%x, StreamType=%d, BitRate=%u, duration=%lld us, has_video=%d",
|
|
offloadInfo.sample_rate, offloadInfo.channel_mask, offloadInfo.format,
|
|
offloadInfo.stream_type, offloadInfo.bit_rate, offloadInfo.duration_us,
|
|
offloadInfo.has_video));
|
|
|
|
err = mAudioSink->Open(mSampleRate,
|
|
numChannels,
|
|
channelMask,
|
|
audioFormat,
|
|
&AudioOffloadPlayer::AudioSinkCallback,
|
|
this,
|
|
(audio_output_flags_t) flags,
|
|
&offloadInfo);
|
|
if (err == OK) {
|
|
// If the playback is offloaded to h/w we pass the
|
|
// HAL some metadata information
|
|
// We don't want to do this for PCM because it will be going
|
|
// through the AudioFlinger mixer before reaching the hardware
|
|
SendMetaDataToHal(mAudioSink, format);
|
|
}
|
|
mStarted = true;
|
|
mPlaying = false;
|
|
|
|
return err;
|
|
}
|
|
|
|
void AudioOffloadPlayer::ChangeState(MediaDecoder::PlayState aState)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
mPlayState = aState;
|
|
|
|
switch (mPlayState) {
|
|
case MediaDecoder::PLAY_STATE_PLAYING:
|
|
Play();
|
|
StartTimeUpdate();
|
|
break;
|
|
|
|
case MediaDecoder::PLAY_STATE_SEEKING: {
|
|
int64_t seekTimeUs
|
|
= mObserver->GetSeekTime();
|
|
SeekTo(seekTimeUs);
|
|
mObserver->ResetSeekTime();
|
|
} break;
|
|
|
|
case MediaDecoder::PLAY_STATE_PAUSED:
|
|
case MediaDecoder::PLAY_STATE_SHUTDOWN:
|
|
// Just pause here during play state shutdown as well to stop playing
|
|
// offload track immediately. Resources will be freed by MediaOmxDecoder
|
|
Pause();
|
|
break;
|
|
|
|
case MediaDecoder::PLAY_STATE_ENDED:
|
|
Pause(true);
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
void AudioOffloadPlayer::Pause(bool aPlayPendingSamples)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(mStarted);
|
|
CHECK(mAudioSink.get());
|
|
|
|
if (aPlayPendingSamples) {
|
|
mAudioSink->Stop();
|
|
} else {
|
|
mAudioSink->Pause();
|
|
}
|
|
mPlaying = false;
|
|
}
|
|
|
|
status_t AudioOffloadPlayer::Play()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(mStarted);
|
|
CHECK(mAudioSink.get());
|
|
|
|
status_t err = OK;
|
|
err = mAudioSink->Start();
|
|
|
|
if (err == OK) {
|
|
mPlaying = true;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
void AudioOffloadPlayer::Reset()
|
|
{
|
|
CHECK(mStarted);
|
|
CHECK(mAudioSink.get());
|
|
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("reset: mPlaying=%d mReachedEOS=%d",
|
|
mPlaying, mReachedEOS));
|
|
|
|
mAudioSink->Stop();
|
|
// If we're closing and have reached EOS, we don't want to flush
|
|
// the track because if it is offloaded there could be a small
|
|
// amount of residual data in the hardware buffer which we must
|
|
// play to give gapless playback.
|
|
// But if we're resetting when paused or before we've reached EOS
|
|
// we can't be doing a gapless playback and there could be a large
|
|
// amount of data queued in the hardware if the track is offloaded,
|
|
// so we must flush to prevent a track switch being delayed playing
|
|
// the buffered data that we don't want now
|
|
if (!mPlaying || !mReachedEOS) {
|
|
mAudioSink->Flush();
|
|
}
|
|
|
|
mAudioSink->Close();
|
|
// Make sure to release any buffer we hold onto so that the
|
|
// source is able to stop().
|
|
|
|
if (mInputBuffer) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Releasing input buffer"));
|
|
|
|
mInputBuffer->release();
|
|
mInputBuffer = nullptr;
|
|
}
|
|
|
|
IPCThreadState::self()->flushCommands();
|
|
StopTimeUpdate();
|
|
|
|
mSeeking = false;
|
|
mSeekTimeUs = 0;
|
|
mReachedEOS = false;
|
|
mStarted = false;
|
|
mPlaying = false;
|
|
mStartPosUs = 0;
|
|
}
|
|
|
|
status_t AudioOffloadPlayer::SeekTo(int64_t aTimeUs)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(mAudioSink.get());
|
|
|
|
android::Mutex::Autolock autoLock(mLock);
|
|
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("SeekTo ( %lld )", aTimeUs));
|
|
|
|
mSeeking = true;
|
|
mReachedEOS = false;
|
|
mPositionTimeMediaUs = -1;
|
|
mSeekTimeUs = aTimeUs;
|
|
mStartPosUs = aTimeUs;
|
|
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaDecoder::SeekingStarted);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
|
|
if (mPlaying) {
|
|
mAudioSink->Pause();
|
|
}
|
|
|
|
mAudioSink->Flush();
|
|
|
|
if (mPlaying) {
|
|
mAudioSink->Start();
|
|
} else {
|
|
mSeekDuringPause = true;
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Fake seek complete during pause"));
|
|
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaDecoder::SeekingStopped);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
double AudioOffloadPlayer::GetMediaTimeSecs()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
return (static_cast<double>(GetMediaTimeUs()) /
|
|
static_cast<double>(USECS_PER_S));
|
|
}
|
|
|
|
int64_t AudioOffloadPlayer::GetMediaTimeUs()
|
|
{
|
|
android::Mutex::Autolock autoLock(mLock);
|
|
|
|
int64_t playPosition = 0;
|
|
if (mSeeking) {
|
|
return mSeekTimeUs;
|
|
}
|
|
|
|
playPosition = GetOutputPlayPositionUs_l();
|
|
if (!mReachedEOS) {
|
|
mPositionTimeMediaUs = playPosition;
|
|
}
|
|
|
|
return mPositionTimeMediaUs;
|
|
}
|
|
|
|
int64_t AudioOffloadPlayer::GetOutputPlayPositionUs_l() const
|
|
{
|
|
CHECK(mAudioSink.get());
|
|
uint32_t playedSamples = 0;
|
|
|
|
mAudioSink->GetPosition(&playedSamples);
|
|
|
|
const int64_t playedUs = (static_cast<int64_t>(playedSamples) * 1000000 ) /
|
|
mSampleRate;
|
|
|
|
// HAL position is relative to the first buffer we sent at mStartPosUs
|
|
const int64_t renderedDuration = mStartPosUs + playedUs;
|
|
return renderedDuration;
|
|
}
|
|
|
|
void AudioOffloadPlayer::NotifyAudioEOS()
|
|
{
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaDecoder::PlaybackEnded);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
}
|
|
|
|
void AudioOffloadPlayer::NotifyPositionChanged()
|
|
{
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaOmxDecoder::PlaybackPositionChanged);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
}
|
|
|
|
void AudioOffloadPlayer::NotifyAudioTearDown()
|
|
{
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaOmxDecoder::AudioOffloadTearDown);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
}
|
|
|
|
// static
|
|
size_t AudioOffloadPlayer::AudioSinkCallback(AudioSink* aAudioSink,
|
|
void* aBuffer,
|
|
size_t aSize,
|
|
void* aCookie,
|
|
AudioSink::cb_event_t aEvent)
|
|
{
|
|
AudioOffloadPlayer* me = (AudioOffloadPlayer*) aCookie;
|
|
|
|
switch (aEvent) {
|
|
|
|
case AudioSink::CB_EVENT_FILL_BUFFER:
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Notify Audio position changed"));
|
|
me->NotifyPositionChanged();
|
|
return me->FillBuffer(aBuffer, aSize);
|
|
|
|
case AudioSink::CB_EVENT_STREAM_END:
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Notify Audio EOS"));
|
|
me->mReachedEOS = true;
|
|
me->NotifyAudioEOS();
|
|
break;
|
|
|
|
case AudioSink::CB_EVENT_TEAR_DOWN:
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Notify Tear down event"));
|
|
me->NotifyAudioTearDown();
|
|
break;
|
|
|
|
default:
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_ERROR, ("Unknown event %d from audio sink",
|
|
aEvent));
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
size_t AudioOffloadPlayer::FillBuffer(void* aData, size_t aSize)
|
|
{
|
|
CHECK(mAudioSink.get());
|
|
|
|
if (mReachedEOS) {
|
|
return 0;
|
|
}
|
|
|
|
bool postSeekComplete = false;
|
|
|
|
size_t sizeDone = 0;
|
|
size_t sizeRemaining = aSize;
|
|
while (sizeRemaining > 0) {
|
|
MediaSource::ReadOptions options;
|
|
bool refreshSeekTime = false;
|
|
|
|
{
|
|
android::Mutex::Autolock autoLock(mLock);
|
|
|
|
if (mSeeking) {
|
|
options.setSeekTo(mSeekTimeUs);
|
|
refreshSeekTime = true;
|
|
|
|
if (mInputBuffer) {
|
|
mInputBuffer->release();
|
|
mInputBuffer = nullptr;
|
|
}
|
|
|
|
mSeeking = false;
|
|
postSeekComplete = true;
|
|
}
|
|
}
|
|
|
|
if (!mInputBuffer) {
|
|
|
|
status_t err;
|
|
err = mSource->read(&mInputBuffer, &options);
|
|
|
|
CHECK((!err && mInputBuffer) || (err && !mInputBuffer));
|
|
|
|
android::Mutex::Autolock autoLock(mLock);
|
|
|
|
if (err != OK) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_ERROR, ("Error while reading media source %d "
|
|
"Ok to receive EOS error at end", err));
|
|
if (!mReachedEOS) {
|
|
// After seek there is a possible race condition if
|
|
// OffloadThread is observing state_stopping_1 before
|
|
// framesReady() > 0. Ensure sink stop is called
|
|
// after last buffer is released. This ensures the
|
|
// partial buffer is written to the driver before
|
|
// stopping one is observed.The drawback is that
|
|
// there will be an unnecessary call to the parser
|
|
// after parser signalled EOS.
|
|
if (sizeDone > 0) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("send Partial buffer down"));
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("skip calling stop till next"
|
|
" fillBuffer"));
|
|
break;
|
|
}
|
|
// no more buffers to push - stop() and wait for STREAM_END
|
|
// don't set mReachedEOS until stream end received
|
|
mAudioSink->Stop();
|
|
}
|
|
break;
|
|
}
|
|
|
|
if(mInputBuffer->range_length() != 0) {
|
|
CHECK(mInputBuffer->meta_data()->findInt64(
|
|
kKeyTime, &mPositionTimeMediaUs));
|
|
}
|
|
|
|
// need to adjust the mStartPosUs for offload decoding since parser
|
|
// might not be able to get the exact seek time requested.
|
|
if (refreshSeekTime) {
|
|
if (postSeekComplete) {
|
|
|
|
if (!mSeekDuringPause) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("FillBuffer posting SEEK_COMPLETE"));
|
|
nsCOMPtr<nsIRunnable> nsEvent = NS_NewRunnableMethod(mObserver,
|
|
&MediaDecoder::SeekingStopped);
|
|
NS_DispatchToMainThread(nsEvent, NS_DISPATCH_NORMAL);
|
|
} else {
|
|
// Callback is already called for seek during pause. Just reset the
|
|
// flag
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Not posting seek complete as its"
|
|
" already faked"));
|
|
mSeekDuringPause = false;
|
|
}
|
|
|
|
NotifyPositionChanged();
|
|
postSeekComplete = false;
|
|
}
|
|
|
|
mStartPosUs = mPositionTimeMediaUs;
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Adjust seek time to: %.2f",
|
|
mStartPosUs / 1E6));
|
|
// clear seek time with mLock locked and once we have valid
|
|
// mPositionTimeMediaUs
|
|
// before clearing mSeekTimeUs check if a new seek request has been
|
|
// received while we were reading from the source with mLock released.
|
|
if (!mSeeking) {
|
|
mSeekTimeUs = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (mInputBuffer->range_length() == 0) {
|
|
mInputBuffer->release();
|
|
mInputBuffer = nullptr;
|
|
continue;
|
|
}
|
|
|
|
size_t copy = sizeRemaining;
|
|
if (copy > mInputBuffer->range_length()) {
|
|
copy = mInputBuffer->range_length();
|
|
}
|
|
|
|
memcpy((char *)aData + sizeDone,
|
|
(const char *)mInputBuffer->data() + mInputBuffer->range_offset(),
|
|
copy);
|
|
|
|
mInputBuffer->set_range(mInputBuffer->range_offset() + copy,
|
|
mInputBuffer->range_length() - copy);
|
|
|
|
sizeDone += copy;
|
|
sizeRemaining -= copy;
|
|
}
|
|
return sizeDone;
|
|
}
|
|
|
|
void AudioOffloadPlayer::SetElementVisibility(bool aIsVisible)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
mIsElementVisible = aIsVisible;
|
|
if (mIsElementVisible) {
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("Element is visible. Start time update"));
|
|
StartTimeUpdate();
|
|
}
|
|
}
|
|
|
|
static void TimeUpdateCallback(nsITimer* aTimer, void* aClosure)
|
|
{
|
|
AudioOffloadPlayer* player = static_cast<AudioOffloadPlayer*>(aClosure);
|
|
player->TimeUpdate();
|
|
}
|
|
|
|
void AudioOffloadPlayer::TimeUpdate()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
TimeStamp now = TimeStamp::Now();
|
|
|
|
// If TIMEUPDATE_MS has passed since the last fire update event fired, fire
|
|
// another timeupdate event.
|
|
if ((mLastFireUpdateTime.IsNull() ||
|
|
now - mLastFireUpdateTime >=
|
|
TimeDuration::FromMilliseconds(TIMEUPDATE_MS))) {
|
|
mLastFireUpdateTime = now;
|
|
NotifyPositionChanged();
|
|
}
|
|
|
|
if (mPlayState != MediaDecoder::PLAY_STATE_PLAYING || !mIsElementVisible) {
|
|
StopTimeUpdate();
|
|
}
|
|
}
|
|
|
|
nsresult AudioOffloadPlayer::StartTimeUpdate()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
if (mTimeUpdateTimer) {
|
|
return NS_OK;
|
|
}
|
|
|
|
mTimeUpdateTimer = do_CreateInstance("@mozilla.org/timer;1");
|
|
return mTimeUpdateTimer->InitWithFuncCallback(TimeUpdateCallback,
|
|
this,
|
|
TIMEUPDATE_MS,
|
|
nsITimer::TYPE_REPEATING_SLACK);
|
|
}
|
|
|
|
nsresult AudioOffloadPlayer::StopTimeUpdate()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
if (!mTimeUpdateTimer) {
|
|
return NS_OK;
|
|
}
|
|
|
|
nsresult rv = mTimeUpdateTimer->Cancel();
|
|
mTimeUpdateTimer = nullptr;
|
|
return rv;
|
|
}
|
|
|
|
MediaDecoderOwner::NextFrameStatus AudioOffloadPlayer::GetNextFrameStatus()
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
if (mPlayState == MediaDecoder::PLAY_STATE_SEEKING) {
|
|
return MediaDecoderOwner::NEXT_FRAME_UNAVAILABLE_BUFFERING;
|
|
} else if (mPlayState == MediaDecoder::PLAY_STATE_ENDED) {
|
|
return MediaDecoderOwner::NEXT_FRAME_UNAVAILABLE;
|
|
} else {
|
|
return MediaDecoderOwner::NEXT_FRAME_AVAILABLE;
|
|
}
|
|
}
|
|
|
|
void AudioOffloadPlayer::SendMetaDataToHal(sp<AudioSink>& aSink,
|
|
const sp<MetaData>& aMeta)
|
|
{
|
|
int32_t sampleRate = 0;
|
|
int32_t bitRate = 0;
|
|
int32_t channelMask = 0;
|
|
int32_t delaySamples = 0;
|
|
int32_t paddingSamples = 0;
|
|
CHECK(aSink.get());
|
|
|
|
AudioParameter param = AudioParameter();
|
|
|
|
if (aMeta->findInt32(kKeySampleRate, &sampleRate)) {
|
|
param.addInt(String8(AUDIO_OFFLOAD_CODEC_SAMPLE_RATE), sampleRate);
|
|
}
|
|
if (aMeta->findInt32(kKeyChannelMask, &channelMask)) {
|
|
param.addInt(String8(AUDIO_OFFLOAD_CODEC_NUM_CHANNEL), channelMask);
|
|
}
|
|
if (aMeta->findInt32(kKeyBitRate, &bitRate)) {
|
|
param.addInt(String8(AUDIO_OFFLOAD_CODEC_AVG_BIT_RATE), bitRate);
|
|
}
|
|
if (aMeta->findInt32(kKeyEncoderDelay, &delaySamples)) {
|
|
param.addInt(String8(AUDIO_OFFLOAD_CODEC_DELAY_SAMPLES), delaySamples);
|
|
}
|
|
if (aMeta->findInt32(kKeyEncoderPadding, &paddingSamples)) {
|
|
param.addInt(String8(AUDIO_OFFLOAD_CODEC_PADDING_SAMPLES), paddingSamples);
|
|
}
|
|
|
|
AUDIO_OFFLOAD_LOG(PR_LOG_DEBUG, ("SendMetaDataToHal: bitRate %d,"
|
|
" sampleRate %d, chanMask %d, delaySample %d, paddingSample %d", bitRate,
|
|
sampleRate, channelMask, delaySamples, paddingSamples));
|
|
|
|
aSink->SetParameters(param.toString());
|
|
return;
|
|
}
|
|
|
|
void AudioOffloadPlayer::SetVolume(double aVolume)
|
|
{
|
|
MOZ_ASSERT(NS_IsMainThread());
|
|
CHECK(mAudioSink.get());
|
|
mAudioSink->SetVolume((float) aVolume);
|
|
}
|
|
|
|
} // namespace mozilla
|