Bug 1336431: P1. Fix coding style. r?gerald
MozReview-Commit-ID: GIwj8WxE9BA
--- a/dom/media/platforms/PlatformDecoderModule.h
+++ b/dom/media/platforms/PlatformDecoderModule.h
@@ -136,18 +136,19 @@ class PlatformDecoderModule
public:
NS_INLINE_DECL_THREADSAFE_REFCOUNTING(PlatformDecoderModule)
// Perform any per-instance initialization.
// This is called on the decode task queue.
virtual nsresult Startup() { return NS_OK; };
// Indicates if the PlatformDecoderModule supports decoding of aMimeType.
- virtual bool SupportsMimeType(const nsACString& aMimeType,
- DecoderDoctorDiagnostics* aDiagnostics) const = 0;
+ virtual bool SupportsMimeType(
+ const nsACString& aMimeType,
+ DecoderDoctorDiagnostics* aDiagnostics) const = 0;
virtual bool Supports(const TrackInfo& aTrackInfo,
DecoderDoctorDiagnostics* aDiagnostics) const
{
// By default, fall back to SupportsMimeType with just the MIME string.
// (So PDMs do not need to override this method -- yet.)
return SupportsMimeType(aTrackInfo.mMimeType, aDiagnostics);
}
@@ -156,21 +157,22 @@ public:
kNeedNone,
kNeedAVCC,
kNeedAnnexB,
};
// Indicates that the decoder requires a specific format.
// The PlatformDecoderModule will convert the demuxed data accordingly before
// feeding it to MediaDataDecoder::Input.
- virtual ConversionRequired DecoderNeedsConversion(const TrackInfo& aConfig) const = 0;
+ virtual ConversionRequired DecoderNeedsConversion(
+ const TrackInfo& aConfig) const = 0;
protected:
- PlatformDecoderModule() {}
- virtual ~PlatformDecoderModule() {}
+ PlatformDecoderModule() { }
+ virtual ~PlatformDecoderModule() { }
friend class H264Converter;
friend class PDMFactory;
friend class dom::RemoteDecoderModule;
friend class EMEDecoderModule;
// Creates a Video decoder. The layers backend is passed in so that
// decoders can determine whether hardware accelerated decoding can be used.
@@ -271,17 +273,21 @@ public:
// completed shutdown. The reader calls Flush() before calling Shutdown(). The
// reader will delete the decoder once the promise is resolved.
// The ShutdownPromise must only ever be resolved.
virtual RefPtr<ShutdownPromise> Shutdown() = 0;
// Called from the state machine task queue or main thread. Decoder needs to
// decide whether or not hardware acceleration is supported after creating.
// It doesn't need to call Init() before calling this function.
- virtual bool IsHardwareAccelerated(nsACString& aFailureReason) const { return false; }
+ virtual bool IsHardwareAccelerated(nsACString& aFailureReason) const
+ {
+ return false;
+ }
+
// Return the name of the MediaDataDecoder, only used for decoding.
// Only return a static const string, as the information may be accessed
// in a non thread-safe fashion.
virtual const char* GetDescriptionName() const = 0;
// Set a hint of seek target time to decoder. Decoder will drop any decoded
// data which pts is smaller than this value. This threshold needs to be clear
--- a/dom/media/platforms/wrappers/H264Converter.cpp
+++ b/dom/media/platforms/wrappers/H264Converter.cpp
@@ -48,18 +48,18 @@ H264Converter::Init()
// We haven't been able to initialize a decoder due to a missing SPS/PPS.
return MediaDataDecoder::InitPromise::CreateAndResolve(
TrackType::kVideoTrack, __func__);
}
RefPtr<MediaDataDecoder::DecodePromise>
H264Converter::Decode(MediaRawData* aSample)
{
- MOZ_RELEASE_ASSERT(!mDecodePromiseRequest.Exists() &&
- !mInitPromiseRequest.Exists(),
+ MOZ_RELEASE_ASSERT(!mDecodePromiseRequest.Exists()
+ && !mInitPromiseRequest.Exists(),
"Can't request a new decode until previous one completed");
if (!mp4_demuxer::AnnexB::ConvertSampleToAVCC(aSample)) {
// We need AVCC content to be able to later parse the SPS.
// This is a no-op if the data is already AVCC.
return DecodePromise::CreateAndReject(
MediaResult(NS_ERROR_OUT_OF_MEMORY, RESULT_DETAIL("ConvertSampleToAVCC")),
__func__);
@@ -92,18 +92,18 @@ H264Converter::Decode(MediaRawData* aSam
RESULT_DETAIL("Unable to create H264 decoder")),
__func__);
}
if (mNeedKeyframe && !aSample->mKeyframe) {
return DecodePromise::CreateAndResolve(DecodedData(), __func__);
}
- if (!mNeedAVCC &&
- !mp4_demuxer::AnnexB::ConvertSampleToAnnexB(aSample, mNeedKeyframe)) {
+ if (!mNeedAVCC
+ && !mp4_demuxer::AnnexB::ConvertSampleToAnnexB(aSample, mNeedKeyframe)) {
return DecodePromise::CreateAndReject(
MediaResult(NS_ERROR_OUT_OF_MEMORY,
RESULT_DETAIL("ConvertSampleToAnnexB")),
__func__);
}
mNeedKeyframe = false;
@@ -171,18 +171,18 @@ H264Converter::CreateDecoder(DecoderDoct
return NS_ERROR_NOT_INITIALIZED;
}
UpdateConfigFromExtraData(mCurrentConfig.mExtraData);
mp4_demuxer::SPSData spsdata;
if (mp4_demuxer::H264::DecodeSPSFromExtraData(mCurrentConfig.mExtraData, spsdata)) {
// Do some format check here.
// WMF H.264 Video Decoder and Apple ATDecoder do not support YUV444 format.
- if (spsdata.profile_idc == 244 /* Hi444PP */ ||
- spsdata.chroma_format_idc == PDMFactory::kYUV444) {
+ if (spsdata.profile_idc == 244 /* Hi444PP */
+ || spsdata.chroma_format_idc == PDMFactory::kYUV444) {
mLastError = NS_ERROR_FAILURE;
if (aDiagnostics) {
aDiagnostics->SetVideoNotSupported();
}
return NS_ERROR_FAILURE;
}
} else {
// SPS was invalid.
@@ -241,18 +241,18 @@ void
H264Converter::OnDecoderInitDone(const TrackType aTrackType)
{
mInitPromiseRequest.Complete();
RefPtr<MediaRawData> sample = mPendingSample.forget();
if (mNeedKeyframe && !sample->mKeyframe) {
mDecodePromise.ResolveIfExists(DecodedData(), __func__);
}
mNeedKeyframe = false;
- if (!mNeedAVCC &&
- !mp4_demuxer::AnnexB::ConvertSampleToAnnexB(sample, mNeedKeyframe)) {
+ if (!mNeedAVCC
+ && !mp4_demuxer::AnnexB::ConvertSampleToAnnexB(sample, mNeedKeyframe)) {
mDecodePromise.RejectIfExists(
MediaResult(NS_ERROR_OUT_OF_MEMORY,
RESULT_DETAIL("ConvertSampleToAnnexB")),
__func__);
return;
}
RefPtr<H264Converter> self = this;
mDecoder->Decode(sample)
@@ -278,42 +278,43 @@ H264Converter::OnDecoderInitFailed(const
__func__);
}
nsresult
H264Converter::CheckForSPSChange(MediaRawData* aSample)
{
RefPtr<MediaByteBuffer> extra_data =
mp4_demuxer::AnnexB::ExtractExtraData(aSample);
- if (!mp4_demuxer::AnnexB::HasSPS(extra_data) ||
- mp4_demuxer::AnnexB::CompareExtraData(extra_data,
- mCurrentConfig.mExtraData)) {
+ if (!mp4_demuxer::AnnexB::HasSPS(extra_data)
+ || mp4_demuxer::AnnexB::CompareExtraData(extra_data,
+ mCurrentConfig.mExtraData)) {
return NS_OK;
}
- if (MediaPrefs::MediaDecoderCheckRecycling() &&
- mDecoder->SupportDecoderRecycling()) {
+ if (MediaPrefs::MediaDecoderCheckRecycling()
+ && mDecoder->SupportDecoderRecycling()) {
// Do not recreate the decoder, reuse it.
UpdateConfigFromExtraData(extra_data);
mNeedKeyframe = true;
return NS_OK;
}
// The SPS has changed, signal to flush the current decoder and create a
// new one.
mDecoder->Flush();
Shutdown();
return CreateDecoderAndInit(aSample);
}
void
H264Converter::UpdateConfigFromExtraData(MediaByteBuffer* aExtraData)
{
mp4_demuxer::SPSData spsdata;
- if (mp4_demuxer::H264::DecodeSPSFromExtraData(aExtraData, spsdata) &&
- spsdata.pic_width > 0 && spsdata.pic_height > 0) {
+ if (mp4_demuxer::H264::DecodeSPSFromExtraData(aExtraData, spsdata)
+ && spsdata.pic_width > 0
+ && spsdata.pic_height > 0) {
mp4_demuxer::H264::EnsureSPSIsSane(spsdata);
mCurrentConfig.mImage.width = spsdata.pic_width;
mCurrentConfig.mImage.height = spsdata.pic_height;
mCurrentConfig.mDisplay.width = spsdata.display_width;
mCurrentConfig.mDisplay.height = spsdata.display_height;
}
mCurrentConfig.mExtraData = aExtraData;
}