Sign Up
Log In
Log In
or
Sign Up
Places
All Projects
Status Monitor
Collapse sidebar
openSUSE:Slowroll
MozillaFirefox
mozilla-silence-no-return-type.patch
Overview
Repositories
Revisions
Requests
Users
Attributes
Meta
File mozilla-silence-no-return-type.patch of Package MozillaFirefox
# HG changeset patch # Parent af0655f894a27ef60aa8438af7939a5ebc498df0 diff --git a/gfx/skia/skia/include/codec/SkEncodedOrigin.h b/gfx/skia/skia/include/codec/SkEncodedOrigin.h --- a/gfx/skia/skia/include/codec/SkEncodedOrigin.h +++ b/gfx/skia/skia/include/codec/SkEncodedOrigin.h @@ -36,16 +36,17 @@ static inline SkMatrix SkEncodedOriginTo case kBottomRight_SkEncodedOrigin: return SkMatrix::MakeAll(-1, 0, w, 0, -1, h, 0, 0, 1); case kBottomLeft_SkEncodedOrigin: return SkMatrix::MakeAll( 1, 0, 0, 0, -1, h, 0, 0, 1); case kLeftTop_SkEncodedOrigin: return SkMatrix::MakeAll( 0, 1, 0, 1, 0, 0, 0, 0, 1); case kRightTop_SkEncodedOrigin: return SkMatrix::MakeAll( 0, -1, w, 1, 0, 0, 0, 0, 1); case kRightBottom_SkEncodedOrigin: return SkMatrix::MakeAll( 0, -1, w, -1, 0, h, 0, 0, 1); case kLeftBottom_SkEncodedOrigin: return SkMatrix::MakeAll( 0, 1, 0, -1, 0, h, 0, 0, 1); } SK_ABORT("Unexpected origin"); + SkUNREACHABLE; } /** * Return true if the encoded origin includes a 90 degree rotation, in which case the width * and height of the source data are swapped relative to a correctly oriented destination. */ static inline bool SkEncodedOriginSwapsWidthHeight(SkEncodedOrigin origin) { return origin >= kLeftTop_SkEncodedOrigin; diff --git a/gfx/skia/skia/include/private/gpu/ganesh/GrTypesPriv.h b/gfx/skia/skia/include/private/gpu/ganesh/GrTypesPriv.h --- a/gfx/skia/skia/include/private/gpu/ganesh/GrTypesPriv.h +++ b/gfx/skia/skia/include/private/gpu/ganesh/GrTypesPriv.h @@ -320,16 +320,17 @@ static inline bool GrTextureTypeHasRestr return false; case GrTextureType::kRectangle: return true; case GrTextureType::kExternal: return true; default: SK_ABORT("Unexpected texture type"); } + SkUNREACHABLE; } ////////////////////////////////////////////////////////////////////////////// /** * Types used to describe format of vertices in arrays. */ enum GrVertexAttribType { diff --git a/gfx/skia/skia/src/core/SkDescriptor.cpp b/gfx/skia/skia/src/core/SkDescriptor.cpp --- a/gfx/skia/skia/src/core/SkDescriptor.cpp +++ b/gfx/skia/skia/src/core/SkDescriptor.cpp @@ -21,16 +21,17 @@ std::unique_ptr<SkDescriptor> SkDescript SkASSERT(length >= sizeof(SkDescriptor) && SkAlign4(length) == length); void* allocation = ::operator new(length); return std::unique_ptr<SkDescriptor>(new (allocation) SkDescriptor{}); } void SkDescriptor::operator delete(void* p) { ::operator delete(p); } void* SkDescriptor::operator new(size_t) { SK_ABORT("Descriptors are created with placement new."); + SkUNREACHABLE; } void SkDescriptor::flatten(SkWriteBuffer& buffer) const { buffer.writePad32(static_cast<const void*>(this), this->fLength); } void* SkDescriptor::addEntry(uint32_t tag, size_t length, const void* data) { SkASSERT(tag); diff --git a/gfx/skia/skia/src/core/SkGeometry.h b/gfx/skia/skia/src/core/SkGeometry.h --- a/gfx/skia/skia/src/core/SkGeometry.h +++ b/gfx/skia/skia/src/core/SkGeometry.h @@ -276,28 +276,30 @@ static inline bool SkCubicIsDegenerate(S case SkCubicType::kLocalCusp: case SkCubicType::kCuspAtInfinity: return false; case SkCubicType::kQuadratic: case SkCubicType::kLineOrPoint: return true; } SK_ABORT("Invalid SkCubicType"); + SkUNREACHABLE; } static inline const char* SkCubicTypeName(SkCubicType type) { switch (type) { case SkCubicType::kSerpentine: return "kSerpentine"; case SkCubicType::kLoop: return "kLoop"; case SkCubicType::kLocalCusp: return "kLocalCusp"; case SkCubicType::kCuspAtInfinity: return "kCuspAtInfinity"; case SkCubicType::kQuadratic: return "kQuadratic"; case SkCubicType::kLineOrPoint: return "kLineOrPoint"; } SK_ABORT("Invalid SkCubicType"); + SkUNREACHABLE; } /** Returns the cubic classification. t[],s[] are set to the two homogeneous parameter values at which points the lines L & M intersect with K, sorted from smallest to largest and oriented so positive values of the implicit are on the "left" side. For a serpentine curve they are the inflection points. For a loop they are the double point. For a local cusp, they are both equal and denote the cusp point. diff --git a/gfx/skia/skia/src/core/SkTextBlob.cpp b/gfx/skia/skia/src/core/SkTextBlob.cpp --- a/gfx/skia/skia/src/core/SkTextBlob.cpp +++ b/gfx/skia/skia/src/core/SkTextBlob.cpp @@ -199,16 +199,17 @@ unsigned SkTextBlob::ScalarsPerGlyph(Gly } void SkTextBlob::operator delete(void* p) { sk_free(p); } void* SkTextBlob::operator new(size_t) { SK_ABORT("All blobs are created by placement new."); + SkUNREACHABLE; } void* SkTextBlob::operator new(size_t, void* p) { return p; } SkTextBlobRunIterator::SkTextBlobRunIterator(const SkTextBlob* blob) : fCurrentRun(SkTextBlob::RunRecord::First(blob)) { diff --git a/gfx/skia/skia/src/core/SkTypeface_remote.h b/gfx/skia/skia/src/core/SkTypeface_remote.h --- a/gfx/skia/skia/src/core/SkTypeface_remote.h +++ b/gfx/skia/skia/src/core/SkTypeface_remote.h @@ -90,83 +90,97 @@ public: SkTypefaceID remoteTypefaceID() const {return fTypefaceID;} int glyphCount() const {return fGlyphCount;} bool isLogging() const {return fIsLogging;} protected: - int onGetUPEM() const override { SK_ABORT("Should never be called."); } + int onGetUPEM() const override { SK_ABORT("Should never be called."); SkUNREACHABLE; } std::unique_ptr<SkStreamAsset> onOpenStream(int* ttcIndex) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } sk_sp<SkTypeface> onMakeClone(const SkFontArguments& args) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } bool onGlyphMaskNeedsCurrentColor() const override { return fGlyphMaskNeedsCurrentColor; } int onGetVariationDesignPosition(SkFontArguments::VariationPosition::Coordinate coordinates[], int coordinateCount) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } int onGetVariationDesignParameters(SkFontParameters::Variation::Axis parameters[], int parameterCount) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } void onGetFamilyName(SkString* familyName) const override { // Used by SkStrikeCache::DumpMemoryStatistics. *familyName = ""; } bool onGetPostScriptName(SkString*) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } SkTypeface::LocalizedStrings* onCreateFamilyNameIterator() const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } int onGetTableTags(SkFontTableTag tags[]) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } size_t onGetTableData(SkFontTableTag, size_t offset, size_t length, void* data) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } std::unique_ptr<SkScalerContext> onCreateScalerContext( const SkScalerContextEffects& effects, const SkDescriptor* desc) const override { return std::make_unique<SkScalerContextProxy>( sk_ref_sp(const_cast<SkTypefaceProxy*>(this)), effects, desc, fDiscardableManager); } void onFilterRec(SkScalerContextRec* rec) const override { // The rec filtering is already applied by the server when generating // the glyphs. } void onGetFontDescriptor(SkFontDescriptor*, bool*) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } void getGlyphToUnicodeMap(SkUnichar*) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } void getPostScriptGlyphNames(SkString*) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } std::unique_ptr<SkAdvancedTypefaceMetrics> onGetAdvancedMetrics() const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } void onCharsToGlyphs(const SkUnichar* chars, int count, SkGlyphID glyphs[]) const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } int onCountGlyphs() const override { return this->glyphCount(); } void* onGetCTFontRef() const override { SK_ABORT("Should never be called."); + SkUNREACHABLE; } private: const SkTypefaceID fTypefaceID; const int fGlyphCount; const bool fIsLogging; const bool fGlyphMaskNeedsCurrentColor; sk_sp<SkStrikeClient::DiscardableHandleManager> fDiscardableManager; diff --git a/gfx/skia/skia/src/effects/imagefilters/SkBlurImageFilter.cpp b/gfx/skia/skia/src/effects/imagefilters/SkBlurImageFilter.cpp --- a/gfx/skia/skia/src/effects/imagefilters/SkBlurImageFilter.cpp +++ b/gfx/skia/skia/src/effects/imagefilters/SkBlurImageFilter.cpp @@ -810,16 +810,17 @@ sk_sp<SkSpecialImage> cpu_blur( SkASSERT(0 <= sigma && sigma <= 2183); // should be guaranteed after map_sigma if (PassMaker* maker = GaussPass::MakeMaker(sigma, &alloc)) { return maker; } if (PassMaker* maker = TentPass::MakeMaker(sigma, &alloc)) { return maker; } SK_ABORT("Sigma is out of range."); + SkUNREACHABLE; }; PassMaker* makerX = makeMaker(sigma.x()); PassMaker* makerY = makeMaker(sigma.y()); if (makerX->window() <= 1 && makerY->window() <= 1) { return copy_image_with_bounds(ctx, input, srcBounds, dstBounds); } diff --git a/gfx/skia/skia/src/fonts/SkFontMgr_indirect.cpp b/gfx/skia/skia/src/fonts/SkFontMgr_indirect.cpp --- a/gfx/skia/skia/src/fonts/SkFontMgr_indirect.cpp +++ b/gfx/skia/skia/src/fonts/SkFontMgr_indirect.cpp @@ -64,16 +64,17 @@ int SkFontMgr_Indirect::onCountFamilies( } void SkFontMgr_Indirect::onGetFamilyName(int index, SkString* familyName) const { SK_ABORT("Not implemented"); } SkFontStyleSet* SkFontMgr_Indirect::onCreateStyleSet(int index) const { SK_ABORT("Not implemented"); + SkUNREACHABLE; } SkFontStyleSet* SkFontMgr_Indirect::onMatchFamily(const char familyName[]) const { return new SkStyleSet_Indirect(this, -1, fProxy->matchName(familyName)); } SkTypeface* SkFontMgr_Indirect::createTypefaceFromFontId(const SkFontIdentity& id) const { if (id.fDataId == SkFontIdentity::kInvalidDataId) { diff --git a/gfx/skia/skia/src/ports/SkFontMgr_FontConfigInterface.cpp b/gfx/skia/skia/src/ports/SkFontMgr_FontConfigInterface.cpp --- a/gfx/skia/skia/src/ports/SkFontMgr_FontConfigInterface.cpp +++ b/gfx/skia/skia/src/ports/SkFontMgr_FontConfigInterface.cpp @@ -157,28 +157,32 @@ public: SkFontMgr_FCI(sk_sp<SkFontConfigInterface> fci) : fFCI(std::move(fci)) , fCache(kMaxSize) {} protected: int onCountFamilies() const override { SK_ABORT("Not implemented."); + SkUNREACHABLE; } void onGetFamilyName(int index, SkString* familyName) const override { SK_ABORT("Not implemented."); + SkUNREACHABLE; } SkFontStyleSet* onCreateStyleSet(int index) const override { SK_ABORT("Not implemented."); + SkUNREACHABLE; } SkFontStyleSet* onMatchFamily(const char familyName[]) const override { SK_ABORT("Not implemented."); + SkUNREACHABLE; } SkTypeface* onMatchFamilyStyle(const char requestedFamilyName[], const SkFontStyle& requestedStyle) const override { SkAutoMutexExclusive ama(fMutex); SkFontConfigInterface::FontIdentity identity; diff --git a/gfx/skia/skia/src/sksl/SkSLCompiler.cpp b/gfx/skia/skia/src/sksl/SkSLCompiler.cpp --- a/gfx/skia/skia/src/sksl/SkSLCompiler.cpp +++ b/gfx/skia/skia/src/sksl/SkSLCompiler.cpp @@ -272,16 +272,17 @@ std::unique_ptr<Expression> Compiler::co case Symbol::Kind::kType: { // go through DSLType so we report errors on private types dsl::DSLModifiers modifiers; dsl::DSLType dslType(result->name(), &modifiers, pos); return TypeReference::Convert(*fContext, pos, &dslType.skslType()); } default: SK_ABORT("unsupported symbol type %d\n", (int) result->kind()); + SkUNREACHABLE; } } bool Compiler::optimizeModuleBeforeMinifying(ProgramKind kind, Module& module) { SkASSERT(this->errorCount() == 0); auto m = SkSL::ModuleLoader::Get(); diff --git a/gfx/skia/skia/src/sksl/SkSLOperator.cpp b/gfx/skia/skia/src/sksl/SkSLOperator.cpp --- a/gfx/skia/skia/src/sksl/SkSLOperator.cpp +++ b/gfx/skia/skia/src/sksl/SkSLOperator.cpp @@ -46,17 +46,17 @@ OperatorPrecedence Operator::getBinaryPr case Kind::SLASHEQ: // fall through case Kind::PERCENTEQ: // fall through case Kind::SHLEQ: // fall through case Kind::SHREQ: // fall through case Kind::BITWISEANDEQ: // fall through case Kind::BITWISEXOREQ: // fall through case Kind::BITWISEOREQ: return OperatorPrecedence::kAssignment; case Kind::COMMA: return OperatorPrecedence::kSequence; - default: SK_ABORT("unsupported binary operator"); + default: SkUNREACHABLE; } } const char* Operator::operatorName() const { switch (this->kind()) { case Kind::PLUS: return " + "; case Kind::MINUS: return " - "; case Kind::STAR: return " * "; diff --git a/gfx/skia/skia/src/sksl/ir/SkSLType.h b/gfx/skia/skia/src/sksl/ir/SkSLType.h --- a/gfx/skia/skia/src/sksl/ir/SkSLType.h +++ b/gfx/skia/skia/src/sksl/ir/SkSLType.h @@ -417,16 +417,17 @@ public: * Returns the number of scalars needed to hold this type. */ virtual size_t slotCount() const { return 0; } virtual const std::vector<Field>& fields() const { SK_ABORT("Internal error: not a struct"); + SkUNREACHABLE; } /** * For generic types, returns the types that this generic type can substitute for. */ virtual SkSpan<const Type* const> coercibleTypes() const { SkDEBUGFAIL("Internal error: not a generic type"); return {}; diff --git a/gfx/skia/skia/src/utils/SkShadowUtils.cpp b/gfx/skia/skia/src/utils/SkShadowUtils.cpp --- a/gfx/skia/skia/src/utils/SkShadowUtils.cpp +++ b/gfx/skia/skia/src/utils/SkShadowUtils.cpp @@ -135,16 +135,17 @@ struct SpotVerticesFactory { } return false; case OccluderType::kDirectional: case OccluderType::kDirectionalTransparent: *translate = that.fOffset - fOffset; return true; } SK_ABORT("Uninitialized occluder type?"); + SkUNREACHABLE; } sk_sp<SkVertices> makeVertices(const SkPath& path, const SkMatrix& ctm, SkVector* translate) const { bool transparent = fOccluderType == OccluderType::kPointTransparent || fOccluderType == OccluderType::kDirectionalTransparent; bool directional = fOccluderType == OccluderType::kDirectional || fOccluderType == OccluderType::kDirectionalTransparent; diff --git a/intl/icu/source/i18n/number_rounding.cpp b/intl/icu/source/i18n/number_rounding.cpp --- a/intl/icu/source/i18n/number_rounding.cpp +++ b/intl/icu/source/i18n/number_rounding.cpp @@ -282,27 +282,29 @@ Precision IncrementPrecision::withMinFra } FractionPrecision Precision::constructFraction(int32_t minFrac, int32_t maxFrac) { FractionSignificantSettings settings; settings.fMinFrac = static_cast<digits_t>(minFrac); settings.fMaxFrac = static_cast<digits_t>(maxFrac); settings.fMinSig = -1; settings.fMaxSig = -1; + settings.fPriority = UNUM_ROUNDING_PRIORITY_RELAXED; PrecisionUnion union_; union_.fracSig = settings; return {RND_FRACTION, union_}; } Precision Precision::constructSignificant(int32_t minSig, int32_t maxSig) { FractionSignificantSettings settings; settings.fMinFrac = -1; settings.fMaxFrac = -1; settings.fMinSig = static_cast<digits_t>(minSig); settings.fMaxSig = static_cast<digits_t>(maxSig); + settings.fPriority = UNUM_ROUNDING_PRIORITY_RELAXED; PrecisionUnion union_; union_.fracSig = settings; return {RND_SIGNIFICANT, union_}; } Precision Precision::constructFractionSignificant( const FractionPrecision &base, diff --git a/js/src/irregexp/imported/regexp-parser.cc b/js/src/irregexp/imported/regexp-parser.cc --- a/js/src/irregexp/imported/regexp-parser.cc +++ b/js/src/irregexp/imported/regexp-parser.cc @@ -2764,16 +2764,17 @@ bool MayContainStrings(ClassSetOperandTy return false; case ClassSetOperandType::kCharacterClassEscape: case ClassSetOperandType::kClassStringDisjunction: return operand->AsClassSetOperand()->has_strings(); case ClassSetOperandType::kNestedClass: if (operand->IsClassRanges()) return false; return operand->AsClassSetExpression()->may_contain_strings(); } + UNREACHABLE(); } } // namespace template <class CharT> void RegExpParserImpl<CharT>::AddMaybeSimpleCaseFoldedRange( ZoneList<CharacterRange>* ranges, CharacterRange new_range) { DCHECK(unicode_sets()); diff --git a/third_party/libwebrtc/api/adaptation/resource.cc b/third_party/libwebrtc/api/adaptation/resource.cc --- a/third_party/libwebrtc/api/adaptation/resource.cc +++ b/third_party/libwebrtc/api/adaptation/resource.cc @@ -17,16 +17,17 @@ namespace webrtc { const char* ResourceUsageStateToString(ResourceUsageState usage_state) { switch (usage_state) { case ResourceUsageState::kOveruse: return "kOveruse"; case ResourceUsageState::kUnderuse: return "kUnderuse"; } RTC_CHECK_NOTREACHED(); + return nullptr; } ResourceListener::~ResourceListener() {} Resource::Resource() {} Resource::~Resource() {} diff --git a/third_party/libwebrtc/api/rtp_parameters.cc b/third_party/libwebrtc/api/rtp_parameters.cc --- a/third_party/libwebrtc/api/rtp_parameters.cc +++ b/third_party/libwebrtc/api/rtp_parameters.cc @@ -27,16 +27,17 @@ const char* DegradationPreferenceToStrin case DegradationPreference::MAINTAIN_FRAMERATE: return "maintain-framerate"; case DegradationPreference::MAINTAIN_RESOLUTION: return "maintain-resolution"; case DegradationPreference::BALANCED: return "balanced"; } RTC_CHECK_NOTREACHED(); + return ""; } const double kDefaultBitratePriority = 1.0; RtcpFeedback::RtcpFeedback() = default; RtcpFeedback::RtcpFeedback(RtcpFeedbackType type) : type(type) {} RtcpFeedback::RtcpFeedback(RtcpFeedbackType type, RtcpFeedbackMessageType message_type) diff --git a/third_party/libwebrtc/api/video/video_frame_buffer.cc b/third_party/libwebrtc/api/video/video_frame_buffer.cc --- a/third_party/libwebrtc/api/video/video_frame_buffer.cc +++ b/third_party/libwebrtc/api/video/video_frame_buffer.cc @@ -101,16 +101,18 @@ const char* VideoFrameBufferTypeToString return "kI210"; case VideoFrameBuffer::Type::kI410: return "kI410"; case VideoFrameBuffer::Type::kNV12: return "kNV12"; default: RTC_DCHECK_NOTREACHED(); } + RTC_DCHECK_NOTREACHED(); + return nullptr; } int I420BufferInterface::ChromaWidth() const { return (width() + 1) / 2; } int I420BufferInterface::ChromaHeight() const { return (height() + 1) / 2; diff --git a/third_party/libwebrtc/api/video_codecs/video_codec.cc b/third_party/libwebrtc/api/video_codecs/video_codec.cc --- a/third_party/libwebrtc/api/video_codecs/video_codec.cc +++ b/third_party/libwebrtc/api/video_codecs/video_codec.cc @@ -126,16 +126,17 @@ const char* CodecTypeToPayloadString(Vid case kVideoCodecMultiplex: return kPayloadNameMultiplex; case kVideoCodecGeneric: return kPayloadNameGeneric; case kVideoCodecH265: return kPayloadNameH265; } RTC_CHECK_NOTREACHED(); + return ""; } VideoCodecType PayloadStringToCodecType(const std::string& name) { if (absl::EqualsIgnoreCase(name, kPayloadNameVp8)) return kVideoCodecVP8; if (absl::EqualsIgnoreCase(name, kPayloadNameVp9)) return kVideoCodecVP9; if (absl::EqualsIgnoreCase(name, kPayloadNameAv1) || diff --git a/third_party/libwebrtc/api/video_codecs/video_encoder_software_fallback_wrapper.cc b/third_party/libwebrtc/api/video_codecs/video_encoder_software_fallback_wrapper.cc --- a/third_party/libwebrtc/api/video_codecs/video_encoder_software_fallback_wrapper.cc +++ b/third_party/libwebrtc/api/video_codecs/video_encoder_software_fallback_wrapper.cc @@ -183,16 +183,17 @@ class VideoEncoderSoftwareFallbackWrappe [[fallthrough]]; case EncoderState::kMainEncoderUsed: return encoder_.get(); case EncoderState::kFallbackDueToFailure: case EncoderState::kForcedFallback: return fallback_encoder_.get(); } RTC_CHECK_NOTREACHED(); + return nullptr; } // Updates encoder with last observed parameters, such as callbacks, rates, // etc. void PrimeEncoder(VideoEncoder* encoder) const; // Settings used in the last InitEncode call and used if a dynamic fallback to // software is required. @@ -363,16 +364,17 @@ int32_t VideoEncoderSoftwareFallbackWrap case EncoderState::kMainEncoderUsed: { return EncodeWithMainEncoder(frame, frame_types); } case EncoderState::kFallbackDueToFailure: case EncoderState::kForcedFallback: return fallback_encoder_->Encode(frame, frame_types); } RTC_CHECK_NOTREACHED(); + return WEBRTC_VIDEO_CODEC_ERROR; } int32_t VideoEncoderSoftwareFallbackWrapper::EncodeWithMainEncoder( const VideoFrame& frame, const std::vector<VideoFrameType>* frame_types) { int32_t ret = encoder_->Encode(frame, frame_types); // If requested, try a software fallback. bool fallback_requested = (ret == WEBRTC_VIDEO_CODEC_FALLBACK_SOFTWARE); diff --git a/third_party/libwebrtc/call/adaptation/video_stream_adapter.cc b/third_party/libwebrtc/call/adaptation/video_stream_adapter.cc --- a/third_party/libwebrtc/call/adaptation/video_stream_adapter.cc +++ b/third_party/libwebrtc/call/adaptation/video_stream_adapter.cc @@ -164,16 +164,17 @@ const char* Adaptation::StatusToString(A case Status::kInsufficientInput: return "kInsufficientInput"; case Status::kAdaptationDisabled: return "kAdaptationDisabled"; case Status::kRejectedByConstraint: return "kRejectedByConstraint"; } RTC_CHECK_NOTREACHED(); + return ""; } Adaptation::Adaptation(int validation_id, VideoSourceRestrictions restrictions, VideoAdaptationCounters counters, VideoStreamInputState input_state) : validation_id_(validation_id), status_(Status::kValid), @@ -386,16 +387,17 @@ VideoStreamAdapter::RestrictionsOrState case DegradationPreference::MAINTAIN_RESOLUTION: { // Scale up framerate. return IncreaseFramerate(input_state, current_restrictions_); } case DegradationPreference::DISABLED: return Adaptation::Status::kAdaptationDisabled; } RTC_CHECK_NOTREACHED(); + return Adaptation::Status::kAdaptationDisabled; } Adaptation VideoStreamAdapter::GetAdaptationDown() { RTC_DCHECK_RUN_ON(&sequence_checker_); VideoStreamInputState input_state = input_state_provider_->InputState(); ++adaptation_validation_id_; RestrictionsOrState restrictions_or_state = GetAdaptationDownStep(input_state, current_restrictions_); @@ -468,16 +470,17 @@ VideoStreamAdapter::GetAdaptationDownSte } case DegradationPreference::MAINTAIN_RESOLUTION: { return DecreaseFramerate(input_state, current_restrictions); } case DegradationPreference::DISABLED: return Adaptation::Status::kAdaptationDisabled; } RTC_CHECK_NOTREACHED(); + return Adaptation::Status::kAdaptationDisabled; } VideoStreamAdapter::RestrictionsOrState VideoStreamAdapter::DecreaseResolution( const VideoStreamInputState& input_state, const RestrictionsWithCounters& current_restrictions) { int target_pixels = GetLowerResolutionThan(input_state.frame_size_pixels().value()); // Use single active stream if set, this stream could be lower than the input. @@ -621,16 +624,18 @@ Adaptation VideoStreamAdapter::GetAdaptD case DegradationPreference::MAINTAIN_FRAMERATE: return GetAdaptationDown(); case DegradationPreference::BALANCED: { return RestrictionsOrStateToAdaptation( GetAdaptDownResolutionStepForBalanced(input_state), input_state); } } RTC_CHECK_NOTREACHED(); + return RestrictionsOrStateToAdaptation( + Adaptation::Status::kAdaptationDisabled, input_state); } VideoStreamAdapter::RestrictionsOrState VideoStreamAdapter::GetAdaptDownResolutionStepForBalanced( const VideoStreamInputState& input_state) const { // Adapt twice if the first adaptation did not decrease resolution. auto first_step = GetAdaptationDownStep(input_state, current_restrictions_); if (!absl::holds_alternative<RestrictionsWithCounters>(first_step)) { diff --git a/third_party/libwebrtc/call/rtp_payload_params.cc b/third_party/libwebrtc/call/rtp_payload_params.cc --- a/third_party/libwebrtc/call/rtp_payload_params.cc +++ b/third_party/libwebrtc/call/rtp_payload_params.cc @@ -405,17 +405,17 @@ absl::optional<FrameDependencyStructure> return structure; } case VideoCodecType::kVideoCodecAV1: case VideoCodecType::kVideoCodecH264: case VideoCodecType::kVideoCodecH265: case VideoCodecType::kVideoCodecMultiplex: return absl::nullopt; } - RTC_DCHECK_NOTREACHED() << "Unsupported codec."; + RTC_CHECK_NOTREACHED(); } void RtpPayloadParams::GenericToGeneric(int64_t shared_frame_id, bool is_keyframe, RTPVideoHeader* rtp_video_header) { RTPVideoHeader::GenericDescriptorInfo& generic = rtp_video_header->generic.emplace(); diff --git a/third_party/libwebrtc/call/video_send_stream.cc b/third_party/libwebrtc/call/video_send_stream.cc --- a/third_party/libwebrtc/call/video_send_stream.cc +++ b/third_party/libwebrtc/call/video_send_stream.cc @@ -25,16 +25,17 @@ const char* StreamTypeToString(VideoSend case VideoSendStream::StreamStats::StreamType::kMedia: return "media"; case VideoSendStream::StreamStats::StreamType::kRtx: return "rtx"; case VideoSendStream::StreamStats::StreamType::kFlexfec: return "flexfec"; } RTC_CHECK_NOTREACHED(); + return ""; } } // namespace VideoSendStream::StreamStats::StreamStats() = default; VideoSendStream::StreamStats::~StreamStats() = default; std::string VideoSendStream::StreamStats::ToString() const { diff --git a/third_party/libwebrtc/media/base/codec.cc b/third_party/libwebrtc/media/base/codec.cc --- a/third_party/libwebrtc/media/base/codec.cc +++ b/third_party/libwebrtc/media/base/codec.cc @@ -200,16 +200,17 @@ bool Codec::Matches(const Codec& codec) (codec.bitrate == 0 || bitrate <= 0 || bitrate == codec.bitrate) && ((codec.channels < 2 && channels < 2) || channels == codec.channels); case Type::kVideo: return IsSameCodecSpecific(name, params, codec.name, codec.params); } + return false; // unreached }; return matches_id && matches_type_specific(); } bool Codec::MatchesRtpCodec(const webrtc::RtpCodec& codec_capability) const { webrtc::RtpCodecParameters codec_parameters = ToCodecParameters(); diff --git a/third_party/libwebrtc/modules/audio_processing/agc2/clipping_predictor.cc b/third_party/libwebrtc/modules/audio_processing/agc2/clipping_predictor.cc --- a/third_party/libwebrtc/modules/audio_processing/agc2/clipping_predictor.cc +++ b/third_party/libwebrtc/modules/audio_processing/agc2/clipping_predictor.cc @@ -373,12 +373,12 @@ std::unique_ptr<ClippingPredictor> Creat config.reference_window_delay, config.clipping_threshold, /*adaptive_step_estimation=*/true); case ClippingPredictorMode::kFixedStepClippingPeakPrediction: return std::make_unique<ClippingPeakPredictor>( num_channels, config.window_length, config.reference_window_length, config.reference_window_delay, config.clipping_threshold, /*adaptive_step_estimation=*/false); } - RTC_DCHECK_NOTREACHED(); + RTC_CHECK_NOTREACHED(); } } // namespace webrtc diff --git a/third_party/libwebrtc/modules/audio_processing/agc2/input_volume_stats_reporter.cc b/third_party/libwebrtc/modules/audio_processing/agc2/input_volume_stats_reporter.cc --- a/third_party/libwebrtc/modules/audio_processing/agc2/input_volume_stats_reporter.cc +++ b/third_party/libwebrtc/modules/audio_processing/agc2/input_volume_stats_reporter.cc @@ -43,16 +43,17 @@ int ComputeAverageUpdate(int sum_updates constexpr absl::string_view MetricNamePrefix( InputVolumeType input_volume_type) { switch (input_volume_type) { case InputVolumeType::kApplied: return "WebRTC.Audio.Apm.AppliedInputVolume."; case InputVolumeType::kRecommended: return "WebRTC.Audio.Apm.RecommendedInputVolume."; } + RTC_CHECK_NOTREACHED(); } metrics::Histogram* CreateVolumeHistogram(InputVolumeType input_volume_type) { char buffer[64]; rtc::SimpleStringBuilder builder(buffer); builder << MetricNamePrefix(input_volume_type) << "OnChange"; return metrics::HistogramFactoryGetCountsLinear(/*name=*/builder.str(), /*min=*/1, diff --git a/third_party/libwebrtc/modules/audio_processing/agc2/rnn_vad/rnn_fc.cc b/third_party/libwebrtc/modules/audio_processing/agc2/rnn_vad/rnn_fc.cc --- a/third_party/libwebrtc/modules/audio_processing/agc2/rnn_vad/rnn_fc.cc +++ b/third_party/libwebrtc/modules/audio_processing/agc2/rnn_vad/rnn_fc.cc @@ -55,16 +55,18 @@ std::vector<float> PreprocessWeights(rtc rtc::FunctionView<float(float)> GetActivationFunction( ActivationFunction activation_function) { switch (activation_function) { case ActivationFunction::kTansigApproximated: return ::rnnoise::TansigApproximated; case ActivationFunction::kSigmoidApproximated: return ::rnnoise::SigmoidApproximated; } + // supposed to be never reached apparently therefore returning bogus + return ::rnnoise::TansigApproximated; } } // namespace FullyConnectedLayer::FullyConnectedLayer( const int input_size, const int output_size, const rtc::ArrayView<const int8_t> bias, diff --git a/third_party/libwebrtc/modules/audio_processing/audio_processing_impl.cc b/third_party/libwebrtc/modules/audio_processing/audio_processing_impl.cc --- a/third_party/libwebrtc/modules/audio_processing/audio_processing_impl.cc +++ b/third_party/libwebrtc/modules/audio_processing/audio_processing_impl.cc @@ -94,16 +94,17 @@ GainControl::Mode Agc1ConfigModeToInterf case Agc1Config::kAdaptiveAnalog: return GainControl::kAdaptiveAnalog; case Agc1Config::kAdaptiveDigital: return GainControl::kAdaptiveDigital; case Agc1Config::kFixedDigital: return GainControl::kFixedDigital; } RTC_CHECK_NOTREACHED(); + return GainControl::kAdaptiveAnalog; } bool MinimizeProcessingForUnusedOutput() { return !field_trial::IsEnabled("WebRTC-MutedStateKillSwitch"); } // Maximum lengths that frame of samples being passed from the render side to // the capture side can have (does not apply to AEC3). @@ -161,17 +162,17 @@ int AudioFormatValidityToErrorCode(Audio case AudioFormatValidity::kValidAndSupported: return AudioProcessing::kNoError; case AudioFormatValidity::kValidButUnsupportedSampleRate: // fall-through case AudioFormatValidity::kInvalidSampleRate: return AudioProcessing::kBadSampleRateError; case AudioFormatValidity::kInvalidChannelCount: return AudioProcessing::kBadNumberChannelsError; } - RTC_DCHECK(false); + RTC_CHECK_NOTREACHED(); } // Returns an AudioProcessing::Error together with the best possible option for // output audio content. std::pair<int, FormatErrorOutputOption> ChooseErrorOutputOption( const StreamConfig& input_config, const StreamConfig& output_config) { AudioFormatValidity input_validity = ValidateAudioFormat(input_config); @@ -2416,16 +2417,17 @@ void AudioProcessingImpl::InitializeNois case NoiseSuppresionConfig::kModerate: return NsConfig::SuppressionLevel::k12dB; case NoiseSuppresionConfig::kHigh: return NsConfig::SuppressionLevel::k18dB; case NoiseSuppresionConfig::kVeryHigh: return NsConfig::SuppressionLevel::k21dB; } RTC_CHECK_NOTREACHED(); + return NsConfig::SuppressionLevel::k6dB; }; NsConfig cfg; cfg.target_level = map_level(config_.noise_suppression.level); submodules_.noise_suppressor = std::make_unique<NoiseSuppressor>( cfg, proc_sample_rate_hz(), num_proc_channels()); } } diff --git a/third_party/libwebrtc/modules/audio_processing/include/audio_processing.cc b/third_party/libwebrtc/modules/audio_processing/include/audio_processing.cc --- a/third_party/libwebrtc/modules/audio_processing/include/audio_processing.cc +++ b/third_party/libwebrtc/modules/audio_processing/include/audio_processing.cc @@ -27,28 +27,30 @@ std::string NoiseSuppressionLevelToStrin case AudioProcessing::Config::NoiseSuppression::Level::kModerate: return "Moderate"; case AudioProcessing::Config::NoiseSuppression::Level::kHigh: return "High"; case AudioProcessing::Config::NoiseSuppression::Level::kVeryHigh: return "VeryHigh"; } RTC_CHECK_NOTREACHED(); + return ""; } std::string GainController1ModeToString(const Agc1Config::Mode& mode) { switch (mode) { case Agc1Config::Mode::kAdaptiveAnalog: return "AdaptiveAnalog"; case Agc1Config::Mode::kAdaptiveDigital: return "AdaptiveDigital"; case Agc1Config::Mode::kFixedDigital: return "FixedDigital"; } RTC_CHECK_NOTREACHED(); + return ""; } } // namespace constexpr int AudioProcessing::kNativeSampleRatesHz[]; void CustomProcessing::SetRuntimeSetting( AudioProcessing::RuntimeSetting setting) {} diff --git a/third_party/libwebrtc/modules/audio_processing/transient/transient_suppressor_impl.cc b/third_party/libwebrtc/modules/audio_processing/transient/transient_suppressor_impl.cc --- a/third_party/libwebrtc/modules/audio_processing/transient/transient_suppressor_impl.cc +++ b/third_party/libwebrtc/modules/audio_processing/transient/transient_suppressor_impl.cc @@ -48,16 +48,17 @@ std::string GetVadModeLabel(TransientSup switch (vad_mode) { case TransientSuppressor::VadMode::kDefault: return "default"; case TransientSuppressor::VadMode::kRnnVad: return "RNN VAD"; case TransientSuppressor::VadMode::kNoVad: return "no VAD"; } + RTC_CHECK_NOTREACHED(); } } // namespace TransientSuppressorImpl::TransientSuppressorImpl(VadMode vad_mode, int sample_rate_hz, int detector_rate_hz, int num_channels) diff --git a/third_party/libwebrtc/modules/congestion_controller/goog_cc/goog_cc_network_control.cc b/third_party/libwebrtc/modules/congestion_controller/goog_cc/goog_cc_network_control.cc --- a/third_party/libwebrtc/modules/congestion_controller/goog_cc/goog_cc_network_control.cc +++ b/third_party/libwebrtc/modules/congestion_controller/goog_cc/goog_cc_network_control.cc @@ -90,16 +90,18 @@ BandwidthLimitedCause GetBandwidthLimite // Probes may not be sent in this state. return BandwidthLimitedCause::kLossLimitedBwe; case LossBasedState::kIncreasing: // Probes may be sent in this state. return BandwidthLimitedCause::kLossLimitedBweIncreasing; case LossBasedState::kDelayBasedEstimate: return BandwidthLimitedCause::kDelayBasedLimited; } + // just return something by default + return BandwidthLimitedCause::kLossLimitedBwe; } } // namespace GoogCcNetworkController::GoogCcNetworkController(NetworkControllerConfig config, GoogCcConfig goog_cc_config) : key_value_config_(config.key_value_config ? config.key_value_config : &trial_based_config_), diff --git a/third_party/libwebrtc/modules/desktop_capture/linux/wayland/screencast_portal.cc b/third_party/libwebrtc/modules/desktop_capture/linux/wayland/screencast_portal.cc --- a/third_party/libwebrtc/modules/desktop_capture/linux/wayland/screencast_portal.cc +++ b/third_party/libwebrtc/modules/desktop_capture/linux/wayland/screencast_portal.cc @@ -39,16 +39,17 @@ ScreenCastPortal::CaptureSourceType Scre switch (type) { case CaptureType::kScreen: return ScreenCastPortal::CaptureSourceType::kScreen; case CaptureType::kWindow: return ScreenCastPortal::CaptureSourceType::kWindow; case CaptureType::kAnyScreenContent: return ScreenCastPortal::CaptureSourceType::kAnyScreenContent; } + RTC_CHECK_NOTREACHED(); } ScreenCastPortal::ScreenCastPortal(CaptureType type, PortalNotifier* notifier) : ScreenCastPortal(type, notifier, OnProxyRequested, OnSourcesRequestResponseSignal, this) {} diff --git a/third_party/libwebrtc/modules/pacing/bitrate_prober.cc b/third_party/libwebrtc/modules/pacing/bitrate_prober.cc --- a/third_party/libwebrtc/modules/pacing/bitrate_prober.cc +++ b/third_party/libwebrtc/modules/pacing/bitrate_prober.cc @@ -64,16 +64,17 @@ bool BitrateProber::ReadyToSetActiveStat return false; case ProbingState::kInactive: // If config_.min_packet_size > 0, a "large enough" packet must be sent // first, before a probe can be generated and sent. Otherwise, send the // probe asap. return packet_size >= std::min(RecommendedMinProbeSize(), config_.min_packet_size.Get()); } + RTC_CHECK_NOTREACHED(); } void BitrateProber::OnIncomingPacket(DataSize packet_size) { if (ReadyToSetActiveState(packet_size)) { next_probe_time_ = Timestamp::MinusInfinity(); probing_state_ = ProbingState::kActive; } } diff --git a/third_party/libwebrtc/modules/rtp_rtcp/source/create_video_rtp_depacketizer.cc b/third_party/libwebrtc/modules/rtp_rtcp/source/create_video_rtp_depacketizer.cc --- a/third_party/libwebrtc/modules/rtp_rtcp/source/create_video_rtp_depacketizer.cc +++ b/third_party/libwebrtc/modules/rtp_rtcp/source/create_video_rtp_depacketizer.cc @@ -36,11 +36,12 @@ std::unique_ptr<VideoRtpDepacketizer> Cr case kVideoCodecH265: // TODO(bugs.webrtc.org/13485): Implement VideoRtpDepacketizerH265. return nullptr; case kVideoCodecGeneric: case kVideoCodecMultiplex: return std::make_unique<VideoRtpDepacketizerGeneric>(); } RTC_CHECK_NOTREACHED(); + return nullptr; } } // namespace webrtc diff --git a/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender.cc b/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender.cc --- a/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender.cc +++ b/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender.cc @@ -133,16 +133,17 @@ bool IsNonVolatile(RTPExtensionType type #if defined(WEBRTC_MOZILLA_BUILD) case kRtpExtensionCsrcAudioLevel: // TODO: Mozilla implement for CsrcAudioLevel RTC_CHECK(false); return false; #endif } RTC_CHECK_NOTREACHED(); + return false; } bool HasBweExtension(const RtpHeaderExtensionMap& extensions_map) { return extensions_map.IsRegistered(kRtpExtensionTransportSequenceNumber) || extensions_map.IsRegistered(kRtpExtensionTransportSequenceNumber02) || extensions_map.IsRegistered(kRtpExtensionAbsoluteSendTime) || extensions_map.IsRegistered(kRtpExtensionTransmissionTimeOffset); } diff --git a/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender_audio.cc b/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender_audio.cc --- a/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender_audio.cc +++ b/third_party/libwebrtc/modules/rtp_rtcp/source/rtp_sender_audio.cc @@ -42,16 +42,17 @@ namespace { case AudioFrameType::kEmptyFrame: return "empty"; case AudioFrameType::kAudioFrameSpeech: return "audio_speech"; case AudioFrameType::kAudioFrameCN: return "audio_cn"; } RTC_CHECK_NOTREACHED(); + return ""; } } // namespace RTPSenderAudio::RTPSenderAudio(Clock* clock, RTPSender* rtp_sender) : clock_(clock), rtp_sender_(rtp_sender), absolute_capture_time_sender_(clock) { diff --git a/third_party/libwebrtc/modules/video_coding/codecs/vp8/default_temporal_layers.cc b/third_party/libwebrtc/modules/video_coding/codecs/vp8/default_temporal_layers.cc --- a/third_party/libwebrtc/modules/video_coding/codecs/vp8/default_temporal_layers.cc +++ b/third_party/libwebrtc/modules/video_coding/codecs/vp8/default_temporal_layers.cc @@ -105,16 +105,17 @@ size_t BufferToIndex(Vp8BufferReference return 0; case Vp8FrameConfig::Vp8BufferReference::kGolden: return 1; case Vp8FrameConfig::Vp8BufferReference::kAltref: return 2; case Vp8FrameConfig::Vp8BufferReference::kNone: RTC_CHECK_NOTREACHED(); } + RTC_CHECK_NOTREACHED(); } } // namespace constexpr size_t DefaultTemporalLayers::kNumReferenceBuffers; std::vector<DefaultTemporalLayers::DependencyInfo> DefaultTemporalLayers::GetDependencyInfo(size_t num_layers) { diff --git a/third_party/libwebrtc/modules/video_coding/codecs/vp8/temporal_layers_checker.cc b/third_party/libwebrtc/modules/video_coding/codecs/vp8/temporal_layers_checker.cc --- a/third_party/libwebrtc/modules/video_coding/codecs/vp8/temporal_layers_checker.cc +++ b/third_party/libwebrtc/modules/video_coding/codecs/vp8/temporal_layers_checker.cc @@ -25,16 +25,17 @@ TemporalLayersChecker::CreateTemporalLay case Vp8TemporalLayersType::kFixedPattern: return std::make_unique<DefaultTemporalLayersChecker>( num_temporal_layers); case Vp8TemporalLayersType::kBitrateDynamic: // Conference mode temporal layering for screen content in base stream. return std::make_unique<TemporalLayersChecker>(num_temporal_layers); } RTC_CHECK_NOTREACHED(); + return nullptr; } TemporalLayersChecker::TemporalLayersChecker(int num_temporal_layers) : num_temporal_layers_(num_temporal_layers), sequence_number_(0), last_sync_sequence_number_(0), last_tl0_sequence_number_(0) {} diff --git a/third_party/libwebrtc/video/adaptation/video_stream_encoder_resource_manager.cc b/third_party/libwebrtc/video/adaptation/video_stream_encoder_resource_manager.cc --- a/third_party/libwebrtc/video/adaptation/video_stream_encoder_resource_manager.cc +++ b/third_party/libwebrtc/video/adaptation/video_stream_encoder_resource_manager.cc @@ -59,16 +59,17 @@ bool IsFramerateScalingEnabled(Degradati std::string ToString(VideoAdaptationReason reason) { switch (reason) { case VideoAdaptationReason::kQuality: return "quality"; case VideoAdaptationReason::kCpu: return "cpu"; } RTC_CHECK_NOTREACHED(); + return ""; } std::vector<bool> GetActiveLayersFlags(const VideoCodec& codec) { std::vector<bool> flags; if (codec.codecType == VideoCodecType::kVideoCodecVP9) { flags.resize(codec.VP9().numberOfSpatialLayers); for (size_t i = 0; i < flags.size(); ++i) { flags[i] = codec.spatialLayers[i].active;
Locations
Projects
Search
Status Monitor
Help
OpenBuildService.org
Documentation
API Documentation
Code of Conduct
Contact
Support
@OBShq
Terms
openSUSE Build Service is sponsored by
The Open Build Service is an
openSUSE project
.
Sign Up
Log In
Places
Places
All Projects
Status Monitor