diff options
Diffstat (limited to 'chromium/third_party/webrtc/modules/audio_coding/neteq/audio_decoder_impl.cc')
-rw-r--r-- | chromium/third_party/webrtc/modules/audio_coding/neteq/audio_decoder_impl.cc | 516 |
1 files changed, 516 insertions, 0 deletions
diff --git a/chromium/third_party/webrtc/modules/audio_coding/neteq/audio_decoder_impl.cc b/chromium/third_party/webrtc/modules/audio_coding/neteq/audio_decoder_impl.cc new file mode 100644 index 00000000000..6c7269a35fe --- /dev/null +++ b/chromium/third_party/webrtc/modules/audio_coding/neteq/audio_decoder_impl.cc @@ -0,0 +1,516 @@ +/* + * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. + * + * Use of this source code is governed by a BSD-style license + * that can be found in the LICENSE file in the root of the source + * tree. An additional intellectual property rights grant can be found + * in the file PATENTS. All contributing project authors may + * be found in the AUTHORS file in the root of the source tree. + */ + +#include "webrtc/modules/audio_coding/neteq/audio_decoder_impl.h" + +#include <assert.h> +#include <string.h> // memmove + +#ifdef WEBRTC_CODEC_CELT +#include "webrtc/modules/audio_coding/codecs/celt/include/celt_interface.h" +#endif +#include "webrtc/modules/audio_coding/codecs/cng/include/webrtc_cng.h" +#include "webrtc/modules/audio_coding/codecs/g711/include/g711_interface.h" +#ifdef WEBRTC_CODEC_G722 +#include "webrtc/modules/audio_coding/codecs/g722/include/g722_interface.h" +#endif +#ifdef WEBRTC_CODEC_ILBC +#include "webrtc/modules/audio_coding/codecs/ilbc/interface/ilbc.h" +#endif +#ifdef WEBRTC_CODEC_ISACFX +#include "webrtc/modules/audio_coding/codecs/isac/fix/interface/isacfix.h" +#endif +#ifdef WEBRTC_CODEC_ISAC +#include "webrtc/modules/audio_coding/codecs/isac/main/interface/isac.h" +#endif +#ifdef WEBRTC_CODEC_OPUS +#include "webrtc/modules/audio_coding/codecs/opus/interface/opus_interface.h" +#endif +#ifdef WEBRTC_CODEC_PCM16 +#include "webrtc/modules/audio_coding/codecs/pcm16b/include/pcm16b.h" +#endif + +namespace webrtc { + +// PCMu +int AudioDecoderPcmU::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcG711_DecodeU( + state_, reinterpret_cast<int16_t*>(const_cast<uint8_t*>(encoded)), + static_cast<int16_t>(encoded_len), decoded, &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderPcmU::PacketDuration(const uint8_t* encoded, + size_t encoded_len) { + // One encoded byte per sample per channel. + return static_cast<int>(encoded_len / channels_); +} + +// PCMa +int AudioDecoderPcmA::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcG711_DecodeA( + state_, reinterpret_cast<int16_t*>(const_cast<uint8_t*>(encoded)), + static_cast<int16_t>(encoded_len), decoded, &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderPcmA::PacketDuration(const uint8_t* encoded, + size_t encoded_len) { + // One encoded byte per sample per channel. + return static_cast<int>(encoded_len / channels_); +} + +// PCM16B +#ifdef WEBRTC_CODEC_PCM16 +AudioDecoderPcm16B::AudioDecoderPcm16B(enum NetEqDecoder type) + : AudioDecoder(type) { + assert(type == kDecoderPCM16B || + type == kDecoderPCM16Bwb || + type == kDecoderPCM16Bswb32kHz || + type == kDecoderPCM16Bswb48kHz); +} + +int AudioDecoderPcm16B::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcPcm16b_DecodeW16( + state_, reinterpret_cast<int16_t*>(const_cast<uint8_t*>(encoded)), + static_cast<int16_t>(encoded_len), decoded, &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderPcm16B::PacketDuration(const uint8_t* encoded, + size_t encoded_len) { + // Two encoded byte per sample per channel. + return static_cast<int>(encoded_len / (2 * channels_)); +} + +AudioDecoderPcm16BMultiCh::AudioDecoderPcm16BMultiCh( + enum NetEqDecoder type) + : AudioDecoderPcm16B(kDecoderPCM16B) { // This will be changed below. + codec_type_ = type; // Changing to actual type here. + switch (codec_type_) { + case kDecoderPCM16B_2ch: + case kDecoderPCM16Bwb_2ch: + case kDecoderPCM16Bswb32kHz_2ch: + case kDecoderPCM16Bswb48kHz_2ch: + channels_ = 2; + break; + case kDecoderPCM16B_5ch: + channels_ = 5; + break; + default: + assert(false); + } +} +#endif + +// iLBC +#ifdef WEBRTC_CODEC_ILBC +AudioDecoderIlbc::AudioDecoderIlbc() : AudioDecoder(kDecoderILBC) { + WebRtcIlbcfix_DecoderCreate(reinterpret_cast<iLBC_decinst_t**>(&state_)); +} + +AudioDecoderIlbc::~AudioDecoderIlbc() { + WebRtcIlbcfix_DecoderFree(static_cast<iLBC_decinst_t*>(state_)); +} + +int AudioDecoderIlbc::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcIlbcfix_Decode(static_cast<iLBC_decinst_t*>(state_), + reinterpret_cast<const int16_t*>(encoded), + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderIlbc::DecodePlc(int num_frames, int16_t* decoded) { + return WebRtcIlbcfix_NetEqPlc(static_cast<iLBC_decinst_t*>(state_), + decoded, num_frames); +} + +int AudioDecoderIlbc::Init() { + return WebRtcIlbcfix_Decoderinit30Ms(static_cast<iLBC_decinst_t*>(state_)); +} +#endif + +// iSAC float +#ifdef WEBRTC_CODEC_ISAC +AudioDecoderIsac::AudioDecoderIsac() : AudioDecoder(kDecoderISAC) { + WebRtcIsac_Create(reinterpret_cast<ISACStruct**>(&state_)); + WebRtcIsac_SetDecSampRate(static_cast<ISACStruct*>(state_), 16000); +} + +AudioDecoderIsac::~AudioDecoderIsac() { + WebRtcIsac_Free(static_cast<ISACStruct*>(state_)); +} + +int AudioDecoderIsac::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcIsac_Decode(static_cast<ISACStruct*>(state_), + reinterpret_cast<const uint16_t*>(encoded), + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderIsac::DecodeRedundant(const uint8_t* encoded, + size_t encoded_len, int16_t* decoded, + SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcIsac_DecodeRcu(static_cast<ISACStruct*>(state_), + reinterpret_cast<const uint16_t*>(encoded), + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderIsac::DecodePlc(int num_frames, int16_t* decoded) { + return WebRtcIsac_DecodePlc(static_cast<ISACStruct*>(state_), + decoded, num_frames); +} + +int AudioDecoderIsac::Init() { + return WebRtcIsac_DecoderInit(static_cast<ISACStruct*>(state_)); +} + +int AudioDecoderIsac::IncomingPacket(const uint8_t* payload, + size_t payload_len, + uint16_t rtp_sequence_number, + uint32_t rtp_timestamp, + uint32_t arrival_timestamp) { + return WebRtcIsac_UpdateBwEstimate(static_cast<ISACStruct*>(state_), + reinterpret_cast<const uint16_t*>(payload), + static_cast<int32_t>(payload_len), + rtp_sequence_number, + rtp_timestamp, + arrival_timestamp); +} + +int AudioDecoderIsac::ErrorCode() { + return WebRtcIsac_GetErrorCode(static_cast<ISACStruct*>(state_)); +} + +// iSAC SWB +AudioDecoderIsacSwb::AudioDecoderIsacSwb() : AudioDecoderIsac() { + codec_type_ = kDecoderISACswb; + WebRtcIsac_SetDecSampRate(static_cast<ISACStruct*>(state_), 32000); +} + +// iSAC FB +AudioDecoderIsacFb::AudioDecoderIsacFb() : AudioDecoderIsacSwb() { + codec_type_ = kDecoderISACfb; +} +#endif + +// iSAC fix +#ifdef WEBRTC_CODEC_ISACFX +AudioDecoderIsacFix::AudioDecoderIsacFix() : AudioDecoder(kDecoderISAC) { + WebRtcIsacfix_Create(reinterpret_cast<ISACFIX_MainStruct**>(&state_)); +} + +AudioDecoderIsacFix::~AudioDecoderIsacFix() { + WebRtcIsacfix_Free(static_cast<ISACFIX_MainStruct*>(state_)); +} + +int AudioDecoderIsacFix::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcIsacfix_Decode(static_cast<ISACFIX_MainStruct*>(state_), + reinterpret_cast<const uint16_t*>(encoded), + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderIsacFix::Init() { + return WebRtcIsacfix_DecoderInit(static_cast<ISACFIX_MainStruct*>(state_)); +} + +int AudioDecoderIsacFix::IncomingPacket(const uint8_t* payload, + size_t payload_len, + uint16_t rtp_sequence_number, + uint32_t rtp_timestamp, + uint32_t arrival_timestamp) { + return WebRtcIsacfix_UpdateBwEstimate( + static_cast<ISACFIX_MainStruct*>(state_), + reinterpret_cast<const uint16_t*>(payload), + static_cast<int32_t>(payload_len), + rtp_sequence_number, rtp_timestamp, arrival_timestamp); +} + +int AudioDecoderIsacFix::ErrorCode() { + return WebRtcIsacfix_GetErrorCode(static_cast<ISACFIX_MainStruct*>(state_)); +} +#endif + +// G.722 +#ifdef WEBRTC_CODEC_G722 +AudioDecoderG722::AudioDecoderG722() : AudioDecoder(kDecoderG722) { + WebRtcG722_CreateDecoder(reinterpret_cast<G722DecInst**>(&state_)); +} + +AudioDecoderG722::~AudioDecoderG722() { + WebRtcG722_FreeDecoder(static_cast<G722DecInst*>(state_)); +} + +int AudioDecoderG722::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcG722_Decode( + static_cast<G722DecInst*>(state_), + const_cast<int16_t*>(reinterpret_cast<const int16_t*>(encoded)), + static_cast<int16_t>(encoded_len), decoded, &temp_type); + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderG722::Init() { + return WebRtcG722_DecoderInit(static_cast<G722DecInst*>(state_)); +} + +int AudioDecoderG722::PacketDuration(const uint8_t* encoded, + size_t encoded_len) { + // 1/2 encoded byte per sample per channel. + return static_cast<int>(2 * encoded_len / channels_); +} + +AudioDecoderG722Stereo::AudioDecoderG722Stereo() + : AudioDecoderG722(), + state_left_(state_), // Base member |state_| is used for left channel. + state_right_(NULL) { + channels_ = 2; + // |state_left_| already created by the base class AudioDecoderG722. + WebRtcG722_CreateDecoder(reinterpret_cast<G722DecInst**>(&state_right_)); +} + +AudioDecoderG722Stereo::~AudioDecoderG722Stereo() { + // |state_left_| will be freed by the base class AudioDecoderG722. + WebRtcG722_FreeDecoder(static_cast<G722DecInst*>(state_right_)); +} + +int AudioDecoderG722Stereo::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + // De-interleave the bit-stream into two separate payloads. + uint8_t* encoded_deinterleaved = new uint8_t[encoded_len]; + SplitStereoPacket(encoded, encoded_len, encoded_deinterleaved); + // Decode left and right. + int16_t ret = WebRtcG722_Decode( + static_cast<G722DecInst*>(state_left_), + reinterpret_cast<int16_t*>(encoded_deinterleaved), + static_cast<int16_t>(encoded_len / 2), decoded, &temp_type); + if (ret >= 0) { + int decoded_len = ret; + ret = WebRtcG722_Decode( + static_cast<G722DecInst*>(state_right_), + reinterpret_cast<int16_t*>(&encoded_deinterleaved[encoded_len / 2]), + static_cast<int16_t>(encoded_len / 2), &decoded[decoded_len], &temp_type); + if (ret == decoded_len) { + decoded_len += ret; + // Interleave output. + for (int k = decoded_len / 2; k < decoded_len; k++) { + int16_t temp = decoded[k]; + memmove(&decoded[2 * k - decoded_len + 2], + &decoded[2 * k - decoded_len + 1], + (decoded_len - k - 1) * sizeof(int16_t)); + decoded[2 * k - decoded_len + 1] = temp; + } + ret = decoded_len; // Return total number of samples. + } + } + *speech_type = ConvertSpeechType(temp_type); + delete [] encoded_deinterleaved; + return ret; +} + +int AudioDecoderG722Stereo::Init() { + int ret = WebRtcG722_DecoderInit(static_cast<G722DecInst*>(state_right_)); + if (ret != 0) { + return ret; + } + return AudioDecoderG722::Init(); +} + +// Split the stereo packet and place left and right channel after each other +// in the output array. +void AudioDecoderG722Stereo::SplitStereoPacket(const uint8_t* encoded, + size_t encoded_len, + uint8_t* encoded_deinterleaved) { + assert(encoded); + // Regroup the 4 bits/sample so |l1 l2| |r1 r2| |l3 l4| |r3 r4| ..., + // where "lx" is 4 bits representing left sample number x, and "rx" right + // sample. Two samples fit in one byte, represented with |...|. + for (size_t i = 0; i + 1 < encoded_len; i += 2) { + uint8_t right_byte = ((encoded[i] & 0x0F) << 4) + (encoded[i + 1] & 0x0F); + encoded_deinterleaved[i] = (encoded[i] & 0xF0) + (encoded[i + 1] >> 4); + encoded_deinterleaved[i + 1] = right_byte; + } + + // Move one byte representing right channel each loop, and place it at the + // end of the bytestream vector. After looping the data is reordered to: + // |l1 l2| |l3 l4| ... |l(N-1) lN| |r1 r2| |r3 r4| ... |r(N-1) r(N)|, + // where N is the total number of samples. + for (size_t i = 0; i < encoded_len / 2; i++) { + uint8_t right_byte = encoded_deinterleaved[i + 1]; + memmove(&encoded_deinterleaved[i + 1], &encoded_deinterleaved[i + 2], + encoded_len - i - 2); + encoded_deinterleaved[encoded_len - 1] = right_byte; + } +} +#endif + +// CELT +#ifdef WEBRTC_CODEC_CELT +AudioDecoderCelt::AudioDecoderCelt(enum NetEqDecoder type) + : AudioDecoder(type) { + assert(type == kDecoderCELT_32 || type == kDecoderCELT_32_2ch); + if (type == kDecoderCELT_32) { + channels_ = 1; + } else { + channels_ = 2; + } + WebRtcCelt_CreateDec(reinterpret_cast<CELT_decinst_t**>(&state_), + static_cast<int>(channels_)); +} + +AudioDecoderCelt::~AudioDecoderCelt() { + WebRtcCelt_FreeDec(static_cast<CELT_decinst_t*>(state_)); +} + +int AudioDecoderCelt::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default to speech. + int ret = WebRtcCelt_DecodeUniversal(static_cast<CELT_decinst_t*>(state_), + encoded, static_cast<int>(encoded_len), + decoded, &temp_type); + *speech_type = ConvertSpeechType(temp_type); + if (ret < 0) { + return -1; + } + // Return the total number of samples. + return ret * static_cast<int>(channels_); +} + +int AudioDecoderCelt::Init() { + return WebRtcCelt_DecoderInit(static_cast<CELT_decinst_t*>(state_)); +} + +bool AudioDecoderCelt::HasDecodePlc() const { return true; } + +int AudioDecoderCelt::DecodePlc(int num_frames, int16_t* decoded) { + int ret = WebRtcCelt_DecodePlc(static_cast<CELT_decinst_t*>(state_), + decoded, num_frames); + if (ret < 0) { + return -1; + } + // Return the total number of samples. + return ret * static_cast<int>(channels_); +} +#endif + +// Opus +#ifdef WEBRTC_CODEC_OPUS +AudioDecoderOpus::AudioDecoderOpus(enum NetEqDecoder type) + : AudioDecoder(type) { + if (type == kDecoderOpus_2ch) { + channels_ = 2; + } else { + channels_ = 1; + } + WebRtcOpus_DecoderCreate(reinterpret_cast<OpusDecInst**>(&state_), + static_cast<int>(channels_)); +} + +AudioDecoderOpus::~AudioDecoderOpus() { + WebRtcOpus_DecoderFree(static_cast<OpusDecInst*>(state_)); +} + +int AudioDecoderOpus::Decode(const uint8_t* encoded, size_t encoded_len, + int16_t* decoded, SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcOpus_DecodeNew(static_cast<OpusDecInst*>(state_), encoded, + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + if (ret > 0) + ret *= static_cast<int16_t>(channels_); // Return total number of samples. + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderOpus::DecodeRedundant(const uint8_t* encoded, + size_t encoded_len, int16_t* decoded, + SpeechType* speech_type) { + int16_t temp_type = 1; // Default is speech. + int16_t ret = WebRtcOpus_DecodeFec(static_cast<OpusDecInst*>(state_), encoded, + static_cast<int16_t>(encoded_len), decoded, + &temp_type); + if (ret > 0) + ret *= static_cast<int16_t>(channels_); // Return total number of samples. + *speech_type = ConvertSpeechType(temp_type); + return ret; +} + +int AudioDecoderOpus::Init() { + return WebRtcOpus_DecoderInitNew(static_cast<OpusDecInst*>(state_)); +} + +int AudioDecoderOpus::PacketDuration(const uint8_t* encoded, + size_t encoded_len) { + return WebRtcOpus_DurationEst(static_cast<OpusDecInst*>(state_), + encoded, static_cast<int>(encoded_len)); +} + +int AudioDecoderOpus::PacketDurationRedundant(const uint8_t* encoded, + size_t encoded_len) const { + return WebRtcOpus_FecDurationEst(encoded, static_cast<int>(encoded_len)); +} + +bool AudioDecoderOpus::PacketHasFec(const uint8_t* encoded, + size_t encoded_len) const { + int fec; + fec = WebRtcOpus_PacketHasFec(encoded, static_cast<int>(encoded_len)); + return (fec == 1); +} +#endif + +AudioDecoderCng::AudioDecoderCng(enum NetEqDecoder type) + : AudioDecoder(type) { + assert(type == kDecoderCNGnb || type == kDecoderCNGwb || + kDecoderCNGswb32kHz || type == kDecoderCNGswb48kHz); + WebRtcCng_CreateDec(reinterpret_cast<CNG_dec_inst**>(&state_)); + assert(state_); +} + +AudioDecoderCng::~AudioDecoderCng() { + if (state_) { + WebRtcCng_FreeDec(static_cast<CNG_dec_inst*>(state_)); + } +} + +int AudioDecoderCng::Init() { + assert(state_); + return WebRtcCng_InitDec(static_cast<CNG_dec_inst*>(state_)); +} + +} // namespace webrtc |