Index: media/filters/audio_renderer_algorithm.cc |
diff --git a/media/filters/audio_renderer_algorithm.cc b/media/filters/audio_renderer_algorithm.cc |
index 97f0811384159ff35868bb45f10eabdc7a94d86b..0ceb3e58a843ea99de6b8aee23904284c9ad8f7e 100644 |
--- a/media/filters/audio_renderer_algorithm.cc |
+++ b/media/filters/audio_renderer_algorithm.cc |
@@ -12,41 +12,78 @@ |
#include "media/audio/audio_util.h" |
#include "media/base/audio_buffer.h" |
#include "media/base/audio_bus.h" |
+#include "media/filters/wsola_internals.h" |
namespace media { |
-// The starting size in frames for |audio_buffer_|. Previous usage maintained a |
-// queue of 16 AudioBuffers, each of 512 frames. This worked well, so we |
-// maintain this number of frames. |
-static const int kStartingBufferSizeInFrames = 16 * 512; |
+ |
+// Waveform Similarity Overlap-and-add (WSOLA). |
+// |
+// One WSOLA iteration |
+// |
+// 1) Extract |target_block_| as input frames at indices |
+// [|target_block_index_|, |target_block_index_| + |ola_window_size_|). |
+// Note that |target_block_| is the "natural" continuation of the output. |
+// |
+// 2) Extract |search_block_| as input frames at indices |
+// [|search_block_index_|, |
+// |search_block_index_| + |num_candidate_blocks_| + |ola_window_size_|). |
+// |
+// 3) Find a block within the |search_block_| that is most similar |
+// to |target_block_|. Let |optimal_index| be the index of such block and |
+// write it to |optimal_block_|. |
+// |
+// 4) Update: |
+// |optimal_block_| = |transition_window_| * |target_block_| + |
+// (1 - |transition_window_|) * |optimal_block_|. |
+// |
+// 5) Overlap-and-add |optimal_block_| to the |wsola_output_|. |
+// |
+// 6) Update: |
+// |target_block_| = |optimal_index| + |ola_window_size_| / 2. |
+// |output_index_| = |output_index_| + |ola_window_size_| / 2, |
+// |search_block_center_offset_| = |output_index_| * |playback_rate_|, and |
+// |search_block_index_| = |search_block_center_offset_| - |
+// |search_block_center_offset_|. |
// The maximum size in frames for the |audio_buffer_|. Arbitrarily determined. |
// This number represents 3 seconds of 96kHz/16 bit 7.1 surround sound. |
static const int kMaxBufferSizeInFrames = 3 * 96000; |
-// Duration of audio segments used for crossfading (in seconds). |
-static const double kWindowDuration = 0.08; |
- |
-// Duration of crossfade between audio segments (in seconds). |
-static const double kCrossfadeDuration = 0.008; |
- |
// Max/min supported playback rates for fast/slow audio. Audio outside of these |
// ranges are muted. |
// Audio at these speeds would sound better under a frequency domain algorithm. |
static const float kMinPlaybackRate = 0.5f; |
static const float kMaxPlaybackRate = 4.0f; |
+// Overlap-and-add window size in milliseconds. |
+static const int kOlaWindowSizeMs = 20; |
+ |
+// Size of search interval in milliseconds. The search interval is |
+// [-delta delta] around |output_index_| * |playback_rate_|. So the search |
+// interval is 2 * delta. |
+static const int kWsolaSearchIntervalMs = 30; |
+ |
+// The starting size in frames for |audio_buffer_|. Previous usage maintained a |
+// queue of 16 AudioBuffers, each of 512 frames. This worked well, so we |
+// maintain this number of frames. |
+static const int kStartingBufferSizeInFrames = 16 * 512; |
+ |
AudioRendererAlgorithm::AudioRendererAlgorithm() |
: channels_(0), |
samples_per_second_(0), |
playback_rate_(0), |
- frames_in_crossfade_(0), |
- index_into_window_(0), |
- crossfade_frame_number_(0), |
muted_(false), |
muted_partial_frame_(0), |
- window_size_(0), |
- capacity_(kStartingBufferSizeInFrames) { |
+ capacity_(kStartingBufferSizeInFrames), |
+ output_time_(0), |
+ search_block_center_offset_(0), |
+ search_block_index_(0), |
+ num_candidate_blocks_(0), |
+ target_block_index_(0), |
+ ola_window_size_(0), |
+ ola_hop_size_(0), |
+ num_complete_frames_(0) { |
} |
AudioRendererAlgorithm::~AudioRendererAlgorithm() {} |
@@ -58,16 +95,59 @@ void AudioRendererAlgorithm::Initialize(float initial_playback_rate, |
channels_ = params.channels(); |
samples_per_second_ = params.sample_rate(); |
SetPlaybackRate(initial_playback_rate); |
- |
- window_size_ = samples_per_second_ * kWindowDuration; |
- frames_in_crossfade_ = samples_per_second_ * kCrossfadeDuration; |
- crossfade_buffer_ = AudioBus::Create(channels_, frames_in_crossfade_); |
+ num_candidate_blocks_ = (kWsolaSearchIntervalMs * samples_per_second_) / 1000; |
+ ola_window_size_ = kOlaWindowSizeMs * samples_per_second_ / 1000; |
+ |
+ // Make sure window size in an even number. |
+ ola_window_size_ += ola_window_size_ & 1; |
+ ola_hop_size_ = ola_window_size_ / 2; |
+ |
+ // |num_candidate_blocks_| / 2 is the offset of the center of the search |
+ // block to the center of the first (left most) candidate block. The offset |
+ // of the center of a candidate block to its left most point is |
+ // |ola_window_size_| / 2 - 1. Note that |ola_window_size_| is even and in |
+ // our convention the center belongs to the left half, so we need to subtract |
+ // one frame to get the correct offset. |
+ // |
+ // Search Block |
+ // <-------------------------------------------> |
+ // |
+ // |ola_window_size_| / 2 - 1 |
+ // <---- |
+ // |
+ // |num_candidate_blocks_| / 2 |
+ // <---------------- |
+ // center |
+ // X----X----------------X---------------X-----X |
+ // <----------> <----------> |
+ // Candidate ... Candidate |
+ // 1, ... |num_candidate_blocks_| |
+ search_block_center_offset_ = num_candidate_blocks_ / 2 + |
+ (ola_window_size_ / 2 - 1); |
+ |
+ ola_window_.reset(new float[ola_window_size_]); |
+ internal::GetSymmetricHanningWindow(ola_window_size_, ola_window_.get()); |
+ |
+ transition_window_.reset(new float[ola_window_size_ * 2]); |
+ internal::GetSymmetricHanningWindow(2 * ola_window_size_, |
+ transition_window_.get()); |
+ |
+ wsola_output_ = AudioBus::Create(channels_, ola_window_size_ + ola_hop_size_); |
+ wsola_output_->Zero(); // Initialize for overlap-and-add of the first block. |
+ |
+ // Auxiliary containers. |
+ optimal_block_ = AudioBus::Create(channels_, ola_window_size_); |
+ search_block_ = AudioBus::Create( |
+ channels_, num_candidate_blocks_ + (ola_window_size_ - 1)); |
+ target_block_ = AudioBus::Create(channels_, ola_window_size_); |
} |
int AudioRendererAlgorithm::FillBuffer(AudioBus* dest, int requested_frames) { |
if (playback_rate_ == 0) |
return 0; |
+ DCHECK_EQ(channels_, dest->channels()); |
+ |
// Optimize the |muted_| case to issue a single clear instead of performing |
// the full crossfade and clearing each crossfaded frame. |
if (muted_) { |
@@ -93,12 +173,12 @@ int AudioRendererAlgorithm::FillBuffer(AudioBus* dest, int requested_frames) { |
return frames_to_render; |
} |
- int slower_step = ceil(window_size_ * playback_rate_); |
- int faster_step = ceil(window_size_ / playback_rate_); |
+ int slower_step = ceil(ola_window_size_ * playback_rate_); |
+ int faster_step = ceil(ola_window_size_ / playback_rate_); |
// Optimize the most common |playback_rate_| ~= 1 case to use a single copy |
// instead of copying frame by frame. |
- if (window_size_ <= faster_step && slower_step >= window_size_) { |
+ if (ola_window_size_ <= faster_step && slower_step >= ola_window_size_) { |
const int frames_to_copy = |
std::min(audio_buffer_.frames(), requested_frames); |
const int frames_read = audio_buffer_.ReadFrames(frames_to_copy, 0, dest); |
@@ -106,277 +186,200 @@ int AudioRendererAlgorithm::FillBuffer(AudioBus* dest, int requested_frames) { |
return frames_read; |
} |
- int total_frames_rendered = 0; |
- while (total_frames_rendered < requested_frames) { |
- if (index_into_window_ >= window_size_) |
- ResetWindow(); |
- |
- int rendered_frames = 0; |
- if (window_size_ > faster_step) { |
- rendered_frames = |
- OutputFasterPlayback(dest, |
- total_frames_rendered, |
- requested_frames - total_frames_rendered, |
- window_size_, |
- faster_step); |
- } else if (slower_step < window_size_) { |
- rendered_frames = |
- OutputSlowerPlayback(dest, |
- total_frames_rendered, |
- requested_frames - total_frames_rendered, |
- slower_step, |
- window_size_); |
- } else { |
- NOTREACHED(); |
- } |
+ int rendered_frames = 0; |
+ do { |
+ rendered_frames += WriteCompletedFramesTo( |
+ requested_frames - rendered_frames, rendered_frames, dest); |
+ } while (rendered_frames < requested_frames && RunOneWsolaIteration()); |
+ return rendered_frames; |
+} |
- if (rendered_frames == 0) |
- break; |
+void AudioRendererAlgorithm::SetPlaybackRate(float new_rate) { |
+ DCHECK_GE(new_rate, 0); |
+ playback_rate_ = new_rate; |
+ muted_ = |
+ playback_rate_ < kMinPlaybackRate || playback_rate_ > kMaxPlaybackRate; |
+} |
- total_frames_rendered += rendered_frames; |
- } |
- return total_frames_rendered; |
+void AudioRendererAlgorithm::FlushBuffers() { |
+ // Clear the queue of decoded packets (releasing the buffers). |
+ audio_buffer_.Clear(); |
+ output_time_ = 0; |
+ search_block_index_ = 0; |
+ target_block_index_ = 0; |
+ wsola_output_->Zero(); |
+ num_complete_frames_ = 0; |
} |
-void AudioRendererAlgorithm::ResetWindow() { |
- DCHECK_LE(index_into_window_, window_size_); |
- index_into_window_ = 0; |
- crossfade_frame_number_ = 0; |
+base::TimeDelta AudioRendererAlgorithm::GetTime() { |
+ return audio_buffer_.current_time(); |
} |
-int AudioRendererAlgorithm::OutputFasterPlayback(AudioBus* dest, |
- int dest_offset, |
- int requested_frames, |
- int input_step, |
- int output_step) { |
- // Ensure we don't run into OOB read/write situation. |
- CHECK_GT(input_step, output_step); |
- DCHECK_LT(index_into_window_, window_size_); |
- DCHECK_GT(playback_rate_, 1.0); |
- DCHECK(!muted_); |
- |
- if (audio_buffer_.frames() < 1) |
- return 0; |
+void AudioRendererAlgorithm::EnqueueBuffer( |
+ const scoped_refptr<AudioBuffer>& buffer_in) { |
+ DCHECK(!buffer_in->end_of_stream()); |
+ audio_buffer_.Append(buffer_in); |
+} |
- // The audio data is output in a series of windows. For sped-up playback, |
- // the window is comprised of the following phases: |
- // |
- // a) Output raw data. |
- // b) Save bytes for crossfade in |crossfade_buffer_|. |
- // c) Drop data. |
- // d) Output crossfaded audio leading up to the next window. |
- // |
- // The duration of each phase is computed below based on the |window_size_| |
- // and |playback_rate_|. |
- DCHECK_LE(frames_in_crossfade_, output_step); |
+bool AudioRendererAlgorithm::IsQueueFull() { |
+ return audio_buffer_.frames() >= capacity_; |
+} |
- // This is the index of the end of phase a, beginning of phase b. |
- int outtro_crossfade_begin = output_step - frames_in_crossfade_; |
+void AudioRendererAlgorithm::IncreaseQueueCapacity() { |
+ capacity_ = std::min(2 * capacity_, kMaxBufferSizeInFrames); |
+} |
- // This is the index of the end of phase b, beginning of phase c. |
- int outtro_crossfade_end = output_step; |
+bool AudioRendererAlgorithm::CanPerformWsola() const { |
+ const int search_block_size = num_candidate_blocks_ + (ola_window_size_ - 1); |
+ const int frames = audio_buffer_.frames(); |
+ return target_block_index_ + ola_window_size_ <= frames && |
+ search_block_index_ + search_block_size <= frames; |
+} |
- // This is the index of the end of phase c, beginning of phase d. |
- // This phase continues until |index_into_window_| reaches |window_size_|, at |
- // which point the window restarts. |
- int intro_crossfade_begin = input_step - frames_in_crossfade_; |
+bool AudioRendererAlgorithm::RunOneWsolaIteration() { |
+ if (!CanPerformWsola()) |
+ return false; |
- // a) Output raw frames if we haven't reached the crossfade section. |
- if (index_into_window_ < outtro_crossfade_begin) { |
- // Read as many frames as we can and return the count. If it's not enough, |
- // we will get called again. |
- const int frames_to_copy = |
- std::min(requested_frames, outtro_crossfade_begin - index_into_window_); |
- int copied = audio_buffer_.ReadFrames(frames_to_copy, dest_offset, dest); |
- index_into_window_ += copied; |
- return copied; |
- } |
- |
- // b) Save outtro crossfade frames into intermediate buffer, but do not output |
- // anything to |dest|. |
- if (index_into_window_ < outtro_crossfade_end) { |
- // This phase only applies if there are bytes to crossfade. |
- DCHECK_GT(frames_in_crossfade_, 0); |
- int crossfade_start = index_into_window_ - outtro_crossfade_begin; |
- int crossfade_count = outtro_crossfade_end - index_into_window_; |
- int copied = audio_buffer_.ReadFrames( |
- crossfade_count, crossfade_start, crossfade_buffer_.get()); |
- index_into_window_ += copied; |
- |
- // Did we get all the frames we need? If not, return and let subsequent |
- // calls try to get the rest. |
- if (copied != crossfade_count) |
- return 0; |
- } |
+ GetOptimalBlock(); |
- // c) Drop frames until we reach the intro crossfade section. |
- if (index_into_window_ < intro_crossfade_begin) { |
- // Check if there is enough data to skip all the frames needed. If not, |
- // return 0 and let subsequent calls try to skip it all. |
- int seek_frames = intro_crossfade_begin - index_into_window_; |
- if (audio_buffer_.frames() < seek_frames) |
- return 0; |
- audio_buffer_.SeekFrames(seek_frames); |
+ // Overlap-and-add. |
+ for (int k = 0; k < channels_; ++k) { |
+ const float* const ch_opt_frame = optimal_block_->channel(k); |
+ float* ch_output = wsola_output_->channel(k) + num_complete_frames_; |
+ for (int n = 0; n < ola_hop_size_; ++n) { |
+ ch_output[n] = ch_output[n] * ola_window_[ola_hop_size_ + n] + |
+ ch_opt_frame[n] * ola_window_[n]; |
+ } |
- // We've dropped all the frames that need to be dropped. |
- index_into_window_ += seek_frames; |
+ // Copy the second half to the output. |
+ memcpy(&ch_output[ola_hop_size_], &ch_opt_frame[ola_hop_size_], |
+ sizeof(*ch_opt_frame) * ola_hop_size_); |
} |
- // d) Crossfade and output a frame, as long as we have data. |
- if (audio_buffer_.frames() < 1) |
- return 0; |
- DCHECK_GT(frames_in_crossfade_, 0); |
- DCHECK_LT(index_into_window_, window_size_); |
- |
- int offset_into_buffer = index_into_window_ - intro_crossfade_begin; |
- int copied = audio_buffer_.ReadFrames(1, dest_offset, dest); |
- DCHECK_EQ(copied, 1); |
- CrossfadeFrame(crossfade_buffer_.get(), |
- offset_into_buffer, |
- dest, |
- dest_offset, |
- offset_into_buffer); |
- index_into_window_ += copied; |
- return copied; |
+ num_complete_frames_ += ola_hop_size_; |
+ UpdateOutputTime(ola_hop_size_); |
+ RemoveOldInputFrames(); |
+ return true; |
} |
-int AudioRendererAlgorithm::OutputSlowerPlayback(AudioBus* dest, |
- int dest_offset, |
- int requested_frames, |
- int input_step, |
- int output_step) { |
- // Ensure we don't run into OOB read/write situation. |
- CHECK_LT(input_step, output_step); |
- DCHECK_LT(index_into_window_, window_size_); |
- DCHECK_LT(playback_rate_, 1.0); |
- DCHECK_NE(playback_rate_, 0); |
- DCHECK(!muted_); |
- |
- if (audio_buffer_.frames() < 1) |
- return 0; |
- |
- // The audio data is output in a series of windows. For slowed down playback, |
- // the window is comprised of the following phases: |
- // |
- // a) Output raw data. |
- // b) Output and save bytes for crossfade in |crossfade_buffer_|. |
- // c) Output* raw data. |
- // d) Output* crossfaded audio leading up to the next window. |
- // |
- // * Phases c) and d) do not progress |audio_buffer_|'s cursor so that the |
- // |audio_buffer_|'s cursor is in the correct place for the next window. |
- // |
- // The duration of each phase is computed below based on the |window_size_| |
- // and |playback_rate_|. |
- DCHECK_LE(frames_in_crossfade_, input_step); |
- |
- // This is the index of the end of phase a, beginning of phase b. |
- int intro_crossfade_begin = input_step - frames_in_crossfade_; |
+void AudioRendererAlgorithm::UpdateOutputTime(float time_change) { |
+ output_time_ += time_change; |
+ // Center of the search region, in frames. |
+ const int search_block_center_index = static_cast<int>( |
+ output_time_ * playback_rate_ + 0.5f); |
+ search_block_index_ = search_block_center_index - search_block_center_offset_; |
+} |
- // This is the index of the end of phase b, beginning of phase c. |
- int intro_crossfade_end = input_step; |
+void AudioRendererAlgorithm::RemoveOldInputFrames() { |
+ const int earliest_used_index = std::min(target_block_index_, |
+ search_block_index_); |
+ if (earliest_used_index <= 0) |
+ return; // Nothing to remove. |
- // This is the index of the end of phase c, beginning of phase d. |
- // This phase continues until |index_into_window_| reaches |window_size_|, at |
- // which point the window restarts. |
- int outtro_crossfade_begin = output_step - frames_in_crossfade_; |
+ // Remove frames from input and adjust indices accordingly. |
+ audio_buffer_.SeekFrames(earliest_used_index); |
+ target_block_index_ -= earliest_used_index; |
- // a) Output raw frames. |
- if (index_into_window_ < intro_crossfade_begin) { |
- // Read as many frames as we can and return the count. If it's not enough, |
- // we will get called again. |
- const int frames_to_copy = |
- std::min(requested_frames, intro_crossfade_begin - index_into_window_); |
- int copied = audio_buffer_.ReadFrames(frames_to_copy, dest_offset, dest); |
- index_into_window_ += copied; |
- return copied; |
- } |
+ // Adjust output index. |
+ float output_time_change = earliest_used_index / playback_rate_; |
+ CHECK_GE(output_time_, output_time_change); |
+ UpdateOutputTime(-output_time_change); |
+} |
- // b) Save the raw frames for the intro crossfade section, then copy the |
- // same frames to |dest|. |
- if (index_into_window_ < intro_crossfade_end) { |
- const int frames_to_copy = |
- std::min(requested_frames, intro_crossfade_end - index_into_window_); |
- int offset = index_into_window_ - intro_crossfade_begin; |
- int copied = audio_buffer_.ReadFrames( |
- frames_to_copy, offset, crossfade_buffer_.get()); |
- crossfade_buffer_->CopyPartialFramesTo(offset, copied, dest_offset, dest); |
- index_into_window_ += copied; |
- return copied; |
- } |
+int AudioRendererAlgorithm::WriteCompletedFramesTo( |
+ int requested_frames, int dest_offset, AudioBus* dest) { |
+ int rendered_frames = std::min(num_complete_frames_, requested_frames); |
- // c) Output a raw frame into |dest| without advancing the |audio_buffer_| |
- // cursor. |
- int audio_buffer_offset = index_into_window_ - intro_crossfade_end; |
- DCHECK_GE(audio_buffer_offset, 0); |
- if (audio_buffer_.frames() <= audio_buffer_offset) |
- return 0; |
- int copied = |
- audio_buffer_.PeekFrames(1, audio_buffer_offset, dest_offset, dest); |
- DCHECK_EQ(1, copied); |
- |
- // d) Crossfade the next frame of |crossfade_buffer_| into |dest| if we've |
- // reached the outtro crossfade section of the window. |
- if (index_into_window_ >= outtro_crossfade_begin) { |
- int offset_into_crossfade_buffer = |
- index_into_window_ - outtro_crossfade_begin; |
- CrossfadeFrame(dest, |
- dest_offset, |
- crossfade_buffer_.get(), |
- offset_into_crossfade_buffer, |
- offset_into_crossfade_buffer); |
- } |
+ if (rendered_frames == 0) |
+ return 0; // There is nothing to read from |wsola_output_|, return. |
- index_into_window_ += copied; |
- return copied; |
-} |
+ wsola_output_->CopyPartialFramesTo(0, rendered_frames, dest_offset, dest); |
-void AudioRendererAlgorithm::CrossfadeFrame(AudioBus* intro, |
- int intro_offset, |
- AudioBus* outtro, |
- int outtro_offset, |
- int fade_offset) { |
- float crossfade_ratio = |
- static_cast<float>(fade_offset) / frames_in_crossfade_; |
- for (int channel = 0; channel < channels_; ++channel) { |
- outtro->channel(channel)[outtro_offset] = |
- (1.0f - crossfade_ratio) * intro->channel(channel)[intro_offset] + |
- (crossfade_ratio) * outtro->channel(channel)[outtro_offset]; |
+ // Remove the frames which are read. |
+ int frames_to_move = wsola_output_->frames() - rendered_frames; |
+ for (int k = 0; k < channels_; ++k) { |
+ float* ch = wsola_output_->channel(k); |
+ memmove(ch, &ch[rendered_frames], sizeof(*ch) * frames_to_move); |
} |
+ num_complete_frames_ -= rendered_frames; |
+ return rendered_frames; |
} |
-void AudioRendererAlgorithm::SetPlaybackRate(float new_rate) { |
- DCHECK_GE(new_rate, 0); |
- playback_rate_ = new_rate; |
- muted_ = |
- playback_rate_ < kMinPlaybackRate || playback_rate_ > kMaxPlaybackRate; |
- |
- ResetWindow(); |
-} |
- |
-void AudioRendererAlgorithm::FlushBuffers() { |
- ResetWindow(); |
- |
- // Clear the queue of decoded packets (releasing the buffers). |
- audio_buffer_.Clear(); |
-} |
+bool AudioRendererAlgorithm::TargetIsWithinSearchRegion() const { |
+ const int search_block_size = num_candidate_blocks_ + (ola_window_size_ - 1); |
-base::TimeDelta AudioRendererAlgorithm::GetTime() { |
- return audio_buffer_.current_time(); |
+ return target_block_index_ >= search_block_index_ && |
+ target_block_index_ + ola_window_size_ <= |
+ search_block_index_ + search_block_size; |
} |
-void AudioRendererAlgorithm::EnqueueBuffer( |
- const scoped_refptr<AudioBuffer>& buffer_in) { |
- DCHECK(!buffer_in->end_of_stream()); |
- audio_buffer_.Append(buffer_in); |
-} |
+void AudioRendererAlgorithm::GetOptimalBlock() { |
+ int optimal_index = 0; |
+ |
+ // An interval around last optimal block which is excluded from the search. |
+ // This is to reduce the buzzy sound. The number 160 is rather arbitrary and |
+ // derived heuristically. |
+ const int kExcludeIntervalLengthFrames = 160; |
+ if (TargetIsWithinSearchRegion()) { |
+ optimal_index = target_block_index_; |
+ PeekAudioWithZeroPrepend(optimal_index, optimal_block_.get()); |
+ } else { |
+ PeekAudioWithZeroPrepend(target_block_index_, target_block_.get()); |
+ PeekAudioWithZeroPrepend(search_block_index_, search_block_.get()); |
+ int last_optimal = target_block_index_ - ola_hop_size_ - |
+ search_block_index_; |
+ internal::Interval exclude_iterval = std::make_pair( |
+ last_optimal - kExcludeIntervalLengthFrames / 2, |
+ last_optimal + kExcludeIntervalLengthFrames / 2); |
+ |
+ // |optimal_index| is in frames and it is relative to the beginning of the |
+ // |search_block_|. |
+ optimal_index = internal::OptimalIndex( |
+ search_block_.get(), target_block_.get(), exclude_iterval); |
+ |
+ // Translate |index| w.r.t. the beginning of |audio_buffer_| and extract the |
+ // optimal block. |
+ optimal_index += search_block_index_; |
+ PeekAudioWithZeroPrepend(optimal_index, optimal_block_.get()); |
+ |
+ // Make a transition from target block to the optimal block if different. |
+ // Target block has the best continuation to the current output. |
+ // Optimal block is the most similar block to the target, however, it might |
+ // introduce some discontinuity when over-lap-added. Therefore, we combine |
+ // them for a smoother transition. The length of transition window is twice |
+ // as that of the optimal-block which makes it like a weighting function |
+ // where target-block has higher weight close to zero (weight of 1 at index |
+ // 0) and lower weight close the end. |
+ for (int k = 0; k < channels_; ++k) { |
+ float* ch_opt = optimal_block_->channel(k); |
+ const float* const ch_target = target_block_->channel(k); |
+ for (int n = 0; n < ola_window_size_; ++n) { |
+ ch_opt[n] = ch_opt[n] * transition_window_[n] + ch_target[n] * |
+ transition_window_[ola_window_size_ + n]; |
+ } |
+ } |
+ } |
-bool AudioRendererAlgorithm::IsQueueFull() { |
- return audio_buffer_.frames() >= capacity_; |
+ // Next target is one hop ahead of the current optimal. |
+ target_block_index_ = optimal_index + ola_hop_size_; |
} |
-void AudioRendererAlgorithm::IncreaseQueueCapacity() { |
- capacity_ = std::min(2 * capacity_, kMaxBufferSizeInFrames); |
+void AudioRendererAlgorithm::PeekAudioWithZeroPrepend( |
+ int read_offset_frames, AudioBus* dest) { |
+ CHECK_LE(read_offset_frames + dest->frames(), audio_buffer_.frames()); |
+ |
+ int write_offset = 0; |
+ int num_frames_to_read = dest->frames(); |
+ if (read_offset_frames < 0) { |
+ int num_zero_frames_appended = std::min(-read_offset_frames, |
+ num_frames_to_read); |
+ read_offset_frames = 0; |
+ num_frames_to_read -= num_zero_frames_appended; |
+ write_offset = num_zero_frames_appended; |
+ dest->ZeroFrames(num_zero_frames_appended); |
+ } |
+ audio_buffer_.PeekFrames(num_frames_to_read, read_offset_frames, |
+ write_offset, dest); |
} |
} // namespace media |