X-Git-Url: https://main.carlh.net/gitweb/?a=blobdiff_plain;f=libs%2Fardour%2Faudioregion.cc;h=44e238fb7ef736caae9be50d3c41ced2a1492840;hb=3dc7728038860bda6eb4b0de1f4a3e14ec9e86cc;hp=2eb09ae192c1aa4c933083500101e8e51f874b49;hpb=e5e12acc5698090f2c0c614385e457cc0b46fbb0;p=ardour.git diff --git a/libs/ardour/audioregion.cc b/libs/ardour/audioregion.cc index 2eb09ae192..44e238fb7e 100644 --- a/libs/ardour/audioregion.cc +++ b/libs/ardour/audioregion.cc @@ -25,6 +25,7 @@ #include #include +#include #include @@ -40,13 +41,20 @@ #include "ardour/session.h" #include "ardour/dB.h" #include "ardour/debug.h" +#include "ardour/event_type_map.h" #include "ardour/playlist.h" #include "ardour/audiofilesource.h" #include "ardour/region_factory.h" #include "ardour/runtime_functions.h" #include "ardour/transient_detector.h" +#include "ardour/parameter_descriptor.h" #include "ardour/progress.h" +#include "ardour/sndfilesource.h" +#ifdef HAVE_COREAUDIO +#include "ardour/coreaudiosource.h" +#endif // HAVE_COREAUDIO + #include "i18n.h" #include @@ -70,8 +78,6 @@ namespace ARDOUR { } } -static const double VERY_SMALL_SIGNAL = 0.0000001; //-140dB - /* Curve manipulations */ static void @@ -103,20 +109,18 @@ generate_db_fade (boost::shared_ptr dst, double len, int nu //generate a fade-out curve by successively applying a gain drop float fade_speed = dB_to_coefficient(dB_drop / (float) num_steps); + float coeff = GAIN_COEFF_UNITY; for (int i = 1; i < (num_steps-1); i++) { - float coeff = 1.0; - for (int j = 0; j < i; j++) { - coeff *= fade_speed; - } + coeff *= fade_speed; dst->fast_simple_add (len*(double)i/(double)num_steps, coeff); } - dst->fast_simple_add (len, VERY_SMALL_SIGNAL); + dst->fast_simple_add (len, GAIN_COEFF_SMALL); } static void -merge_curves (boost::shared_ptr dst, - boost::shared_ptr curve1, +merge_curves (boost::shared_ptr dst, + boost::shared_ptr curve1, boost::shared_ptr curve2) { Evoral::ControlList::EventList::size_type size = curve1->size(); @@ -125,13 +129,13 @@ merge_curves (boost::shared_ptr dst, if (size != curve2->size()) { return; } - + Evoral::ControlList::const_iterator c1 = curve1->begin(); int count = 0; for (Evoral::ControlList::const_iterator c2 = curve2->begin(); c2!=curve2->end(); c2++ ) { float v1 = accurate_coefficient_to_dB((*c1)->value); float v2 = accurate_coefficient_to_dB((*c2)->value); - + double interp = v1 * ( 1.0-( (double)count / (double)size) ); interp += v2 * ( (double)count / (double)size ); @@ -209,7 +213,7 @@ AudioRegion::register_properties () , _fade_in (Properties::fade_in, boost::shared_ptr (new AutomationList (*other->_fade_in.val()))) \ , _inverse_fade_in (Properties::fade_in, boost::shared_ptr (new AutomationList (*other->_inverse_fade_in.val()))) \ , _fade_out (Properties::fade_in, boost::shared_ptr (new AutomationList (*other->_fade_out.val()))) \ - , _inverse_fade_out (Properties::fade_in, boost::shared_ptr (new AutomationList (*other->_inverse_fade_out.val()))) + , _inverse_fade_out (Properties::fade_in, boost::shared_ptr (new AutomationList (*other->_inverse_fade_out.val()))) \ /* a Session will reset these to its chosen defaults by calling AudioRegion::set_default_fade() */ void @@ -368,7 +372,7 @@ void AudioRegion::connect_to_analysis_changed () { for (SourceList::const_iterator i = _sources.begin(); i != _sources.end(); ++i) { - (*i)->AnalysisChanged.connect_same_thread (*this, boost::bind (&AudioRegion::invalidate_transients, this)); + (*i)->AnalysisChanged.connect_same_thread (*this, boost::bind (&AudioRegion::maybe_invalidate_transients, this)); } } @@ -416,7 +420,7 @@ AudioRegion::set_envelope_active (bool yn) * @param chan_n Channel. * @param frames_per_pixel Number of samples to use to generate one peak value. */ - + ARDOUR::framecnt_t AudioRegion::read_peaks (PeakData *buf, framecnt_t npeaks, framecnt_t offset, framecnt_t cnt, uint32_t chan_n, double frames_per_pixel) const { @@ -426,15 +430,16 @@ AudioRegion::read_peaks (PeakData *buf, framecnt_t npeaks, framecnt_t offset, fr if (audio_source(chan_n)->read_peaks (buf, npeaks, offset, cnt, frames_per_pixel)) { return 0; - } else { - if (_scale_amplitude != 1.0f) { - for (framecnt_t n = 0; n < npeaks; ++n) { - buf[n].max *= _scale_amplitude; - buf[n].min *= _scale_amplitude; - } + } + + if (_scale_amplitude != 1.0f) { + for (framecnt_t n = 0; n < npeaks; ++n) { + buf[n].max *= _scale_amplitude; + buf[n].min *= _scale_amplitude; } - return cnt; } + + return npeaks; } /** @param buf Buffer to write data to (existing data will be overwritten). @@ -484,7 +489,7 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, */ assert (cnt >= 0); - + if (n_channels() == 0) { return 0; } @@ -514,29 +519,29 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, of any fade out that we are dealing with */ frameoffset_t fade_out_offset = 0; - + /* Amount (length) of fade out that we are dealing with in this read */ framecnt_t fade_out_limit = 0; framecnt_t fade_interval_start = 0; /* Fade in */ - + if (_fade_in_active && _session.config.get_use_region_fades()) { - + framecnt_t fade_in_length = (framecnt_t) _fade_in->back()->when; /* see if this read is within the fade in */ - + if (internal_offset < fade_in_length) { fade_in_limit = min (to_read, fade_in_length - internal_offset); } } - + /* Fade out */ - + if (_fade_out_active && _session.config.get_use_region_fades()) { - + /* see if some part of this read is within the fade out */ /* ................. >| REGION @@ -558,7 +563,7 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, fade_interval_start = max (internal_offset, _length - framecnt_t (_fade_out->back()->when)); framecnt_t fade_interval_end = min(internal_offset + to_read, _length.val()); - + if (fade_interval_end > fade_interval_start) { /* (part of the) the fade out is in this buffer */ fade_out_limit = fade_interval_end - fade_interval_start; @@ -596,7 +601,7 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, /* APPLY FADES TO THE DATA IN mixdown_buffer AND MIX THE RESULTS INTO * buf. The key things to realize here: (1) the fade being applied is - * (as of April 26th 2012) just the inverse of the fade in curve (2) + * (as of April 26th 2012) just the inverse of the fade in curve (2) * "buf" contains data from lower regions already. So this operation * fades out the existing material. */ @@ -609,26 +614,26 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, /* explicit inverse fade in curve (e.g. for constant * power), so we have to fetch it. */ - + _inverse_fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit); - + /* Fade the data from lower layers out */ for (framecnt_t n = 0; n < fade_in_limit; ++n) { buf[n] *= gain_buffer[n]; } - + /* refill gain buffer with the fade in */ - + _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit); - + } else { - + /* no explicit inverse fade in, so just use (1 - fade * in) for the fade out of lower layers */ - + _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit); - + for (framecnt_t n = 0; n < fade_in_limit; ++n) { buf[n] *= 1 - gain_buffer[n]; } @@ -649,27 +654,27 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, if (opaque()) { if (_inverse_fade_out) { - + _inverse_fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit); - + /* Fade the data from lower levels in */ for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) { buf[m] *= gain_buffer[n]; } - + /* fetch the actual fade out */ _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit); - + } else { /* no explicit inverse fade out (which is * actually a fade in), so just use (1 - fade * out) for the fade in of lower layers */ - + _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit); - + for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) { buf[m] *= 1 - gain_buffer[n]; } @@ -685,7 +690,7 @@ AudioRegion::read_at (Sample *buf, Sample *mixdown_buffer, float *gain_buffer, buf[m] += mixdown_buffer[m] * gain_buffer[n]; } } - + /* MIX OR COPY THE REGION BODY FROM mixdown_buffer INTO buf */ framecnt_t const N = to_read - fade_in_limit - fade_out_limit; @@ -726,7 +731,7 @@ AudioRegion::read_from_sources (SourceList const & srcs, framecnt_t limit, Sampl if (to_read == 0) { return 0; } - + if (chan_n < n_channels()) { boost::shared_ptr src = boost::dynamic_pointer_cast (srcs[chan_n]); @@ -752,7 +757,7 @@ AudioRegion::read_from_sources (SourceList const & srcs, framecnt_t limit, Sampl } } else { - + /* use silence */ memset (buf, 0, sizeof (Sample) * to_read); } @@ -766,7 +771,7 @@ AudioRegion::get_basic_state () { XMLNode& node (Region::state ()); char buf[64]; - LocaleGuard lg (X_("POSIX")); + LocaleGuard lg (); snprintf (buf, sizeof (buf), "%u", (uint32_t) _sources.size()); node.add_property ("channels", buf); @@ -779,7 +784,7 @@ AudioRegion::state () { XMLNode& node (get_basic_state()); XMLNode *child; - LocaleGuard lg (X_("POSIX")); + LocaleGuard lg (); child = node.add_child ("Envelope"); @@ -789,8 +794,8 @@ AudioRegion::state () // so, if they are both at 1.0f, that means the default region. if (_envelope->size() == 2 && - _envelope->front()->value == 1.0f && - _envelope->back()->value==1.0f) { + _envelope->front()->value == GAIN_COEFF_UNITY && + _envelope->back()->value==GAIN_COEFF_UNITY) { if (_envelope->front()->when == 0 && _envelope->back()->when == _length) { default_env = true; } @@ -835,8 +840,8 @@ int AudioRegion::_set_state (const XMLNode& node, int version, PropertyChange& what_changed, bool send) { const XMLNodeList& nlist = node.children(); - const XMLProperty *prop; - LocaleGuard lg (X_("POSIX")); + XMLProperty const * prop; + LocaleGuard lg (); boost::shared_ptr the_playlist (_playlist.lock()); suspend_property_changes (); @@ -866,7 +871,7 @@ AudioRegion::_set_state (const XMLNode& node, int version, PropertyChange& what_ for (XMLNodeConstIterator niter = nlist.begin(); niter != nlist.end(); ++niter) { XMLNode *child; - XMLProperty *prop; + XMLProperty const * prop; child = (*niter); @@ -914,7 +919,7 @@ AudioRegion::_set_state (const XMLNode& node, int version, PropertyChange& what_ _fade_out->set_state (*grandchild, version); } } - + if ((prop = child->property ("active")) != 0) { if (string_is_affirmative (prop->value())) { set_fade_out_active (true); @@ -922,13 +927,13 @@ AudioRegion::_set_state (const XMLNode& node, int version, PropertyChange& what_ set_fade_out_active (false); } } - - } else if (child->name() == "InverseFadeIn") { + + } else if ( (child->name() == "InverseFadeIn") || (child->name() == "InvFadeIn") ) { XMLNode* grandchild = child->child ("AutomationList"); if (grandchild) { _inverse_fade_in->set_state (*grandchild, version); } - } else if (child->name() == "InverseFadeOut") { + } else if ( (child->name() == "InverseFadeOut") || (child->name() == "InvFadeOut") ) { XMLNode* grandchild = child->child ("AutomationList"); if (grandchild) { _inverse_fade_out->set_state (*grandchild, version); @@ -957,6 +962,36 @@ AudioRegion::set_state (const XMLNode& node, int version) return _set_state (node, version, what_changed, true); } +void +AudioRegion::fade_range (framepos_t start, framepos_t end) +{ + framepos_t s, e; + + switch (coverage (start, end)) { + case Evoral::OverlapStart: + trim_front(start); + s = _position; + e = end; + set_fade_in (FadeConstantPower, e - s); + break; + case Evoral::OverlapEnd: + trim_end(end); + s = start; + e = _position + _length; + set_fade_out (FadeConstantPower, e - s); + break; + case Evoral::OverlapInternal: + /* needs addressing, perhaps. Difficult to do if we can't + * control one edge of the fade relative to the relevant edge + * of the region, which we cannot - fades are currently assumed + * to start/end at the start/end of the region + */ + break; + default: + return; + } +} + void AudioRegion::set_fade_in_shape (FadeShape shape) { @@ -983,9 +1018,10 @@ AudioRegion::set_fade_in (boost::shared_ptr f) void AudioRegion::set_fade_in (FadeShape shape, framecnt_t len) { - boost::shared_ptr c1 (new Evoral::ControlList (FadeInAutomation)); - boost::shared_ptr c2 (new Evoral::ControlList (FadeInAutomation)); - boost::shared_ptr c3 (new Evoral::ControlList (FadeInAutomation)); + const ARDOUR::ParameterDescriptor desc(FadeInAutomation); + boost::shared_ptr c1 (new Evoral::ControlList (FadeInAutomation, desc)); + boost::shared_ptr c2 (new Evoral::ControlList (FadeInAutomation, desc)); + boost::shared_ptr c3 (new Evoral::ControlList (FadeInAutomation, desc)); _fade_in->freeze (); _fade_in->clear (); @@ -995,8 +1031,8 @@ AudioRegion::set_fade_in (FadeShape shape, framecnt_t len) switch (shape) { case FadeLinear: - _fade_in->fast_simple_add (0.0, VERY_SMALL_SIGNAL); - _fade_in->fast_simple_add (len, 1.0); + _fade_in->fast_simple_add (0.0, GAIN_COEFF_SMALL); + _fade_in->fast_simple_add (len, GAIN_COEFF_UNITY); reverse_curve (_inverse_fade_in.val(), _fade_in.val()); break; @@ -1017,15 +1053,15 @@ AudioRegion::set_fade_in (FadeShape shape, framecnt_t len) break; case FadeConstantPower: - _fade_in->fast_simple_add (0.0, VERY_SMALL_SIGNAL); + _fade_in->fast_simple_add (0.0, GAIN_COEFF_SMALL); for (int i = 1; i < num_steps; ++i) { const float dist = i / (num_steps + 1.f); _fade_in->fast_simple_add (len * dist, sin (dist * M_PI / 2.0)); } - _fade_in->fast_simple_add (len, 1.0); + _fade_in->fast_simple_add (len, GAIN_COEFF_UNITY); reverse_curve (_inverse_fade_in.val(), _fade_in.val()); break; - + case FadeSymmetric: //start with a nearly linear cuve _fade_in->fast_simple_add (0, 1); @@ -1034,9 +1070,9 @@ AudioRegion::set_fade_in (FadeShape shape, framecnt_t len) const double breakpoint = 0.7; //linear for first 70% for (int i = 2; i < 9; ++i) { const float coeff = (1.f - breakpoint) * powf (0.5, i); - _fade_in->fast_simple_add (len * (breakpoint + ((1.0 - breakpoint) * (double)i / 9.0)), coeff); + _fade_in->fast_simple_add (len * (breakpoint + ((GAIN_COEFF_UNITY - breakpoint) * (double)i / 9.0)), coeff); } - _fade_in->fast_simple_add (len, VERY_SMALL_SIGNAL); + _fade_in->fast_simple_add (len, GAIN_COEFF_SMALL); reverse_curve (c3, _fade_in.val()); _fade_in->copy_events (*c3); reverse_curve (_inverse_fade_in.val(), _fade_in.val()); @@ -1065,8 +1101,9 @@ AudioRegion::set_fade_out (boost::shared_ptr f) void AudioRegion::set_fade_out (FadeShape shape, framecnt_t len) { - boost::shared_ptr c1 (new Evoral::ControlList (FadeOutAutomation)); - boost::shared_ptr c2 (new Evoral::ControlList (FadeOutAutomation)); + const ARDOUR::ParameterDescriptor desc(FadeOutAutomation); + boost::shared_ptr c1 (new Evoral::ControlList (FadeOutAutomation, desc)); + boost::shared_ptr c2 (new Evoral::ControlList (FadeOutAutomation, desc)); _fade_out->freeze (); _fade_out->clear (); @@ -1076,17 +1113,17 @@ AudioRegion::set_fade_out (FadeShape shape, framecnt_t len) switch (shape) { case FadeLinear: - _fade_out->fast_simple_add (0.0, 1.0); - _fade_out->fast_simple_add (len, VERY_SMALL_SIGNAL); + _fade_out->fast_simple_add (0.0, GAIN_COEFF_UNITY); + _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL); reverse_curve (_inverse_fade_out.val(), _fade_out.val()); break; - - case FadeFast: + + case FadeFast: generate_db_fade (_fade_out.val(), len, num_steps, -60); generate_inverse_power_curve (_inverse_fade_out.val(), _fade_out.val()); break; - - case FadeSlow: + + case FadeSlow: generate_db_fade (c1, len, num_steps, -1); //start off with a slow fade generate_db_fade (c2, len, num_steps, -80); //end with a fast fade merge_curves (_fade_out.val(), c1, c2); @@ -1096,15 +1133,15 @@ AudioRegion::set_fade_out (FadeShape shape, framecnt_t len) case FadeConstantPower: //constant-power fades use a sin/cos relationship //the cutoff is abrupt but it has the benefit of being symmetrical - _fade_out->fast_simple_add (0.0, 1.0); + _fade_out->fast_simple_add (0.0, GAIN_COEFF_UNITY); for (int i = 1; i < num_steps; ++i) { const float dist = i / (num_steps + 1.f); _fade_out->fast_simple_add (len * dist, cos (dist * M_PI / 2.0)); } - _fade_out->fast_simple_add (len, VERY_SMALL_SIGNAL); + _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL); reverse_curve (_inverse_fade_out.val(), _fade_out.val()); break; - + case FadeSymmetric: //start with a nearly linear cuve _fade_out->fast_simple_add (0, 1); @@ -1113,9 +1150,9 @@ AudioRegion::set_fade_out (FadeShape shape, framecnt_t len) const double breakpoint = 0.7; //linear for first 70% for (int i = 2; i < 9; ++i) { const float coeff = (1.f - breakpoint) * powf (0.5, i); - _fade_out->fast_simple_add (len * (breakpoint + ((1.0 - breakpoint) * (double)i / 9.0)), coeff); + _fade_out->fast_simple_add (len * (breakpoint + ((GAIN_COEFF_UNITY - breakpoint) * (double)i / 9.0)), coeff); } - _fade_out->fast_simple_add (len, VERY_SMALL_SIGNAL); + _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL); reverse_curve (_inverse_fade_out.val(), _fade_out.val()); break; } @@ -1134,7 +1171,7 @@ AudioRegion::set_fade_in_length (framecnt_t len) if (len > _length) { len = _length - 1; } - + if (len < 64) { len = 64; } @@ -1165,7 +1202,7 @@ AudioRegion::set_fade_out_length (framecnt_t len) bool changed = _fade_out->extend_to (len); if (changed) { - + if (_inverse_fade_out) { _inverse_fade_out->extend_to (len); } @@ -1234,8 +1271,8 @@ AudioRegion::set_default_envelope () { _envelope->freeze (); _envelope->clear (); - _envelope->fast_simple_add (0, 1.0f); - _envelope->fast_simple_add (_length, 1.0f); + _envelope->fast_simple_add (0, GAIN_COEFF_UNITY); + _envelope->fast_simple_add (_length, GAIN_COEFF_UNITY); _envelope->thaw (); } @@ -1415,14 +1452,14 @@ AudioRegion::normalize (float max_amplitude, float target_dB) { gain_t target = dB_to_coefficient (target_dB); - if (target == 1.0f) { + if (target == GAIN_COEFF_UNITY) { /* do not normalize to precisely 1.0 (0 dBFS), to avoid making it appear that we may have clipped. */ target -= FLT_EPSILON; } - if (max_amplitude == 0.0f) { + if (max_amplitude < GAIN_COEFF_SMALL) { /* don't even try */ return; } @@ -1529,93 +1566,164 @@ AudioRegion::audio_source (uint32_t n) const return boost::dynamic_pointer_cast(source(n)); } -int -AudioRegion::adjust_transients (frameoffset_t delta) +uint32_t +AudioRegion::get_related_audio_file_channel_count () const { - for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) { - (*x) = (*x) + delta; - } + uint32_t chan_count = 0; + for (SourceList::const_iterator i = _sources.begin(); i != _sources.end(); ++i) { - send_change (PropertyChange (Properties::valid_transients)); + boost::shared_ptr sndf = boost::dynamic_pointer_cast(*i); + if (sndf ) { - return 0; + if (sndf->channel_count() > chan_count) { + chan_count = sndf->channel_count(); + } + } +#ifdef HAVE_COREAUDIO + else { + boost::shared_ptr cauf = boost::dynamic_pointer_cast(*i); + if (cauf) { + if (cauf->channel_count() > chan_count) { + chan_count = cauf->channel_count(); + } + } + } +#endif // HAVE_COREAUDIO + } + + return chan_count; } -int -AudioRegion::update_transient (framepos_t old_position, framepos_t new_position) +void +AudioRegion::clear_transients () // yet unused { - for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) { - if ((*x) == old_position) { - (*x) = new_position; - send_change (PropertyChange (Properties::valid_transients)); + _user_transients.clear (); + _valid_transients = false; + send_change (PropertyChange (Properties::valid_transients)); +} - break; +void +AudioRegion::add_transient (framepos_t where) +{ + if (where < first_frame () || where >= last_frame ()) { + return; + } + where -= _position; + + if (!_valid_transients) { + _transient_user_start = _start; + _valid_transients = true; + } + frameoffset_t offset = _transient_user_start - _start; + + if (where < offset) { + if (offset <= 0) { + return; } + // region start changed (extend to front), shift points and offset + for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) { + (*x) += offset; + } + _transient_user_start -= offset; + offset = 0; } - return 0; + const framepos_t p = where - offset; + _user_transients.push_back(p); + send_change (PropertyChange (Properties::valid_transients)); } void -AudioRegion::add_transient (framepos_t where) +AudioRegion::update_transient (framepos_t old_position, framepos_t new_position) { - _transients.push_back(where); - _valid_transients = true; + bool changed = false; + if (!_onsets.empty ()) { + const framepos_t p = old_position - _position; + AnalysisFeatureList::iterator x = std::find (_onsets.begin (), _onsets.end (), p); + if (x != _transients.end ()) { + (*x) = new_position - _position; + changed = true; + } + } - send_change (PropertyChange (Properties::valid_transients)); + if (_valid_transients) { + const frameoffset_t offset = _position + _transient_user_start - _start; + const framepos_t p = old_position - offset; + AnalysisFeatureList::iterator x = std::find (_user_transients.begin (), _user_transients.end (), p); + if (x != _transients.end ()) { + (*x) = new_position - offset; + changed = true; + } + } + + if (changed) { + send_change (PropertyChange (Properties::valid_transients)); + } } void AudioRegion::remove_transient (framepos_t where) { - _transients.remove(where); - _valid_transients = true; + bool changed = false; + if (!_onsets.empty ()) { + const framepos_t p = where - _position; + AnalysisFeatureList::iterator i = std::find (_onsets.begin (), _onsets.end (), p); + if (i != _transients.end ()) { + _onsets.erase (i); + changed = true; + } + } - send_change (PropertyChange (Properties::valid_transients)); + if (_valid_transients) { + const framepos_t p = where - (_position + _transient_user_start - _start); + AnalysisFeatureList::iterator i = std::find (_user_transients.begin (), _user_transients.end (), p); + if (i != _transients.end ()) { + _transients.erase (i); + changed = true; + } + } + + if (changed) { + send_change (PropertyChange (Properties::valid_transients)); + } } -int -AudioRegion::set_transients (AnalysisFeatureList& results) +void +AudioRegion::set_onsets (AnalysisFeatureList& results) { - _transients.clear(); - _transients = results; - _valid_transients = true; - + _onsets.clear(); + _onsets = results; send_change (PropertyChange (Properties::valid_transients)); - - return 0; } -int -AudioRegion::get_transients (AnalysisFeatureList& results, bool force_new) +void +AudioRegion::build_transients () { + _transients.clear (); + _transient_analysis_start = _transient_analysis_end = 0; + boost::shared_ptr pl = playlist(); if (!pl) { - return -1; - } - - if (_valid_transients && !force_new) { - results = _transients; - return 0; + return; } + /* check analyzed sources first */ SourceList::iterator s; - for (s = _sources.begin() ; s != _sources.end(); ++s) { if (!(*s)->has_been_analysed()) { +#ifndef NDEBUG cerr << "For " << name() << " source " << (*s)->name() << " has not been analyzed\n"; +#endif break; } } if (s == _sources.end()) { /* all sources are analyzed, merge data from each one */ - for (s = _sources.begin() ; s != _sources.end(); ++s) { /* find the set of transients within the bounds of this region */ - AnalysisFeatureList::iterator low = lower_bound ((*s)->transients.begin(), (*s)->transients.end(), _start); @@ -1625,23 +1733,19 @@ AudioRegion::get_transients (AnalysisFeatureList& results, bool force_new) _start + _length); /* and add them */ - - results.insert (results.end(), low, high); + _transients.insert (_transients.end(), low, high); } - TransientDetector::cleanup_transients (results, pl->session().frame_rate(), 3.0); + TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0); /* translate all transients to current position */ - - for (AnalysisFeatureList::iterator x = results.begin(); x != results.end(); ++x) { + for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) { (*x) -= _start; - (*x) += _position; } - _transients = results; - _valid_transients = true; - - return 0; + _transient_analysis_start = _start; + _transient_analysis_end = _start + _length; + return; } /* no existing/complete transient info */ @@ -1650,71 +1754,88 @@ AudioRegion::get_transients (AnalysisFeatureList& results, bool force_new) if (!Config->get_auto_analyse_audio()) { if (!analyse_dialog_shown) { - pl->session().Dialog (_("\ + pl->session().Dialog (string_compose (_("\ You have requested an operation that requires audio analysis.\n\n\ You currently have \"auto-analyse-audio\" disabled, which means \ that transient data must be generated every time it is required.\n\n\ If you are doing work that will require transient data on a \ regular basis, you should probably enable \"auto-analyse-audio\" \ -then quit ardour and restart.\n\n\ +in Preferences > Audio > Regions, then quit %1 and restart.\n\n\ This dialog will not display again. But you may notice a slight delay \ in this and future transient-detection operations.\n\ -")); +"), PROGRAM_NAME)); analyse_dialog_shown = true; } } - TransientDetector t (pl->session().frame_rate()); - bool existing_results = !results.empty(); - - _transients.clear (); - _valid_transients = false; - - for (uint32_t i = 0; i < n_channels(); ++i) { + try { + TransientDetector t (pl->session().frame_rate()); + for (uint32_t i = 0; i < n_channels(); ++i) { - AnalysisFeatureList these_results; + AnalysisFeatureList these_results; - t.reset (); + t.reset (); - if (t.run ("", this, i, these_results)) { - return -1; - } - - /* translate all transients to give absolute position */ + /* this produces analysis result relative to current position + * ::read() sample 0 is at _position */ + if (t.run ("", this, i, these_results)) { + return; + } - for (AnalysisFeatureList::iterator i = these_results.begin(); i != these_results.end(); ++i) { - (*i) += _position; + /* merge */ + _transients.insert (_transients.end(), these_results.begin(), these_results.end()); } - - /* merge */ - - _transients.insert (_transients.end(), these_results.begin(), these_results.end()); + } catch (...) { + error << string_compose(_("Transient Analysis failed for %1."), _("Audio Region")) << endmsg; + return; } - if (!results.empty()) { - if (existing_results) { - - /* merge our transients into the existing ones, then clean up - those. - */ - - results.insert (results.end(), _transients.begin(), _transients.end()); - TransientDetector::cleanup_transients (results, pl->session().frame_rate(), 3.0); - } - - /* make sure ours are clean too */ + TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0); + _transient_analysis_start = _start; + _transient_analysis_end = _start + _length; +} - TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0); +/* Transient analysis uses ::read() which is relative to _start, + * at the time of analysis and spans _length samples. + * + * This is true for RhythmFerret::run_analysis and the + * TransientDetector here. + * + * We store _start and length in _transient_analysis_start, + * _transient_analysis_end in case the region is trimmed or split after analysis. + * + * Various methods (most notably Playlist::find_next_transient and + * RhythmFerret::do_split_action) span multiple regions and *merge/combine* + * Analysis results. + * We therefore need to translate the analysis timestamps to absolute session-time + * and include the _position of the region. + * + * Note: we should special case the AudioRegionView. The region-view itself + * is located at _position (currently ARV subtracts _position again) + */ +void +AudioRegion::get_transients (AnalysisFeatureList& results) +{ + boost::shared_ptr pl = playlist(); + if (!playlist ()) { + return; + } - } else { + Region::merge_features (results, _user_transients, _position + _transient_user_start - _start); - TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0); - results = _transients; + if (!_onsets.empty ()) { + // onsets are invalidated when start or length changes + merge_features (results, _onsets, _position); + return; } - _valid_transients = true; + if ((_transient_analysis_start == _transient_analysis_end) + || _transient_analysis_start > _start + || _transient_analysis_end < _start + _length) { + build_transients (); + } - return 0; + merge_features (results, _transients, _position + _transient_analysis_start - _start); } /** Find areas of `silence' within a region. @@ -1725,55 +1846,70 @@ in this and future transient-detection operations.\n\ */ AudioIntervalResult -AudioRegion::find_silence (Sample threshold, framecnt_t min_length, InterThreadInfo& itt) const +AudioRegion::find_silence (Sample threshold, framecnt_t min_length, framecnt_t fade_length, InterThreadInfo& itt) const { framecnt_t const block_size = 64 * 1024; boost::scoped_array loudest (new Sample[block_size]); boost::scoped_array buf (new Sample[block_size]); + assert (fade_length >= 0); + assert (min_length > 0); + framepos_t pos = _start; - framepos_t const end = _start + _length - 1; + framepos_t const end = _start + _length; AudioIntervalResult silent_periods; - bool in_silence = false; - frameoffset_t silence_start = 0; + bool in_silence = true; + frameoffset_t silence_start = _start; while (pos < end && !itt.cancel) { + framecnt_t cur_samples = 0; + framecnt_t const to_read = min (end - pos, block_size); /* fill `loudest' with the loudest absolute sample at each instant, across all channels */ memset (loudest.get(), 0, sizeof (Sample) * block_size); + for (uint32_t n = 0; n < n_channels(); ++n) { - read_raw_internal (buf.get(), pos, block_size, n); - for (framecnt_t i = 0; i < block_size; ++i) { + cur_samples = read_raw_internal (buf.get(), pos, to_read, n); + for (framecnt_t i = 0; i < cur_samples; ++i) { loudest[i] = max (loudest[i], abs (buf[i])); } } /* now look for silence */ - for (framecnt_t i = 0; i < block_size; ++i) { + for (framecnt_t i = 0; i < cur_samples; ++i) { bool const silence = abs (loudest[i]) < threshold; if (silence && !in_silence) { /* non-silence to silence */ in_silence = true; - silence_start = pos + i; + silence_start = pos + i + fade_length; } else if (!silence && in_silence) { /* silence to non-silence */ in_silence = false; - if (pos + i - 1 - silence_start >= min_length) { - silent_periods.push_back (std::make_pair (silence_start, pos + i - 1)); + frameoffset_t silence_end = pos + i - 1 - fade_length; + + if (silence_end - silence_start >= min_length) { + silent_periods.push_back (std::make_pair (silence_start, silence_end)); } } } - pos += block_size; - itt.progress = (end-pos)/(double)_length; + pos += cur_samples; + itt.progress = (end - pos) / (double)_length; + + if (cur_samples == 0) { + assert (pos >= end); + break; + } } - if (in_silence && end - 1 - silence_start >= min_length) { + if (in_silence && !itt.cancel) { /* last block was silent, so finish off the last period */ - silent_periods.push_back (std::make_pair (silence_start, end)); + if (end - 1 - silence_start >= min_length + fade_length) { + silent_periods.push_back (std::make_pair (silence_start, end - 1)); + } } itt.done = true; @@ -1806,7 +1942,7 @@ AudioRegion::get_single_other_xfade_region (bool start) const } else { rl = pl->regions_at (last_frame()); } - + RegionList::iterator i; boost::shared_ptr other; uint32_t n = 0; @@ -1859,6 +1995,6 @@ AudioRegion::verify_xfade_bounds (framecnt_t len, bool start) } return min (length(), min (maxlen, len)); - + }