#include <set>
#include <boost/scoped_array.hpp>
+#include <boost/shared_ptr.hpp>
-#include <glibmm/thread.h>
+#include <glibmm/threads.h>
#include "pbd/basename.h"
#include "pbd/xml++.h"
#include "evoral/Curve.hpp"
#include "ardour/audioregion.h"
-#include "ardour/debug.h"
#include "ardour/session.h"
#include "ardour/dB.h"
+#include "ardour/debug.h"
+#include "ardour/event_type_map.h"
#include "ardour/playlist.h"
#include "ardour/audiofilesource.h"
#include "ardour/region_factory.h"
#include "ardour/runtime_functions.h"
#include "ardour/transient_detector.h"
+#include "ardour/parameter_descriptor.h"
#include "ardour/progress.h"
-#include "i18n.h"
+#include "ardour/sndfilesource.h"
+#ifdef HAVE_COREAUDIO
+#include "ardour/coreaudiosource.h"
+#endif // HAVE_COREAUDIO
+
+#include "pbd/i18n.h"
#include <locale.h>
using namespace std;
PBD::PropertyDescriptor<bool> fade_in_active;
PBD::PropertyDescriptor<bool> fade_out_active;
PBD::PropertyDescriptor<float> scale_amplitude;
+ PBD::PropertyDescriptor<boost::shared_ptr<AutomationList> > fade_in;
+ PBD::PropertyDescriptor<boost::shared_ptr<AutomationList> > inverse_fade_in;
+ PBD::PropertyDescriptor<boost::shared_ptr<AutomationList> > fade_out;
+ PBD::PropertyDescriptor<boost::shared_ptr<AutomationList> > inverse_fade_out;
+ PBD::PropertyDescriptor<boost::shared_ptr<AutomationList> > envelope;
}
}
-static const double VERY_SMALL_SIGNAL = 0.0000001; //-140dB
-
/* Curve manipulations */
static void
reverse_curve (boost::shared_ptr<Evoral::ControlList> dst, boost::shared_ptr<const Evoral::ControlList> src)
{
size_t len = src->back()->when;
-
- for (Evoral::ControlList::const_iterator it = src->begin(); it!=src->end(); it++) {
- dst->add ( len - (*it)->when, (*it)->value );
+ for (Evoral::ControlList::const_reverse_iterator it = src->rbegin(); it!=src->rend(); it++) {
+ dst->fast_simple_add (len - (*it)->when, (*it)->value);
}
}
static void
generate_inverse_power_curve (boost::shared_ptr<Evoral::ControlList> dst, boost::shared_ptr<const Evoral::ControlList> src)
{
- //calc inverse curve using sum of squares
+ // calc inverse curve using sum of squares
for (Evoral::ControlList::const_iterator it = src->begin(); it!=src->end(); ++it ) {
float value = (*it)->value;
value = 1 - powf(value,2);
}
}
-/*
-static void
-generate_inverse_coefficient_curve (boost::shared_ptr<Evoral::ControlList> dst, boost::shared_ptr<const Evoral::ControlList> src)
-{
- //calc inverse gain coefficient curve
- for (Evoral::ControlList::const_iterator it = src->begin(); it!=src->end(); ++it ) {
- float value = 1.0 - (*it)->value;
- dst->fast_simple_add ( (*it)->when, value );
- }
-}
-*/
-
static void
generate_db_fade (boost::shared_ptr<Evoral::ControlList> dst, double len, int num_steps, float dB_drop)
{
+ dst->clear ();
dst->fast_simple_add (0, 1);
//generate a fade-out curve by successively applying a gain drop
float fade_speed = dB_to_coefficient(dB_drop / (float) num_steps);
+ float coeff = GAIN_COEFF_UNITY;
for (int i = 1; i < (num_steps-1); i++) {
- float coeff = 1.0;
- for (int j = 0; j < i; j++) {
- coeff *= fade_speed;
- }
+ coeff *= fade_speed;
dst->fast_simple_add (len*(double)i/(double)num_steps, coeff);
}
- dst->fast_simple_add (len, VERY_SMALL_SIGNAL);
+ dst->fast_simple_add (len, GAIN_COEFF_SMALL);
}
static void
-merge_curves (boost::shared_ptr<Evoral::ControlList> dst,
- boost::shared_ptr<const Evoral::ControlList> curve1,
+merge_curves (boost::shared_ptr<Evoral::ControlList> dst,
+ boost::shared_ptr<const Evoral::ControlList> curve1,
boost::shared_ptr<const Evoral::ControlList> curve2)
{
Evoral::ControlList::EventList::size_type size = curve1->size();
if (size != curve2->size()) {
return;
}
-
+
Evoral::ControlList::const_iterator c1 = curve1->begin();
int count = 0;
for (Evoral::ControlList::const_iterator c2 = curve2->begin(); c2!=curve2->end(); c2++ ) {
float v1 = accurate_coefficient_to_dB((*c1)->value);
float v2 = accurate_coefficient_to_dB((*c2)->value);
-
+
double interp = v1 * ( 1.0-( (double)count / (double)size) );
interp += v2 * ( (double)count / (double)size );
interp = dB_to_coefficient(interp);
- dst->add ( (*c1)->when, interp );
+ dst->fast_simple_add ( (*c1)->when, interp );
c1++;
count++;
}
DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for fade-out-active = %1\n", Properties::fade_out_active.property_id));
Properties::scale_amplitude.property_id = g_quark_from_static_string (X_("scale-amplitude"));
DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for scale-amplitude = %1\n", Properties::scale_amplitude.property_id));
+ Properties::fade_in.property_id = g_quark_from_static_string (X_("FadeIn"));
+ DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for FadeIn = %1\n", Properties::fade_in.property_id));
+ Properties::inverse_fade_in.property_id = g_quark_from_static_string (X_("InverseFadeIn"));
+ DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for InverseFadeIn = %1\n", Properties::inverse_fade_in.property_id));
+ Properties::fade_out.property_id = g_quark_from_static_string (X_("FadeOut"));
+ DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for FadeOut = %1\n", Properties::fade_out.property_id));
+ Properties::inverse_fade_out.property_id = g_quark_from_static_string (X_("InverseFadeOut"));
+ DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for InverseFadeOut = %1\n", Properties::inverse_fade_out.property_id));
+ Properties::envelope.property_id = g_quark_from_static_string (X_("Envelope"));
+ DEBUG_TRACE (DEBUG::Properties, string_compose ("quark for Envelope = %1\n", Properties::envelope.property_id));
}
void
add_property (_fade_in_active);
add_property (_fade_out_active);
add_property (_scale_amplitude);
+ add_property (_fade_in);
+ add_property (_inverse_fade_in);
+ add_property (_fade_out);
+ add_property (_inverse_fade_out);
+ add_property (_envelope);
}
#define AUDIOREGION_STATE_DEFAULT \
, _default_fade_out (Properties::default_fade_out, true) \
, _fade_in_active (Properties::fade_in_active, true) \
, _fade_out_active (Properties::fade_out_active, true) \
- , _scale_amplitude (Properties::scale_amplitude, 1.0)
+ , _scale_amplitude (Properties::scale_amplitude, 1.0) \
+ , _fade_in (Properties::fade_in, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter (FadeInAutomation)))) \
+ , _inverse_fade_in (Properties::inverse_fade_in, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter (FadeInAutomation)))) \
+ , _fade_out (Properties::fade_out, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter (FadeOutAutomation)))) \
+ , _inverse_fade_out (Properties::inverse_fade_out, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter (FadeOutAutomation))))
#define AUDIOREGION_COPY_STATE(other) \
_envelope_active (Properties::envelope_active, other->_envelope_active) \
, _default_fade_out (Properties::default_fade_out, other->_default_fade_out) \
, _fade_in_active (Properties::fade_in_active, other->_fade_in_active) \
, _fade_out_active (Properties::fade_out_active, other->_fade_out_active) \
- , _scale_amplitude (Properties::scale_amplitude, other->_scale_amplitude)
+ , _scale_amplitude (Properties::scale_amplitude, other->_scale_amplitude) \
+ , _fade_in (Properties::fade_in, boost::shared_ptr<AutomationList> (new AutomationList (*other->_fade_in.val()))) \
+ , _inverse_fade_in (Properties::fade_in, boost::shared_ptr<AutomationList> (new AutomationList (*other->_inverse_fade_in.val()))) \
+ , _fade_out (Properties::fade_in, boost::shared_ptr<AutomationList> (new AutomationList (*other->_fade_out.val()))) \
+ , _inverse_fade_out (Properties::fade_in, boost::shared_ptr<AutomationList> (new AutomationList (*other->_inverse_fade_out.val()))) \
/* a Session will reset these to its chosen defaults by calling AudioRegion::set_default_fade() */
void
AudioRegion::AudioRegion (Session& s, framepos_t start, framecnt_t len, std::string name)
: Region (s, start, len, name, DataType::AUDIO)
, AUDIOREGION_STATE_DEFAULT
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter(EnvelopeAutomation))))
, _automatable (s)
- , _fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _inverse_fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _inverse_fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _envelope (new AutomationList(Evoral::Parameter(EnvelopeAutomation)))
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
init ();
assert (_sources.size() == _master_sources.size());
AudioRegion::AudioRegion (const SourceList& srcs)
: Region (srcs)
, AUDIOREGION_STATE_DEFAULT
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList (Evoral::Parameter(EnvelopeAutomation))))
, _automatable(srcs[0]->session())
- , _fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _inverse_fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _inverse_fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _envelope (new AutomationList(Evoral::Parameter(EnvelopeAutomation)))
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
init ();
assert (_sources.size() == _master_sources.size());
AudioRegion::AudioRegion (boost::shared_ptr<const AudioRegion> other)
: Region (other)
, AUDIOREGION_COPY_STATE (other)
- , _automatable (other->session())
- , _fade_in (new AutomationList (*other->_fade_in))
- , _inverse_fade_in (new AutomationList(*other->_inverse_fade_in))
- , _fade_out (new AutomationList (*other->_fade_out))
- , _inverse_fade_out (new AutomationList (*other->_inverse_fade_out))
/* As far as I can see, the _envelope's times are relative to region position, and have nothing
to do with sources (and hence _start). So when we copy the envelope, we just use the supplied offset.
*/
- , _envelope (new AutomationList (*other->_envelope, 0, other->_length))
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList (*other->_envelope.val(), 0, other->_length)))
+ , _automatable (other->session())
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
/* don't use init here, because we got fade in/out from the other region
*/
assert (_sources.size() == _master_sources.size());
}
-AudioRegion::AudioRegion (boost::shared_ptr<const AudioRegion> other, framecnt_t offset)
+AudioRegion::AudioRegion (boost::shared_ptr<const AudioRegion> other, MusicFrame offset)
: Region (other, offset)
, AUDIOREGION_COPY_STATE (other)
- , _automatable (other->session())
- , _fade_in (new AutomationList (*other->_fade_in))
- , _inverse_fade_in (new AutomationList(*other->_inverse_fade_in))
- , _fade_out (new AutomationList (*other->_fade_out))
- , _inverse_fade_out (new AutomationList (*other->_inverse_fade_out))
/* As far as I can see, the _envelope's times are relative to region position, and have nothing
to do with sources (and hence _start). So when we copy the envelope, we just use the supplied offset.
*/
- , _envelope (new AutomationList (*other->_envelope, offset, other->_length))
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList (*other->_envelope.val(), offset.frame, other->_length)))
+ , _automatable (other->session())
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
/* don't use init here, because we got fade in/out from the other region
*/
AudioRegion::AudioRegion (boost::shared_ptr<const AudioRegion> other, const SourceList& srcs)
: Region (boost::static_pointer_cast<const Region>(other), srcs)
, AUDIOREGION_COPY_STATE (other)
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList (*other->_envelope.val())))
, _automatable (other->session())
- , _fade_in (new AutomationList (*other->_fade_in))
- , _inverse_fade_in (new AutomationList(*other->_inverse_fade_in))
- , _fade_out (new AutomationList (*other->_fade_out))
- , _inverse_fade_out (new AutomationList (*other->_inverse_fade_out))
- , _envelope (new AutomationList (*other->_envelope))
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
/* make-a-sort-of-copy-with-different-sources constructor (used by audio filter) */
AudioRegion::AudioRegion (SourceList& srcs)
: Region (srcs)
, AUDIOREGION_STATE_DEFAULT
+ , _envelope (Properties::envelope, boost::shared_ptr<AutomationList> (new AutomationList(Evoral::Parameter(EnvelopeAutomation))))
, _automatable(srcs[0]->session())
- , _fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _inverse_fade_in (new AutomationList(Evoral::Parameter(FadeInAutomation)))
- , _fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _inverse_fade_out (new AutomationList(Evoral::Parameter(FadeOutAutomation)))
- , _envelope (new AutomationList(Evoral::Parameter(EnvelopeAutomation)))
, _fade_in_suspended (0)
, _fade_out_suspended (0)
- , _fade_in_is_xfade (false)
- , _fade_out_is_xfade (false)
{
init ();
AudioRegion::connect_to_analysis_changed ()
{
for (SourceList::const_iterator i = _sources.begin(); i != _sources.end(); ++i) {
- (*i)->AnalysisChanged.connect_same_thread (*this, boost::bind (&AudioRegion::invalidate_transients, this));
+ (*i)->AnalysisChanged.connect_same_thread (*this, boost::bind (&AudioRegion::maybe_invalidate_transients, this));
}
}
}
}
+/** @param buf Buffer to put peak data in.
+ * @param npeaks Number of peaks to read (ie the number of PeakDatas in buf)
+ * @param offset Start position, as an offset from the start of this region's source.
+ * @param cnt Number of samples to read.
+ * @param chan_n Channel.
+ * @param frames_per_pixel Number of samples to use to generate one peak value.
+ */
+
ARDOUR::framecnt_t
-AudioRegion::read_peaks (PeakData *buf, framecnt_t npeaks, framecnt_t offset, framecnt_t cnt, uint32_t chan_n, double samples_per_unit) const
+AudioRegion::read_peaks (PeakData *buf, framecnt_t npeaks, framecnt_t offset, framecnt_t cnt, uint32_t chan_n, double frames_per_pixel) const
{
if (chan_n >= _sources.size()) {
return 0;
}
- if (audio_source(chan_n)->read_peaks (buf, npeaks, offset, cnt, samples_per_unit)) {
+ if (audio_source(chan_n)->read_peaks (buf, npeaks, offset, cnt, frames_per_pixel)) {
return 0;
- } else {
- if (_scale_amplitude != 1.0f) {
- for (framecnt_t n = 0; n < npeaks; ++n) {
- buf[n].max *= _scale_amplitude;
- buf[n].min *= _scale_amplitude;
- }
+ }
+
+ if (_scale_amplitude != 1.0f) {
+ for (framecnt_t n = 0; n < npeaks; ++n) {
+ buf[n].max *= _scale_amplitude;
+ buf[n].min *= _scale_amplitude;
}
- return cnt;
}
+
+ return npeaks;
}
/** @param buf Buffer to write data to (existing data will be overwritten).
*/
assert (cnt >= 0);
-
+
if (n_channels() == 0) {
return 0;
}
- if (muted()) {
- return 0; /* read nothing */
- }
-
-
/* WORK OUT WHERE TO GET DATA FROM */
framecnt_t to_read;
/* COMPUTE DETAILS OF ANY FADES INVOLVED IN THIS READ */
- /* Amount of fade in that we are dealing with in this read */
+ /* Amount (length) of fade in that we are dealing with in this read */
framecnt_t fade_in_limit = 0;
/* Offset from buf / mixdown_buffer of the start
of any fade out that we are dealing with
*/
frameoffset_t fade_out_offset = 0;
-
- /* Amount of fade in that we are dealing with in this read */
+
+ /* Amount (length) of fade out that we are dealing with in this read */
framecnt_t fade_out_limit = 0;
framecnt_t fade_interval_start = 0;
/* Fade in */
-
+
if (_fade_in_active && _session.config.get_use_region_fades()) {
-
+
framecnt_t fade_in_length = (framecnt_t) _fade_in->back()->when;
-
+
/* see if this read is within the fade in */
-
+
if (internal_offset < fade_in_length) {
fade_in_limit = min (to_read, fade_in_length - internal_offset);
}
}
-
+
/* Fade out */
-
+
if (_fade_out_active && _session.config.get_use_region_fades()) {
-
+
/* see if some part of this read is within the fade out */
/* ................. >| REGION
- _length
-
- { } FADE
- fade_out_length
- ^
- _length - fade_out_length
- |--------------|
- ^internal_offset
- ^internal_offset + to_read
-
- we need the intersection of [internal_offset,internal_offset+to_read] with
- [_length - fade_out_length, _length]
-
- */
-
+ * _length
+ *
+ * { } FADE
+ * fade_out_length
+ * ^
+ * _length - fade_out_length
+ *
+ * |--------------|
+ * ^internal_offset
+ * ^internal_offset + to_read
+ *
+ * we need the intersection of [internal_offset,internal_offset+to_read] with
+ * [_length - fade_out_length, _length]
+ *
+ */
fade_interval_start = max (internal_offset, _length - framecnt_t (_fade_out->back()->when));
framecnt_t fade_interval_end = min(internal_offset + to_read, _length.val());
-
+
if (fade_interval_end > fade_interval_start) {
/* (part of the) the fade out is in this buffer */
fade_out_limit = fade_interval_end - fade_interval_start;
/* READ DATA FROM THE SOURCE INTO mixdown_buffer.
We can never read directly into buf, since it may contain data
- from a transparent region `below' this one in the stack; we
- must always mix.
+ from a region `below' this one in the stack, and our fades (if they exist)
+ may need to mix with the existing data.
*/
if (read_from_sources (_sources, _length, mixdown_buffer, position, to_read, chan_n) != to_read) {
apply_gain_to_buffer (mixdown_buffer, to_read, _scale_amplitude);
}
-
/* APPLY FADES TO THE DATA IN mixdown_buffer AND MIX THE RESULTS INTO
* buf. The key things to realize here: (1) the fade being applied is
- * (as of April 26th 2012) just the inverse of the fade in curve (2)
+ * (as of April 26th 2012) just the inverse of the fade in curve (2)
* "buf" contains data from lower regions already. So this operation
* fades out the existing material.
*/
if (fade_in_limit != 0) {
- if (_inverse_fade_in) {
- /* explicit inverse fade in curve (e.g. for constant
- * power), so we have to fetch it.
- */
+ if (opaque()) {
+ if (_inverse_fade_in) {
- _inverse_fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
+ /* explicit inverse fade in curve (e.g. for constant
+ * power), so we have to fetch it.
+ */
- /* Fade the data from lower layers out */
- for (framecnt_t n = 0; n < fade_in_limit; ++n) {
- buf[n] *= gain_buffer[n];
- }
+ _inverse_fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
- /* refill gain buffer with the fade in */
+ /* Fade the data from lower layers out */
+ for (framecnt_t n = 0; n < fade_in_limit; ++n) {
+ buf[n] *= gain_buffer[n];
+ }
- _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
+ /* refill gain buffer with the fade in */
- } else {
+ _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
+
+ } else {
- /* no explicit inverse fade in, so just use (1 - fade
- * in) for the fade out of lower layers
- */
+ /* no explicit inverse fade in, so just use (1 - fade
+ * in) for the fade out of lower layers
+ */
- _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
+ _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
- for (framecnt_t n = 0; n < fade_in_limit; ++n) {
- buf[n] *= 1 - gain_buffer[n];
+ for (framecnt_t n = 0; n < fade_in_limit; ++n) {
+ buf[n] *= 1 - gain_buffer[n];
+ }
}
+ } else {
+ _fade_in->curve().get_vector (internal_offset, internal_offset + fade_in_limit, gain_buffer, fade_in_limit);
}
/* Mix our newly-read data in, with the fade */
framecnt_t const curve_offset = fade_interval_start - (_length - _fade_out->back()->when);
- if (_inverse_fade_out) {
+ if (opaque()) {
+ if (_inverse_fade_out) {
- _inverse_fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
+ _inverse_fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
- /* Fade the data from lower levels out */
- for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) {
- buf[m] *= gain_buffer[n];
- }
+ /* Fade the data from lower levels in */
+ for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) {
+ buf[m] *= gain_buffer[n];
+ }
- /* fetch the actual fade out */
+ /* fetch the actual fade out */
- _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
+ _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
- } else {
-
- /* no explicit inverse fade out, so just use (1 - fade
- * out) for the fade in of lower layers
- */
+ } else {
- _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
-
- for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) {
- buf[m] *= 1 - gain_buffer[n];
+ /* no explicit inverse fade out (which is
+ * actually a fade in), so just use (1 - fade
+ * out) for the fade in of lower layers
+ */
+
+ _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
+
+ for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) {
+ buf[m] *= 1 - gain_buffer[n];
+ }
}
+ } else {
+ _fade_out->curve().get_vector (curve_offset, curve_offset + fade_out_limit, gain_buffer, fade_out_limit);
}
- /* Mix our newly-read data out, with the fade */
+ /* Mix our newly-read data with whatever was already there,
+ with the fade out applied to our data.
+ */
for (framecnt_t n = 0, m = fade_out_offset; n < fade_out_limit; ++n, ++m) {
buf[m] += mixdown_buffer[m] * gain_buffer[n];
}
}
-
- /* MIX THE REGION BODY FROM mixdown_buffer INTO buf */
+ /* MIX OR COPY THE REGION BODY FROM mixdown_buffer INTO buf */
- mix_buffers_no_gain (buf + fade_in_limit, mixdown_buffer + fade_in_limit, to_read - fade_in_limit - fade_out_limit);
+ framecnt_t const N = to_read - fade_in_limit - fade_out_limit;
+ if (N > 0) {
+ if (opaque ()) {
+ DEBUG_TRACE (DEBUG::AudioPlayback, string_compose ("Region %1 memcpy into buf @ %2 + %3, from mixdown buffer @ %4 + %5, len = %6 cnt was %7\n",
+ name(), buf, fade_in_limit, mixdown_buffer, fade_in_limit, N, cnt));
+ memcpy (buf + fade_in_limit, mixdown_buffer + fade_in_limit, N * sizeof (Sample));
+ } else {
+ mix_buffers_no_gain (buf + fade_in_limit, mixdown_buffer + fade_in_limit, N);
+ }
+ }
return to_read;
}
if (to_read == 0) {
return 0;
}
-
+
if (chan_n < n_channels()) {
boost::shared_ptr<AudioSource> src = boost::dynamic_pointer_cast<AudioSource> (srcs[chan_n]);
*/
if (Config->get_replicate_missing_region_channels()) {
- /* track is N-channel, this region has less channels, so use a relevant channel
- */
- uint32_t channel = n_channels() % chan_n;
+ /* copy an existing channel's data in for this non-existant one */
+
+ uint32_t channel = chan_n % n_channels();
boost::shared_ptr<AudioSource> src = boost::dynamic_pointer_cast<AudioSource> (srcs[channel]);
if (src->read (buf, _start + internal_offset, to_read) != to_read) {
return 0; /* "read nothing" */
}
+
+ } else {
+
+ /* use silence */
+ memset (buf, 0, sizeof (Sample) * to_read);
}
}
}
XMLNode&
-AudioRegion::state ()
+AudioRegion::get_basic_state ()
{
XMLNode& node (Region::state ());
- XMLNode *child;
- char buf[64];
- LocaleGuard lg (X_("POSIX"));
+ LocaleGuard lg;
- snprintf (buf, sizeof (buf), "%u", (uint32_t) _sources.size());
- node.add_property ("channels", buf);
+ node.set_property ("channels", (uint32_t)_sources.size());
- Stateful::add_properties (node);
+ return node;
+}
+
+XMLNode&
+AudioRegion::state ()
+{
+ XMLNode& node (get_basic_state());
+ XMLNode *child;
+ LocaleGuard lg;
child = node.add_child ("Envelope");
// so, if they are both at 1.0f, that means the default region.
if (_envelope->size() == 2 &&
- _envelope->front()->value == 1.0f &&
- _envelope->back()->value==1.0f) {
+ _envelope->front()->value == GAIN_COEFF_UNITY &&
+ _envelope->back()->value==GAIN_COEFF_UNITY) {
if (_envelope->front()->when == 0 && _envelope->back()->when == _length) {
default_env = true;
}
}
if (default_env) {
- child->add_property ("default", "yes");
+ child->set_property ("default", "yes");
} else {
child->add_child_nocopy (_envelope->get_state ());
}
child = node.add_child (X_("FadeIn"));
- child->add_property ("is-xfade", (_fade_in_is_xfade ? "yes" : "no"));
if (_default_fade_in) {
- child->add_property ("default", "yes");
+ child->set_property ("default", "yes");
} else {
child->add_child_nocopy (_fade_in->get_state ());
}
if (_inverse_fade_in) {
- child = node.add_child (X_("InvFadeIn"));
+ child = node.add_child (X_("InverseFadeIn"));
child->add_child_nocopy (_inverse_fade_in->get_state ());
}
child = node.add_child (X_("FadeOut"));
- child->add_property ("is-xfade", (_fade_out_is_xfade ? "yes" : "no"));
if (_default_fade_out) {
- child->add_property ("default", "yes");
+ child->set_property ("default", "yes");
} else {
child->add_child_nocopy (_fade_out->get_state ());
}
if (_inverse_fade_out) {
- child = node.add_child (X_("InvFadeOut"));
+ child = node.add_child (X_("InverseFadeOut"));
child->add_child_nocopy (_inverse_fade_out->get_state ());
}
AudioRegion::_set_state (const XMLNode& node, int version, PropertyChange& what_changed, bool send)
{
const XMLNodeList& nlist = node.children();
- const XMLProperty *prop;
- LocaleGuard lg (X_("POSIX"));
+ LocaleGuard lg;
boost::shared_ptr<Playlist> the_playlist (_playlist.lock());
suspend_property_changes ();
Region::_set_state (node, version, what_changed, false);
- if ((prop = node.property ("scale-gain")) != 0) {
- float a = atof (prop->value().c_str());
- if (a != _scale_amplitude) {
- _scale_amplitude = a;
+ float val;
+ if (node.get_property ("scale-gain", val)) {
+ if (val != _scale_amplitude) {
+ _scale_amplitude = val;
what_changed.add (Properties::scale_amplitude);
}
}
for (XMLNodeConstIterator niter = nlist.begin(); niter != nlist.end(); ++niter) {
XMLNode *child;
- XMLProperty *prop;
+ XMLProperty const * prop;
child = (*niter);
_fade_in->clear ();
- if (((prop = child->property ("default")) != 0 && string_is_affirmative (prop->value())) || (prop = child->property ("steepness")) != 0) {
+ bool is_default;
+ if ((child->get_property ("default", is_default) && is_default) || (prop = child->property ("steepness")) != 0) {
set_default_fade_in ();
} else {
XMLNode* grandchild = child->child ("AutomationList");
}
}
- if ((prop = child->property ("is-xfade")) != 0) {
- _fade_in_is_xfade = string_is_affirmative (prop->value());
- } else {
- _fade_in_is_xfade = false;
- }
-
- if ((prop = child->property ("active")) != 0) {
- if (string_is_affirmative (prop->value())) {
- set_fade_in_active (true);
- } else {
- set_fade_in_active (false);
- }
+ bool is_active;
+ if (child->get_property ("active", is_active)) {
+ set_fade_in_active (is_active);
}
} else if (child->name() == "FadeOut") {
_fade_out->clear ();
- if (((prop = child->property ("default")) != 0 && (string_is_affirmative (prop->value()))) || (prop = child->property ("steepness")) != 0) {
+ bool is_default;
+ if ((child->get_property ("default", is_default) && is_default) || (prop = child->property ("steepness")) != 0) {
set_default_fade_out ();
} else {
XMLNode* grandchild = child->child ("AutomationList");
}
}
- if ((prop = child->property ("is-xfade")) != 0) {
- _fade_out_is_xfade = string_is_affirmative (prop->value());
- } else {
- _fade_out_is_xfade = false;
- }
-
- if ((prop = child->property ("active")) != 0) {
- if (string_is_affirmative (prop->value())) {
- set_fade_out_active (true);
- } else {
- set_fade_out_active (false);
- }
+ bool is_active;
+ if (child->get_property ("active", is_active)) {
+ set_fade_out_active (is_active);
}
- } else if (child->name() == "InvFadeIn") {
+ } else if ( (child->name() == "InverseFadeIn") || (child->name() == "InvFadeIn") ) {
XMLNode* grandchild = child->child ("AutomationList");
if (grandchild) {
_inverse_fade_in->set_state (*grandchild, version);
}
- } else if (child->name() == "InvFadeOut") {
+ } else if ( (child->name() == "InverseFadeOut") || (child->name() == "InvFadeOut") ) {
XMLNode* grandchild = child->child ("AutomationList");
if (grandchild) {
_inverse_fade_out->set_state (*grandchild, version);
return _set_state (node, version, what_changed, true);
}
+void
+AudioRegion::fade_range (framepos_t start, framepos_t end)
+{
+ framepos_t s, e;
+
+ switch (coverage (start, end)) {
+ case Evoral::OverlapStart:
+ trim_front(start);
+ s = _position;
+ e = end;
+ set_fade_in (FadeConstantPower, e - s);
+ break;
+ case Evoral::OverlapEnd:
+ trim_end(end);
+ s = start;
+ e = _position + _length;
+ set_fade_out (FadeConstantPower, e - s);
+ break;
+ case Evoral::OverlapInternal:
+ /* needs addressing, perhaps. Difficult to do if we can't
+ * control one edge of the fade relative to the relevant edge
+ * of the region, which we cannot - fades are currently assumed
+ * to start/end at the start/end of the region
+ */
+ break;
+ default:
+ return;
+ }
+}
+
void
AudioRegion::set_fade_in_shape (FadeShape shape)
{
AudioRegion::set_fade_in (boost::shared_ptr<AutomationList> f)
{
_fade_in->freeze ();
- *_fade_in = *f;
+ *(_fade_in.val()) = *f;
_fade_in->thaw ();
_default_fade_in = false;
void
AudioRegion::set_fade_in (FadeShape shape, framecnt_t len)
{
- boost::shared_ptr<Evoral::ControlList> c1 (new Evoral::ControlList (FadeInAutomation));
- boost::shared_ptr<Evoral::ControlList> c2 (new Evoral::ControlList (FadeInAutomation));
- boost::shared_ptr<Evoral::ControlList> c3 (new Evoral::ControlList (FadeInAutomation));
+ const ARDOUR::ParameterDescriptor desc(FadeInAutomation);
+ boost::shared_ptr<Evoral::ControlList> c1 (new Evoral::ControlList (FadeInAutomation, desc));
+ boost::shared_ptr<Evoral::ControlList> c2 (new Evoral::ControlList (FadeInAutomation, desc));
+ boost::shared_ptr<Evoral::ControlList> c3 (new Evoral::ControlList (FadeInAutomation, desc));
_fade_in->freeze ();
_fade_in->clear ();
_inverse_fade_in->clear ();
+ const int num_steps = 32;
+
switch (shape) {
case FadeLinear:
- _fade_in->fast_simple_add (0.0, 0.0);
- _fade_in->fast_simple_add (len, 1.0);
- reverse_curve (_inverse_fade_in, _fade_in);
+ _fade_in->fast_simple_add (0.0, GAIN_COEFF_SMALL);
+ _fade_in->fast_simple_add (len, GAIN_COEFF_UNITY);
+ reverse_curve (_inverse_fade_in.val(), _fade_in.val());
break;
case FadeFast:
- generate_db_fade (_fade_in, len, 10, -60);
- reverse_curve (c1, _fade_in);
+ generate_db_fade (_fade_in.val(), len, num_steps, -60);
+ reverse_curve (c1, _fade_in.val());
_fade_in->copy_events (*c1);
- generate_inverse_power_curve (_inverse_fade_in, _fade_in);
+ generate_inverse_power_curve (_inverse_fade_in.val(), _fade_in.val());
break;
case FadeSlow:
- generate_db_fade (c1, len, 10, -1); // start off with a slow fade
- generate_db_fade (c2, len, 10, -80); // end with a fast fade
- merge_curves (_fade_in, c1, c2);
- generate_inverse_power_curve (_inverse_fade_in, _fade_in);
+ generate_db_fade (c1, len, num_steps, -1); // start off with a slow fade
+ generate_db_fade (c2, len, num_steps, -80); // end with a fast fade
+ merge_curves (_fade_in.val(), c1, c2);
+ reverse_curve (c3, _fade_in.val());
+ _fade_in->copy_events (*c3);
+ generate_inverse_power_curve (_inverse_fade_in.val(), _fade_in.val());
break;
case FadeConstantPower:
- for (int i = 0; i < 9; ++i) {
- float dist = (float) i / 10.0f;
- _fade_in->fast_simple_add (len*dist, sin (dist*M_PI/2));
+ _fade_in->fast_simple_add (0.0, GAIN_COEFF_SMALL);
+ for (int i = 1; i < num_steps; ++i) {
+ const float dist = i / (num_steps + 1.f);
+ _fade_in->fast_simple_add (len * dist, sin (dist * M_PI / 2.0));
}
- _fade_in->fast_simple_add (len, 1.0);
- reverse_curve (_inverse_fade_in, _fade_in);
+ _fade_in->fast_simple_add (len, GAIN_COEFF_UNITY);
+ reverse_curve (_inverse_fade_in.val(), _fade_in.val());
break;
-
+
case FadeSymmetric:
- // starts kind of like a constant power but has a slower fadeout
- // however it is NOT constant power and there will be a level drop in the middle of the crossfade
- c1->fast_simple_add (0.0, 1.0);
- for ( int i = 1; i < 9; i++ ) {
- float dist = (float)i/10.0;
- c1->fast_simple_add ((len * dist), cos(dist*M_PI/10.0));
+ //start with a nearly linear cuve
+ _fade_in->fast_simple_add (0, 1);
+ _fade_in->fast_simple_add (0.5 * len, 0.6);
+ //now generate a fade-out curve by successively applying a gain drop
+ const double breakpoint = 0.7; //linear for first 70%
+ for (int i = 2; i < 9; ++i) {
+ const float coeff = (1.f - breakpoint) * powf (0.5, i);
+ _fade_in->fast_simple_add (len * (breakpoint + ((GAIN_COEFF_UNITY - breakpoint) * (double)i / 9.0)), coeff);
}
- c1->fast_simple_add (len, VERY_SMALL_SIGNAL);
-
- //curve 2 is a slow fade at end
- generate_db_fade (c2, len, 10, -30 );
-
- merge_curves (c3, c1, c2);
- reverse_curve (_fade_in, c3);
- reverse_curve (_inverse_fade_in, _fade_in );
+ _fade_in->fast_simple_add (len, GAIN_COEFF_SMALL);
+ reverse_curve (c3, _fade_in.val());
+ _fade_in->copy_events (*c3);
+ reverse_curve (_inverse_fade_in.val(), _fade_in.val());
break;
}
+ _fade_in->set_interpolation(Evoral::ControlList::Curved);
+ _inverse_fade_in->set_interpolation(Evoral::ControlList::Curved);
+
_default_fade_in = false;
_fade_in->thaw ();
send_change (PropertyChange (Properties::fade_in));
AudioRegion::set_fade_out (boost::shared_ptr<AutomationList> f)
{
_fade_out->freeze ();
- *_fade_out = *f;
+ *(_fade_out.val()) = *f;
_fade_out->thaw ();
_default_fade_out = false;
- send_change (PropertyChange (Properties::fade_in));
+ send_change (PropertyChange (Properties::fade_out));
}
void
AudioRegion::set_fade_out (FadeShape shape, framecnt_t len)
{
- boost::shared_ptr<Evoral::ControlList> c1 (new Evoral::ControlList (FadeOutAutomation));
- boost::shared_ptr<Evoral::ControlList> c2 (new Evoral::ControlList (FadeOutAutomation));
+ const ARDOUR::ParameterDescriptor desc(FadeOutAutomation);
+ boost::shared_ptr<Evoral::ControlList> c1 (new Evoral::ControlList (FadeOutAutomation, desc));
+ boost::shared_ptr<Evoral::ControlList> c2 (new Evoral::ControlList (FadeOutAutomation, desc));
_fade_out->freeze ();
_fade_out->clear ();
_inverse_fade_out->clear ();
+ const int num_steps = 32;
+
switch (shape) {
case FadeLinear:
- _fade_out->fast_simple_add (0.0, 1.0);
- _fade_out->fast_simple_add (len, VERY_SMALL_SIGNAL);
- reverse_curve (_inverse_fade_out, _fade_out);
+ _fade_out->fast_simple_add (0.0, GAIN_COEFF_UNITY);
+ _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL);
+ reverse_curve (_inverse_fade_out.val(), _fade_out.val());
break;
-
- case FadeFast:
- generate_db_fade (_fade_out, len, 10, -60 );
- generate_inverse_power_curve (_inverse_fade_out, _fade_out);
+
+ case FadeFast:
+ generate_db_fade (_fade_out.val(), len, num_steps, -60);
+ generate_inverse_power_curve (_inverse_fade_out.val(), _fade_out.val());
break;
-
- case FadeSlow:
- generate_db_fade (c1, len, 10, -1 ); //start off with a slow fade
- generate_db_fade (c2, len, 10, -80 ); //end with a fast fade
- merge_curves (_fade_out, c1, c2);
- generate_inverse_power_curve (_inverse_fade_out, _fade_out);
+
+ case FadeSlow:
+ generate_db_fade (c1, len, num_steps, -1); //start off with a slow fade
+ generate_db_fade (c2, len, num_steps, -80); //end with a fast fade
+ merge_curves (_fade_out.val(), c1, c2);
+ generate_inverse_power_curve (_inverse_fade_out.val(), _fade_out.val());
break;
case FadeConstantPower:
//constant-power fades use a sin/cos relationship
//the cutoff is abrupt but it has the benefit of being symmetrical
- _fade_out->fast_simple_add (0.0, 1.0);
- for (int i = 1; i < 9; i++ ) {
- float dist = (float)i/10.0;
- _fade_out->fast_simple_add ((len * dist), cos(dist*M_PI/2));
+ _fade_out->fast_simple_add (0.0, GAIN_COEFF_UNITY);
+ for (int i = 1; i < num_steps; ++i) {
+ const float dist = i / (num_steps + 1.f);
+ _fade_out->fast_simple_add (len * dist, cos (dist * M_PI / 2.0));
}
- _fade_out->fast_simple_add (len, VERY_SMALL_SIGNAL);
- reverse_curve (_inverse_fade_out, _fade_out);
+ _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL);
+ reverse_curve (_inverse_fade_out.val(), _fade_out.val());
break;
-
+
case FadeSymmetric:
- //starts kind of like a constant power but has a slower fadeout
- //however it is NOT constant power and there will be a level drop in the middle of the crossfade
- c1->fast_simple_add (0.0, 1.0);
- for ( int i = 1; i < 9; i++ ) {
- float dist = (float)i/10.0;
- c1->fast_simple_add ((len * dist), cos(dist*M_PI/10.0)); //cheesy way of making a flat line
+ //start with a nearly linear cuve
+ _fade_out->fast_simple_add (0, 1);
+ _fade_out->fast_simple_add (0.5 * len, 0.6);
+ //now generate a fade-out curve by successively applying a gain drop
+ const double breakpoint = 0.7; //linear for first 70%
+ for (int i = 2; i < 9; ++i) {
+ const float coeff = (1.f - breakpoint) * powf (0.5, i);
+ _fade_out->fast_simple_add (len * (breakpoint + ((GAIN_COEFF_UNITY - breakpoint) * (double)i / 9.0)), coeff);
}
- c1->fast_simple_add (len, VERY_SMALL_SIGNAL);
-
- //curve 2 is a slow fade at end
- generate_db_fade (c2, len, 10, -30);
-
- merge_curves (_fade_out, c1, c2);
- reverse_curve (_inverse_fade_out, _fade_out);
+ _fade_out->fast_simple_add (len, GAIN_COEFF_SMALL);
+ reverse_curve (_inverse_fade_out.val(), _fade_out.val());
break;
}
+ _fade_out->set_interpolation(Evoral::ControlList::Curved);
+ _inverse_fade_out->set_interpolation(Evoral::ControlList::Curved);
+
_default_fade_out = false;
_fade_out->thaw ();
send_change (PropertyChange (Properties::fade_out));
if (len > _length) {
len = _length - 1;
}
-
+
if (len < 64) {
len = 64;
}
_inverse_fade_in->extend_to (len);
}
- if (_session.config.get_xfade_model() == FullCrossfade &&
- _session.config.get_auto_xfade() &&
- _fade_in_is_xfade) {
-
- /* trim a single other region below us to the new start
- of the fade.
- */
-
- boost::shared_ptr<Region> other = get_single_other_xfade_region (true);
- if (other) {
- other->trim_end (position() + len);
- }
- }
-
_default_fade_in = false;
send_change (PropertyChange (Properties::fade_in));
}
bool changed = _fade_out->extend_to (len);
if (changed) {
-
+
if (_inverse_fade_out) {
_inverse_fade_out->extend_to (len);
}
_default_fade_out = false;
-
- if (_session.config.get_xfade_model() == FullCrossfade &&
- _session.config.get_auto_xfade() &&
- _fade_out_is_xfade) {
-
- /* trim a single other region below us to the new start
- of the fade.
- */
-
- boost::shared_ptr<Region> other = get_single_other_xfade_region (false);
- if (other) {
- other->trim_front (last_frame() - len);
- }
- }
send_change (PropertyChange (Properties::fade_out));
}
AudioRegion::set_default_fade_in ()
{
_fade_in_suspended = 0;
- _fade_in_is_xfade = false;
- set_fade_in (FadeLinear, 64);
+ set_fade_in (Config->get_default_fade_shape(), 64);
}
void
AudioRegion::set_default_fade_out ()
{
_fade_out_suspended = 0;
- _fade_out_is_xfade = false;
- set_fade_out (FadeLinear, 64);
+ set_fade_out (Config->get_default_fade_shape(), 64);
}
void
{
_envelope->freeze ();
_envelope->clear ();
- _envelope->fast_simple_add (0, 1.0f);
- _envelope->fast_simple_add (_length, 1.0f);
+ _envelope->fast_simple_add (0, GAIN_COEFF_UNITY);
+ _envelope->fast_simple_add (_length, GAIN_COEFF_UNITY);
_envelope->thaw ();
}
send_change (PropertyChange (Properties::scale_amplitude));
}
-/** @return the maximum (linear) amplitude of the region, or a -ve
- * number if the Progress object reports that the process was cancelled.
- */
double
AudioRegion::maximum_amplitude (Progress* p) const
{
return maxamp;
}
+double
+AudioRegion::rms (Progress* p) const
+{
+ framepos_t fpos = _start;
+ framepos_t const fend = _start + _length;
+ uint32_t const n_chan = n_channels ();
+ double rms = 0;
+
+ framecnt_t const blocksize = 64 * 1024;
+ Sample buf[blocksize];
+
+ framecnt_t total = 0;
+
+ if (n_chan == 0 || fend == fpos) {
+ return 0;
+ }
+
+ while (fpos < fend) {
+ framecnt_t const to_read = min (fend - fpos, blocksize);
+ for (uint32_t c = 0; c < n_chan; ++c) {
+ if (read_raw_internal (buf, fpos, to_read, c) != to_read) {
+ return 0;
+ }
+ for (framepos_t i = 0; i < to_read; ++i) {
+ rms += buf[i] * buf[i];
+ }
+ }
+ total += to_read;
+ fpos += to_read;
+ if (p) {
+ p->set_progress (float (fpos - _start) / _length);
+ if (p->cancelled ()) {
+ return -1;
+ }
+ }
+ }
+ return sqrt (rms / (double)(total * n_chan));
+}
+
/** Normalize using a given maximum amplitude and target, so that region
* _scale_amplitude becomes target / max_amplitude.
*/
{
gain_t target = dB_to_coefficient (target_dB);
- if (target == 1.0f) {
+ if (target == GAIN_COEFF_UNITY) {
/* do not normalize to precisely 1.0 (0 dBFS), to avoid making it appear
that we may have clipped.
*/
target -= FLT_EPSILON;
}
- if (max_amplitude == 0.0f) {
+ if (max_amplitude < GAIN_COEFF_SMALL) {
/* don't even try */
return;
}
return boost::dynamic_pointer_cast<AudioSource>(source(n));
}
-int
-AudioRegion::adjust_transients (frameoffset_t delta)
+uint32_t
+AudioRegion::get_related_audio_file_channel_count () const
{
- for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) {
- (*x) = (*x) + delta;
- }
+ uint32_t chan_count = 0;
+ for (SourceList::const_iterator i = _sources.begin(); i != _sources.end(); ++i) {
- send_change (PropertyChange (Properties::valid_transients));
+ boost::shared_ptr<SndFileSource> sndf = boost::dynamic_pointer_cast<SndFileSource>(*i);
+ if (sndf ) {
- return 0;
+ if (sndf->channel_count() > chan_count) {
+ chan_count = sndf->channel_count();
+ }
+ }
+#ifdef HAVE_COREAUDIO
+ else {
+ boost::shared_ptr<CoreAudioSource> cauf = boost::dynamic_pointer_cast<CoreAudioSource>(*i);
+ if (cauf) {
+ if (cauf->channel_count() > chan_count) {
+ chan_count = cauf->channel_count();
+ }
+ }
+ }
+#endif // HAVE_COREAUDIO
+ }
+
+ return chan_count;
}
-int
-AudioRegion::update_transient (framepos_t old_position, framepos_t new_position)
+void
+AudioRegion::clear_transients () // yet unused
+{
+ _user_transients.clear ();
+ _valid_transients = false;
+ send_change (PropertyChange (Properties::valid_transients));
+}
+
+void
+AudioRegion::add_transient (framepos_t where)
{
- for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) {
- if ((*x) == old_position) {
- (*x) = new_position;
- send_change (PropertyChange (Properties::valid_transients));
+ if (where < first_frame () || where >= last_frame ()) {
+ return;
+ }
+ where -= _position;
- break;
+ if (!_valid_transients) {
+ _transient_user_start = _start;
+ _valid_transients = true;
+ }
+ frameoffset_t offset = _transient_user_start - _start;
+
+ if (where < offset) {
+ if (offset <= 0) {
+ return;
}
+ // region start changed (extend to front), shift points and offset
+ for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) {
+ (*x) += offset;
+ }
+ _transient_user_start -= offset;
+ offset = 0;
}
- return 0;
+ const framepos_t p = where - offset;
+ _user_transients.push_back(p);
+ send_change (PropertyChange (Properties::valid_transients));
}
void
-AudioRegion::add_transient (framepos_t where)
+AudioRegion::update_transient (framepos_t old_position, framepos_t new_position)
{
- _transients.push_back(where);
- _valid_transients = true;
+ bool changed = false;
+ if (!_onsets.empty ()) {
+ const framepos_t p = old_position - _position;
+ AnalysisFeatureList::iterator x = std::find (_onsets.begin (), _onsets.end (), p);
+ if (x != _transients.end ()) {
+ (*x) = new_position - _position;
+ changed = true;
+ }
+ }
- send_change (PropertyChange (Properties::valid_transients));
+ if (_valid_transients) {
+ const frameoffset_t offset = _position + _transient_user_start - _start;
+ const framepos_t p = old_position - offset;
+ AnalysisFeatureList::iterator x = std::find (_user_transients.begin (), _user_transients.end (), p);
+ if (x != _transients.end ()) {
+ (*x) = new_position - offset;
+ changed = true;
+ }
+ }
+
+ if (changed) {
+ send_change (PropertyChange (Properties::valid_transients));
+ }
}
void
AudioRegion::remove_transient (framepos_t where)
{
- _transients.remove(where);
- _valid_transients = true;
+ bool changed = false;
+ if (!_onsets.empty ()) {
+ const framepos_t p = where - _position;
+ AnalysisFeatureList::iterator i = std::find (_onsets.begin (), _onsets.end (), p);
+ if (i != _transients.end ()) {
+ _onsets.erase (i);
+ changed = true;
+ }
+ }
- send_change (PropertyChange (Properties::valid_transients));
+ if (_valid_transients) {
+ const framepos_t p = where - (_position + _transient_user_start - _start);
+ AnalysisFeatureList::iterator i = std::find (_user_transients.begin (), _user_transients.end (), p);
+ if (i != _transients.end ()) {
+ _transients.erase (i);
+ changed = true;
+ }
+ }
+
+ if (changed) {
+ send_change (PropertyChange (Properties::valid_transients));
+ }
}
-int
-AudioRegion::set_transients (AnalysisFeatureList& results)
+void
+AudioRegion::set_onsets (AnalysisFeatureList& results)
{
- _transients.clear();
- _transients = results;
- _valid_transients = true;
-
+ _onsets.clear();
+ _onsets = results;
send_change (PropertyChange (Properties::valid_transients));
-
- return 0;
}
-int
-AudioRegion::get_transients (AnalysisFeatureList& results, bool force_new)
+void
+AudioRegion::build_transients ()
{
+ _transients.clear ();
+ _transient_analysis_start = _transient_analysis_end = 0;
+
boost::shared_ptr<Playlist> pl = playlist();
if (!pl) {
- return -1;
- }
-
- if (_valid_transients && !force_new) {
- results = _transients;
- return 0;
+ return;
}
+ /* check analyzed sources first */
SourceList::iterator s;
-
for (s = _sources.begin() ; s != _sources.end(); ++s) {
if (!(*s)->has_been_analysed()) {
+#ifndef NDEBUG
cerr << "For " << name() << " source " << (*s)->name() << " has not been analyzed\n";
+#endif
break;
}
}
if (s == _sources.end()) {
/* all sources are analyzed, merge data from each one */
-
for (s = _sources.begin() ; s != _sources.end(); ++s) {
/* find the set of transients within the bounds of this region */
-
AnalysisFeatureList::iterator low = lower_bound ((*s)->transients.begin(),
(*s)->transients.end(),
_start);
_start + _length);
/* and add them */
-
- results.insert (results.end(), low, high);
+ _transients.insert (_transients.end(), low, high);
}
- TransientDetector::cleanup_transients (results, pl->session().frame_rate(), 3.0);
+ TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0);
/* translate all transients to current position */
-
- for (AnalysisFeatureList::iterator x = results.begin(); x != results.end(); ++x) {
+ for (AnalysisFeatureList::iterator x = _transients.begin(); x != _transients.end(); ++x) {
(*x) -= _start;
- (*x) += _position;
}
- _transients = results;
- _valid_transients = true;
-
- return 0;
+ _transient_analysis_start = _start;
+ _transient_analysis_end = _start + _length;
+ return;
}
/* no existing/complete transient info */
if (!Config->get_auto_analyse_audio()) {
if (!analyse_dialog_shown) {
- pl->session().Dialog (_("\
+ pl->session().Dialog (string_compose (_("\
You have requested an operation that requires audio analysis.\n\n\
You currently have \"auto-analyse-audio\" disabled, which means \
that transient data must be generated every time it is required.\n\n\
If you are doing work that will require transient data on a \
regular basis, you should probably enable \"auto-analyse-audio\" \
-then quit ardour and restart.\n\n\
+in Preferences > Audio > Regions, then quit %1 and restart.\n\n\
This dialog will not display again. But you may notice a slight delay \
in this and future transient-detection operations.\n\
-"));
+"), PROGRAM_NAME));
analyse_dialog_shown = true;
}
}
- TransientDetector t (pl->session().frame_rate());
- bool existing_results = !results.empty();
-
- _transients.clear ();
- _valid_transients = false;
+ try {
+ TransientDetector t (pl->session().frame_rate());
+ for (uint32_t i = 0; i < n_channels(); ++i) {
- for (uint32_t i = 0; i < n_channels(); ++i) {
+ AnalysisFeatureList these_results;
- AnalysisFeatureList these_results;
+ t.reset ();
- t.reset ();
-
- if (t.run ("", this, i, these_results)) {
- return -1;
- }
-
- /* translate all transients to give absolute position */
+ /* this produces analysis result relative to current position
+ * ::read() sample 0 is at _position */
+ if (t.run ("", this, i, these_results)) {
+ return;
+ }
- for (AnalysisFeatureList::iterator i = these_results.begin(); i != these_results.end(); ++i) {
- (*i) += _position;
+ /* merge */
+ _transients.insert (_transients.end(), these_results.begin(), these_results.end());
}
-
- /* merge */
-
- _transients.insert (_transients.end(), these_results.begin(), these_results.end());
+ } catch (...) {
+ error << string_compose(_("Transient Analysis failed for %1."), _("Audio Region")) << endmsg;
+ return;
}
- if (!results.empty()) {
- if (existing_results) {
-
- /* merge our transients into the existing ones, then clean up
- those.
- */
-
- results.insert (results.end(), _transients.begin(), _transients.end());
- TransientDetector::cleanup_transients (results, pl->session().frame_rate(), 3.0);
- }
-
- /* make sure ours are clean too */
+ TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0);
+ _transient_analysis_start = _start;
+ _transient_analysis_end = _start + _length;
+}
- TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0);
+/* Transient analysis uses ::read() which is relative to _start,
+ * at the time of analysis and spans _length samples.
+ *
+ * This is true for RhythmFerret::run_analysis and the
+ * TransientDetector here.
+ *
+ * We store _start and length in _transient_analysis_start,
+ * _transient_analysis_end in case the region is trimmed or split after analysis.
+ *
+ * Various methods (most notably Playlist::find_next_transient and
+ * RhythmFerret::do_split_action) span multiple regions and *merge/combine*
+ * Analysis results.
+ * We therefore need to translate the analysis timestamps to absolute session-time
+ * and include the _position of the region.
+ *
+ * Note: we should special case the AudioRegionView. The region-view itself
+ * is located at _position (currently ARV subtracts _position again)
+ */
+void
+AudioRegion::get_transients (AnalysisFeatureList& results)
+{
+ boost::shared_ptr<Playlist> pl = playlist();
+ if (!playlist ()) {
+ return;
+ }
- } else {
+ Region::merge_features (results, _user_transients, _position + _transient_user_start - _start);
- TransientDetector::cleanup_transients (_transients, pl->session().frame_rate(), 3.0);
- results = _transients;
+ if (!_onsets.empty ()) {
+ // onsets are invalidated when start or length changes
+ merge_features (results, _onsets, _position);
+ return;
}
- _valid_transients = true;
+ if ((_transient_analysis_start == _transient_analysis_end)
+ || _transient_analysis_start > _start
+ || _transient_analysis_end < _start + _length) {
+ build_transients ();
+ }
- return 0;
+ merge_features (results, _transients, _position + _transient_analysis_start - _start);
}
/** Find areas of `silence' within a region.
*/
AudioIntervalResult
-AudioRegion::find_silence (Sample threshold, framecnt_t min_length, InterThreadInfo& itt) const
+AudioRegion::find_silence (Sample threshold, framecnt_t min_length, framecnt_t fade_length, InterThreadInfo& itt) const
{
framecnt_t const block_size = 64 * 1024;
boost::scoped_array<Sample> loudest (new Sample[block_size]);
boost::scoped_array<Sample> buf (new Sample[block_size]);
+ assert (fade_length >= 0);
+ assert (min_length > 0);
+
framepos_t pos = _start;
- framepos_t const end = _start + _length - 1;
+ framepos_t const end = _start + _length;
AudioIntervalResult silent_periods;
- bool in_silence = false;
- frameoffset_t silence_start = 0;
+ bool in_silence = true;
+ frameoffset_t silence_start = _start;
while (pos < end && !itt.cancel) {
+ framecnt_t cur_samples = 0;
+ framecnt_t const to_read = min (end - pos, block_size);
/* fill `loudest' with the loudest absolute sample at each instant, across all channels */
memset (loudest.get(), 0, sizeof (Sample) * block_size);
+
for (uint32_t n = 0; n < n_channels(); ++n) {
- read_raw_internal (buf.get(), pos, block_size, n);
- for (framecnt_t i = 0; i < block_size; ++i) {
+ cur_samples = read_raw_internal (buf.get(), pos, to_read, n);
+ for (framecnt_t i = 0; i < cur_samples; ++i) {
loudest[i] = max (loudest[i], abs (buf[i]));
}
}
/* now look for silence */
- for (framecnt_t i = 0; i < block_size; ++i) {
+ for (framecnt_t i = 0; i < cur_samples; ++i) {
bool const silence = abs (loudest[i]) < threshold;
if (silence && !in_silence) {
/* non-silence to silence */
in_silence = true;
- silence_start = pos + i;
+ silence_start = pos + i + fade_length;
} else if (!silence && in_silence) {
/* silence to non-silence */
in_silence = false;
- if (pos + i - 1 - silence_start >= min_length) {
- silent_periods.push_back (std::make_pair (silence_start, pos + i - 1));
+ frameoffset_t silence_end = pos + i - 1 - fade_length;
+
+ if (silence_end - silence_start >= min_length) {
+ silent_periods.push_back (std::make_pair (silence_start, silence_end));
}
}
}
- pos += block_size;
- itt.progress = (end-pos)/(double)_length;
+ pos += cur_samples;
+ itt.progress = (end - pos) / (double)_length;
+
+ if (cur_samples == 0) {
+ assert (pos >= end);
+ break;
+ }
}
- if (in_silence && end - 1 - silence_start >= min_length) {
+ if (in_silence && !itt.cancel) {
/* last block was silent, so finish off the last period */
- silent_periods.push_back (std::make_pair (silence_start, end));
+ if (end - 1 - silence_start >= min_length + fade_length) {
+ silent_periods.push_back (std::make_pair (silence_start, end - 1));
+ }
}
itt.done = true;
Evoral::Range<framepos_t>
AudioRegion::body_range () const
{
- return Evoral::Range<framepos_t> (first_frame() + _fade_in->back()->when, last_frame() - _fade_out->back()->when);
-}
-
-void
-AudioRegion::set_fade_in_is_xfade (bool yn)
-{
- _fade_in_is_xfade = yn;
-}
-
-void
-AudioRegion::set_fade_out_is_xfade (bool yn)
-{
- _fade_out_is_xfade = yn;
+ return Evoral::Range<framepos_t> (first_frame() + _fade_in->back()->when + 1, last_frame() - _fade_out->back()->when);
}
boost::shared_ptr<Region>
} else {
rl = pl->regions_at (last_frame());
}
-
+
RegionList::iterator i;
boost::shared_ptr<Region> other;
uint32_t n = 0;
framecnt_t
AudioRegion::verify_xfade_bounds (framecnt_t len, bool start)
{
+ /* this is called from a UI to check on whether a new proposed
+ length for an xfade is legal or not. it returns the legal
+ length corresponding to @a len which may be shorter than or
+ equal to @a len itself.
+ */
+
boost::shared_ptr<Region> other = get_single_other_xfade_region (start);
framecnt_t maxlen;
if (!other) {
- /* zero or > 2 regions here, don't care about len */
- return len;
+ /* zero or > 2 regions here, don't care about len, but
+ it can't be longer than the region itself.
+ */
+ return min (length(), len);
}
/* we overlap a single region. clamp the length of an xfade to
maxlen = last_frame() - other->earliest_possible_position();
}
- return min (maxlen, len);
-
-}
-
-extern "C" {
+ return min (length(), min (maxlen, len));
- int region_read_peaks_from_c (void *arg, uint32_t npeaks, uint32_t start, uint32_t cnt, intptr_t data, uint32_t n_chan, double samples_per_unit)
-{
- return ((AudioRegion *) arg)->read_peaks ((PeakData *) data, (framecnt_t) npeaks, (framepos_t) start, (framecnt_t) cnt, n_chan,samples_per_unit);
-}
-
-uint32_t region_length_from_c (void *arg)
-{
-
- return ((AudioRegion *) arg)->length();
-}
-
-uint32_t sourcefile_length_from_c (void *arg, double zoom_factor)
-{
- return ( (AudioRegion *) arg)->audio_source()->available_peaks (zoom_factor) ;
}
-} /* extern "C" */