This source file includes following definitions.
- m_imagData
- m_inverseContext
- m_imagData
- initialize
- cleanup
- doFFT
- doInverseFFT
- realData
- imagData
- contextForSize
#include "config.h"
#if ENABLE(WEB_AUDIO)
#if OS(ANDROID) && USE(WEBAUDIO_OPENMAX_DL_FFT)
#include "platform/audio/FFTFrame.h"
#include "platform/audio/AudioArray.h"
#include "wtf/MathExtras.h"
#include <dl/sp/api/armSP.h>
#include <dl/sp/api/omxSP.h>
namespace WebCore {
#if !ASSERT_DISABLED
const unsigned kMaxFFTPow2Size = 15;
#endif
FFTFrame::FFTFrame(unsigned fftSize)
: m_FFTSize(fftSize)
, m_log2FFTSize(static_cast<unsigned>(log2(fftSize)))
, m_forwardContext(0)
, m_inverseContext(0)
, m_complexData(fftSize)
, m_realData(fftSize / 2)
, m_imagData(fftSize / 2)
{
ASSERT(1UL << m_log2FFTSize == m_FFTSize);
m_forwardContext = contextForSize(m_log2FFTSize);
m_inverseContext = contextForSize(m_log2FFTSize);
}
FFTFrame::FFTFrame()
: m_FFTSize(0)
, m_log2FFTSize(0)
, m_forwardContext(0)
, m_inverseContext(0)
{
}
FFTFrame::FFTFrame(const FFTFrame& frame)
: m_FFTSize(frame.m_FFTSize)
, m_log2FFTSize(frame.m_log2FFTSize)
, m_forwardContext(0)
, m_inverseContext(0)
, m_complexData(frame.m_FFTSize)
, m_realData(frame.m_FFTSize / 2)
, m_imagData(frame.m_FFTSize / 2)
{
m_forwardContext = contextForSize(m_log2FFTSize);
m_inverseContext = contextForSize(m_log2FFTSize);
unsigned nbytes = sizeof(float) * (m_FFTSize / 2);
memcpy(realData(), frame.realData(), nbytes);
memcpy(imagData(), frame.imagData(), nbytes);
}
void FFTFrame::initialize()
{
}
void FFTFrame::cleanup()
{
}
FFTFrame::~FFTFrame()
{
if (m_forwardContext)
free(m_forwardContext);
if (m_inverseContext)
free(m_inverseContext);
}
void FFTFrame::doFFT(const float* data)
{
ASSERT(m_forwardContext);
if (m_forwardContext) {
AudioFloatArray complexFFT(m_FFTSize + 2);
omxSP_FFTFwd_RToCCS_F32(data, complexFFT.data(), m_forwardContext);
unsigned len = m_FFTSize / 2;
const float* c = complexFFT.data();
float* real = m_realData.data();
float* imag = m_imagData.data();
for (unsigned k = 1; k < len; ++k) {
int index = 2 * k;
real[k] = c[index];
imag[k] = c[index + 1];
}
real[0] = c[0];
imag[0] = c[m_FFTSize];
}
}
void FFTFrame::doInverseFFT(float* data)
{
ASSERT(m_inverseContext);
if (m_inverseContext) {
AudioFloatArray fftDataArray(m_FFTSize + 2);
unsigned len = m_FFTSize / 2;
float* fftData = fftDataArray.data();
const float* real = m_realData.data();
const float* imag = m_imagData.data();
for (unsigned k = 1; k < len; ++k) {
int index = 2 * k;
fftData[index] = real[k];
fftData[index + 1] = imag[k];
}
fftData[0] = real[0];
fftData[1] = 0;
fftData[m_FFTSize] = imag[0];
fftData[m_FFTSize + 1] = 0;
omxSP_FFTInv_CCSToR_F32(fftData, data, m_inverseContext);
}
}
float* FFTFrame::realData() const
{
return const_cast<float*>(m_realData.data());
}
float* FFTFrame::imagData() const
{
return const_cast<float*>(m_imagData.data());
}
OMXFFTSpec_R_F32* FFTFrame::contextForSize(unsigned log2FFTSize)
{
ASSERT(log2FFTSize);
ASSERT(log2FFTSize <= kMaxFFTPow2Size);
int bufSize;
OMXResult status = omxSP_FFTGetBufSize_R_F32(log2FFTSize, &bufSize);
if (status == OMX_Sts_NoErr) {
OMXFFTSpec_R_F32* context = static_cast<OMXFFTSpec_R_F32*>(malloc(bufSize));
omxSP_FFTInit_R_F32(context, log2FFTSize);
return context;
}
return 0;
}
}
#endif
#endif