android - Android 音频编码失败
问题描述
以下代码尝试在 Android 上使用 AAC 将浮点正弦信号编码为 mp4 文件。但是,它失败了。要么我得到一个扭曲的音调,要么什么都没有。mp4文件的长度也是错误的。
我会很感激一些帮助解决这个问题。谢谢!
public class MyAudioTest {
MediaMuxer mux = null;
int SAMPLING_RATE = 44100;
int BUFFER_SIZE = 4096;
int timeOutMicroSeconds = 10000;
double sampleRateHz = 44100;
double pha = 0.0;
double deltapha = 2.0 * Math.PI * 440.0 / 44100;
double maxpha = 2.0 * Math.PI;
int SIZE_OF_FLOAT = 4;
int NUM_SAMPLES = 441;
int numBlocks = 0;
private byte[] GetFloatArrayAsByteArray(float[] sampleBlock)
{
ByteBuffer buffer = ByteBuffer.allocate(sampleBlock.length * SIZE_OF_FLOAT);
buffer.order(ByteOrder.nativeOrder());
buffer.asFloatBuffer().put(sampleBlock);
return buffer.array();
}
private short NextSineSample() {
float val = (float) Math.sin(pha);
pha = pha + deltapha;
if (pha >= maxpha) pha = pha - maxpha;
return val;
}
public void AudioEncoderTest() {
Log.i("AUDIO", "Starting audio encoder test");
try {
mux = new MediaMuxer("/mnt/sdcard/testing.mp4", MediaMuxer.OutputFormat.MUXER_OUTPUT_MPEG_4);
MediaFormat outputFormat = MediaFormat.createAudioFormat(MediaFormat.MIMETYPE_AUDIO_AAC, SAMPLING_RATE, 1);
outputFormat.setInteger(MediaFormat.KEY_AAC_PROFILE, MediaCodecInfo.CodecProfileLevel.AACObjectLC);
outputFormat.setInteger(MediaFormat.KEY_BIT_RATE, 96000);
outputFormat.setString(MediaFormat.KEY_MIME, MediaFormat.MIMETYPE_AUDIO_AAC);
outputFormat.setInteger(MediaFormat.KEY_CHANNEL_COUNT, 1);
outputFormat.setInteger(MediaFormat.KEY_SAMPLE_RATE, SAMPLING_RATE);
outputFormat.setInteger(MediaFormat.KEY_PCM_ENCODING, AudioFormat.ENCODING_PCM_FLOAT);
MediaCodec codec = MediaCodec.createEncoderByType(MediaFormat.MIMETYPE_AUDIO_AAC);
codec.configure(outputFormat, null, null, MediaCodec.CONFIGURE_FLAG_ENCODE);
codec.start();
ByteBuffer[] codecInputBuffers = codec.getInputBuffers();
ByteBuffer[] codecOutputBuffers = codec.getOutputBuffers();
MediaCodec.BufferInfo outBuffInfo = new MediaCodec.BufferInfo();
byte[] tempBuffer;
float[] sampleBlock = new float[NUM_SAMPLES];
boolean hasMoreData = true;
double presentationTimeUs = 0;
int audioTrackIdx = 0;
int totalBytesRead = 0;
int totalSamplesProcessed = 0;
do {
int inputBufIndex = 0;
while (inputBufIndex != -1 && hasMoreData) {
inputBufIndex = codec.dequeueInputBuffer(timeOutMicroSeconds);
if (inputBufIndex >= 0) {
ByteBuffer dstBuf = codecInputBuffers[inputBufIndex];
dstBuf.clear();
for (int i = 0; i < NUM_SAMPLES; i++) {
sampleBlock[i] = NextSineSample();
}
numBlocks++;
int bytesRead = NUM_SAMPLES * SIZE_OF_FLOAT;
tempBuffer = GetFloatArrayAsByteArray(sampleBlock);
totalSamplesProcessed = totalSamplesProcessed + NUM_SAMPLES;
if (totalSamplesProcessed >= SAMPLING_RATE * 2) {
bytesRead = -1;
}
if (bytesRead == -1) { // -1 implies EOS
hasMoreData = false;
Log.i("AUDIO", "No more input for encoder");
codec.queueInputBuffer(inputBufIndex, 0, 0, (long) presentationTimeUs, MediaCodec.BUFFER_FLAG_END_OF_STREAM);
} else {
totalBytesRead += bytesRead;
dstBuf.put(tempBuffer, 0, bytesRead);
Log.i("AUDIO", "Encoding data. Block: " + numBlocks);
codec.queueInputBuffer(inputBufIndex, 0, bytesRead, (long) presentationTimeUs, 0);
presentationTimeUs = 1000000l * (totalBytesRead / SIZE_OF_FLOAT) / SAMPLING_RATE;
Log.i("AUDIOTIME", "Block: " + numBlocks + " | Presentation time: " + presentationTimeUs);
}
}
}
// Drain audio
int outputBufIndex = 0;
while (outputBufIndex != MediaCodec.INFO_TRY_AGAIN_LATER)
{
outputBufIndex = codec.dequeueOutputBuffer(outBuffInfo, timeOutMicroSeconds);
if (outputBufIndex >= 0)
{
ByteBuffer encodedData = codecOutputBuffers[outputBufIndex];
encodedData.position(outBuffInfo.offset);
encodedData.limit(outBuffInfo.offset + outBuffInfo.size);
if ((outBuffInfo.flags & MediaCodec.BUFFER_FLAG_CODEC_CONFIG) != 0 && outBuffInfo.size != 0) {
codec.releaseOutputBuffer(outputBufIndex, false);
} else {
Log.i("AUDIO", "Writing encoded data to mux");
mux.writeSampleData(audioTrackIdx, codecOutputBuffers[outputBufIndex], outBuffInfo);
codec.releaseOutputBuffer(outputBufIndex, false);
}
} else if (outputBufIndex == MediaCodec.INFO_OUTPUT_FORMAT_CHANGED) {
outputFormat = codec.getOutputFormat();
Log.v("AUDIO", "Output format changed - " + outputFormat);
audioTrackIdx = mux.addTrack(outputFormat);
Log.i("AUDIO", "Start mux");
mux.start();
} else if (outputBufIndex == MediaCodec.INFO_OUTPUT_BUFFERS_CHANGED) {
Log.e("AUDIO", "Output buffers changed during encode!");
} else if (outputBufIndex != MediaCodec.INFO_TRY_AGAIN_LATER) {
Log.e("AUDIO", "Unknown return code from dequeueOutputBuffer - " + outputBufIndex);
}
}
} while (outBuffInfo.flags != MediaCodec.BUFFER_FLAG_END_OF_STREAM);
mux.stop();
mux.release();
} catch (FileNotFoundException e) {
e.printStackTrace();
} catch (IOException e) {
e.printStackTrace();
}
Log.i("AUDIO", "Finished audio encoder test. Blocks: " + numBlocks);
}
}
解决方案
也许字节序很重要。要在 MediaCodec 的输入缓冲区中排队的字节数组必须存储所有 PCM 样本值,native byte order
如下所示。
//private byte[] GetFloatArrayAsByteArray(float[] samples) throws IOException {
// ByteArrayOutputStream bas = new ByteArrayOutputStream();
// DataOutputStream ds = new DataOutputStream(bas);
// for (float f : samples)
// ds.writeFloat(f);
// return bas.toByteArray();
//}
private byte[] GetFloatArrayAsByteArray(float[] sampleBlock)
{
ByteBuffer buffer = ByteBuffer.allocate(sampleBlock.length * SIZE_OF_FLOAT);
buffer.order(ByteOrder.nativeOrder());
buffer.asFloatBuffer().put(sampleBlock);
return buffer.array();
}
由于 PCM_FLOAT 样本值的有效范围介于 -1.0F 和 1.0F 之间,因此您需要对这些值进行标准化。
float[] sampleBlock = new float[NUM_SAMPLES];
:
for (int i = 0; i < NUM_SAMPLES; i++) {
sampleBlock[i] = NextSineSample() / 32767.0F;
}
目前尚不清楚 mp4 文件的长度是如何出错的。如果它能正常播放 2 秒的音调,那就没问题了。
推荐阅读
- regex - OADC SMS 号码的正则表达式
- angular - 基于 URL 的 Angular 主题
- php - Laravel:类stdClass的对象无法转换为字符串
- python - 网页抓取,与空间文本在同一分割
- javascript - 如何有效地隐藏链接中的 Href?
- highcharts - Highmap colorAxis 过滤器
- hibernate - 为什么 Propagation.NOT_SUPPORTED 会更改生成的 select 语句?
- xml - 如何在 XSLT 中创建一个增加 1 个 foreach 循环的计数器,我已经嵌套了 foreach 循环
- c# - 如何使用 ASP.NET Core 中的模型验证来验证所需的查询字符串字段
- html - HTML