/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package android.media; import android.annotation.IntDef; import android.annotation.NonNull; import android.annotation.Nullable; import android.media.MediaCodec; import android.media.MediaCodec.BufferInfo; import dalvik.system.CloseGuard; import java.io.FileDescriptor; import java.io.IOException; import java.io.RandomAccessFile; import java.lang.annotation.Retention; import java.lang.annotation.RetentionPolicy; import java.nio.ByteBuffer; import java.util.Map; /** * MediaMuxer facilitates muxing elementary streams. Currently MediaMuxer supports MP4, Webm * and 3GP file as the output. It also supports muxing B-frames in MP4 since Android Nougat. *
* It is generally used like this: * *
* MediaMuxer muxer = new MediaMuxer("temp.mp4", OutputFormat.MUXER_OUTPUT_MPEG_4); * // More often, the MediaFormat will be retrieved from MediaCodec.getOutputFormat() * // or MediaExtractor.getTrackFormat(). * MediaFormat audioFormat = new MediaFormat(...); * MediaFormat videoFormat = new MediaFormat(...); * int audioTrackIndex = muxer.addTrack(audioFormat); * int videoTrackIndex = muxer.addTrack(videoFormat); * ByteBuffer inputBuffer = ByteBuffer.allocate(bufferSize); * boolean finished = false; * BufferInfo bufferInfo = new BufferInfo(); * * muxer.start(); * while(!finished) { * // getInputBuffer() will fill the inputBuffer with one frame of encoded * // sample from either MediaCodec or MediaExtractor, set isAudioSample to * // true when the sample is audio data, set up all the fields of bufferInfo, * // and return true if there are no more samples. * finished = getInputBuffer(inputBuffer, isAudioSample, bufferInfo); * if (!finished) { * int currentTrackIndex = isAudioSample ? audioTrackIndex : videoTrackIndex; * muxer.writeSampleData(currentTrackIndex, inputBuffer, bufferInfo); * } * }; * muxer.stop(); * muxer.release(); **
Per-frame metadata is useful in carrying extra information that correlated with video or audio to facilitate offline processing, e.g. gyro signals from the sensor could help video stabilization when doing offline processing. Metaadata track is only supported in MP4 container. When adding a new metadata track, track's mime format must start with prefix "application/", e.g. "applicaton/gyro". Metadata's format/layout will be defined by the application. Writing metadata is nearly the same as writing video/audio data except that the data will not be from mediacodec. Application just needs to pass the bytebuffer that contains the metadata and also the associated timestamp to the {@link #writeSampleData} api. The timestamp must be in the same time base as video and audio. The generated MP4 file uses TextMetaDataSampleEntry defined in section 12.3.3.2 of the ISOBMFF to signal the metadata's mime format. When using{@link android.media.MediaExtractor} to extract the file with metadata track, the mime format of the metadata will be extracted into {@link android.media.MediaFormat}.
MediaMuxer muxer = new MediaMuxer("temp.mp4", OutputFormat.MUXER_OUTPUT_MPEG_4); // SetUp Video/Audio Tracks. MediaFormat audioFormat = new MediaFormat(...); MediaFormat videoFormat = new MediaFormat(...); int audioTrackIndex = muxer.addTrack(audioFormat); int videoTrackIndex = muxer.addTrack(videoFormat); // Setup Metadata Track MediaFormat metadataFormat = new MediaFormat(...); metadataFormat.setString(KEY_MIME, "application/gyro"); int metadataTrackIndex = muxer.addTrack(metadataFormat); muxer.start(); while(..) { // Allocate bytebuffer and write gyro data(x,y,z) into it. ByteBuffer metaData = ByteBuffer.allocate(bufferSize); metaData.putFloat(x); metaData.putFloat(y); metaData.putFloat(z); BufferInfo metaInfo = new BufferInfo(); // Associate this metadata with the video frame by setting // the same timestamp as the video frame. metaInfo.presentationTimeUs = currentVideoTrackTimeUs; metaInfo.offset = 0; metaInfo.flags = 0; metaInfo.size = bufferSize; muxer.writeSampleData(metadataTrackIndex, metaData, metaInfo); }; muxer.stop(); muxer.release(); }
The following table summarizes the feature support in different API version and containers. For API version numbers, see {@link android.os.Build.VERSION_CODES}.
Symbol | Meaning |
---|---|
● | Supported |
○ | Not supported |
▧ | Supported in MP4/WebM/3GP |
⁕ | Only Supported in MP4 |
Feature | SDK Version | |||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
18 | 19 | 20 | 21 | 22 | 23 | 24 | 25 | 26+ | ||||||||||||||||
MP4 container | ● | ● | ● | ● | ● | ● | ● | ● | ● | WebM container | ○ | ○ | ○ | ● | ● | ● | ● | ● | ● | 3GP container | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ● | Muxing B-Frames(bi-directional predicted frames) | ○ | ○ | ○ | ○ | ○ | ○ | ⁕ | ⁕ | ⁕ | Muxing Single Video/Audio Track | ▧ | ▧ | ▧ | ▧ | ▧ | ▧ | ▧ | ▧ | ▧ | Muxing Multiple Video/Audio Tracks | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ⁕ | Muxing Metadata Tracks | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ○ | ⁕ |
This method should be called before {@link #start}. Calling this * method will not rotate the video frame when muxer is generating the file, * but add a composition matrix containing the rotation angle in the output * video if the output format is * {@link OutputFormat#MUXER_OUTPUT_MPEG_4} so that a video player can * choose the proper orientation for playback. Note that some video players * may choose to ignore the composition matrix in a video during playback. * By default, the rotation degree is 0.
* @param degrees the angle to be rotated clockwise in degrees. * The supported angles are 0, 90, 180, and 270 degrees. * @throws IllegalArgumentException if degree is not supported. * @throws IllegalStateException If this method is called after {@link #start}. */ public void setOrientationHint(int degrees) { if (degrees != 0 && degrees != 90 && degrees != 180 && degrees != 270) { throw new IllegalArgumentException("Unsupported angle: " + degrees); } if (mState == MUXER_STATE_INITIALIZED) { nativeSetOrientationHint(mNativeObject, degrees); } else { throw new IllegalStateException("Can't set rotation degrees due" + " to wrong state."); } } /** * Set and store the geodata (latitude and longitude) in the output file. * This method should be called before {@link #start}. The geodata is stored * in udta box if the output format is * {@link OutputFormat#MUXER_OUTPUT_MPEG_4}, and is ignored for other output * formats. The geodata is stored according to ISO-6709 standard. * * @param latitude Latitude in degrees. Its value must be in the range [-90, * 90]. * @param longitude Longitude in degrees. Its value must be in the range * [-180, 180]. * @throws IllegalArgumentException If the given latitude or longitude is out * of range. * @throws IllegalStateException If this method is called after {@link #start}. */ public void setLocation(float latitude, float longitude) { int latitudex10000 = (int) (latitude * 10000 + 0.5); int longitudex10000 = (int) (longitude * 10000 + 0.5); if (latitudex10000 > 900000 || latitudex10000 < -900000) { String msg = "Latitude: " + latitude + " out of range."; throw new IllegalArgumentException(msg); } if (longitudex10000 > 1800000 || longitudex10000 < -1800000) { String msg = "Longitude: " + longitude + " out of range"; throw new IllegalArgumentException(msg); } if (mState == MUXER_STATE_INITIALIZED && mNativeObject != 0) { nativeSetLocation(mNativeObject, latitudex10000, longitudex10000); } else { throw new IllegalStateException("Can't set location due to wrong state."); } } /** * Starts the muxer. *Make sure this is called after {@link #addTrack} and before * {@link #writeSampleData}.
* @throws IllegalStateException If this method is called after {@link #start} * or Muxer is released */ public void start() { if (mNativeObject == 0) { throw new IllegalStateException("Muxer has been released!"); } if (mState == MUXER_STATE_INITIALIZED) { nativeStart(mNativeObject); mState = MUXER_STATE_STARTED; } else { throw new IllegalStateException("Can't start due to wrong state."); } } /** * Stops the muxer. *Once the muxer stops, it can not be restarted.
* @throws IllegalStateException if muxer is in the wrong state. */ public void stop() { if (mState == MUXER_STATE_STARTED) { nativeStop(mNativeObject); mState = MUXER_STATE_STOPPED; } else { throw new IllegalStateException("Can't stop due to wrong state."); } } @Override protected void finalize() throws Throwable { try { if (mCloseGuard != null) { mCloseGuard.warnIfOpen(); } if (mNativeObject != 0) { nativeRelease(mNativeObject); mNativeObject = 0; } } finally { super.finalize(); } } /** * Adds a track with the specified format. ** The following table summarizes support for specific format keys across android releases. * Keys marked with '+:' are required. * *
OS Version(s) | *{@code MediaFormat} keys used for * | ||
---|---|---|---|
All Tracks | *Audio Tracks | *Video Tracks | *|
{@link android.os.Build.VERSION_CODES#JELLY_BEAN_MR2} | *+: {@link MediaFormat#KEY_MIME} | *+: {@link MediaFormat#KEY_SAMPLE_RATE}, * +: {@link MediaFormat#KEY_CHANNEL_COUNT}, * +: codec-specific dataAAC |
* +: {@link MediaFormat#KEY_WIDTH}, * +: {@link MediaFormat#KEY_HEIGHT}, * no {@code KEY_ROTATION}, * use {@link #setOrientationHint setOrientationHint()}.mp4, * +: codec-specific dataAVC, MPEG4 |
*
{@link android.os.Build.VERSION_CODES#KITKAT} | *|||
{@link android.os.Build.VERSION_CODES#KITKAT_WATCH} | *|||
{@link android.os.Build.VERSION_CODES#LOLLIPOP} | *as above, plus * +: codec-specific dataVorbis & .webm |
* ||
{@link android.os.Build.VERSION_CODES#LOLLIPOP_MR1} | *|||
{@link android.os.Build.VERSION_CODES#M} | *as above, plus * {@link MediaFormat#KEY_BIT_RATE}AAC |
* ||
{@link android.os.Build.VERSION_CODES#N} | *as above, plus * * {@link MediaFormat#KEY_BIT_RATE}MPEG4, * {@link MediaFormat#KEY_HDR_STATIC_INFO}#, .webm, * {@link MediaFormat#KEY_COLOR_STANDARD}#, * {@link MediaFormat#KEY_COLOR_TRANSFER}#, * {@link MediaFormat#KEY_COLOR_RANGE}#, * +: codec-specific dataHEVC, * codec-specific dataVP9 |
* ||
* Notes: |
* |||
* Note that the codec-specific data for the track must be specified using * this method. Furthermore, codec-specific data must not be passed/specified via the * {@link #writeSampleData writeSampleData()} call. * |
*
* The following table summarizes codec support for containers across android releases: * *
OS Version(s) | *Codec support * | ||
---|---|---|---|
{@linkplain OutputFormat#MUXER_OUTPUT_MPEG_4 MP4} | *{@linkplain OutputFormat#MUXER_OUTPUT_WEBM WEBM} | *||
{@link android.os.Build.VERSION_CODES#JELLY_BEAN_MR2} | *{@link MediaFormat#MIMETYPE_AUDIO_AAC AAC}, * {@link MediaFormat#MIMETYPE_AUDIO_AMR_NB NB-AMR}, * {@link MediaFormat#MIMETYPE_AUDIO_AMR_WB WB-AMR}, * {@link MediaFormat#MIMETYPE_VIDEO_H263 H.263}, * {@link MediaFormat#MIMETYPE_VIDEO_MPEG4 MPEG-4}, * {@link MediaFormat#MIMETYPE_VIDEO_AVC AVC} (H.264) |
* Not supported | *|
{@link android.os.Build.VERSION_CODES#KITKAT} | *|||
{@link android.os.Build.VERSION_CODES#KITKAT_WATCH} | *|||
{@link android.os.Build.VERSION_CODES#LOLLIPOP} | *{@link MediaFormat#MIMETYPE_AUDIO_VORBIS Vorbis}, * {@link MediaFormat#MIMETYPE_VIDEO_VP8 VP8} |
* ||
{@link android.os.Build.VERSION_CODES#LOLLIPOP_MR1} | *|||
{@link android.os.Build.VERSION_CODES#M} | *|||
{@link android.os.Build.VERSION_CODES#N} | *as above, plus * {@link MediaFormat#MIMETYPE_VIDEO_HEVC HEVC} (H.265) |
* as above, plus * {@link MediaFormat#MIMETYPE_VIDEO_VP9 VP9} |
*
The application needs to make sure that the samples are written into * the right tracks. Also, it needs to make sure the samples for each track * are written in chronological order (e.g. in the order they are provided * by the encoder.)
* @param byteBuf The encoded sample. * @param trackIndex The track index for this sample. * @param bufferInfo The buffer information related to this sample. * @throws IllegalArgumentException if trackIndex, byteBuf or bufferInfo is invalid. * @throws IllegalStateException if muxer is in wrong state. * MediaMuxer uses the flags provided in {@link MediaCodec.BufferInfo}, * to signal sync frames. */ public void writeSampleData(int trackIndex, @NonNull ByteBuffer byteBuf, @NonNull BufferInfo bufferInfo) { if (trackIndex < 0 || trackIndex > mLastTrackIndex) { throw new IllegalArgumentException("trackIndex is invalid"); } if (byteBuf == null) { throw new IllegalArgumentException("byteBuffer must not be null"); } if (bufferInfo == null) { throw new IllegalArgumentException("bufferInfo must not be null"); } if (bufferInfo.size < 0 || bufferInfo.offset < 0 || (bufferInfo.offset + bufferInfo.size) > byteBuf.capacity() || bufferInfo.presentationTimeUs < 0) { throw new IllegalArgumentException("bufferInfo must specify a" + " valid buffer offset, size and presentation time"); } if (mNativeObject == 0) { throw new IllegalStateException("Muxer has been released!"); } if (mState != MUXER_STATE_STARTED) { throw new IllegalStateException("Can't write, muxer is not started"); } nativeWriteSampleData(mNativeObject, trackIndex, byteBuf, bufferInfo.offset, bufferInfo.size, bufferInfo.presentationTimeUs, bufferInfo.flags); } /** * Make sure you call this when you're done to free up any resources * instead of relying on the garbage collector to do this for you at * some point in the future. */ public void release() { if (mState == MUXER_STATE_STARTED) { stop(); } if (mNativeObject != 0) { nativeRelease(mNativeObject); mNativeObject = 0; mCloseGuard.close(); } mState = MUXER_STATE_UNINITIALIZED; } }