java.lang.Object
↳androidx.media3.extractor.DtsUtil
Gradle dependencies
compile group: 'androidx.media3', name: 'media3-extractor', version: '1.5.0-alpha01'
- groupId: androidx.media3
- artifactId: media3-extractor
- version: 1.5.0-alpha01
Artifact androidx.media3:media3-extractor:1.5.0-alpha01 it located at Google repository (https://maven.google.com/)
Overview
Utility methods for parsing DTS frames.
Summary
Methods |
---|
public static int | getDtsFrameSize(byte[] data[])
Returns the size in bytes of the given DTS Core frame. |
public static int | getFrameType(int word)
Returns the DtsUtil.FrameType if word is a DTS sync word, otherwise DtsUtil.FRAME_TYPE_UNKNOWN. |
public static int | parseDtsAudioSampleCount(byte[] data[])
Returns the number of audio samples represented by the given DTS Core frame. |
public static int | parseDtsAudioSampleCount(java.nio.ByteBuffer buffer)
Like DtsUtil.parseDtsAudioSampleCount(byte[]) but reads from a java.nio.ByteBuffer . |
public static Format | parseDtsFormat(byte[] frame[], java.lang.String trackId, java.lang.String language, int roleFlags, DrmInitData drmInitData)
Returns the DTS format given data containing the DTS Core frame according to ETSI TS
102 114 V1.6.1 (2019-08) subsections 5.3/5.4. |
public static DtsUtil.DtsHeader | parseDtsHdHeader(byte[] header[])
Parses the DtsUtil.DtsHeader data from the extension substream header of a DTS-HD frame
according to ETSI TS 102 114 V1.6.1 (2019-08), Section 7.5.2. |
public static int | parseDtsHdHeaderSize(byte[] headerPrefix[])
Returns the size of the extension substream header in a DTS-HD frame according to ETSI TS 102
114 V1.6.1 (2019-08), Section 7.5.2. |
public static DtsUtil.DtsHeader | parseDtsUhdHeader(byte[] header[], java.util.concurrent.atomic.AtomicInteger uhdAudioChunkId)
Parses the DtsUtil.DtsHeader data from the headers of a DTS-UHD(Profile 2) frame according to
ETSI TS 103 491 V1.2.1 (2019-05), Section 6.4.3. |
public static int | parseDtsUhdHeaderSize(byte[] headerPrefix[])
Returns the size of frame header in a DTS-UHD(Profile 2) frame according to ETSI TS 103 491
V1.2.1 (2019-05), Section 6.4.3. |
from java.lang.Object | clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait |
Fields
public static final int
FRAME_TYPE_UNKNOWNRepresents a DTS frame for which type is unknown.
public static final int
FRAME_TYPE_CORERepresents a DTS core frame.
public static final int
FRAME_TYPE_EXTENSION_SUBSTREAMRepresents a DTS extension substream frame.
public static final int
FRAME_TYPE_UHD_SYNCRepresents a DTS UHD sync frame.
public static final int
FRAME_TYPE_UHD_NON_SYNCRepresents a DTS UHD non-sync frame.
public static final int
DTS_MAX_RATE_BYTES_PER_SECONDMaximum rate for a DTS audio stream, in bytes per second.
DTS allows an 'open' bitrate, but we assume the maximum listed value: 1536 kbit/s.
public static final int
DTS_HD_MAX_RATE_BYTES_PER_SECONDMaximum rate for a DTS-HD audio stream, in bytes per second.
public static final int
DTS_EXPRESS_MAX_RATE_BITS_PER_SECONDMaximum bit-rate for a DTS Express audio stream, in bits per second.
Methods
public static int
getFrameType(int word)
Returns the DtsUtil.FrameType if word is a DTS sync word, otherwise DtsUtil.FRAME_TYPE_UNKNOWN.
public static
Format parseDtsFormat(byte[] frame[], java.lang.String trackId, java.lang.String language, int roleFlags,
DrmInitData drmInitData)
Returns the DTS format given data containing the DTS Core frame according to ETSI TS
102 114 V1.6.1 (2019-08) subsections 5.3/5.4.
Parameters:
frame: The DTS Core frame to parse.
trackId: The track identifier to set on the format.
language: The language to set on the format.
roleFlags: The role flags to set on the format.
drmInitData: DrmInitData to be included in the format.
Returns:
The DTS format parsed from data in the header.
public static int
parseDtsAudioSampleCount(byte[] data[])
Returns the number of audio samples represented by the given DTS Core frame.
Parameters:
data: The frame to parse.
Returns:
The number of audio samples represented by the frame.
public static int
parseDtsAudioSampleCount(java.nio.ByteBuffer buffer)
Like DtsUtil.parseDtsAudioSampleCount(byte[]) but reads from a java.nio.ByteBuffer
. The
buffer's position is not modified.
Parameters:
buffer: The java.nio.ByteBuffer
from which to read.
Returns:
The number of audio samples represented by the syncframe.
public static int
getDtsFrameSize(byte[] data[])
Returns the size in bytes of the given DTS Core frame.
Parameters:
data: The frame to parse.
Returns:
The frame's size in bytes.
Parses the DtsUtil.DtsHeader data from the extension substream header of a DTS-HD frame
according to ETSI TS 102 114 V1.6.1 (2019-08), Section 7.5.2.
Parameters:
header: The DTS-HD extension substream header to parse.
Returns:
The DtsUtil.DtsHeader data extracted from the header.
public static int
parseDtsHdHeaderSize(byte[] headerPrefix[])
Returns the size of the extension substream header in a DTS-HD frame according to ETSI TS 102
114 V1.6.1 (2019-08), Section 7.5.2.
Parameters:
headerPrefix: A byte array containing at least the first 55 bits of a DTS-HD frame.
Returns:
Size of the DTS-HD frame header in bytes.
public static
DtsUtil.DtsHeader parseDtsUhdHeader(byte[] header[], java.util.concurrent.atomic.AtomicInteger uhdAudioChunkId)
Parses the DtsUtil.DtsHeader data from the headers of a DTS-UHD(Profile 2) frame according to
ETSI TS 103 491 V1.2.1 (2019-05), Section 6.4.3.
Parameters:
header: The DTS-UHD header to parse.
uhdAudioChunkId: An java.util.concurrent.atomic.AtomicInteger
containing the last read UHD audio chunk ID
from a synchronized frame, or zero if unset. This parameter is both an input and output
parameter. In synchronized frames, the input value is not used; instead, the parameter is
set to the current UHD audio chunk ID, which becomes the output value. For non-synchronized
frames, it is used without any modification.
Returns:
The DtsUtil.DtsHeader data extracted from the header.
public static int
parseDtsUhdHeaderSize(byte[] headerPrefix[])
Returns the size of frame header in a DTS-UHD(Profile 2) frame according to ETSI TS 103 491
V1.2.1 (2019-05), Section 6.4.3.
Parameters:
headerPrefix: A byte array containing at least the first 47 bits of a DTS-UHD frame.
Returns:
Size of the DTS-UHD frame header in bytes.
Source
/*
* Copyright (C) 2016 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package androidx.media3.extractor;
import static java.lang.annotation.ElementType.TYPE_USE;
import static java.lang.annotation.RetentionPolicy.SOURCE;
import androidx.annotation.IntDef;
import androidx.annotation.Nullable;
import androidx.annotation.StringDef;
import androidx.media3.common.C;
import androidx.media3.common.DrmInitData;
import androidx.media3.common.Format;
import androidx.media3.common.MimeTypes;
import androidx.media3.common.ParserException;
import androidx.media3.common.util.ParsableBitArray;
import androidx.media3.common.util.UnstableApi;
import androidx.media3.common.util.Util;
import java.lang.annotation.Documented;
import java.lang.annotation.Retention;
import java.lang.annotation.Target;
import java.nio.ByteBuffer;
import java.util.Arrays;
import java.util.concurrent.atomic.AtomicInteger;
/** Utility methods for parsing DTS frames. */
@UnstableApi
public final class DtsUtil {
/** Information parsed from a DTS frame header. */
public static final class DtsHeader {
/** The mime type of the DTS bitstream. */
public final @DtsAudioMimeType String mimeType;
/** The audio sampling rate in Hertz, or {@link C#RATE_UNSET_INT} if unknown. */
public final int sampleRate;
/** The number of channels, or {@link C#LENGTH_UNSET} if unknown. */
public final int channelCount;
/** The size of the DTS frame (compressed), in bytes. */
public final int frameSize;
/** The duration of the DTS frame in microseconds, or {@link C#TIME_UNSET} if unknown. */
public final long frameDurationUs;
/** The bitrate of compressed stream. */
public final int bitrate;
private DtsHeader(
String mimeType,
int channelCount,
int sampleRate,
int frameSize,
long frameDurationUs,
int bitrate) {
this.mimeType = mimeType;
this.channelCount = channelCount;
this.sampleRate = sampleRate;
this.frameSize = frameSize;
this.frameDurationUs = frameDurationUs;
this.bitrate = bitrate;
}
}
/**
* The possible MIME types for DTS that can be used.
*
* <p>One of:
*
* <ul>
* <li>{@link MimeTypes#AUDIO_DTS}
* <li>{@link MimeTypes#AUDIO_DTS_EXPRESS}
* <li>{@link MimeTypes#AUDIO_DTS_X}
* </ul>
*/
@Documented
@Retention(SOURCE)
@Target(TYPE_USE)
@StringDef({MimeTypes.AUDIO_DTS, MimeTypes.AUDIO_DTS_EXPRESS, MimeTypes.AUDIO_DTS_X})
public @interface DtsAudioMimeType {}
/**
* Frame types for a DTS stream.
*
* <p>One of:
*
* <ul>
* <li>{@link #FRAME_TYPE_UNKNOWN}
* <li>{@link #FRAME_TYPE_CORE}
* <li>{@link #FRAME_TYPE_EXTENSION_SUBSTREAM}
* <li>{@link #FRAME_TYPE_UHD_SYNC}
* <li>{@link #FRAME_TYPE_UHD_NON_SYNC}
* </ul>
*/
@Documented
@Retention(SOURCE)
@Target(TYPE_USE)
@IntDef({
FRAME_TYPE_UNKNOWN,
FRAME_TYPE_CORE,
FRAME_TYPE_EXTENSION_SUBSTREAM,
FRAME_TYPE_UHD_SYNC,
FRAME_TYPE_UHD_NON_SYNC
})
public @interface FrameType {}
/** Represents a DTS frame for which type is unknown. */
public static final int FRAME_TYPE_UNKNOWN = 0;
/** Represents a DTS core frame. */
public static final int FRAME_TYPE_CORE = 1;
/** Represents a DTS extension substream frame. */
public static final int FRAME_TYPE_EXTENSION_SUBSTREAM = 2;
/** Represents a DTS UHD sync frame. */
public static final int FRAME_TYPE_UHD_SYNC = 3;
/** Represents a DTS UHD non-sync frame. */
public static final int FRAME_TYPE_UHD_NON_SYNC = 4;
/**
* Maximum rate for a DTS audio stream, in bytes per second.
*
* <p>DTS allows an 'open' bitrate, but we assume the maximum listed value: 1536 kbit/s.
*/
public static final int DTS_MAX_RATE_BYTES_PER_SECOND = 1536 * 1000 / 8;
/** Maximum rate for a DTS-HD audio stream, in bytes per second. */
public static final int DTS_HD_MAX_RATE_BYTES_PER_SECOND = 18000 * 1000 / 8;
/** Maximum bit-rate for a DTS Express audio stream, in bits per second. */
public static final int DTS_EXPRESS_MAX_RATE_BITS_PER_SECOND = 768000;
/**
* DTS Core Syncword (in different Endianness). See ETSI TS 102 114 V1.6.1 (2019-08), Section 5.3.
*/
private static final int SYNC_VALUE_BE = 0x7FFE8001;
private static final int SYNC_VALUE_14B_BE = 0x1FFFE800;
private static final int SYNC_VALUE_LE = 0xFE7F0180;
private static final int SYNC_VALUE_14B_LE = 0xFF1F00E8;
/**
* DTS Extension Substream Syncword (in different Endianness). See ETSI TS 102 114 (V1.6.1)
* Section 7.4.1.
*/
private static final int SYNC_VALUE_EXTSS_BE = 0x64582025;
private static final int SYNC_VALUE_EXTSS_LE = 0x25205864;
/**
* DTS UHD FTOC Sync words (in different Endianness). See ETSI TS 103 491 (V1.2.1) Section
* 6.4.4.1.
*/
private static final int SYNC_VALUE_UHD_FTOC_SYNC_BE = 0x40411BF2;
private static final int SYNC_VALUE_UHD_FTOC_SYNC_LE = 0xF21B4140;
private static final int SYNC_VALUE_UHD_FTOC_NONSYNC_BE = 0x71C442E8;
private static final int SYNC_VALUE_UHD_FTOC_NONSYNC_LE = 0xE842C471;
private static final byte FIRST_BYTE_BE = (byte) (SYNC_VALUE_BE >>> 24);
private static final byte FIRST_BYTE_14B_BE = (byte) (SYNC_VALUE_14B_BE >>> 24);
private static final byte FIRST_BYTE_LE = (byte) (SYNC_VALUE_LE >>> 24);
private static final byte FIRST_BYTE_14B_LE = (byte) (SYNC_VALUE_14B_LE >>> 24);
private static final byte FIRST_BYTE_EXTSS_BE = (byte) (SYNC_VALUE_EXTSS_BE >>> 24);
private static final byte FIRST_BYTE_EXTSS_LE = (byte) (SYNC_VALUE_EXTSS_LE >>> 24);
private static final byte FIRST_BYTE_UHD_FTOC_SYNC_BE =
(byte) (SYNC_VALUE_UHD_FTOC_SYNC_BE >>> 24);
private static final byte FIRST_BYTE_UHD_FTOC_SYNC_LE =
(byte) (SYNC_VALUE_UHD_FTOC_SYNC_LE >>> 24);
private static final byte FIRST_BYTE_UHD_FTOC_NONSYNC_BE =
(byte) (SYNC_VALUE_UHD_FTOC_NONSYNC_BE >>> 24);
private static final byte FIRST_BYTE_UHD_FTOC_NONSYNC_LE =
(byte) (SYNC_VALUE_UHD_FTOC_NONSYNC_LE >>> 24);
/** Maps AMODE to the number of channels. See ETSI TS 102 114 table 5-4. */
private static final int[] CHANNELS_BY_AMODE =
new int[] {1, 2, 2, 2, 2, 3, 3, 4, 4, 5, 6, 6, 6, 7, 8, 8};
/** Maps SFREQ to the sampling frequency in Hz. See ETSI TS 102 114 table 5-5. */
private static final int[] SAMPLE_RATE_BY_SFREQ =
new int[] {
-1, 8_000, 16_000, 32_000, -1, -1, 11_025, 22_050, 44_100, -1, -1, 12_000, 24_000, 48_000,
-1, -1
};
/** Maps RATE to 2 * bitrate in kbit/s. See ETSI TS 102 114 table 5-7. */
private static final int[] TWICE_BITRATE_KBPS_BY_RATE =
new int[] {
64, 112, 128, 192, 224, 256, 384, 448, 512, 640, 768, 896, 1_024, 1_152, 1_280, 1_536,
1_920, 2_048, 2_304, 2_560, 2_688, 2_816, 2_823, 2_944, 3_072, 3_840, 4_096, 6_144, 7_680
};
/**
* Maps MaxSampleRate index to sampling frequency in Hz. See ETSI TS 102 114 V1.6.1 (2019-08)
* Table 7-9.
*/
private static final int[] SAMPLE_RATE_BY_INDEX =
new int[] {
8_000, 16_000, 32_000, 64_000, 128_000, 22_050, 44_100, 88_200, 176_400, 352_800, 12_000,
24_000, 48_000, 96_000, 192_000, 384_000
};
/**
* Payload length table for DTS UHD FTOC messages. See ETSI TS 103 491 V1.2.1 (2019-05), Section
* 6.4.3.
*/
private static final int[] UHD_FTOC_PAYLOAD_LENGTH_TABLE = new int[] {5, 8, 10, 12};
/** Metadata chunk size length table for DTS UHD. See ETSI TS 103 491 V1.2.1, Table 6-20. */
private static final int[] UHD_METADATA_CHUNK_SIZE_LENGTH_TABLE = new int[] {6, 9, 12, 15};
/** Audio chunk ID length table for DTS UHD. See ETSI TS 103 491 V1.2.1, Section 6.4.14.4. */
private static final int[] UHD_AUDIO_CHUNK_ID_LENGTH_TABLE = new int[] {2, 4, 6, 8};
/** Audio chunk size length table for DTS UHD. See ETSI TS 103 491 V1.2.1, Section 6.4.14.4. */
private static final int[] UHD_AUDIO_CHUNK_SIZE_LENGTH_TABLE = new int[] {9, 11, 13, 16};
/** Header size length table for DTS UHD. See ETSI TS 103 491 V1.2.1 (2019-05), Section 6.4.3. */
private static final int[] UHD_HEADER_SIZE_LENGTH_TABLE = new int[] {5, 8, 10, 12};
/**
* Returns the {@link FrameType} if {@code word} is a DTS sync word, otherwise {@link
* #FRAME_TYPE_UNKNOWN}.
*/
public static @FrameType int getFrameType(int word) {
if (word == SYNC_VALUE_BE
|| word == SYNC_VALUE_LE
|| word == SYNC_VALUE_14B_BE
|| word == SYNC_VALUE_14B_LE) {
return FRAME_TYPE_CORE;
} else if (word == SYNC_VALUE_EXTSS_BE || word == SYNC_VALUE_EXTSS_LE) {
return FRAME_TYPE_EXTENSION_SUBSTREAM;
} else if (word == SYNC_VALUE_UHD_FTOC_SYNC_BE || word == SYNC_VALUE_UHD_FTOC_SYNC_LE) {
return FRAME_TYPE_UHD_SYNC;
} else if (word == SYNC_VALUE_UHD_FTOC_NONSYNC_BE || word == SYNC_VALUE_UHD_FTOC_NONSYNC_LE) {
return FRAME_TYPE_UHD_NON_SYNC;
}
return FRAME_TYPE_UNKNOWN;
}
/**
* Returns the DTS format given {@code data} containing the DTS Core frame according to ETSI TS
* 102 114 V1.6.1 (2019-08) subsections 5.3/5.4.
*
* @param frame The DTS Core frame to parse.
* @param trackId The track identifier to set on the format.
* @param language The language to set on the format.
* @param roleFlags The role flags to set on the format.
* @param drmInitData {@link DrmInitData} to be included in the format.
* @return The DTS format parsed from data in the header.
*/
public static Format parseDtsFormat(
byte[] frame,
@Nullable String trackId,
@Nullable String language,
@C.RoleFlags int roleFlags,
@Nullable DrmInitData drmInitData) {
ParsableBitArray frameBits = getNormalizedFrame(frame);
frameBits.skipBits(32 + 1 + 5 + 1 + 7 + 14); // SYNC, FTYPE, SHORT, CPF, NBLKS, FSIZE
int amode = frameBits.readBits(6);
int channelCount = CHANNELS_BY_AMODE[amode];
int sfreq = frameBits.readBits(4);
int sampleRate = SAMPLE_RATE_BY_SFREQ[sfreq];
int rate = frameBits.readBits(5);
int bitrate =
rate >= TWICE_BITRATE_KBPS_BY_RATE.length
? Format.NO_VALUE
: TWICE_BITRATE_KBPS_BY_RATE[rate] * 1000 / 2;
frameBits.skipBits(10); // MIX, DYNF, TIMEF, AUXF, HDCD, EXT_AUDIO_ID, EXT_AUDIO, ASPF
channelCount += frameBits.readBits(2) > 0 ? 1 : 0; // LFF
return new Format.Builder()
.setId(trackId)
.setSampleMimeType(MimeTypes.AUDIO_DTS)
.setAverageBitrate(bitrate)
.setChannelCount(channelCount)
.setSampleRate(sampleRate)
.setDrmInitData(drmInitData)
.setLanguage(language)
.setRoleFlags(roleFlags)
.build();
}
/**
* Returns the number of audio samples represented by the given DTS Core frame.
*
* @param data The frame to parse.
* @return The number of audio samples represented by the frame.
*/
public static int parseDtsAudioSampleCount(byte[] data) {
int nblks;
switch (data[0]) {
case FIRST_BYTE_LE:
nblks = ((data[5] & 0x01) << 6) | ((data[4] & 0xFC) >> 2);
break;
case FIRST_BYTE_14B_LE:
nblks = ((data[4] & 0x07) << 4) | ((data[7] & 0x3C) >> 2);
break;
case FIRST_BYTE_14B_BE:
nblks = ((data[5] & 0x07) << 4) | ((data[6] & 0x3C) >> 2);
break;
default:
// We blindly assume FIRST_BYTE_BE if none of the others match.
nblks = ((data[4] & 0x01) << 6) | ((data[5] & 0xFC) >> 2);
}
return (nblks + 1) * 32;
}
/**
* Like {@link #parseDtsAudioSampleCount(byte[])} but reads from a {@link ByteBuffer}. The
* buffer's position is not modified.
*
* @param buffer The {@link ByteBuffer} from which to read.
* @return The number of audio samples represented by the syncframe.
*/
public static int parseDtsAudioSampleCount(ByteBuffer buffer) {
if ((buffer.getInt(0) == SYNC_VALUE_UHD_FTOC_SYNC_LE)
|| (buffer.getInt(0) == SYNC_VALUE_UHD_FTOC_NONSYNC_LE)) {
// Check for DTS:X Profile 2 sync or non sync word and return 1024 if found. This is the only
// audio sample count that is used by DTS:X Streaming Encoder.
return 1024;
} else if (buffer.getInt(0) == SYNC_VALUE_EXTSS_LE) {
// Check for DTS Express sync word and return 4096 if found. This is the only audio sample
// count that is used by DTS Streaming Encoder.
return 4096;
}
// See ETSI TS 102 114 subsection 5.4.1.
int position = buffer.position();
int nblks;
switch (buffer.get(position)) {
case FIRST_BYTE_LE:
nblks = ((buffer.get(position + 5) & 0x01) << 6) | ((buffer.get(position + 4) & 0xFC) >> 2);
break;
case FIRST_BYTE_14B_LE:
nblks = ((buffer.get(position + 4) & 0x07) << 4) | ((buffer.get(position + 7) & 0x3C) >> 2);
break;
case FIRST_BYTE_14B_BE:
nblks = ((buffer.get(position + 5) & 0x07) << 4) | ((buffer.get(position + 6) & 0x3C) >> 2);
break;
default:
// We blindly assume FIRST_BYTE_BE if none of the others match.
nblks = ((buffer.get(position + 4) & 0x01) << 6) | ((buffer.get(position + 5) & 0xFC) >> 2);
}
return (nblks + 1) * 32;
}
/**
* Returns the size in bytes of the given DTS Core frame.
*
* @param data The frame to parse.
* @return The frame's size in bytes.
*/
public static int getDtsFrameSize(byte[] data) {
int fsize;
boolean uses14BitPerWord = false;
switch (data[0]) {
case FIRST_BYTE_14B_BE:
fsize = (((data[6] & 0x03) << 12) | ((data[7] & 0xFF) << 4) | ((data[8] & 0x3C) >> 2)) + 1;
uses14BitPerWord = true;
break;
case FIRST_BYTE_LE:
fsize = (((data[4] & 0x03) << 12) | ((data[7] & 0xFF) << 4) | ((data[6] & 0xF0) >> 4)) + 1;
break;
case FIRST_BYTE_14B_LE:
fsize = (((data[7] & 0x03) << 12) | ((data[6] & 0xFF) << 4) | ((data[9] & 0x3C) >> 2)) + 1;
uses14BitPerWord = true;
break;
default:
// We blindly assume FIRST_BYTE_BE if none of the others match.
fsize = (((data[5] & 0x03) << 12) | ((data[6] & 0xFF) << 4) | ((data[7] & 0xF0) >> 4)) + 1;
}
// If the frame is stored in 14-bit mode, adjust the frame size to reflect the actual byte size.
return uses14BitPerWord ? fsize * 16 / 14 : fsize;
}
/**
* Parses the {@link DtsHeader} data from the extension substream header of a DTS-HD frame
* according to ETSI TS 102 114 V1.6.1 (2019-08), Section 7.5.2.
*
* @param header The DTS-HD extension substream header to parse.
* @return The {@link DtsHeader} data extracted from the header.
*/
public static DtsHeader parseDtsHdHeader(byte[] header) throws ParserException {
ParsableBitArray headerBits = getNormalizedFrame(header);
headerBits.skipBits(32 + 8); // SYNCEXTSSH, UserDefinedBits
int extensionSubstreamIndex = headerBits.readBits(2); // nExtSSIndex
int headerSizeInBits; // nuBits4Header
int extensionSubstreamFrameSizeBits; // nuBits4ExSSFsize
if (!headerBits.readBit()) { // bHeaderSizeType
headerSizeInBits = 8;
extensionSubstreamFrameSizeBits = 16;
} else {
headerSizeInBits = 12;
extensionSubstreamFrameSizeBits = 20;
}
headerBits.skipBits(headerSizeInBits); // nuExtSSHeaderSize
int extensionSubstreamFrameSize =
headerBits.readBits(extensionSubstreamFrameSizeBits) + 1; // nuExtSSFsize
int assetsCount; // nuNumAssets
int referenceClockCode; // nuRefClockCode
int extensionSubstreamFrameDurationCode; // nuExSSFrameDurationCode
boolean staticFieldsPresent = headerBits.readBit(); // bStaticFieldsPresent
if (staticFieldsPresent) {
referenceClockCode = headerBits.readBits(2);
extensionSubstreamFrameDurationCode = 512 * (headerBits.readBits(3) + 1);
if (headerBits.readBit()) { // bTimeStampFlag
headerBits.skipBits(32 + 4); // nuTimeStamp, nLSB
}
int audioPresentationsCount = headerBits.readBits(3) + 1; // nuNumAudioPresnt
assetsCount = headerBits.readBits(3) + 1;
if (audioPresentationsCount != 1 || assetsCount != 1) {
throw ParserException.createForUnsupportedContainerFeature(
/* message= */ "Multiple audio presentations or assets not supported");
}
// We've already asserted audioPresentationsCount = 1.
int activeExtensionSubstreamMask =
headerBits.readBits(extensionSubstreamIndex + 1); // nuActiveExSSMask
for (int i = 0; i < extensionSubstreamIndex + 1; i++) {
if (((activeExtensionSubstreamMask >> i) & 0x1) == 1) {
headerBits.skipBits(8); // nuActiveAssetMask
}
}
if (headerBits.readBit()) { // bMixMetadataEnbl
headerBits.skipBits(2); // nuMixMetadataAdjLevel
int mixerOutputMaskBits = (headerBits.readBits(2) + 1) << 2; // nuBits4MixOutMask
int mixerOutputConfigurationCount = headerBits.readBits(2) + 1; // nuNumMixOutConfigs
// Output Mixing Configuration Loop
for (int i = 0; i < mixerOutputConfigurationCount; i++) {
headerBits.skipBits(mixerOutputMaskBits); // nuMixOutChMask
}
}
} else {
// Assignments below are placeholders and will never be used as they are only relevant when
// staticFieldsPresent == true. Initialised here to keep the compiler happy.
referenceClockCode = C.INDEX_UNSET;
extensionSubstreamFrameDurationCode = 0;
}
// We've already asserted assetsCount = 1.
headerBits.skipBits(extensionSubstreamFrameSizeBits); // nuAssetFsize
int sampleRate = C.RATE_UNSET_INT;
int channelCount = C.LENGTH_UNSET; // nuTotalNumChs
// Asset descriptor, see ETSI TS 102 114 V1.6.1 (2019-08) Table 7-5.
headerBits.skipBits(9 + 3); // nuAssetDescriptFsize, nuAssetIndex
if (staticFieldsPresent) {
if (headerBits.readBit()) { // bAssetTypeDescrPresent
headerBits.skipBits(4); // nuAssetTypeDescriptor
}
if (headerBits.readBit()) { // bLanguageDescrPresent
headerBits.skipBits(24); // LanguageDescriptor
}
if (headerBits.readBit()) { // bInfoTextPresent
int infoTextByteSize = headerBits.readBits(10) + 1; // nuInfoTextByteSize
headerBits.skipBytes(infoTextByteSize); // InfoTextString
}
headerBits.skipBits(5); // nuBitResolution
sampleRate = SAMPLE_RATE_BY_INDEX[headerBits.readBits(4)]; // nuMaxSampleRate
channelCount = headerBits.readBits(8) + 1;
// Done reading necessary bits, ignoring the rest.
}
long frameDurationUs = C.TIME_UNSET;
if (staticFieldsPresent) {
int referenceClockFrequency;
// ETSI TS 102 114 V1.6.1 (2019-08) Table 7-3.
switch (referenceClockCode) {
case 0:
referenceClockFrequency = 32_000;
break;
case 1:
referenceClockFrequency = 44_100;
break;
case 2:
referenceClockFrequency = 48_000;
break;
default:
throw ParserException.createForMalformedContainer(
/* message= */ "Unsupported reference clock code in DTS HD header: "
+ referenceClockCode,
/* cause= */ null);
}
frameDurationUs =
Util.scaleLargeTimestamp(
extensionSubstreamFrameDurationCode, C.MICROS_PER_SECOND, referenceClockFrequency);
}
return new DtsHeader(
MimeTypes.AUDIO_DTS_EXPRESS,
channelCount,
sampleRate,
extensionSubstreamFrameSize,
frameDurationUs,
/* bitrate= */ 0);
}
/**
* Returns the size of the extension substream header in a DTS-HD frame according to ETSI TS 102
* 114 V1.6.1 (2019-08), Section 7.5.2.
*
* @param headerPrefix A byte array containing at least the first 55 bits of a DTS-HD frame.
* @return Size of the DTS-HD frame header in bytes.
*/
public static int parseDtsHdHeaderSize(byte[] headerPrefix) {
ParsableBitArray headerPrefixBits = getNormalizedFrame(headerPrefix);
headerPrefixBits.skipBits(32 + 8 + 2); // SYNCEXTSSH, UserDefinedBits, nExtSSIndex
// Unpack the num of bits to be used to read header size
int headerBits = headerPrefixBits.readBit() ? 12 : 8; // bHeaderSizeType
// Unpack the substream header size
return headerPrefixBits.readBits(headerBits) + 1; // nuExtSSHeaderSize
}
/**
* Parses the {@link DtsHeader} data from the headers of a DTS-UHD(Profile 2) frame according to
* ETSI TS 103 491 V1.2.1 (2019-05), Section 6.4.3.
*
* @param header The DTS-UHD header to parse.
* @param uhdAudioChunkId An {@link AtomicInteger} containing the last read UHD audio chunk ID
* from a synchronized frame, or zero if unset. This parameter is both an input and output
* parameter. In synchronized frames, the input value is not used; instead, the parameter is
* set to the current UHD audio chunk ID, which becomes the output value. For non-synchronized
* frames, it is used without any modification.
* @return The {@link DtsHeader} data extracted from the header.
*/
public static DtsHeader parseDtsUhdHeader(byte[] header, AtomicInteger uhdAudioChunkId)
throws ParserException {
ParsableBitArray headerBits = getNormalizedFrame(header);
int syncWord = headerBits.readBits(32);
boolean syncFrameFlag = syncWord == SYNC_VALUE_UHD_FTOC_SYNC_BE;
int ftocPayloadInBytes =
parseUnsignedVarInt(
headerBits, UHD_FTOC_PAYLOAD_LENGTH_TABLE, /* extractAndAddFlag= */ true)
+ 1;
// ETSI TS 103 491 V1.2.1, Section 6.4.5.
int sampleRate = C.RATE_UNSET_INT; // m_unAudioSamplRate
long frameDurationUs = C.TIME_UNSET;
if (syncFrameFlag) {
// ETSI TS 103 491 V1.2.1, Section 6.4.6.1.
if (!headerBits.readBit()) { // m_bFullChannelBasedMixFlag
throw ParserException.createForUnsupportedContainerFeature(
/* message= */ "Only supports full channel mask-based audio presentation");
}
// ETSI TS 103 491 V1.2.1, Section 6.4.6.2.
checkCrc(header, ftocPayloadInBytes);
int baseDurationIndex = headerBits.readBits(2);
int baseDuration; // m_unBaseDuration
// ETSI TS 103 491 V1.2.1 (2019-05) Table 6-13.
switch (baseDurationIndex) {
case 0:
baseDuration = 512;
break;
case 1:
baseDuration = 480;
break;
case 2:
baseDuration = 384;
break;
default:
throw ParserException.createForMalformedContainer(
/* message= */ "Unsupported base duration index in DTS UHD header: "
+ baseDurationIndex,
/* cause= */ null);
}
int frameDurationInClockPeriods =
baseDuration * (headerBits.readBits(3) + 1); // m_unFrameDuration
int clockRateIndex = headerBits.readBits(2);
int clockRateHertz; // m_unClockRateInHz
switch (clockRateIndex) {
case 0:
clockRateHertz = 32_000;
break;
case 1:
clockRateHertz = 44_100;
break;
case 2:
clockRateHertz = 48_000;
break;
default:
throw ParserException.createForMalformedContainer(
/* message= */ "Unsupported clock rate index in DTS UHD header: " + clockRateIndex,
/* cause= */ null);
}
// Skip time stamp information if present, see section 5.2.3.2.
if (headerBits.readBit()) { // m_bParamPresent
// m_bUpdateFlag == true as m_bSyncFramePredefValueExists is set to false in the encoder.
headerBits.skipBits(32 + 4); // m_TimeStamp
}
int sampleRateMultiplier = (1 << headerBits.readBits(2));
sampleRate = clockRateHertz * sampleRateMultiplier;
frameDurationUs =
Util.scaleLargeTimestamp(
frameDurationInClockPeriods, C.MICROS_PER_SECOND, clockRateHertz);
}
// ETSI TS 103 491 V1.2.1, Table 6-20.
// m_bFullChannelBasedMixFlag == true as we throw unsupported container feature otherwise.
int chunkPayloadBytes = 0;
int numOfMetadataChunks = syncFrameFlag ? 1 : 0; // Metadata chunks
for (int i = 0; i < numOfMetadataChunks; i++) {
int metadataChunkSize =
parseUnsignedVarInt(
headerBits, UHD_METADATA_CHUNK_SIZE_LENGTH_TABLE, /* extractAndAddFlag= */ true);
chunkPayloadBytes += metadataChunkSize;
}
// See ETSI TS 103 491 V1.2.1, Section 6.4.14.4.
// m_bFullChannelBasedMixFlag == true as we throw unsupported container feature otherwise.
int numAudioChunks = 1;
for (int i = 0; i < numAudioChunks; i++) {
// If syncFrameFlag is true the audio chunk ID will be present.
if (syncFrameFlag) {
uhdAudioChunkId.set(
parseUnsignedVarInt(
headerBits, UHD_AUDIO_CHUNK_ID_LENGTH_TABLE, /* extractAndAddFlag= */ true));
}
int audioChunkSize =
uhdAudioChunkId.get() != 0
? parseUnsignedVarInt(
headerBits, UHD_AUDIO_CHUNK_SIZE_LENGTH_TABLE, /* extractAndAddFlag= */ true)
: 0;
chunkPayloadBytes += audioChunkSize;
}
int frameSize = ftocPayloadInBytes + chunkPayloadBytes;
return new DtsHeader(
MimeTypes.AUDIO_DTS_X,
// To determine the actual number of channels from a bit stream, we need to read the
// metadata chunk bytes. If defining a constant channel count causes problems, we can
// consider adding additional parsing logic for UHD frames.
// For now, using the estimated number of channels for DTS UHD bitstreams as 2.
/* channelCount= */ 2,
sampleRate,
frameSize,
frameDurationUs,
/* bitrate= */ 0);
}
/**
* Returns the size of frame header in a DTS-UHD(Profile 2) frame according to ETSI TS 103 491
* V1.2.1 (2019-05), Section 6.4.3.
*
* @param headerPrefix A byte array containing at least the first 47 bits of a DTS-UHD frame.
* @return Size of the DTS-UHD frame header in bytes.
*/
public static int parseDtsUhdHeaderSize(byte[] headerPrefix) {
ParsableBitArray headerPrefixBits = getNormalizedFrame(headerPrefix);
headerPrefixBits.skipBits(32); // SYNC
return parseUnsignedVarInt(
headerPrefixBits, UHD_HEADER_SIZE_LENGTH_TABLE, /* extractAndAddFlag= */ true)
+ 1;
}
/**
* Check if calculated and extracted CRC-16 words match. See ETSI TS 103 491 V1.2.1, Table 6-8.
*/
private static void checkCrc(byte[] frame, int sizeInBytes) throws ParserException {
int initialValue = 0xFFFF;
int extractedCrc =
(((frame[sizeInBytes - 2] << 8) & initialValue) | (frame[sizeInBytes - 1] & 0xFF));
int calculatedCrc = Util.crc16(frame, /* start= */ 0, /* end= */ sizeInBytes - 2, initialValue);
if (extractedCrc != calculatedCrc) {
throw ParserException.createForMalformedContainer(
/* message= */ "CRC check failed", /* cause= */ null);
}
}
/**
* Helper function for the DTS UHD header parsing. Used to extract a field of variable length. See
* ETSI TS 103 491 V1.2.1, Section 5.2.3.1.
*/
private static int parseUnsignedVarInt(
ParsableBitArray frameBits, int[] lengths, boolean extractAndAddFlag) {
int index = 0;
for (int i = 0; i < 3; i++) {
if (frameBits.readBit()) {
index++;
} else {
break;
}
}
int value = 0;
if (extractAndAddFlag) {
for (int i = 0; i < index; i++) {
value += (1 << lengths[i]);
}
}
return value + frameBits.readBits(lengths[index]);
}
private static ParsableBitArray getNormalizedFrame(byte[] frame) {
if (frame[0] == FIRST_BYTE_BE
|| frame[0] == FIRST_BYTE_EXTSS_BE
|| frame[0] == FIRST_BYTE_UHD_FTOC_SYNC_BE
|| frame[0] == FIRST_BYTE_UHD_FTOC_NONSYNC_BE) {
// The frame is already 16-bit mode, big endian.
return new ParsableBitArray(frame);
}
// Data is not normalized, but we don't want to modify frame.
frame = Arrays.copyOf(frame, frame.length);
if (isLittleEndianFrameHeader(frame)) {
// Change endianness.
for (int i = 0; i < frame.length - 1; i += 2) {
byte temp = frame[i];
frame[i] = frame[i + 1];
frame[i + 1] = temp;
}
}
ParsableBitArray frameBits = new ParsableBitArray(frame);
if (frame[0] == (byte) (SYNC_VALUE_14B_BE >> 24)) {
// Discard the 2 most significant bits of each 16 bit word.
ParsableBitArray scratchBits = new ParsableBitArray(frame);
while (scratchBits.bitsLeft() >= 16) {
scratchBits.skipBits(2);
frameBits.putInt(scratchBits.readBits(14), 14);
}
}
frameBits.reset(frame);
return frameBits;
}
private static boolean isLittleEndianFrameHeader(byte[] frameHeader) {
return frameHeader[0] == FIRST_BYTE_LE
|| frameHeader[0] == FIRST_BYTE_14B_LE
|| frameHeader[0] == FIRST_BYTE_EXTSS_LE
|| frameHeader[0] == FIRST_BYTE_UHD_FTOC_SYNC_LE
|| frameHeader[0] == FIRST_BYTE_UHD_FTOC_NONSYNC_LE;
}
private DtsUtil() {}
}