/**@class android.graphics.ImageFormat
@extends java.lang.Object

*/
var ImageFormat = {

/***/
UNKNOWN : "0",
/** RGB format used for pictures encoded as RGB_565. See
 {@link android.hardware.Camera.Parameters#setPictureFormat(int)}.
*/
RGB_565 : "4",
/** <p>Android YUV format.</p>

 <p>This format is exposed to software decoders and applications.</p>

 <p>YV12 is a 4:2:0 YCrCb planar format comprised of a WxH Y plane followed
 by (W/2) x (H/2) Cr and Cb planes.</p>

 <p>This format assumes
 <ul>
 <li>an even width</li>
 <li>an even height</li>
 <li>a horizontal stride multiple of 16 pixels</li>
 <li>a vertical stride equal to the height</li>
 </ul>
 </p>

 <pre> y_size = stride * height
 c_stride = ALIGN(stride/2, 16)
 c_size = c_stride * height/2
 size = y_size + c_size * 2
 cr_offset = y_size
 cb_offset = y_size + c_size</pre>

 <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
 recommended for YUV output instead.</p>

 <p>For the older camera API, this format is guaranteed to be supported for
 {@link android.hardware.Camera} preview images since API level 12; for earlier API versions,
 check {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.

 <p>Note that for camera preview callback use (see
 {@link android.hardware.Camera#setPreviewCallback}), the
 <var>stride</var> value is the smallest possible; that is, it is equal
 to:

 <pre>stride = ALIGN(width, 16)</pre>

 @see android.hardware.Camera.Parameters#setPreviewCallback
 @see android.hardware.Camera.Parameters#setPreviewFormat
 @see android.hardware.Camera.Parameters#getSupportedPreviewFormats
 </p>
*/
YV12 : "842094169",
/** <p>Android Y8 format.</p>

 <p>Y8 is a YUV planar format comprised of a WxH Y plane only, with each pixel
 being represented by 8 bits. It is equivalent to just the Y plane from {@link #YV12}
 format.</p>

 <p>This format assumes
 <ul>
 <li>an even width</li>
 <li>an even height</li>
 <li>a horizontal stride multiple of 16 pixels</li>
 </ul>
 </p>

 <pre> size = stride * height </pre>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.hardware.camera2.CameraDevice} (if
 supported) through a {@link android.media.ImageReader} object. The
 {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
 single plane containing the pixel data. The pixel stride is always 1 in
 {@link android.media.Image.Plane#getPixelStride()}, and the
 {@link android.media.Image.Plane#getRowStride()} describes the vertical
 neighboring pixel distance (in bytes) between adjacent rows.</p>

 @see android.media.Image
 @see android.media.ImageReader
 @see android.hardware.camera2.CameraDevice
*/
Y8 : "538982489",
/** <p>Android Y16 format.</p>

 Y16 is a YUV planar format comprised of a WxH Y plane, with each pixel
 being represented by 16 bits. It is just like {@link #Y8}, but has 16
 bits per pixel (little endian).</p>

 <p>This format assumes
 <ul>
 <li>an even width</li>
 <li>an even height</li>
 <li>a horizontal stride multiple of 16 pixels</li>
 </ul>
 </p>

 <pre> y_size = stride * height </pre>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.hardware.camera2.CameraDevice}
 through a {@link android.media.ImageReader} object if this format is
 supported by {@link android.hardware.camera2.CameraDevice}.</p>

 @see android.media.Image
 @see android.media.ImageReader
 @see android.hardware.camera2.CameraDevice

 @hide
*/
Y16 : "540422489",
/** YCbCr format, used for video.

 <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
 recommended for YUV output instead.</p>

 <p>Whether this format is supported by the old camera API can be determined by
 {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>

*/
NV16 : "16",
/** YCrCb format used for images, which uses the NV21 encoding format.

 <p>This is the default format
 for {@link android.hardware.Camera} preview images, when not otherwise set with
 {@link android.hardware.Camera.Parameters#setPreviewFormat(int)}.</p>

 <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
 recommended for YUV output instead.</p>
*/
NV21 : "17",
/** YCbCr format used for images, which uses YUYV (YUY2) encoding format.

 <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
 recommended for YUV output instead.</p>

 <p>This is an alternative format for {@link android.hardware.Camera} preview images. Whether
 this format is supported by the camera hardware can be determined by
 {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>
*/
YUY2 : "20",
/** Compressed JPEG format.

 <p>This format is always supported as an output format for the
 {@link android.hardware.camera2} API, and as a picture format for the older
 {@link android.hardware.Camera} API</p>
*/
JPEG : "256",
/** Depth augmented compressed JPEG format.

 <p>JPEG compressed main image along with XMP embedded depth metadata
 following ISO 16684-1:2011(E).</p>
*/
DEPTH_JPEG : "1768253795",
/** <p>Multi-plane Android YUV 420 format</p>

 <p>This format is a generic YCbCr format, capable of describing any 4:2:0
 chroma-subsampled planar or semiplanar buffer (but not fully interleaved),
 with 8 bits per color sample.</p>

 <p>Images in this format are always represented by three separate buffers
 of data, one for each color plane. Additional information always
 accompanies the buffers, describing the row stride and the pixel stride
 for each plane.</p>

 <p>The order of planes in the array returned by
 {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
 plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>

 <p>The Y-plane is guaranteed not to be interleaved with the U/V planes
 (in particular, pixel stride is always 1 in
 {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}).</p>

 <p>The U/V planes are guaranteed to have the same row stride and pixel stride
 (in particular,
 {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
 == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
 {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
 == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
 ).</p>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.hardware.camera2.CameraDevice}
 through a {@link android.media.ImageReader} object.</p>

 @see android.media.Image
 @see android.media.ImageReader
 @see android.hardware.camera2.CameraDevice
*/
YUV_420_888 : "35",
/** <p>Multi-plane Android YUV 422 format</p>

 <p>This format is a generic YCbCr format, capable of describing any 4:2:2
 chroma-subsampled (planar, semiplanar or interleaved) format,
 with 8 bits per color sample.</p>

 <p>Images in this format are always represented by three separate buffers
 of data, one for each color plane. Additional information always
 accompanies the buffers, describing the row stride and the pixel stride
 for each plane.</p>

 <p>The order of planes in the array returned by
 {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
 plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>

 <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
 stride greater than 1 in
 {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>

 <p>The U/V planes are guaranteed to have the same row stride and pixel stride
 (in particular,
 {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
 == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
 {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
 == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
 ).</p>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.media.MediaCodec}
 through {@link android.media.MediaCodec#getOutputImage} object.</p>

 @see android.media.Image
 @see android.media.MediaCodec
*/
YUV_422_888 : "39",
/** <p>Multi-plane Android YUV 444 format</p>

 <p>This format is a generic YCbCr format, capable of describing any 4:4:4
 (planar, semiplanar or interleaved) format,
 with 8 bits per color sample.</p>

 <p>Images in this format are always represented by three separate buffers
 of data, one for each color plane. Additional information always
 accompanies the buffers, describing the row stride and the pixel stride
 for each plane.</p>

 <p>The order of planes in the array returned by
 {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
 plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>

 <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
 stride greater than 1 in
 {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>

 <p>The U/V planes are guaranteed to have the same row stride and pixel stride
 (in particular,
 {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
 == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
 {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
 == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
 ).</p>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.media.MediaCodec}
 through {@link android.media.MediaCodec#getOutputImage} object.</p>

 @see android.media.Image
 @see android.media.MediaCodec
*/
YUV_444_888 : "40",
/** <p>Multi-plane Android RGB format</p>

 <p>This format is a generic RGB format, capable of describing most RGB formats,
 with 8 bits per color sample.</p>

 <p>Images in this format are always represented by three separate buffers
 of data, one for each color plane. Additional information always
 accompanies the buffers, describing the row stride and the pixel stride
 for each plane.</p>

 <p>The order of planes in the array returned by
 {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
 plane #0 is always R (red), plane #1 is always G (green), and plane #2 is always B
 (blue).</p>

 <p>All three planes are guaranteed to have the same row strides and pixel strides.</p>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.media.MediaCodec}
 through {@link android.media.MediaCodec#getOutputImage} object.</p>

 @see android.media.Image
 @see android.media.MediaCodec
*/
FLEX_RGB_888 : "41",
/** <p>Multi-plane Android RGBA format</p>

 <p>This format is a generic RGBA format, capable of describing most RGBA formats,
 with 8 bits per color sample.</p>

 <p>Images in this format are always represented by four separate buffers
 of data, one for each color plane. Additional information always
 accompanies the buffers, describing the row stride and the pixel stride
 for each plane.</p>

 <p>The order of planes in the array returned by
 {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
 plane #0 is always R (red), plane #1 is always G (green), plane #2 is always B (blue),
 and plane #3 is always A (alpha). This format may represent pre-multiplied or
 non-premultiplied alpha.</p>

 <p>All four planes are guaranteed to have the same row strides and pixel strides.</p>

 <p>For example, the {@link android.media.Image} object can provide data
 in this format from a {@link android.media.MediaCodec}
 through {@link android.media.MediaCodec#getOutputImage} object.</p>

 @see android.media.Image
 @see android.media.MediaCodec
*/
FLEX_RGBA_8888 : "42",
/** <p>General raw camera sensor image format, usually representing a
 single-channel Bayer-mosaic image. Each pixel color sample is stored with
 16 bits of precision.</p>

 <p>The layout of the color mosaic, the maximum and minimum encoding
 values of the raw pixel data, the color space of the image, and all other
 needed information to interpret a raw sensor image must be queried from
 the {@link android.hardware.camera2.CameraDevice} which produced the
 image.</p>
*/
RAW_SENSOR : "32",
/** <p>Private raw camera sensor image format, a single channel image with
 implementation depedent pixel layout.</p>

 <p>RAW_PRIVATE is a format for unprocessed raw image buffers coming from an
 image sensor. The actual structure of buffers of this format is
 implementation-dependent.</p>

*/
RAW_PRIVATE : "36",
/** <p>
 Android 10-bit raw format
 </p>
 <p>
 This is a single-plane, 10-bit per pixel, densely packed (in each row),
 unprocessed format, usually representing raw Bayer-pattern images coming
 from an image sensor.
 </p>
 <p>
 In an image buffer with this format, starting from the first pixel of
 each row, each 4 consecutive pixels are packed into 5 bytes (40 bits).
 Each one of the first 4 bytes contains the top 8 bits of each pixel, The
 fifth byte contains the 2 least significant bits of the 4 pixels, the
 exact layout data for each 4 consecutive pixels is illustrated below
 ({@code Pi[j]} stands for the jth bit of the ith pixel):
 </p>
 <table>
 <thead>
 <tr>
 <th align="center"></th>
 <th align="center">bit 7</th>
 <th align="center">bit 6</th>
 <th align="center">bit 5</th>
 <th align="center">bit 4</th>
 <th align="center">bit 3</th>
 <th align="center">bit 2</th>
 <th align="center">bit 1</th>
 <th align="center">bit 0</th>
 </tr>
 </thead> <tbody>
 <tr>
 <td align="center">Byte 0:</td>
 <td align="center">P0[9]</td>
 <td align="center">P0[8]</td>
 <td align="center">P0[7]</td>
 <td align="center">P0[6]</td>
 <td align="center">P0[5]</td>
 <td align="center">P0[4]</td>
 <td align="center">P0[3]</td>
 <td align="center">P0[2]</td>
 </tr>
 <tr>
 <td align="center">Byte 1:</td>
 <td align="center">P1[9]</td>
 <td align="center">P1[8]</td>
 <td align="center">P1[7]</td>
 <td align="center">P1[6]</td>
 <td align="center">P1[5]</td>
 <td align="center">P1[4]</td>
 <td align="center">P1[3]</td>
 <td align="center">P1[2]</td>
 </tr>
 <tr>
 <td align="center">Byte 2:</td>
 <td align="center">P2[9]</td>
 <td align="center">P2[8]</td>
 <td align="center">P2[7]</td>
 <td align="center">P2[6]</td>
 <td align="center">P2[5]</td>
 <td align="center">P2[4]</td>
 <td align="center">P2[3]</td>
 <td align="center">P2[2]</td>
 </tr>
 <tr>
 <td align="center">Byte 3:</td>
 <td align="center">P3[9]</td>
 <td align="center">P3[8]</td>
 <td align="center">P3[7]</td>
 <td align="center">P3[6]</td>
 <td align="center">P3[5]</td>
 <td align="center">P3[4]</td>
 <td align="center">P3[3]</td>
 <td align="center">P3[2]</td>
 </tr>
 <tr>
 <td align="center">Byte 4:</td>
 <td align="center">P3[1]</td>
 <td align="center">P3[0]</td>
 <td align="center">P2[1]</td>
 <td align="center">P2[0]</td>
 <td align="center">P1[1]</td>
 <td align="center">P1[0]</td>
 <td align="center">P0[1]</td>
 <td align="center">P0[0]</td>
 </tr>
 </tbody>
 </table>
 <p>
 This format assumes
 <ul>
 <li>a width multiple of 4 pixels</li>
 <li>an even height</li>
 </ul>
 </p>

 <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
 not pixels.

 <p>
 Since this is a densely packed format, the pixel stride is always 0. The
 application must use the pixel data layout defined in above table to
 access each row data. When row stride is equal to {@code width * (10 / 8)}, there
 will be no padding bytes at the end of each row, the entire image data is
 densely packed. When stride is larger than {@code width * (10 / 8)}, padding
 bytes will be present at the end of each row.
 </p>
 <p>
 For example, the {@link android.media.Image} object can provide data in
 this format from a {@link android.hardware.camera2.CameraDevice} (if
 supported) through a {@link android.media.ImageReader} object. The
 {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
 single plane containing the pixel data. The pixel stride is always 0 in
 {@link android.media.Image.Plane#getPixelStride()}, and the
 {@link android.media.Image.Plane#getRowStride()} describes the vertical
 neighboring pixel distance (in bytes) between adjacent rows.
 </p>

 @see android.media.Image
 @see android.media.ImageReader
 @see android.hardware.camera2.CameraDevice
*/
RAW10 : "37",
/** <p>
 Android 12-bit raw format
 </p>
 <p>
 This is a single-plane, 12-bit per pixel, densely packed (in each row),
 unprocessed format, usually representing raw Bayer-pattern images coming
 from an image sensor.
 </p>
 <p>
 In an image buffer with this format, starting from the first pixel of each
 row, each two consecutive pixels are packed into 3 bytes (24 bits). The first
 and second byte contains the top 8 bits of first and second pixel. The third
 byte contains the 4 least significant bits of the two pixels, the exact layout
 data for each two consecutive pixels is illustrated below (Pi[j] stands for
 the jth bit of the ith pixel):
 </p>
 <table>
 <thead>
 <tr>
 <th align="center"></th>
 <th align="center">bit 7</th>
 <th align="center">bit 6</th>
 <th align="center">bit 5</th>
 <th align="center">bit 4</th>
 <th align="center">bit 3</th>
 <th align="center">bit 2</th>
 <th align="center">bit 1</th>
 <th align="center">bit 0</th>
 </tr>
 </thead> <tbody>
 <tr>
 <td align="center">Byte 0:</td>
 <td align="center">P0[11]</td>
 <td align="center">P0[10]</td>
 <td align="center">P0[ 9]</td>
 <td align="center">P0[ 8]</td>
 <td align="center">P0[ 7]</td>
 <td align="center">P0[ 6]</td>
 <td align="center">P0[ 5]</td>
 <td align="center">P0[ 4]</td>
 </tr>
 <tr>
 <td align="center">Byte 1:</td>
 <td align="center">P1[11]</td>
 <td align="center">P1[10]</td>
 <td align="center">P1[ 9]</td>
 <td align="center">P1[ 8]</td>
 <td align="center">P1[ 7]</td>
 <td align="center">P1[ 6]</td>
 <td align="center">P1[ 5]</td>
 <td align="center">P1[ 4]</td>
 </tr>
 <tr>
 <td align="center">Byte 2:</td>
 <td align="center">P1[ 3]</td>
 <td align="center">P1[ 2]</td>
 <td align="center">P1[ 1]</td>
 <td align="center">P1[ 0]</td>
 <td align="center">P0[ 3]</td>
 <td align="center">P0[ 2]</td>
 <td align="center">P0[ 1]</td>
 <td align="center">P0[ 0]</td>
 </tr>
 </tbody>
 </table>
 <p>
 This format assumes
 <ul>
 <li>a width multiple of 4 pixels</li>
 <li>an even height</li>
 </ul>
 </p>

 <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
 not pixels.

 <p>
 Since this is a densely packed format, the pixel stride is always 0. The
 application must use the pixel data layout defined in above table to
 access each row data. When row stride is equal to {@code width * (12 / 8)}, there
 will be no padding bytes at the end of each row, the entire image data is
 densely packed. When stride is larger than {@code width * (12 / 8)}, padding
 bytes will be present at the end of each row.
 </p>
 <p>
 For example, the {@link android.media.Image} object can provide data in
 this format from a {@link android.hardware.camera2.CameraDevice} (if
 supported) through a {@link android.media.ImageReader} object. The
 {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
 single plane containing the pixel data. The pixel stride is always 0 in
 {@link android.media.Image.Plane#getPixelStride()}, and the
 {@link android.media.Image.Plane#getRowStride()} describes the vertical
 neighboring pixel distance (in bytes) between adjacent rows.
 </p>

 @see android.media.Image
 @see android.media.ImageReader
 @see android.hardware.camera2.CameraDevice
*/
RAW12 : "38",
/** <p>Android dense depth image format.</p>

 <p>Each pixel is 16 bits, representing a depth ranging measurement from a depth camera or
 similar sensor. The 16-bit sample consists of a confidence value and the actual ranging
 measurement.</p>

 <p>The confidence value is an estimate of correctness for this sample.  It is encoded in the
 3 most significant bits of the sample, with a value of 0 representing 100% confidence, a
 value of 1 representing 0% confidence, a value of 2 representing 1/7, a value of 3
 representing 2/7, and so on.</p>

 <p>As an example, the following sample extracts the range and confidence from the first pixel
 of a DEPTH16-format {@link android.media.Image}, and converts the confidence to a
 floating-point value between 0 and 1.f inclusive, with 1.f representing maximum confidence:

 <pre>
    ShortBuffer shortDepthBuffer = img.getPlanes()[0].getBuffer().asShortBuffer();
    short depthSample = shortDepthBuffer.get()
    short depthRange = (short) (depthSample & 0x1FFF);
    short depthConfidence = (short) ((depthSample >> 13) & 0x7);
    float depthPercentage = depthConfidence == 0 ? 1.f : (depthConfidence - 1) / 7.f;
 </pre>
 </p>

 <p>This format assumes
 <ul>
 <li>an even width</li>
 <li>an even height</li>
 <li>a horizontal stride multiple of 16 pixels</li>
 </ul>
 </p>

 <pre> y_size = stride * height </pre>

 When produced by a camera, the units for the range are millimeters.
*/
DEPTH16 : "1144402265",
/** Android sparse depth point cloud format.

 <p>A variable-length list of 3D points plus a confidence value, with each point represented
 by four floats; first the X, Y, Z position coordinates, and then the confidence value.</p>

 <p>The number of points is {@code (size of the buffer in bytes) / 16}.

 <p>The coordinate system and units of the position values depend on the source of the point
 cloud data. The confidence value is between 0.f and 1.f, inclusive, with 0 representing 0%
 confidence and 1.f representing 100% confidence in the measured position values.</p>

 <p>As an example, the following code extracts the first depth point in a DEPTH_POINT_CLOUD
 format {@link android.media.Image}:
 <pre>
    FloatBuffer floatDepthBuffer = img.getPlanes()[0].getBuffer().asFloatBuffer();
    float x = floatDepthBuffer.get();
    float y = floatDepthBuffer.get();
    float z = floatDepthBuffer.get();
    float confidence = floatDepthBuffer.get();
 </pre>

 For camera devices that support the
 {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_DEPTH_OUTPUT DEPTH_OUTPUT}
 capability, DEPTH_POINT_CLOUD coordinates have units of meters, and the coordinate system is
 defined by the camera's pose transforms:
 {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_TRANSLATION} and
 {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_ROTATION}. That means the origin is
 the optical center of the camera device, and the positive Z axis points along the camera's optical axis,
 toward the scene.
*/
DEPTH_POINT_CLOUD : "257",
/** Unprocessed implementation-dependent raw
 depth measurements, opaque with 16 bit
 samples.

 @hide
*/
RAW_DEPTH : "4098",
/** Android private opaque image format.
 <p>
 The choices of the actual format and pixel data layout are entirely up to
 the device-specific and framework internal implementations, and may vary
 depending on use cases even for the same device. The buffers of this
 format can be produced by components like
 {@link android.media.ImageWriter ImageWriter} , and interpreted correctly
 by consumers like {@link android.hardware.camera2.android.graphics.CameraDevice
 android.graphics.CameraDevice} based on the device/framework private information. However,
 these buffers are not directly accessible to the application.
 </p>
 <p>
 When an {@link android.media.Image Image} of this format is obtained from
 an {@link android.media.ImageReader ImageReader} or
 {@link android.media.ImageWriter ImageWriter}, the
 {@link android.media.Image#getPlanes() getPlanes()} method will return an
 empty {@link android.media.Image.Plane Plane} array.
 </p>
 <p>
 If a buffer of this format is to be used as an OpenGL ES texture, the
 framework will assume that sampling the texture will always return an
 alpha value of 1.0 (i.e. the buffer contains only opaque pixel values).
 </p>
*/
PRIVATE : "34",
/** Compressed HEIC format.

 <p>This format defines the HEIC brand of High Efficiency Image File
 Format as described in ISO/IEC 23008-12.</p>
*/
HEIC : "1212500294",
/**Use this function to retrieve the number of bits per pixel of an
 ImageFormat.
@param {Number} format
@return {Number} the number of bits per pixel of the given format or -1 if the
         format doesn't exist or is not supported.
*/
getBitsPerPixel : function(  ) {},

/**Determine whether or not this is a public-visible {@code format}.

 <p>In particular, {@code @hide} formats will return {@code false}.</p>

 <p>Any other formats (including UNKNOWN) will return {@code false}.</p>
@param {Number} format an integer format
@return {Boolean} a boolean
@hide 
*/
isPublicFormat : function(  ) {},


};