实例介绍
ISO IEC 14496-1-2010
ISO/EC14496-1:2010(E) Contents Page Foreword Introduction cope Normative references Additional references 2 3456789 Terms and definitions 2 Abbreviated terms 10 Conventions∴… 11 Streaming Framework… Syntactic Description Language.,,,,,…,…,, …99 Profiles .110 AnnexA( informative) Time Base reconstruction…………,,……,…,………………….112 Annex B(informative) Registration procedure 115 Annex c( informative) The Qos Management Model for ISO/EC14496 Content.…,,…,…………,119 Annex D(informative)Conversion Between Time and Date Conventions ... ∴,120 Annex E(informative)Graphical Representation of Object Descriptor and Sync Layer Syntax... 122 Annex F(informative)Elementary Stream Interface. a 130 Annex G(informative)Upstream Walkthroug 132 Annex H(informative) Scene and Object Description Carrousel 137 Annex I (normative)Usage of ITU-T Recommendation H. 264 ISO/EC 14496-10 AVC 138 Annex J( informative) Patent statements……… 41 Bibliography...... A icAnn nn4n All rights reserved lII ISO/IEC14496-1:2010 0(E) Foreword iso (the International Organization for Standardization) and iEc (the International Electrotechnical Commission) form the specialized system for worldwide standardization. National bodies that are members of Iso or IEC participate in the development of International Standards through technical committees established by the respective organization to deal with particular fields of technical activity. ISO and IEC technical committees collaborate in fields of mutual interest. Other international organizations, governmental and non-governmental, in liaison with Iso and IEC, also take part in the work. In the field of information technology: ISO and lEC have established a joint technical committee, ISO/EC JTC 1 International Standards are drafted in accordance with the rules given in the ISo/EC Directives, Part 2 The main task of the joint technical committee is to prepare International Standards. Draft International Standards adopted by the joint technical committee are circulated to national bodies for voting. Publication as an International Standard requires approval by at least 75 of the national bodies casting a vote ISO/EC 14496-1 was prepared by Joint Technical Committee ISO/EC JTC1, information technology, Subcommittee SC 29, Coding of audio, picture, multimedia and hypermedia information This fourth edition cancels and replaces the third edition(ISO/EC 14496-1: 2004), which has been technically revised incorporates the Amendments ISO/EC 14496-1: 2004/Amd 1: 2005 ISO/EC 14496-1: 2004/Amd. 2: 2007, ISO/EC 14496-1: 2004/Amd. 3: 2007 and Technical Corrigenda so/EC14496-1:2004cor.1:2006and|sOEc14496-1:2004/cor.2:2007, ISo/EC 14496 consists of the following parts, under the general title Information technology- Coding of audio-visual objects Part 1: Systems -Part 2: visual - Part 3: audio Part 4: Conformance testing Part 5 Reference software Part 6: Delivery Multime dia Integration Frame work( DMIF Part 7: Optimized reference software for coding of audio-visual objects Part 8: Carriage of/so/Ec 14496 contents over /P networks Part 9: Reference hardware description Part 10: Advanced video Coding Part 11. Scene description and application engine Part 12/so base media file format Part 13: Intellectual Property Management and Protection(PMP)extensions O ISO/EC 2010-All rights reserved sOEC144961:2010(E Part 14: MP4 file format Part 15: Advanced Video Coding(AvC) file format Part 16: Animation Framework eXtension(AFX Part 17. Streaming text format Part 18: Font compression and streaming Part 19: Synthesized texture stream Part 20: Lightweight Application Scene Representation (LASeR) and Simple Aggregation Format (SAF) Part 21: MPEG- Graphics Frame work e XtensionS (GFX) Part 22: Open Font Format Part 23: Symbolic Music Representation Part 24: Audio and systems interaction Part 25: 3D Graphics Compression Mode Part 26: audio conformance Part 27: 3D Graphics conformance A icAnn nn4n All rights reserved ISO/IEC14496-1:2010 0(E) o Introduction 0.1 Overview ISO/EC 14496 specifies a system for the communication of interactive audio-visual scenes. This specification cludes the following elements a) The coded representation of natural or synthetic, two-dimensional(2D)or three-dimensional 3D)objects that can be manifested audibly and/or visually(audio-visual objects)(specified in Parts 2, 3, 10, 11, 16, 19,20,23and25of|sO/EC14496 b) The coded representation of the spatio-temporal positioning of audio-visual objects as well as their behavior in response to interaction(scene description, specified in Parts 1 1 and 20 of ISO/EC 14496) c)The coded representation of information related to the management of data streams(synchronization dentification, description and association of stream content, specified in this Part and in Part 24 of soEC14496) d) a generic interface to the data stream delivery layer functionality (specified in Part 6 of ISO/EC 14496) e) An application engine for programmatic control of the player: format, delivery of downloadable Java byte code as well as its execution lifecycle and behavior through APIs(specified in Parts 11 and 21 of ISO/EC 14496) f) A file format to contain the media information of an Iso/EC 14496 presentation in a flexible, extensible format to facilitate interchange, management, editing, and presentation of the media specified in Part 12 (ISO File Format), Part 14(MP4 File Format)and Part 15(AVC File Format)of ISO/EC 14496 g) The coded representation of font data and of information related to the management of text streams and font data streams(specified in Parts 17, 18 and 22 of ISO/EC 14496) The overall operation of a system communicating audio-Visual scenes can be paraphrased as follows At the sending terminal, the audio-visual scene information is compressed, supplemented with synchronization information and passed to a delivery layer that multiplexes it into one or more coded binary streams that are transmitted or stored. At the receiving terminal, these streams are demultiplexed and decompressed. The audio-visual objects are composed according to the scene description and synchronization information and presented to the end user. The end user may have the option to interact with this presentation Interaction information can be processed locally or transmitted back to the sending terminal so/EC 14496 defines the syntax and semantics of the bitstreams that convey such scene information, as well as the details of their decoding processes This part of ISONEC 14496 specifies the following tools a terminal model for time and buffer management A coded representation of metadata for the identification, description and logical dependencies of the elementary streams(object descriptors and other descriptors) A coded representation of descriptive audio-visual content information [object content information(OCl) An interface to intellectual property management and protection(IPMP)systems A coded representation of synchronization information(sync layer- SL) A multiplexed representation of individual elementary streams in a single stream(M4 Mux) O ISO/EC 2010-All rights reserved ISO/EC14496-1:2010(E) These various elements are described functionally in this clause and specified in the normative clauses that fo‖lDW 0.2 Architecture The information representation specified in ISo/EC 14496 describes the means to create an interactive audio-visual scene in terms of coded audio-visual information and associated scene description information The entity that composes and sends, or receives and presents such a coded representation of an interactive audio-visual scene is generically referred to as an"audio-visual terminal or just"terminal. This terminal may correspond to a stand-alone application or be part of an application system Di isplay and User Interaction Interactive Audiovisual Composition and rendering Upstream Compression In formation ayer AⅤ Object Descriptor Descriptio Information Elementary Streams Elementary Stream Interface SLSL SL SL SL SL ayer SI-Puckeiized streamy DMIF Application Interface M4MuX M4MuX M4MuX Delivery (RTP) AAL2 H223 DAB MPEG-2 LDP ATM PSTN Mux TS IP Multiplexed st Transmission/Storage Mcdium Figure 1- The ISonEC 14496 Terminal Architecture A icAnn nn4n All rights reserved so/EC14496-1:2010(E) The basic operations performed by such a receiver terminal are as follows. Information that allows access to content complying with ISo/EC 14496 is provided as initial session set up information to the terminal. Part 6 of IsonEC 14496 defines the procedures for establishing such session contexts as well as the interface to the delivery layer that generically abstracts the storage or transport medium The initial set up information allows a recursive manner, to locate one or more elementary streams that are part of the coded content representation. Some of these elementary streams may be grouped together using the multiplexing tool described in ISO/EC 14496-1 Elementary streams contain the coded representation of either audio or visual data or scene description information or user interaction data or text or font data. Elementary streams may as well themselves convey information to identify streams, to describe logical dependencies between streams, or to describe information related to the content of the streams. Each elementary stream contains only one type of data Elementary streams are decoded using their respective stream-specific decoders. The audio-visual objects are composed according to the scene description information and presented by the terminal's presentation device(s). All these processes are synchronized according to the systems decoder model (SDM) using the synchronization information provided at the synchronization layer These basic operations are depicted in Figure 1, and are described in more detail below 0.3 Terminal Model: Systems decoder Model The systems decoder model provides an abstract view of the behavior of a terminal complying with ISO/EC 14496-1. Its purpose is to enable a sending terminal to predict how the receiving terminal will behave in terms of buffer management and synchronization when reconstructing the audio-visual information that comprises the presentation. The systems decoder model includes a systems timing model and a systems uffer model which are described briefly in the following Subclauses 0.3.1 Timing Model The timing model defines the mechanisms through which a receiving terminal establishes a notion of time that enables it to process time-dependent events. this model also allows the receiving terminal to establish mechanisms to maintain synchronization both across and within particular audio-visual objects as well as with user interaction events. In order to facilitate these functions at the receiving terminal, the timing model requires that the transmitted data streams contain implicit or explicit timing information. Two sets of timing information are defined in ISO/EC 14496-1: clock references and time stamps. The former convey the sending terminal's time base to the receiving terminal, while the latter convey a notion of relative time for specific events such as the desired decoding or composition time for portions of the encoded audio-visual formation 0.3.2 Buffer model The buffer model enables the sending terminal to monitor and control the buffer resources that are needed to decode each elementary stream in a presentation. The required buffer resources are conveyed to the receiving terminal by means of descriptors at the beginning of the presentation. the terminal can then decide whether or not it is capable of handling this particular presentation. The buffer model allows the sending terminal to specify when information may be removed from these buffers and enables it to schedule data transmission so that the appropriate buffers at the receiving terminal do not overflow or underflow 0. 4 Multiplexing of Streams: The Delivery Layer The term delivery layer is used as a generic abstraction of any existing transport protocol stack that may be used to transmit and/or store content complying with ISO/IEC 14496. The functionality of this layer is not within the scope of ISo/EC 14496-1, and only the interface to this layer is considered. This interface is the DMIF Application Interface(DAI)specified in ISO/EC 14496-6. The dal defines not only an interface for the delivery of streaming data, but also for signaling information required for session and channel set up as well as tear down. a wide variety of delivery mechanisms exist below this interface, with some of them indicated in igure 1. These mechanisms serve for transmission as well as storage of streaming data, i.e., a file is O ISO/EC 2010-All rights reserved ISO/EC14496-1:2010(E) considered to be a particular instance of a delivery layer. For applications where the desired transport facility does not fully address the needs of a service according to the specifications in ISO/EC 14496, a simple multiplexing tool (M4Mux) with low delay and low overhead is defined in ISO/EC 14496-1 0.5 Synchronization of Streams: The Sync Layer Elementary streams are the basic abstraction for any streaming data source. Elementary streams are conveyed as sync layer-packetized (SL-packetized)streams at the DMIF Application Interface. This packetized representation additionally provides timing and synchronization information, as well as fragmentation and random access information. The sync layer(SL) extracts this timing information to enable synchronized decoding and subsequently composition of the elementary stream data 0. 6 The Compression Layer The compression layer receives data in its encoded format and performs the necessary operations to decode this data. The decoded information is then used by the terminal's composition, rendering and presentation subsystems 0.6.1 Object Description Framework The purpose of the object description framework is to identify and describe elementary streams and to associate them appropriately to an audio-visual scene description. object descriptors serve to gain access to ISO/EC 14496 content Object content information and the interface to intellectual property management and protection systems are also part of this framework An object descriptor is a collection of one or more elementary stream descriptors that provide the configuration and other information for the streams that relate to either an audio-visual object, or text or font data, or a scene description Object descriptors are themselves conveyed in elementary streams. Each object descriptor is assigned an identifier(object descriptor ID), which is unique within a defined name scope. This identifier is used to associate audio-visual objects in the scene description with a particular object descriptor and thus the elementary streams related to that particular object Elementary stream descriptors include information about the source of the stream data, in form of a unique numeric identifier(the elementary stream ID)or a URL pointing to a remote source for the stream. Elementary stream descriptors also include information about the encoding format, configuration information for the decoding process and the sync layer packetization, as well as quality of service requirements for the transmission of the stream and intellectual property identification. Dependencies between streams can also be signaled within the elementary stream descriptors. This functionality may be used, for example, in scalable audio or visual object representations to indicate the logical dependency of a stream containing enhancement information, to a stream containing the base information. It can also be used to describe alternative representations for the same content(e.g. the same speech content in various languages) 0.6.1.1 Intellectual Property Management and Protection The intellectual property management and protection(IPMP)framework for ISo/EC 14496 content consists of a normative interface that permits an iso/EC 14496 terminal to host one or more IPMP Systems in the form of monolithic IPMP Systems or modular IPMP Tools. The IPMP interface consists of IPMP elementary streams and IPMP descriptors. IPMP descriptors are carried as part of an object descriptor stream. IPMP elementary streams carry time variant IPMP information that can be associated to multiple object descriptors The IPMP System, or IPMP Tools themselves are non-normative components that provides intellectual property management and protection functions for the terminal. The IPMP Systems or Tools uses the information carried by the IPMP elementary streams and descriptors to make protected ISO/EC 14496 content available to the termina The intellectual property management and protection(IPMP) framework for ISo/EC 14496 content consists of a set of tools that permits an iso/EC 14496 terminal to support iPMP functionality. This functionality is A icAnn nn4n All rights reserved ISO/IEC14496-1:2010 0(E) provided by the following two different complementary technologies, supporting different levels of interoperability. a) The iPMP framework as defined in 7.2.3, consists of a normative interface that permits an ISO/EC 14496 terminal to host one or more IPMP Systems. The IPMP interface consists of IPMP elementary streams and IPMP descriptors. IPMP descriptors are carried as part of an object descriptor stream. IPMP elementary streams carry time variant IPMP information that can be associated to multiple object descriptors. The IPMP System itself is a non-normative component that provides intellectual property management and protection functions for the terminal. the IPMP System uses the information carried by the iPMP elementary streams and descriptors to make protected Iso/Ec 14496 content available to the terminal b) The IPMP framework extension, as specified in ISO/EC 14496-13 allows, in addition to the functionality specified in ISo/EC 14496-1, a finer granularity of governance. ISO/EC 14496-13 provides normative support for individual IPMP components, referred to as IPMP Tools, to be normatively placed at identified points of control within the terminal systems model. Additionally ISo/EC 14496-13 provides normative support for secure communications to be performed between IPMP Tools. ISO/EC 14496-1 also specifies specific normative extensions at the Systems level to support the IPMP functionality described so∥Ec14496-13 An application may choose not to use an IPMP System, thereby offering no management and protection features 0.6.1.2 Object Content Information Object content information(OCI)descriptors convey descriptive information about audio-visual objects. The main content descriptors are: content classification descriptors, keyword descriptors, rating descriptors language descriptors, textual descriptors, and descriptors about the creation of the content. oCl descriptors can be included directly in the related object descriptor or elementary stream descriptor or, if it is time variant, it may be carried in an elementary stream by itself. An oCl stream is organized in a sequence of small synchronized entities called events that contain a set of oci descriptors ocl streams can be as sociated to multiple object descriptors 0.6.2 Scene Description Streams Scene description addresses the organization of audio-visual objects in a scene, in terms of both spatial and temporal attributes. This information allows the composition and rendering of individual audio-visual objects after the respective decoders have reconstructed the streaming data for them For visual data ISO/EC 14496-11 does not mandate particular composition algorithms. Hence, visual composition is im plementation dependent. For audio data, the composition process is defined in a normative manner in lso/EC14496-11and|sO/Ec14496-3 The scene description is represented using a parametric approach(BIFS- Binary Format for Scenes).The description consists of an encoded hierarchy(tree) of nodes with attributes and other information (including event sources and targets ). Leaf nodes in this tree correspond to elementary audio-visual data, whereas intermediate nodes group this material to form audio-visual objects, and perform grouping, transformation, and other such operations on audio-visual objects(scene description nodes). The scene description can evolve over time by using scene description updates. In order to facilitate active user involvement with the presented audio-visual information, ISO/EC 14496-11 provides support for user and object interactions. Interactivity mechanisms are integrated with the scene description information in the form of linked event sources and targets (routes as well as sensors(special nodes that can trigger events based on specific conditions ). These event sources and targets are part of scene description nodes, and thus allow close coupling of dynamic and interactive behavior with the specific scene at hand. ISO/EC 14496-11, however, does not specify a particular user interface or a mechanism that maps user actions(e.g. keyboard key presses or mouse movements)to such events O ISO/EC 2010-All rights reserved 【实例截图】
【核心代码】
标签:
小贴士
感谢您为本站写下的评论,您的评论对其它用户来说具有重要的参考价值,所以请认真填写。
- 类似“顶”、“沙发”之类没有营养的文字,对勤劳贡献的楼主来说是令人沮丧的反馈信息。
- 相信您也不想看到一排文字/表情墙,所以请不要反馈意义不大的重复字符,也请尽量不要纯表情的回复。
- 提问之前请再仔细看一遍楼主的说明,或许是您遗漏了。
- 请勿到处挖坑绊人、招贴广告。既占空间让人厌烦,又没人会搭理,于人于己都无利。
关于好例子网
本站旨在为广大IT学习爱好者提供一个非营利性互相学习交流分享平台。本站所有资源都可以被免费获取学习研究。本站资源来自网友分享,对搜索内容的合法性不具有预见性、识别性、控制性,仅供学习研究,请务必在下载后24小时内给予删除,不得用于其他任何用途,否则后果自负。基于互联网的特殊性,平台无法对用户传输的作品、信息、内容的权属或合法性、安全性、合规性、真实性、科学性、完整权、有效性等进行实质审查;无论平台是否已进行审查,用户均应自行承担因其传输的作品、信息、内容而可能或已经产生的侵权或权属纠纷等法律责任。本站所有资源不代表本站的观点或立场,基于网友分享,根据中国法律《信息网络传播权保护条例》第二十二与二十三条之规定,若资源存在侵权或相关问题请联系本站客服人员,点此联系我们。关于更多版权及免责申明参见 版权及免责申明
网友评论
我要评论