ASF Media Source
Media Foundation provides the ASF media source that an application can use to represent an ASF file in the pipeline layer of the architecture.
In order to play an ASF file, an application can use the ASF media source to feed data into a playback pipeline. In an encoding scenario, application can use the ASF media source as the source to convert into another format or for converting a higher bit-rate file to a lower bit-rate file without changing formats. The ASF media source must be used in the pipeline layer, that is, an application must use the Media Session to control the operation. This level of access allows the applications to get events while the Media Session is in progress. To get lower-level access to the ASF content, an application must use the WMContainer Layer ASF components.
The ASF media source implements the IMFMediaSource interface, which is the generic interface for media sources in Media Foundation. Like any other media source, the ASF media source provides a presentation descriptor that primarily describes the ASF Header Object. In addition, the ASF media source provides stream descriptors that represent each stream in the media content. For more information, see ASF File Structure.
- Creating the ASF Media Source
- Configuration Settings for the ASF Media Source
- ASF Media Source Object Model
- ASF Media Source Services
- Related topics
To create the ASF media source, the application must use the Source Resolver. In order to create the ASF media source, the application must provide the source for which the source resolver creates the ASF media source. The source information must be provided by either specifying the URL of the source file or the byte stream that contains the media. If the application chooses to create the ASF media source by specifying the URL, it must call either IMFSourceResolver::CreateObjectFromURL for synchronous operation or IMFSourceResolver::Begin…EndCreateObjectFromURL for asynchronous operation. The process for creating the media source from a byte stream is similar. The application must call either IMFSourceResolver::CreateObjectFromByteStream for synchronous operation or IMFSourceResolver::Begin…EndCreateObjectFromBytestream for asynchronous operation. These calls must specify MF_RESOLUTION_MEDIASOURCE in dwFlags parameter. For more information about using this flag, see Using the Source Resolver.
If the application specifies a URL, for a local file, the URL string can contain the path of the media file or can be prefixed with the “file: “scheme. The file name extension must be ".asf", ".wm", L".wma", or ".wmv". For an ASF file on the network, the source resolver creates the Network Source, which uses the ASF media source.
During the object creation process, the source resolver looks up the list of scheme handlers and the byte stream handlers in the system registry and loads the closest matching handler that can parse the media content and also create the media source object underneath. Irrespective of the method used to create the media source (URL and byte stream), the source resolver creates a byte stream and reads the contents of source media into the byte stream. For more information, see Scheme Handlers and Byte-Stream Handlers.
For code example about how to create a media source, see Using the Source Resolver.
The source resolver queries for the capabilities of the underlying byte stream and determines the operations are allowed on the newly created media source. One of such capabilities is seeking. If a seek operation is allowed, the application can specify the seeking mode that the media source uses while seeking to a particular point in the presentation.
An application can set the seek mode, for the media source to use, during object creation. After the ASF media source is created with the specified seek mode, it cannot be modified on the media source or dynamically changed during a presentation. Seeking preferences are specified as properties in the application’s call to the relevant source resolver methods that are used to create the media source. These set of properties are set in a property store and passed to the pProps parameter. If these properties are not passed, the media source functions with the default settings. For more information about setting these properties, see Configuring a Media Source.
If seeking is allowed, the ASF media source supports the following seek modes:
- Exact seeking: In this mode, the ASF media source relies on the ASF Index Object of the ASF file. If the file does not have the Index Object, exact seeking is disabled and one of the other modes are used depending on the application-specified properties that are set on the media source.
- Approximate seeking: This mode is requested by the application by passing MFPKEY_ASFMediaSource_ApproxSeek in the property store to the relevant source resolver methods. However, approximate seeking is only used if the byte stream does not support time-based seeking. In this mode, the media source determine a start time that is relatively closer to the seek time. If the ASF file contains an ASF Index Object, it is used to calculate the start time. Approximate seeking is less accurate but faster than exact seeking because the time taken render the first sample at the predetermined start time is faster.
- Iterative seeking: To set this mode, the application must set the MFPKEY_ASFMediaSource_IterativeSeekIfNoIndex property. Iterative seeking is an algorithm to find a position in an ASF file that contains no ASF Index Object. In this mode, the media source determins a rough estimate of the seek point by reading the byte stream offset. It uses a series of approximations, based on the average bit rate, to get progressively closer to the target seek time. (The algorithm is similar to a binary search.) Iterative seeking can take longer than seeking by using the Index Object, so it is disabled by default. If you set this property, use the following properties to set the search parameters: MFPKEY_ASFMediaSource_IterativeSeek_Max_Count MFPKEY_ASFMediaSource_IterativeSeek_Tolerance_In_MilliSecond. These properties set the maximum number of iterations and the tolerance, respectively. The algorithm halts when it reaches the maximum number of iterations, or when it finds a packet whose distance from the seek time is within the specified tolerance.
The ASF media source implements the IMFMediaSource interface and exposes the following interfaces. An application can get a reference to these interfaces by calling IMFMediaSource::QueryInterface on the ASF media source.
|IMFMediaSource||Required for all media sources.|
|IMFMediaEventGenerator||Required for all media sources. Allows applications to get events from the media source through the Media Session.|
|IMFGetService||Queries the media source for the specified service interface.|
|IPropertyStore||Sets and gets properties on the media source. Each property contains a descriptive name and a value.|
|IMFMetadata||Describes the metadata for the ASF file.|
|IMFMetadataProvider||Retrieves a collection of metadata, either for an entire presentation, or for one stream in the presentation.|
|IMFRateSupport||Queries the range of playback rates that are supported, including reverse playback.|
|IMFRateControl||Gets or sets the playback rate.|
|IMFTrustedInput||Gets the ITA for each of the ASF streams contained in the source.|
|IMFPMPClient||Enables a media source to receive a pointer to the IMFPMPHost interface, which is used to create objects in the PMP process.|
|IMFTimecodeTranslate||Converts between Society of Motion Picture and Television Engineers (SMPTE) time codes and 100-nanosecond time units.|
The ASF media source provides various services that are scoped for ASF file. To get a pointer to a particular service, the application must use one of the following service identifiers in its call to MFGetService. For information, see Service Interfaces.
|MF_RATE_CONTROL_SERVICE||By using this identifier, an application can get a pointer to either IMFRateSupport or IMFRateControl interfaces. By using the rate support object implemented by the media source, the application can check if a particular rate is supported by the underlying ASF media file also get the fastest and the slowest rate. By using the rate control object the application can get and set the playback rate. If the application specifies a particular rate for playback, the ASF media source first checks if the requested rate is within the rate limits (determined by fasted and slowest rates) and then sets the rate. This does not check for variable conditions as the bitrate might change any moment depending on the network bandwidth. For more information, Rate Control.|
|MF_METADATA_PROVIDER_SERVICE||By using this identifier, the application can get a pointer to the IMFMetadataProvider interface of the ASF media source. This interface provides access to information about ASF file specifically the ASF Header Objects and the streams contained within the media content. Header information is exposes through presentation descriptor attributes and stream information is provided through stream descriptor attributes. For more information about these attributes, see Media Foundation Attributes for ASF Header Objects. In addition to information that is exposed through attributes, there also exists descriptive metadata, which are set as properties.|
|MF_PROPERTY_HANDLER_SERVICE||By using this identifier, the application can get a pointer to the IPropertyStore interface of the ASF media source. The property store contains all the metadata related to the ASF file. This identifier is new in Windows 7 and replaces MF_METADATA_PROVIDER_SERVICE for getting metadata properties.|
|MFNETSOURCE_STATISTICS_SERVICE||For more information, see Retrieving Network Statistics in Client Logging.|
|MF_TIMECODE_SERVICE||By using this identifier, the application can get a pointer to the media source’s IMFTimecodeTranslate interface. This implementation can be used to perform time code translations such as converting SMPTE time code to 100-nano second units and back.|
The ASF media source provides a time code translation service that allows your application to convert SMPTE time code to the closest presentation time (in 100-nanosecond unit). Conversely, the application can also get the time code for a requested presentation time. The service is available through the IMFTimecodeTranslate interface, which the ASF media source implements. The method calls on this interface pointer are asynchronous that can run from your main application thread without blocking the user interface of your application.
The following steps summarize the procedure for time code translation:
- Get pointer to the IMFTimecodeTranslate interface of the ASF media source by calling MFGetService and specifying the MF_TIMECODE_SERVICE identifier.
- Call IMFTimecodeTranslate::BeginConvertTimecodeToHNS or IMFTimecodeTranslate::BeginConvertHNSToTimecode and specify the time that is to be translated.. For BeginConvertTimecodeToHNS the time code must be specified as a PROPVARIANT variable with VT_I8 data type. The BeginConvertHNSToTimecode method requires the time in 100-nanosecond units as the MFTIME type.
- Complete the asynchronous operation by calling IMFTimecodeTranslate::EndConvertTimecodeToHNS or IMFTimecodeTranslate::EndConvertTimecodeToHNS appropriately.
During the creation of the media source, the source resolver creates a byte stream for the file from which the media source reads the ASF content. In order for the time conversions to be successful, the byte stream associated with the ASF file must have seeking capabilities; otherwise, the application gets the MF_E_BYTESTREAM_NOT_SEEKABLE error from the Begin... call. Another requirement for time conversions is that the ASF file, represented by the ASF media source, must have ASF Index Objects. The presentation times and the time codes are extracted from the ASF Index Objects that maintain all the indexes and the corresponding seek- points for the ASF file. For presentation time-to-time code translation, the ASF file must contain a Simple Index Object; for time code-to-presentation time translation, the ASF file must have an Index Object. The conversion operations rely on the underlying indexer (WMContainer component) to parse and read the Index Object associated with the ASF file. If the file does not contain an Index Object, the application asynchronously receives the MF_E_NO_INDEX error code.
To translate the time requested by the application, the media source enumerates the streams within the file and finds the stream that contains the appropriate ASF Index Object. If such a stream is found, the media source parses the stream’s ASF packets are parsed until the correct time code is reached. After it finds the correct sample it retrieves the corresponding presentation time or the time code from the sample, and returns it to the caller.