<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd"> <html> <head> <title>OpenSL ES for Android</title> </head> <body> <h1>OpenSL ES for Android</h1> This article describes the Android native audio APIs based on the Khronos Group OpenSL ES™ 1.0.1 standard, as of Android API level 9 (Android version 2.3). <p> OpenSL ES provides a C language interface that is also callable from C++, and exposes features similar to the audio portions of these Android APIs callable from Java programming language code: <ul> <li><a href="http://developer.android.com/reference/android/media/MediaPlayer.html"> android.media.MediaPlayer</a> <li><a href="http://developer.android.com/reference/android/media/MediaRecorder.html"> android.media.MediaRecorder</a> </ul> As with all of the Android Native Development Kit (NDK), the primary purpose of OpenSL ES for Android is to facilitate the implementation of shared libraries to be called from Java programming language code via Java Native Interface (JNI). NDK is not intended for writing pure C/C++ applications. That said, OpenSL ES is a full-featured API, and we expect that you should be able to accomplish most of your audio needs using only this API, without up-calls to Java. <p> Note: though based on OpenSL ES, Android native audio at API level 9 is <i>not</i> a conforming implementation of any OpenSL ES 1.0.1 profile (game, music, or phone). This is because Android does not implement all of the features required by any one of the profiles. Any known cases where Android behaves differently than the specification are described in section "Android extensions" below. <h2>Getting started</h2> <h3>Example code</h3> <h4>Recommended</h4> Supported and tested example code, usable as a model for your own code, is located in NDK folder <code>platforms/android-9/samples/native-audio/</code>. <h4>Not recommended</h4> The OpenSL ES 1.0.1 specification contains example code in the appendices (see section "References" below for the link to this specification). However, the examples in Appendix B: Sample Code and Appendix C: Use Case Sample Code use features not supported by Android API level 9. Some examples also contain typographical errors, or use APIs that are likely to change. Proceed with caution in referring to these; though the code may be helpful in understanding the full OpenSL ES standard, it should not be used as is with Android. <h3>Adding OpenSL ES to your application source code</h3> OpenSL ES is a C API, but is callable from both C and C++ code. <p> At a minimum, add the following line to your code: <pre> #include <SLES/OpenSLES.h> </pre> If you use Android extensions, also include these headers: <pre> #include <SLES/OpenSLES_Android.h> #include <SLES/OpenSLES_AndroidConfiguration.h> </pre> <h3>Makefile</h3> Modify your Android.mk as follows: <pre> LOCAL_LDLIBS += libOpenSLES </pre> <h3>Audio content</h3> There are many ways to package audio content for your application, including: <dl> <dt>Resources</dt> <dd> By placing your audio files into the <code>res/raw/</code> folder, they can be accessed easily by the associated APIs for <a href="http://developer.android.com/reference/android/content/res/Resources.html"> Resources</a>. However there is no direct native access to resources, so you will need to write Java programming language code to copy them out before use. </dd> <dt>Assets</dt> <dd> By placing your audio files into the <code>assets/</code> folder, they will be directly accessible by the Android native asset manager APIs. See the header files <code>android/asset_manager.h</code> and <code>android/asset_manager_jni.h</code> for more information on these APIs, which are new for API level 9. The example code located in NDK folder <code>platforms/android-9/samples/native-audio/</code> uses these native asset manager APIs in conjunction with the Android file descriptor data locator. </dd> <dt>Network</dt> <dd> You can use the URI data locator to play audio content directly from the network. However, be sure to read section "Security and permissions" below. </dd> <dt>Local filesystem</dt> <dd> The URI data locator supports the <code>file:</code> scheme for local files, provided the files are accessible by the application. Note that the Android security framework restricts file access via the Linux user ID and group ID mechanism. </dd> <dt>Recorded</dt> <dd>Your application can record audio data from the microphone input, store this content, and then play it back later. The example code uses this method for the "Playback" clip. </dd> <dt>Compiled and linked inline</dt> <dd> You can link your audio content directly into the shared library, and then play it using an audio player with buffer queue data locator. This is most suitable for short PCM format clips. The example code uses this technique for the "Hello" and "Android" clips. The PCM data was converted to hex strings using a <code>bin2c</code> tool (not supplied). </dd> <dt>Synthesis</dt> <dd> Your application can synthesize PCM data on the fly and then play it using an audio player with buffer queue data locator. This is a relatively advanced technique, and the details of audio synthesis are beyond the scope of this article. </dd> </dl> Finding or creating useful audio content for your application is beyond the scope of this article, but see the "References" section below for some suggested web search terms. <p> Note that it is your responsibility to ensure that you are legally permitted to play or record content, and that there may be privacy considerations for recording content. <h3>Debugging</h3> For robustness, we recommend that you examine the <code>SLresult</code> value which is returned by most APIs. Use of <code>assert</code> vs. more advanced error handling logic is a matter of coding style and the particular API; see the Wikipedia article on <a href="http://en.wikipedia.org/wiki/Assertion_(computing)">assert</a> for more information. In the supplied example, we have used <code>assert</code> for "impossible" conditions which would indicate a coding error, and explicit error handling for others which are more likely to occur in production. <p> Many API errors result in a log entry, in addition to the non-zero result code. These log entries provide additional detail which can be especially useful for the more complex APIs such as <code>Engine::CreateAudioPlayer</code>. <p> Use <a href="http://developer.android.com/guide/developing/tools/adb.html"> adb logcat</a>, the <a href="http://developer.android.com/guide/developing/eclipse-adt.html"> Eclipse ADT plugin</a> LogCat pane, or <a href="http://developer.android.com/guide/developing/tools/ddms.html#logcat"> ddms logcat</a> to see the log. <h2>Supported features from OpenSL ES 1.0.1</h2> This section summarizes available features in this API level. In some cases, there are limitations which are described in the next sub-section. <h3>Global entry points</h3> Supported global entry points: <ul> <li><code>slCreateEngine</code> <li><code>slQueryNumSupportedEngineInterfaces</code> <li><code>slQuerySupportedEngineInterfaces</code> </ul> <h3>Objects and interfaces</h3> The following figure indicates objects and interfaces supported by Android's OpenSL ES implementation. A green cell means the feature is supported. <p> <img src="chart1.png" alt="Supported objects and interfaces"> <h3>Limitations</h3> This section details limitations with respect to the supported objects and interfaces from the previous section. <h4>Buffer queue data locator</h4> An audio player or recorder with buffer queue data locator supports PCM data format only. <h4>Device data locator</h4> The only supported use of an I/O device data locator is when it is specified as the data source for <code>Engine::CreateAudioRecorder</code>. It should be initialized using these values, as shown in the example: <pre> SLDataLocator_IODevice loc_dev = {SL_DATALOCATOR_IODEVICE, SL_IODEVICE_AUDIOINPUT, SL_DEFAULTDEVICEID_AUDIOINPUT, NULL}; </pre> <h4>Dynamic interface management</h4> <code>RemoveInterface</code> and <code>ResumeInterface</code> are not supported. <h4>Effect combinations</h4> It is meaningless to have both environmental reverb and preset reverb on the same output mix. <p> The platform may ignore effect requests if it estimates that the CPU load would be too high. <h4>Effect send</h4> <code>SetSendLevel</code> supports a single send level per audio player. <h4>Environmental reverb</h4> Environmental reverb does not support the <code>reflectionsDelay</code>, <code>reflectionsLevel</code>, or <code>reverbDelay</code> fields of <code>struct SLEnvironmentalReverbSettings</code>. <h4>MIME data format</h4> The MIME data format can be used with URI data locator only, and only for player (not recorder). <p> The Android implementation of OpenSL ES requires that <code>mimeType</code> be initialized to either <code>NULL</code> or a valid UTF-8 string, and that <code>containerType</code> be initialized to a valid value. In the absence of other considerations, such as portability to other implementations, or content format which cannot be identified by header, we recommend that you set the <code>mimeType</code> to <code>NULL</code> and <code>containerType</code> to <code>SL_CONTAINERTYPE_UNSPECIFIED</code>. <p> Supported formats include WAV PCM, WAV alaw, WAV ulaw, MP3, Ogg Vorbis, AAC LC, HE-AACv1 (aacPlus), HE-AACv2 (enhanced aacPlus), and AMR [provided these are supported by the overall platform, and AAC formats must be located within an MP4 container]. MIDI is not supported. WMA is not part of the open source release, and compatibility with Android OpenSL ES has not been verified. <p> The Android implementation of OpenSL ES does not support direct playback of DRM or encrypted content; if you want to play this, you will need to convert to cleartext in your application before playing, and enforce any DRM restrictions in your application. <h4>Object</h4> <code>Resume</code>, <code>RegisterCallback</code>, <code>AbortAsyncOperation</code>, <code>SetPriority</code>, <code>GetPriority</code>, and <code>SetLossOfControlInterfaces</code> are not supported. <h4>PCM data format</h4> The PCM data format can be used with buffer queues only. Supported PCM playback configurations are 8-bit unsigned or 16-bit signed, mono or stereo, little endian byte ordering, and these sample rates: 8000, 11025, 12000, 16000, 22050, 24000, 32000, 44100, or 48000 Hz. For recording, the supported configurations are device-dependent, however generally 16000 Hz mono 16-bit signed is usually available. <p> Note that the field <code>samplesPerSec</code> is actually in units of milliHz, despite the misleading name. To avoid accidentally using the wrong value, you should initialize this field using one of the symbolic constants defined for this purpose (such as <code>SL_SAMPLINGRATE_44_1</code> etc.) <h4>Playback rate</h4> A single playback rate range from 500 per mille to 2000 per mille inclusive is supported, with property <code>SL_RATEPROP_NOPITCHCORAUDIO</code>. <h4>Record</h4> The <code>SL_RECORDEVENT_HEADATLIMIT</code> and <code>SL_RECORDEVENT_HEADMOVING</code> events are not supported. <h4>Seek</h4> <code>SetLoop</code> enables whole file looping. The <code>startPos</code> and <code>endPos</code> parameters are ignored. <h4>URI data locator</h4> The URI data locator can be used with MIME data format only, and only for an audio player (not audio recorder). Supported schemes are <code>http:</code> and <code>file:</code>. A missing scheme defaults to the <code>file:</code> scheme. Other schemes such as <code>https:</code>, <code>ftp:</code>, and <code>content:</code> are not supported. <code>rtsp:</code> is not verified. <h3>Data structures</h3> Android API level 9 supports these OpenSL ES 1.0.1 data structures: <ul> <li>SLDataFormat_MIME <li>SLDataFormat_PCM <li>SLDataLocator_BufferQueue <li>SLDataLocator_IODevice <li>SLDataLocator_OutputMix <li>SLDataLocator_URI <li>SLDataSink <li>SLDataSource <li>SLEngineOption <li>SLEnvironmentalReverbSettings <li>SLInterfaceID </ul> <h3>Platform configuration</h3> OpenSL ES for Android is designed for multi-threaded applications, and is thread-safe. <p> OpenSL ES for Android supports a single engine per application, and up to 32 objects. Available device memory and CPU may further restrict the usable number of objects. <p> <code>slCreateEngine</code> recognizes, but ignores, these engine options: <ul> <li><code>SL_ENGINEOPTION_THREADSAFE</code> <li><code>SL_ENGINEOPTION_LOSSOFCONTROL</code> </ul> <h2>Planning for future versions of OpenSL ES</h2> The Android native audio APIs at level 9 are based on Khronos Group OpenSL ES 1.0.1 (see section "References" below). As of the time of this writing, the OpenSL ES working group is preparing a revised version of the standard. The revised version will likely include new features, clarifications, correction of typographical errors, and some incompatibilities. Most of the expected incompatibilities are relatively minor, or are in areas of OpenSL ES not supported by Android API level 9. However, even a small change can be significant for an application developer, so it important to prepare for this. <p> The Android team is committed to preserving future API binary compatibility for developers to the extent feasible. It is our intention to continue to support future binary compatibility of the 1.0.1-based API, even as we add support for later versions of the standard. An application developed with this version should work on future versions of the Android platform, provided that you follow the guidelines listed in section "Planning for binary compatibility" below. <p> Note that future source compatibility will <i>not</i> be a goal. That is, if you upgrade to a newer version of the NDK, you may need to modify your application source code to conform to the new API. We expect that most such changes will be minor; see details below. <h3>Planning for binary compatibility</h3> We recommend that your application follow these guidelines, to improve future binary compatibility: <ul> <li> Use only the documented subset of Android-supported features from OpenSL ES 1.0.1. <li> Do not depend on a particular result code for an unsuccessful operation; be prepared to deal with a different result code. <li> Application callback handlers generally run in a restricted context, and should be written to perform their work quickly and then return as soon as possible. Do not do complex operations within a callback handler. For example, within a buffer queue completion callback, you can enqueue another buffer, but do not create an audio player. <li> Callback handlers should be prepared to be called more or less frequently, to receive additional event types, and should ignore event types that they do not recognize. Callbacks that are configured with an event mask of enabled event types should be prepared to be called with multiple event type bits set simultaneously. Use "&" to test for each event bit rather than a switch case. <li> Use prefetch status and callbacks as a general indication of progress, but do not depend on specific hard-coded fill levels or callback sequence. The meaning of the prefetch status fill level, and the behavior for errors that are detected during prefetch, may change. <li> See section "Buffer queue behavior" below. </ul> <h3>Planning for source compatibility</h3> As mentioned, source code incompatibilities are expected in the next version of OpenSL ES from Khronos Group. Likely areas of change include: <ul> <li>The buffer queue interface is expected to have significant changes, especially in the areas of <code>BufferQueue::Enqueue</code>, the parameter list for <code>slBufferQueueCallback</code>, and the name of field <code>SLBufferQueueState.playIndex</code>. We recommend that your application code use Android simple buffer queues instead, because we do not plan to change that API. In the example code supplied with the NDK, we have used Android simple buffer queues for playback for this reason. (We also use Android simple buffer queue for recording, but that is because standard OpenSL ES 1.0.1 does not support record to a buffer queue data sink.) <li>Addition of <code>const</code> to input parameters passed by reference, and to <code>SLchar *</code> struct fields used as input values. This should not require any changes to your code. <li>Substitution of unsigned types for some parameters that are currently signed. You may need to change a parameter type from <code>SLint32</code> to <code>SLuint32</code> or similar, or add a cast. <li><code>Equalizer::GetPresetName</code> will copy the string to application memory instead of returning a pointer to implementation memory. This will be a significant change, so we recommend that you either avoid calling this method, or isolate your use of it. <li>Additional fields in struct types. For output parameters, these new fields can be ignored, but for input parameters the new fields will need to be initialized. Fortunately, these are expected to all be in areas not supported by API level 9. <li>Interface <a href="http://en.wikipedia.org/wiki/Globally_unique_identifier"> GUIDs</a> will change. Refer to interfaces by symbolic name rather than GUID to avoid a dependency. <li><code>SLchar</code> will change from <code>unsigned char</code> to <code>char</code>. This primarily affects the URI data locator and MIME data format. <li><code>SLDataFormat_MIME.mimeType</code> will be renamed to <code>pMimeType</code>, and <code>SLDataLocator_URI.URI</code> will be renamed to <code>pURI</code>. We recommend that you initialize the <code>SLDataFormat_MIME</code> and <code>SLDataLocator_URI</code> data structures using a brace-enclosed comma-separated list of values, rather than by field name, to isolate your code from this change. In the example code we have used this technique. <li><code>SL_DATAFORMAT_PCM</code> does not permit the application to specify the representation of the data as signed integer, unsigned integer, or floating-point. The Android implementation assumes that 8-bit data is unsigned integer and 16-bit is signed integer. In addition, the field <code>samplesPerSec</code> is a misnomer, as the actual units are milliHz. These issues are expected to be addressed in the next OpenSL ES version, which will introduce a new extended PCM data format that permits the application to explicitly specify the representation, and corrects the field name. As this will be a new data format, and the current PCM data format will still be available (though deprecated), it should not require any immediate changes to your code. </ul> Any actual source code incompatibilities will be explained in detail when the time comes. <h2>Android extensions</h2> The API for Android extensions is defined in <code>SLES/OpenSLES_Android.h</code>. Consult that file for details on these extensions. Unless otherwise noted, all interfaces are "explicit". <p> Note that use these extensions will limit your application's portability to other OpenSL ES implementations. If this is a concern, we advise that you avoid using them, or isolate your use of these with <code>#ifdef</code> etc. <p> The following figure shows which Android-specific interfaces and data locators are available for each object type. <p> <img src="chart2.png" alt="Android extensions"> <h3>Android configuration interface</h3> The Android configuration interface provides a means to set platform-specific parameters for objects. Unlike other OpenSL ES 1.0.1 interfaces, the Android configuration interface is available prior to object realization. This permits the object to be configured and then realized. Header file <code>SLES/OpenSLES_AndroidConfiguration.h</code> documents the available configuration keys and values: <ul> <li>stream type for audio players (default <code>SL_ANDROID_STREAM_MEDIA</code>) <li>record profile for audio recorders (default <code>SL_ANDROID_RECORDING_PRESET_GENERIC</code>) </ul> Here is an example code fragment that sets the Android audio stream type on an audio player: <pre> // CreateAudioPlayer and specify SL_IID_ANDROIDCONFIGURATION // in the required interface ID array. Do not realize player yet. // ... SLAndroidConfigurationItf playerConfig; result = (*playerObject)->GetInterface(playerObject, SL_IID_ANDROIDCONFIGURATION, &playerConfig); assert(SL_RESULT_SUCCESS == result); SLint32 streamType = SL_ANDROID_STREAM_ALARM; result = (*playerConfig)->SetConfiguration(playerConfig, SL_ANDROID_KEY_STREAM_TYPE, &streamType, sizeof(SLint32)); assert(SL_RESULT_SUCCESS == result); // ... // Now realize the player here. </pre> Similar code can be used to configure the preset for an audio recorder. <h3>Android effects interfaces</h3> The Android effect, effect send, and effect capabilities interfaces provide a generic mechanism for an application to query and use device-specific audio effects. A device manufacturer should document any available device-specific audio effects. <p> Portable applications should use the OpenSL ES 1.0.1 APIs for audio effects instead of the Android effect extensions. <h3>Android file descriptor data locator</h3> The Android file descriptor data locator permits the source for an audio player to be specified as an open file descriptor with read access. The data format must be MIME. <p> This is especially useful in conjunction with the native asset manager. <h3>Android simple buffer queue data locator and interface</h3> The Android simple buffer queue data locator and interface are identical to the OpenSL ES 1.0.1 buffer queue locator and interface, except that Android simple buffer queues may be used with both audio players and audio recorders, and are limited to PCM data format. [OpenSL ES 1.0.1 buffer queues are for audio players only, and are not restricted to PCM data format.] <p> For recording, the application should enqueue empty buffers. Upon notification of completion via a registered callback, the filled buffer is available for the application to read. <p> For playback there is no difference. But for future source code compatibility, we suggest that applications use Android simple buffer queues instead of OpenSL ES 1.0.1 buffer queues. <h3>Dynamic interfaces at object creation</h3> For convenience, the Android implementation of OpenSL ES 1.0.1 permits dynamic interfaces to be specified at object creation time, as an alternative to adding these interfaces after object creation with <code>DynamicInterfaceManagement::AddInterface</code>. <h3>Buffer queue behavior</h3> The OpenSL ES 1.0.1 specification requires that "On transition to the <code>SL_PLAYSTATE_STOPPED</code> state the play cursor is returned to the beginning of the currently playing buffer." The Android implementation does not necessarily conform to this requirement. For Android, it is unspecified whether a transition to <code>SL_PLAYSTATE_STOPPED</code> operates as described, or leaves the play cursor unchanged. <p> We recommend that you do not rely on either behavior; after a transition to <code>SL_PLAYSTATE_STOPPED</code>, you should explicitly call <code>BufferQueue::Clear</code>. This will place the buffer queue into a known state. <p> A corollary is that it is unspecified whether buffer queue callbacks are called upon transition to <code>SL_PLAYSTATE_STOPPED</code> or by <code>BufferQueue::Clear</code>. We recommend that you do not rely on either behavior; be prepared to receive a callback in these cases, but also do not depend on receiving one. <p> It is expected that a future version of OpenSL ES will clarify these issues. However, upgrading to that version would result in source code incompatibilities (see section "Planning for source compatibility" above). <h3>Reporting of extensions</h3> <code>Engine::QueryNumSupportedExtensions</code>, <code>Engine::QuerySupportedExtension</code>, <code>Engine::IsExtensionSupported</code> report these extensions: <ul> <li><code>ANDROID_SDK_LEVEL_9</code> (platforms that support a higher API level will report an extension of <code>ANDROID_SDK_LEVEL_#</code> where # is that API level) </ul> <h2>Programming notes</h2> These notes supplement the OpenSL ES 1.0.1 specification, available in the "References" section below. <h3>Objects and interface initialization</h3> Two aspects of the OpenSL ES programming model that may be unfamiliar to new developers are the distinction between objects and interfaces, and the initialization sequence. <p> Briefly, an OpenSL ES object is similar to the object concept in programming languages such as Java and C++, except an OpenSL ES object is <i>only</i> visible via its associated interfaces. This includes the initial interface for all objects, called <code>SLObjectItf</code>. There is no handle for an object itself, only a handle to the <code>SLObjectItf</code> interface of the object. <p> An OpenSL ES object is first "created", which returns an <code>SLObjectItf</code>, then "realized". This is similar to the common programming pattern of first constructing an object (which should never fail other than for lack of memory or invalid parameters), and then completing initialization (which may fail due to lack of resources). The realize step gives the implementation a logical place to allocate additional resources if needed. <p> As part of the API to create an object, an application specifies an array of desired interfaces that it plans to acquire later. Note that this array does <i>not</i> automatically acquire the interfaces; it merely indicates a future intention to acquire them. Interfaces are distinguished as "implicit" or "explicit". An explicit interface <i>must</i> be listed in the array if it will be acquired later. An implicit interface need not be listed in the object create array, but there is no harm in listing it there. OpenSL ES has one more kind of interface called "dynamic", which does not need to be specified in the object create array, and can be added later after the object is created. The Android implementation provides a convenience feature to avoid this complexity; see section "Dynamic interfaces at object creation" above. <p> After the object is created and realized, the application should acquire interfaces for each feature it needs, using <code>GetInterface</code> on the initial <code>SLObjectItf</code>. <p> Finally, the object is available for use via its interfaces, though note that some objects require further setup. In particular, an audio player with URI data source needs a bit more preparation in order to detect connection errors. See the next section "Audio player prefetch" for details. <p> After your application is done with the object, you should explicitly destroy it; see section "Destroy" below. <h3>Audio player prefetch</h3> For an audio player with URI data source, <code>Object::Realize</code> allocates resources but does not connect to the data source (i.e. "prepare") or begin pre-fetching data. These occur once the player state is set to either <code>SL_PLAYSTATE_PAUSED</code> or <code>SL_PLAYSTATE_PLAYING</code>. <p> Note that some information may still be unknown until relatively late in this sequence. In particular, initially <code>Player::GetDuration</code> will return <code>SL_TIME_UNKNOWN</code> and <code>MuteSolo::GetChannelCount</code> will return zero. These APIs will return the proper values once they are known. <p> Other properties that are initially unknown include the sample rate and actual media content type based on examining the content's header (as opposed to the application-specified MIME type and container type). These too, are determined later during prepare / prefetch, but there are no APIs to retrieve them. <p> The prefetch status interface is useful for detecting when all information is available. Or, your application can poll periodically. Note that some information may <i>never</i> be known, for example, the duration of a streaming MP3. <p> The prefetch status interface is also useful for detecting errors. Register a callback and enable at least the <code>SL_PREFETCHEVENT_FILLLEVELCHANGE</code> and <code>SL_PREFETCHEVENT_STATUSCHANGE</code> events. If both of these events are delivered simultaneously, and <code>PrefetchStatus::GetFillLevel</code> reports a zero level, and <code>PrefetchStatus::GetPrefetchStatus</code> reports <code>SL_PREFETCHSTATUS_UNDERFLOW</code>, then this indicates a non-recoverable error in the data source. This includes the inability to connect to the data source because the local filename does not exist or the network URI is invalid. <p> The next version of OpenSL ES is expected to add more explicit support for handling errors in the data source. However, for future binary compatibility, we intend to continue to support the current method for reporting a non-recoverable error. <p> In summary, a recommended code sequence is: <ul> <li>Engine::CreateAudioPlayer <li>Object:Realize <li>Object::GetInterface for SL_IID_PREFETCHSTATUS <li>PrefetchStatus::SetCallbackEventsMask <li>PrefetchStatus::SetFillUpdatePeriod <li>PrefetchStatus::RegisterCallback <li>Object::GetInterface for SL_IID_PLAY <li>Play::SetPlayState to SL_PLAYSTATE_PAUSED or SL_PLAYSTATE_PLAYING <li>preparation and prefetching occur here; during this time your callback will be called with periodic status updates </ul> <h3>Destroy</h3> Be sure to destroy all objects on exit from your application. Objects should be destroyed in reverse order of their creation, as it is not safe to destroy an object that has any dependent objects. For example, destroy in this order: audio players and recorders, output mix, then finally the engine. <p> OpenSL ES does not support automatic garbage collection or <a href="http://en.wikipedia.org/wiki/Reference_counting">reference counting</a> of interfaces. After you call <code>Object::Destroy</code>, all extant interfaces derived from the associated object become <i>undefined</i>. <p> The Android OpenSL ES implementation does not detect the incorrect use of such interfaces. Continuing to use such interfaces after the object is destroyed will cause your application to crash or behave in unpredictable ways. <p> We recommend that you explicitly set both the primary object interface and all associated interfaces to NULL as part of your object destruction sequence, to prevent the accidental misuse of a stale interface handle. <h3>Stereo panning</h3> When <code>Volume::EnableStereoPosition</code> is used to enable stereo panning of a mono source, there is a 3 dB reduction in total <a href="http://en.wikipedia.org/wiki/Sound_power_level"> sound power level</a>. This is needed to permit the total sound power level to remain constant as the source is panned from one channel to the other. Therefore, don't enable stereo positioning if you don't need it. See the Wikipedia article on <a href="http://en.wikipedia.org/wiki/Panning_(audio)">audio panning</a> for more information. <h3>Callbacks and threads</h3> Callback handlers are generally called <i>synchronously</i> with respect to the event, that is, at the moment and location where the event is detected by the implementation. But this point is <i>asynchronous</i> with respect to the application. Thus you should use a mutex or other synchronization mechanism to control access to any variables shared between the application and the callback handler. In the example code, such as for buffer queues, we have omitted this synchronization in the interest of simplicity. However, proper mutual exclusion would be critical for any production code. <p> Callback handlers are called from internal non-application thread(s) which are not attached to the Dalvik virtual machine and thus are ineligible to use JNI. Because these internal threads are critical to the integrity of the OpenSL ES implementation, a callback handler should also not block or perform excessive work. Therefore, if your callback handler needs to use JNI or do anything significant (e.g. beyond an <code>Enqueue</code> or something else simple such as the "Get" family), the handler should instead post an event for another thread to process. <p> Note that the converse is safe: a Dalvik application thread which has entered JNI is allowed to directly call OpenSL ES APIs, including those which block. However, blocking calls are not recommended from the main thread, as they may result in the dreaded "Application Not Responding" (ANR). <h3>Performance</h3> As OpenSL ES is a native C API, non-Dalvik application threads which call OpenSL ES have no Dalvik-related overhead such as garbage collection pauses. However, there is no additional performance benefit to the use of OpenSL ES other than this. In particular, use of OpenSL ES does not result in lower audio latency, higher scheduling priority, etc. than what the platform generally provides. On the other hand, as the Android platform and specific device implementations continue to evolve, an OpenSL ES application can expect to benefit from any future system performance improvements. <h3>Security and permissions</h3> As far as who can do what, security in Android is done at the process level. Java programming language code can't do anything more than native code, nor can native code do anything more than Java code. The only differences between them are what APIs are available that provide functionality that the platform promises to support in the future and across different devices. <p> Applications using OpenSL ES must request whatever permissions they would need for similar non-native APIs. For example, if your application records audio, then it needs the <code>android.permission.RECORD_AUDIO</code> permission. Applications that use audio effects need <code>android.permission.MODIFY_AUDIO_SETTINGS</code>. Applications that play network URI resources need <code>android.permission.NETWORK</code>. <p> Media content parsers and software codecs run within the context of the Android application that calls OpenSL ES (hardware codecs are abstracted, but are device-dependent). Malformed content designed to exploit parser and codec vulnerabilities is a known attack vector. We recommend that you play media only from trustworthy sources, or that you partition your application such that code that handles media from untrustworthy sources runs in a relatively sandboxed environment. For example you could process media from untrustworthy sources in a separate process. Though both processes would still run under the same UID, this separation does make an attack more difficult. <h2>Platform issues</h2> This section describes known issues in the initial platform release which supports these APIs. <h3>Dynamic interface management</h3> <code>DynamicInterfaceManagement::AddInterface</code> does not work. Instead, specify the interface in the array passed to Create, as shown in the example code for environmental reverb. <h2>References and resources</h2> Google Android: <ul> <li><a href="http://developer.android.com/resources/index.html"> Android developer resources</a> <li><a href="http://groups.google.com/group/android-developers"> Android developers discussion group</a> <li><a href="http://developer.android.com/sdk/ndk/index.html">Android NDK</a> <li><a href="http://groups.google.com/group/android-ndk"> Android NDK discussion group</a> (for developers of native code, including OpenSL ES) <li><a href="http://code.google.com/p/android/issues/"> Android open source bug database</a> </ul> Khronos Group: <ul> <li><a href="http://www.khronos.org/opensles/"> Khronos Group OpenSL ES Overview</a> <li><a href="http://www.khronos.org/registry/sles/"> Khronos Group OpenSL ES 1.0.1 specification</a> <li><a href="http://www.khronos.org/message_boards/viewforum.php?f=15"> Khronos Group public message board for OpenSL ES</a> (please limit to non-Android questions) </ul> For convenience, we have included a copy of the OpenSL ES 1.0.1 specification with the NDK in <code>docs/opensles/OpenSL_ES_Specification_1.0.1.pdf</code>. <p> Miscellaneous: <ul> <li><a href="http://en.wikipedia.org/wiki/Java_Native_Interface">JNI</a> <li><a href="http://stackoverflow.com/search?q=android+audio"> Stack Overflow</a> <li>web search for "interactive audio", "game audio", "sound design", "audio programming", "audio content", "audio formats", etc. </ul> </body> </html>