* Mobile Opera11 は Audio をサポートするがイベントが取れない\r
* iframe 内で生成して、Audio Sprite の preset で再生できないか?\r
*/\r
-var X_Audio_Sprite_shouldUse = window.HTMLAudioElement && ( X_UA.iOS || X_UA.AndroidBrowser || X_UA.OperaMobile || X_UA.OperaTablet ),\r
- X_Audio_Sprite_useVideoForMulti = 4 <= X_UA.AndroidBrowser && 534.3 < X_UA.AndroidBrowserWebkit, // ドスパラパッドはビデオのインライン再生が不可 \r
- X_Audio_Sprite_needTouchFirst = !!X_UA.iOS || X_Audio_Sprite_useVideoForMulti,\r
- X_Audio_Sprite_inTouchAction = false,\r
- X_Audio_Sprite_enableMultiTrack = !( X_UA.iOS && !X_Audio_WebAudio_context ) && !( X_UA.AndroidBrowser4 && X_UA.AndroidBrowserWebkit <= 534.3 ),\r
- X_Audio_Sprite_enableVolume = window.HTMLAudioElement && ( !X_UA.iOS && !X_UA.AndroidBrowser && !X_UA.OperaMobile && !X_UA.OperaTablet ),\r
+var X_Audio_Sprite_shouldUse = window.HTMLAudioElement && ( X_UA[ 'iOS' ] || X_UA[ 'AndroidBrowser' ] || X_UA[ 'OperaMobile' ] || X_UA[ 'OperaTablet' ] ), // Flash がない\r
+ X_Audio_Sprite_useVideoForMulti = 4 <= X_UA[ 'AndroidBrowser' ] && 534.3 < X_UA[ 'AndroidBrowserWebkit' ], // ドスパラパッドはビデオのインライン再生が不可 \r
+ X_Audio_Sprite_needTouchAndroid = X_Audio_Sprite_useVideoForMulti, \r
+ X_Audio_Sprite_needTouchFirst = X_UA[ 'iOS' ] || X_Audio_Sprite_needTouchAndroid || ( X_UA[ 'WinPhone' ] && X_UA[ 'IE9' ] ),\r
+ X_Audio_Sprite_enableMultiTrack = !( X_UA[ 'iOS' ] && !X_Audio_WebAudio_context ) && !( X_UA[ 'AndroidBrowser4' ] && X_UA[ 'AndroidBrowserWebkit' ] <= 534.3 ),\r
+ X_Audio_Sprite_enableVolume = window.HTMLAudioElement && ( !X_UA[ 'iOS' ] && !X_UA[ 'AndroidBrowser' ] && !X_UA[ 'OperaMobile' ] && !X_UA[ 'OperaTablet' ] ), // TODO fennec は 25以上\r
X_Audio_Sprite_maxTracks = !X_Audio_Sprite_enableMultiTrack ? 1 : X_Audio_Sprite_useVideoForMulti ? 2 : 9,\r
X_Audio_Sprite_lengthSilence = 10000, // 一番最初の無音部分の長さ\r
X_Audio_Sprite_lengthDistance = 5000, // 音間の無音の長さ\r
presets : {},\r
BGMs : {},\r
tracks : [],\r
- pauseTracks : [], // X.Event.DEACTIVATE によって pause した再生中のトラックたち。\r
+ pauseTracks : [], // X_EVENT_DEACTIVATE によって pause した再生中のトラックたち。\r
volume : 1,\r
bgmTrack : null,\r
bgmPosition : 0,\r
},\r
X_Audio_Sprite_instance;\r
\r
-X.Audio.Sprite = {\r
+X[ 'Audio' ][ 'Sprite' ] = {\r
\r
- shouldUse : X_Audio_Sprite_shouldUse,\r
+ 'shouldUse' : X_Audio_Sprite_shouldUse,\r
\r
- needTouchFirst : X_Audio_Sprite_needTouchFirst,\r
+ 'needTouchFirst' : X_Audio_Sprite_needTouchFirst,\r
\r
- enableMultiTrack : X_Audio_Sprite_enableMultiTrack,\r
+ 'enableMultiTrack' : X_Audio_Sprite_enableMultiTrack,\r
\r
- create : function( setting ){\r
+ 'create' : function( setting ){\r
// close()\r
if( X_Audio_Sprite_instance ){\r
X_Audio_Sprite_instance.close();\r
} else {\r
- X_Audio_Sprite_instance = X_Class_override( new X.EventDispatcher(), X_Audio_Sprite_members );\r
- X_ViewPort.listen( [ X.Event.VIEW_ACTIVATE, X.Event.VIEW_DEACTIVATE ], X_Audio_Sprite_instance, X_Audio_Sprite_handleEvent );\r
+ X_Audio_Sprite_instance = X_Class_override( X_EventDispatcher(), X_Audio_Sprite_members );\r
+ X_ViewPort[ 'listen' ]( [ X_EVENT_VIEW_ACTIVATE, X_EVENT_VIEW_DEACTIVATE ], X_Audio_Sprite_instance, X_Audio_Sprite_handleEvent );\r
};\r
X_Audio_Sprite_instance.setup( setting );\r
return X_Audio_Sprite_instance;\r
* BGM_01 : [ '15.00', '45.500', true, '17.666', '50.999' ],\r
* BGM_02 : [ '56.00', '1:15.230', true ]\r
* }\r
+ * \r
+ * X_EVENT_BACKEND_READY\r
+ * X_EVENT_BACKEND_NONE\r
+ * \r
+ * X_EVENT_READY\r
+ * X_EVENT_MEDIA_LOOPED\r
+ * X_EVENT_MEDIA_ENDED\r
+ * \r
*/\r
\r
X_Audio_Sprite_members = {\r
\r
for( k in setting ){\r
v = setting[ k ];\r
- if( X.Type.isArray( v ) && v !== urls){\r
- v = X.Object.cloneArray( v );\r
+ if( X_Type_isArray( v ) && v !== urls ){\r
+ v = X_Object_cloneArray( v );\r
for( i = v.length; i; ){\r
--i;\r
if( i !== 2 ) v[ i ] = X_AudioWrapper_timeStringToNumber( v[ i ] );\r
if( video || ( i === 1 && X_Audio_Sprite_useVideoForMulti ) ){\r
option[ 'useVideo' ] = true;\r
};\r
- tracks.push( X.Audio.create( urls, X.Object.clone( option ) ) );\r
+ tracks.push( X[ 'Audio' ]( urls, X_Object_clone( option ) ) );\r
};\r
\r
- tracks[ n - 1 ].listenOnce( [ 'backendfound', 'nobackend' ], this, X_Audio_Sprite_handleEvent );\r
+ tracks[ n - 1 ][ 'listenOnce' ]( [ X_EVENT_BACKEND_READY, X_EVENT_BACKEND_NONE ], this, X_Audio_Sprite_handleEvent );\r
\r
X_Audio_Sprite_instance.numTracks = n;\r
},\r
k;\r
\r
while( tracks.length ){\r
- tracks.pop().kill();\r
+ tracks.pop()[ 'kill' ]();\r
};\r
\r
for( k in bgms ){\r
var tracks = X_Audio_Sprite_TEMP.tracks,\r
i = 0, l = tracks.length;\r
for( ; i < l; ++i ){\r
- X_AudioProxy_getAudioWrapper( tracks[ i ] )._rawObject.load();\r
+ if( X_UA[ 'WinPhone' ] ){\r
+ console.log( 'touch -> play()' );\r
+ //tracks[ i ].play( 0, X_Audio_Sprite_lengthSilence, true, 0, X_Audio_Sprite_lengthSilence ).seek( 0 );\r
+ this.pause( i );\r
+ } else {\r
+ X_Audio_getAudioWrapper( tracks[ i ] )[ '_rawObject' ].load();\r
+ };\r
};\r
},\r
\r
bgms = X_Audio_Sprite_TEMP.BGMs,\r
presets = X_Audio_Sprite_TEMP.presets,\r
preset = presets[ name ],\r
- i, k;\r
+ track, i, k;\r
\r
if( preset ){\r
if( bgms[ name ] ){\r
track = X_Audio_Sprite_TEMP.bgmTrack = tracks[ 0 ];\r
};\r
\r
- if( track.listen( 'looped', this, X_Audio_Sprite_handleEvent ).isPlaying() ){\r
+ if( track[ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_BEFORE_LOOP ], this, X_Audio_Sprite_handleEvent ).isPlaying() ){\r
track\r
.state( {\r
loop : true,\r
if( 1 < tracks.length ){\r
track = X_Audio_Sprite_getTrackEnded( X_Audio_Sprite_TEMP.bgmPlaying );\r
track\r
- .listen( 'looped', this, X_Audio_Sprite_handleEvent )\r
+ [ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_BEFORE_LOOP ], this, X_Audio_Sprite_handleEvent )\r
.state( { looped : false } )\r
.play( preset[ 0 ], preset[ 1 ], true, 0, X_Audio_Sprite_lengthSilence );\r
} else {\r
};\r
track = tracks[ 0 ];\r
\r
- if( track.listen( 'looped', this, X_Audio_Sprite_handleEvent ).isPlaying() ){\r
+ if( track[ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_BEFORE_LOOP ], this, X_Audio_Sprite_handleEvent ).isPlaying() ){\r
track\r
.state( {\r
loop : true,\r
X_Audio_Sprite_TEMP.bgmPlaying = false;\r
X_Audio_Sprite_TEMP.bgmTrack = null;\r
};\r
- console.log( 'pause' );\r
track && track.play( 0, X_Audio_Sprite_lengthSilence, true, 0, X_Audio_Sprite_lengthSilence ).seek( 0 );\r
+ this[ 'asyncDispatch' ]( X_EVENT_MEDIA_PAUSED );\r
return this;\r
},\r
\r
},\r
\r
state : function( uid, opt_obj ){\r
- var track = X_Audio_Sprite_TEMP.tracks[ uid ];\r
+ var track = X_Audio_Sprite_TEMP.tracks[ uid ],\r
+ state, start, end;\r
// TODO uid = 0\r
if( opt_obj === undefined ){\r
- return track ? track.state() : { volume : X_Audio_Sprite_TEMP.volume };\r
+ // TODO pause\r
+ if( track ){\r
+ state = track.state();\r
+ start = state.startTime;\r
+ return {\r
+ 'currentTime' : state.currentTime - state.startTime,\r
+ 'playing' : state.startTime <= state.currentTime && state.currentTime <= state.endTime,\r
+ 'duration' : state.endTime - state.startTime,\r
+ 'volume' : X_Audio_Sprite_TEMP.volume\r
+ };\r
+ };\r
+ return { 'volume' : X_Audio_Sprite_TEMP.volume, 'playing' : false };\r
};\r
track && track.state( opt_obj );\r
return this;\r
};\r
\r
function X_Audio_Sprite_handleEvent( e ){\r
- var i, tracks, track;\r
+ var i, tracks, track, _e;\r
+ \r
switch( e.type ){\r
- case 'backendfound' :\r
- this.asyncDispatch( e );\r
- e.target.unlisten( 'nobackend', this, X_Audio_Sprite_handleEvent );\r
+ case X_EVENT_BACKEND_READY :\r
+ _e = {\r
+ 'type' : X_EVENT_BACKEND_READY,\r
+ 'source' : e[ 'source' ],\r
+ 'backendName' : e[ 'backendName' ]\r
+ };\r
\r
- if( e.backendName === 'HTML Audio' ){\r
- e.target.listenOnce( [ X_Audio_HTMLAudio_playTrigger, 'loadeddata' ], this, X_Audio_Sprite_handleEvent );\r
- } else {\r
- e.target.listenOnce( 'canplaythrough', this, X_Audio_Sprite_handleEvent );\r
+ if( X_Audio_Sprite_needTouchFirst ){\r
+ if( e.backendName === 'Web Audio' ){\r
+ _e[ 'needTouchForPlay' ] = true;\r
+ } else {\r
+ _e[ 'needTouchForLoad' ] = true;\r
+ };\r
+ };\r
+ this[ 'asyncDispatch' ]( _e );\r
+ \r
+ e.target\r
+ [ 'unlisten' ]( X_EVENT_BACKEND_NONE, this, X_Audio_Sprite_handleEvent )\r
+ [ 'listenOnce' ]( X_EVENT_READY, this, X_Audio_Sprite_handleEvent );\r
+\r
+ // READY, needTouchForPlay, needTouchForLoad\r
+ if( X_Audio_HTMLAudioWrapper_durationFix ){\r
+ for( i = 0; i < X_Audio_Sprite_TEMP.tracks.length; ++i ){\r
+ X_Audio_Sprite_instance.pause( i );\r
+ };\r
};\r
break;\r
\r
- case 'nobackend' :\r
- this.asyncDispatch( e );\r
- e.target.unlisten( 'backendfound', this, X_Audio_Sprite_handleEvent );\r
+ case X_EVENT_BACKEND_NONE :\r
+ this[ 'asyncDispatch' ]( X_EVENT_BACKEND_NONE );\r
+ e.target[ 'unlisten' ]( X_EVENT_BACKEND_READY, this, X_Audio_Sprite_handleEvent );\r
break;\r
\r
- case 'canplaythrough' :\r
- case X_Audio_HTMLAudio_playTrigger :\r
- case 'loadeddata' :\r
- e.target.unlisten( [ X_Audio_HTMLAudio_playTrigger, 'loadeddata', 'canplaythrough' ], this, X_Audio_Sprite_handleEvent );\r
- \r
- if( X_Audio_Sprite_useVideoForMulti ){\r
+ case X_EVENT_READY :\r
+ console.log( 'X.AudioSprite - Ready!' );\r
+ if( X_Audio_Sprite_needTouchAndroid ){\r
for( i = 0; i < X_Audio_Sprite_TEMP.tracks.length; ++i ){\r
X_Audio_Sprite_instance.pause( i );\r
};\r
- e.target.listenOnce( 'timeupdate', this, X_Audio_Sprite_handleEvent ); // Android 標準ブラウザ\r
+ e.target[ 'listenOnce' ]( X_EVENT_MEDIA_PLAYING, this, this.asyncDispatch, [ X_EVENT_READY ] ); // Android 標準ブラウザ\r
return;\r
};\r
- case 'timeupdate' :\r
- this.asyncDispatch( 'audioSpriteCanPlay' );\r
+ this[ 'asyncDispatch' ]( X_EVENT_READY );\r
+ break;\r
+ \r
+ case X_EVENT_MEDIA_PLAYING :\r
+ ( e.target === X_Audio_Sprite_TEMP.bgmTrack || !e.target.state().looped ) && this[ 'asyncDispatch' ]( X_EVENT_MEDIA_PLAYING );\r
break;\r
\r
- case 'looped' :\r
+ case X_EVENT_MEDIA_BEFORE_LOOP :\r
if( e.target === X_Audio_Sprite_TEMP.bgmTrack ){\r
X_Audio_Sprite_TEMP.bgmLooped = true;\r
+ this[ 'asyncDispatch' ]( X_EVENT_MEDIA_LOOPED ); // TODO uid\r
} else {\r
+ if( e.target.state().looped ){\r
+ //this[ 'asyncDispatch' ]( X_EVENT_MEDIA_LOOPED ); // TODO uid\r
+ } else {\r
+ this[ 'asyncDispatch' ]( X_EVENT_MEDIA_ENDED ); // TODO uid\r
+ };\r
+ \r
// single track | iOS\r
if( X_Audio_Sprite_TEMP.bgmPlaying && !X_Audio_Sprite_TEMP.bgmTrack ){\r
X_Audio_Sprite_TEMP.bgmTrack = e.target;\r
};\r
break;\r
\r
- case X.Event.VIEW_ACTIVATE :\r
+ case X_EVENT_VIEW_ACTIVATE :\r
+ console.log( '■ アクティブ' );\r
// track.play(); or iOS need touch??\r
tracks = X_Audio_Sprite_TEMP.pauseTracks;\r
while( tracks.length ) tracks.pop().play();\r
break;\r
\r
- case X.Event.VIEW_DEACTIVATE :\r
+ case X_EVENT_VIEW_DEACTIVATE :\r
+ console.log( '■ デアクティブ' );\r
// track.pause();\r
tracks = X_Audio_Sprite_TEMP.tracks;\r
i = tracks.length;\r
};\r
break;\r
\r
- case X.Event.KILL_INSTANCE :\r
- X_ViewPort.unlisten( [ X.Event.VIEW_ACTIVATE, X.Event.VIEW_DEACTIVATE ], this, X_Audio_Sprite_handleEvent );\r
+ case X_EVENT_KILL_INSTANCE :\r
+ X_ViewPort[ 'unlisten' ]( [ X_EVENT_VIEW_ACTIVATE, X_EVENT_VIEW_DEACTIVATE ], this, X_Audio_Sprite_handleEvent );\r
this.close();\r
break;\r
};\r