3 * http://uupaa.hatenablog.com/entry/2011/12/12/213233
\r
4 * Mobile Opera11 は Audio をサポートするがイベントが取れない
\r
5 * iframe 内で生成して、Audio Sprite の preset で再生できないか?
\r
7 var X_Audio_Sprite_shouldUse = window.HTMLAudioElement && ( X_UA.iOS || X_UA.AndroidBrowser || X_UA.OperaMobile || X_UA.OperaTablet ),
\r
8 X_Audio_Sprite_useVideoForMulti = 4 <= X_UA.AndroidBrowser && 534.3 < X_UA.AndroidBrowserWebkit, // ドスパラパッドはビデオのインライン再生が不可
\r
9 X_Audio_Sprite_needTouchAndroid = X_Audio_Sprite_useVideoForMulti,
\r
10 X_Audio_Sprite_needTouchFirst = !!X_UA.iOS || X_Audio_Sprite_useVideoForMulti,
\r
11 X_Audio_Sprite_enableMultiTrack = !( X_UA.iOS && !X_Audio_WebAudio_context ) && !( X_UA.AndroidBrowser4 && X_UA.AndroidBrowserWebkit <= 534.3 ),
\r
12 X_Audio_Sprite_enableVolume = window.HTMLAudioElement && ( !X_UA.iOS && !X_UA.AndroidBrowser && !X_UA.OperaMobile && !X_UA.OperaTablet ),
\r
13 X_Audio_Sprite_maxTracks = !X_Audio_Sprite_enableMultiTrack ? 1 : X_Audio_Sprite_useVideoForMulti ? 2 : 9,
\r
14 X_Audio_Sprite_lengthSilence = 10000, // 一番最初の無音部分の長さ
\r
15 X_Audio_Sprite_lengthDistance = 5000, // 音間の無音の長さ
\r
16 X_Audio_Sprite_uid = 0,
\r
17 X_Audio_Sprite_members = {},
\r
18 X_Audio_Sprite_TEMP = {
\r
22 pauseTracks : [], // X.Event.DEACTIVATE によって pause した再生中のトラックたち。
\r
30 X_Audio_Sprite_instance;
\r
34 shouldUse : X_Audio_Sprite_shouldUse,
\r
36 needTouchFirst : X_Audio_Sprite_needTouchFirst,
\r
38 enableMultiTrack : X_Audio_Sprite_enableMultiTrack,
\r
40 create : function( setting ){
\r
42 if( X_Audio_Sprite_instance ){
\r
43 X_Audio_Sprite_instance.close();
\r
45 X_Audio_Sprite_instance = X_Class_override( new X.EventDispatcher(), X_Audio_Sprite_members );
\r
46 X_ViewPort.listen( [ X.Event.VIEW_ACTIVATE, X.Event.VIEW_DEACTIVATE ], X_Audio_Sprite_instance, X_Audio_Sprite_handleEvent );
\r
48 X_Audio_Sprite_instance.setup( setting );
\r
49 return X_Audio_Sprite_instance;
\r
54 // 再生が終わっているもの、終わりかけのものを探す
\r
55 // TODO 終わりかけのもの、と一番古いもの、どちらを再利用するか?これ以上に細かい実装を望む場合は X.Audio.Sprite は使わず自力で実装
\r
56 function X_Audio_Sprite_getTrackEnded(){
\r
57 var tracks = X_Audio_Sprite_TEMP.tracks,
\r
59 i = 0, track, state, last = 1 / 0, _last, index;
\r
61 for( ; i < l; ++i ){
\r
62 track = tracks[ i ];
\r
63 state = track.state();
\r
64 if( !state.playing ) return track;
\r
65 if( track === X_Audio_Sprite_TEMP.bgmTrack ) continue;
\r
66 if( state.currentTime <= X_Audio_Sprite_lengthSilence + X_Audio_Sprite_lengthDistance ) return track;
\r
67 _last = state.endTime - state.currentTime;
\r
73 return tracks[ index ];
\r
78 * urls : [ 'xx.ogg', 'xx.mp3' ],
\r
82 * BGM_01 : [ '15.00', '45.500', true, '17.666', '50.999' ],
\r
83 * BGM_02 : [ '56.00', '1:15.230', true ]
\r
87 X_Audio_Sprite_members = {
\r
89 setup : function( setting ){
\r
91 var tracks = X_Audio_Sprite_TEMP.tracks,
\r
92 bgms = X_Audio_Sprite_TEMP.BGMs,
\r
93 presets = X_Audio_Sprite_TEMP.presets,
\r
94 urls = setting[ 'urls' ],
\r
95 video = setting[ 'useVideo' ],
\r
96 n = video ? 1 : setting[ 'numTracks' ] || 1,
\r
98 volume : setting[ 'volume' ] || 0.5,
\r
101 endTime : X_Audio_Sprite_lengthSilence,
\r
106 n = n <= X_Audio_Sprite_maxTracks ? n : X_Audio_Sprite_maxTracks;
\r
108 for( k in setting ){
\r
110 if( X.Type.isArray( v ) && v !== urls){
\r
111 v = X.Object.cloneArray( v );
\r
112 for( i = v.length; i; ){
\r
114 if( i !== 2 ) v[ i ] = X_AudioWrapper_timeStringToNumber( v[ i ] );
\r
116 if( v[ 2 ] ) bgms[ k ] = v;
\r
121 for( i = 0; i < n; ++i ){
\r
122 if( video || ( i === 1 && X_Audio_Sprite_useVideoForMulti ) ){
\r
123 option[ 'useVideo' ] = true;
\r
125 tracks.push( X.Audio.create( urls, X.Object.clone( option ) ) );
\r
128 tracks[ n - 1 ].listenOnce( [ 'backendfound', 'nobackend' ], this, X_Audio_Sprite_handleEvent );
\r
130 X_Audio_Sprite_instance.numTracks = n;
\r
133 close : function(){
\r
134 var tracks = X_Audio_Sprite_TEMP.tracks,
\r
135 bgms = X_Audio_Sprite_TEMP.BGMs,
\r
136 presets = X_Audio_Sprite_TEMP.presets,
\r
139 while( tracks.length ){
\r
140 tracks.pop().kill();
\r
146 for( k in presets ){
\r
147 delete presets[ k ];
\r
150 X_Audio_Sprite_TEMP.bgmTrack = null;
\r
151 X_Audio_Sprite_TEMP.bgmPosition = 0;
\r
152 X_Audio_Sprite_TEMP.bgmName = '';
\r
153 X_Audio_Sprite_TEMP.bgmLooped = false;
\r
154 X_Audio_Sprite_TEMP.bgmPlaying = false;
\r
158 var tracks = X_Audio_Sprite_TEMP.tracks,
\r
159 i = 0, l = tracks.length;
\r
160 for( ; i < l; ++i ){
\r
161 X_AudioProxy_getAudioWrapper( tracks[ i ] )._rawObject.load();
\r
166 * @return uid Number
\r
168 play : function( name ){
\r
169 var bgm = X_Audio_Sprite_TEMP.bgmTrack,
\r
170 tracks = X_Audio_Sprite_TEMP.tracks,
\r
171 bgms = X_Audio_Sprite_TEMP.BGMs,
\r
172 presets = X_Audio_Sprite_TEMP.presets,
\r
173 preset = presets[ name ],
\r
177 if( bgms[ name ] ){
\r
178 if( name !== X_Audio_Sprite_TEMP.bgmName ){
\r
180 X_Audio_Sprite_TEMP.bgmName = name;
\r
181 X_Audio_Sprite_TEMP.bgmPosition = preset[ 0 ];
\r
182 X_Audio_Sprite_TEMP.bgmPlaying = true;
\r
183 X_Audio_Sprite_TEMP.bgmLooped = false;
\r
188 if( 1 < tracks.length ){
\r
189 track = X_Audio_Sprite_TEMP.bgmTrack = X_Audio_Sprite_getTrackEnded();
\r
191 track = X_Audio_Sprite_TEMP.bgmTrack = tracks[ 0 ];
\r
194 if( track.listen( 'looped', this, X_Audio_Sprite_handleEvent ).isPlaying() ){
\r
198 looped : X_Audio_Sprite_TEMP.bgmLooped,
\r
199 currentTime : X_Audio_Sprite_TEMP.bgmPosition,
\r
200 startTime : preset[ 0 ],
\r
201 endTime : preset[ 1 ],
\r
202 loopStartTime : preset[ 3 ],
\r
203 loopEndTime : preset[ 4 ]
\r
207 .state( { looped : X_Audio_Sprite_TEMP.bgmLooped } )
\r
208 .play( preset[ 0 ], preset[ 1 ], true, preset[ 3 ], preset[ 4 ] )
\r
209 .seek( X_Audio_Sprite_TEMP.bgmPosition );
\r
213 if( 1 < tracks.length ){
\r
214 track = X_Audio_Sprite_getTrackEnded( X_Audio_Sprite_TEMP.bgmPlaying );
\r
216 .listen( 'looped', this, X_Audio_Sprite_handleEvent )
\r
217 .state( { looped : false } )
\r
218 .play( preset[ 0 ], preset[ 1 ], true, 0, X_Audio_Sprite_lengthSilence );
\r
220 // single track, iOS
\r
222 X_Audio_Sprite_TEMP.bgmPosition = bgm.currentTime();
\r
223 console.log( 'bgm position : ' + X_Audio_Sprite_TEMP.bgmPosition + ' isPlay:' + bgm.isPlaying() );
\r
224 X_Audio_Sprite_TEMP.bgmTrack = null;
\r
226 track = tracks[ 0 ];
\r
228 if( track.listen( 'looped', this, X_Audio_Sprite_handleEvent ).isPlaying() ){
\r
233 startTime : preset[ 0 ],
\r
234 endTime : preset[ 1 ],
\r
236 loopEndTime : X_Audio_Sprite_lengthSilence
\r
240 .play( preset[ 0 ], preset[ 1 ], true, 0, X_Audio_Sprite_lengthSilence );
\r
244 return tracks.indexOf( track );
\r
249 pause : function( uid ){
\r
250 var track = X_Audio_Sprite_TEMP.tracks[ uid ];
\r
251 if( X_Audio_Sprite_TEMP.bgmTrack === track ){
\r
252 X_Audio_Sprite_TEMP.bgmPosition = track.currentTime();
\r
253 X_Audio_Sprite_TEMP.bgmPlaying = false;
\r
254 X_Audio_Sprite_TEMP.bgmTrack = null;
\r
256 console.log( 'pause' );
\r
257 track && track.play( 0, X_Audio_Sprite_lengthSilence, true, 0, X_Audio_Sprite_lengthSilence ).seek( 0 );
\r
261 seek : function( uid, position ){
\r
262 var track = X_Audio_Sprite_TEMP.tracks[ uid ],
\r
265 delete track.seekTime;
\r
266 end = X_AudioWrapper_getEndTime( track );
\r
267 position <= end && X_AudioWrapper_getStartTime( track, end ) <= position && track.seek( postion );
\r
272 volume : function( uid, opt_volume ){
\r
276 if( opt_volume === undefined ){
\r
277 return X_Audio_Sprite_TEMP.volume;
\r
279 for( i = X_Audio_Sprite_TEMP.tracks.length; i; ){
\r
280 X_Audio_Sprite_TEMP.tracks[ --i ].volume( opt_volume );
\r
284 track = X_Audio_Sprite_TEMP.tracks[ uid ];
\r
285 if( opt_volume === undefined ){
\r
286 return track ? track.volume() : -1;
\r
288 track && track.volume( opt_volume );
\r
292 state : function( uid, opt_obj ){
\r
293 var track = X_Audio_Sprite_TEMP.tracks[ uid ];
\r
295 if( opt_obj === undefined ){
\r
296 return track ? track.state() : { volume : X_Audio_Sprite_TEMP.volume };
\r
298 track && track.state( opt_obj );
\r
303 function X_Audio_Sprite_handleEvent( e ){
\r
304 var i, tracks, track, _e;
\r
306 case 'backendfound' :
\r
310 backendName : e.backendName
\r
313 if( X_Audio_Sprite_needTouchFirst ){
\r
314 if( e.backendName === 'Web Audio' ){
\r
315 _e[ 'needTouchForPlay' ] = true;
\r
317 _e[ 'needTouchForLoad' ] = true;
\r
320 this.asyncDispatch( _e );
\r
322 e.target.unlisten( 'nobackend', this, X_Audio_Sprite_handleEvent );
\r
324 if( e.backendName === 'HTML Audio' ){
\r
325 // Android 4.4.4 Y!Browser でたまにロードを検出できていない?問題を解決できないかな? -> , 'canplaythrough'
\r
326 // タッチイベント内でloadを開始。ロード完了したら無音の再生を実施。timeupdate が返ったらいつでも再生が出来る状態になる。
\r
327 e.target.listenOnce( [ X_Audio_HTMLAudio_playTrigger, 'loadeddata', 'canplaythrough' ], this, X_Audio_Sprite_handleEvent );
\r
329 e.target.listenOnce( 'canplaythrough', this, X_Audio_Sprite_handleEvent );
\r
331 // READY, needTouchForPlay, needTouchForLoad
\r
332 if( X_Audio_HTMLAudioWrapper_durationFix ){
\r
333 for( i = 0; i < X_Audio_Sprite_TEMP.tracks.length; ++i ){
\r
334 X_Audio_Sprite_instance.pause( i );
\r
340 this.asyncDispatch( e );
\r
341 e.target.unlisten( 'backendfound', this, X_Audio_Sprite_handleEvent );
\r
344 case 'canplaythrough' :
\r
345 case X_Audio_HTMLAudio_playTrigger :
\r
346 case 'loadeddata' :
\r
347 e.target.unlisten( [ X_Audio_HTMLAudio_playTrigger, 'loadeddata', 'canplaythrough' ], this, X_Audio_Sprite_handleEvent );
\r
349 if( X_Audio_Sprite_needTouchAndroid ){
\r
350 for( i = 0; i < X_Audio_Sprite_TEMP.tracks.length; ++i ){
\r
351 X_Audio_Sprite_instance.pause( i );
\r
353 e.target.listenOnce( 'timeupdate', this, X_Audio_Sprite_handleEvent ); // Android 標準ブラウザ
\r
356 case 'timeupdate' :
\r
357 this.asyncDispatch( 'audioSpriteCanPlay' );
\r
361 if( e.target === X_Audio_Sprite_TEMP.bgmTrack ){
\r
362 X_Audio_Sprite_TEMP.bgmLooped = true;
\r
364 e.target.state().looped || this.asyncDispatch( 'ended' ); // uid
\r
366 // single track | iOS
\r
367 if( X_Audio_Sprite_TEMP.bgmPlaying && !X_Audio_Sprite_TEMP.bgmTrack ){
\r
368 X_Audio_Sprite_TEMP.bgmTrack = e.target;
\r
369 this.play( X_Audio_Sprite_TEMP.bgmName );
\r
370 return X.Callback.PREVENT_DEFAULT;
\r
375 case X.Event.VIEW_ACTIVATE :
\r
376 // track.play(); or iOS need touch??
\r
377 tracks = X_Audio_Sprite_TEMP.pauseTracks;
\r
378 while( tracks.length ) tracks.pop().play();
\r
381 case X.Event.VIEW_DEACTIVATE :
\r
383 tracks = X_Audio_Sprite_TEMP.tracks;
\r
386 track = tracks[ --i ];
\r
387 track.isPlaying() && X_Audio_Sprite_TEMP.pauseTracks.push( track.pause() );
\r
391 case X.Event.KILL_INSTANCE :
\r
392 X_ViewPort.unlisten( [ X.Event.VIEW_ACTIVATE, X.Event.VIEW_DEACTIVATE ], this, X_Audio_Sprite_handleEvent );
\r