3 * http://uupaa.hatenablog.com/entry/2011/12/12/213233
\r
4 * Mobile Opera11 は Audio をサポートするがイベントが取れない
\r
5 * iframe 内で生成して、Audio Sprite の preset で再生できないか?
\r
7 var X_AudioSprite_shouldUse = X_HTMLAudio && ( X_UA[ 'iOS' ] || X_UA[ 'AOSP' ] || X_UA[ 'OperaMobile' ] || X_UA[ 'OperaTablet' ] ), // Flash がない
\r
8 X_AudioSprite_useVideoForMulti = //( 3.1 <= X_UA[ 'AOSP' ] < 4 ) ||
\r
9 //( ( 4.2 <= X_UA[ 'AOSP' ] ),
\r
10 // ドスパラパッドはビデオのインライン再生が不可
\r
12 X_AudioSprite_disableMultiTrack = !X_WebAudio && ( X_UA[ 'iOS' ] || 4 <= X_UA[ 'AOSP' ] || X_UA[ 'ChromeWV' ] || ( X_UA[ 'WinPhone' ] && X_UA[ 'IE9' ] ) ),
\r
13 X_AudioSprite_enableVolume = X_HTMLAudio && ( !X_UA[ 'iOS' ] && !X_UA[ 'AOSP' ] && !X_UA[ 'OperaMobile' ] && !X_UA[ 'OperaTablet' ] ), // TODO fennec は 25以上
\r
14 // http://tukumemo.com/html5-audio-sp/
\r
15 // iOS6、Android4.1から同時再生が可能になりました。
\r
16 X_AudioSprite_maxTracks = X_AudioSprite_useVideoForMulti ? 2 : X_AudioSprite_disableMultiTrack ? 1 : 9,
\r
17 X_AudioSprite_lengthSilence = 10000, // 一番最初の無音部分の長さ
\r
18 X_AudioSprite_lengthDistance = 5000, // 音間の無音の長さ
\r
19 X_AudioSprite_uid = 0,
\r
20 X_AudioSprite_members = {},
\r
21 X_AudioSprite_TEMP = {
\r
25 pauseTracks : [], // X_EVENT_DEACTIVATE によって pause した再生中のトラックたち。
\r
33 X_AudioSprite_instance,
\r
34 X_AudioSprite_numTracks,
\r
35 X_AudioSprite_useVideo;
\r
39 * urls : [ 'xx.ogg', 'xx.mp3' ],
\r
43 * BGM_01 : [ '15.00', '45.500', true, '17.666', '50.999' ],
\r
44 * BGM_02 : [ '56.00', '1:15.230', true ]
\r
47 * X_EVENT_BACKEND_READY
\r
48 * X_EVENT_BACKEND_NONE
\r
51 * X_EVENT_MEDIA_LOOPED
\r
52 * X_EVENT_MEDIA_ENDED
\r
54 * @namespace X.AudioSprite
\r
55 * @alias X.AudioSprite
\r
57 X[ 'AudioSprite' ] = function( setting ){
\r
58 var tracks = X_AudioSprite_TEMP.tracks,
\r
59 bgms = X_AudioSprite_TEMP.BGMs,
\r
60 presets = X_AudioSprite_TEMP.presets,
\r
61 urls = setting[ 'urls' ],
\r
62 video = setting[ 'useVideo' ],
\r
63 n = video ? 1 : setting[ 'numTracks' ] || 1,
\r
65 volume : setting[ 'volume' ] || 0.5,
\r
68 endTime : X_AudioSprite_lengthSilence,
\r
73 if( !X_AudioSprite_instance ){
\r
74 X_AudioSprite_instance = X_Class_override( X_EventDispatcher(), X_AudioSprite_members );
\r
75 X_ViewPort[ 'listen' ]( [ X_EVENT_VIEW_ACTIVATE, X_EVENT_VIEW_DEACTIVATE ], X_AudioSprite_instance, X_AudioSprite_handleEvent );
\r
78 n = n <= X_AudioSprite_maxTracks ? n : X_AudioSprite_maxTracks;
\r
81 // Android4.x標準ブラウザ(Chrome系)でブラウザが隠れた場合に音が鳴り続ける問題、ビデオで解決できる?
\r
82 //if( X_AudioSprite_needTouchAndroid && n === 1 ){
\r
86 for( k in setting ){
\r
88 if( X_Type_isArray( v ) && v !== urls ){
\r
89 v = X_Array_copy( v );
\r
90 for( i = v.length; i; ){
\r
92 if( i !== 2 ) v[ i ] = X_Audio_timeStringToNumber( v[ i ] );
\r
94 if( v[ 2 ] ) bgms[ k ] = v;
\r
99 X_Audio_startDetectionBackend( X_Audio_BACKENDS[ 0 ], X_AudioSprite_instance, X_Array_copy( urls ), option );
\r
101 X_AudioSprite_instance[ 'listenOnce' ]( [ X_EVENT_BACKEND_READY, X_EVENT_BACKEND_NONE ], X_AudioSprite_backendHandler );
\r
102 X_AudioSprite_instance[ 'listenOnce' ]( X_EVENT_KILL_INSTANCE, X_AudioSprite_handleEvent );
\r
104 X_AudioSprite_useVideo = video;
\r
105 X_AudioSprite_numTracks = X_AudioSprite_instance[ 'numTracks' ] = n;
\r
107 return X_AudioSprite_instance;
\r
110 X[ 'AudioSprite' ][ 'shouldUse' ] = X_AudioSprite_shouldUse;
\r
111 X[ 'AudioSprite' ][ 'enableMultiTrack' ] = !X_AudioSprite_disableMultiTrack;
\r
113 // 再生が終わっているもの、終わりかけのものを探す
\r
114 // TODO 終わりかけのもの、と一番古いもの、どちらを再利用するか?これ以上に細かい実装を望む場合は X.AudioSprite は使わず自力で実装
\r
115 function X_AudioSprite_getTrackEnded(){
\r
116 var tracks = X_AudioSprite_TEMP.tracks,
\r
118 i = 0, track, state, last = 1 / 0, _last, index;
\r
120 for( ; i < l; ++i ){
\r
121 track = tracks[ i ];
\r
122 state = track.getState();
\r
123 if( !state.playing ) return track;
\r
124 if( track === X_AudioSprite_TEMP.bgmTrack ) continue;
\r
125 if( state.currentTime <= X_AudioSprite_lengthSilence + X_AudioSprite_lengthDistance ) return track;
\r
126 _last = state.endTime - state.currentTime;
\r
127 if( _last < last ){
\r
132 return tracks[ index ];
\r
135 X_AudioSprite_members =
\r
136 /** @lends X.AudioSprite.prototype */
\r
144 * モバイル用タッチイベント中に呼び出す
\r
146 'on1stTouch' : function(){
\r
147 var i = 0, l = X_AudioSprite_TEMP.tracks.length;
\r
149 for( ; i < l; ++i ){
\r
150 //if( X_UA[ 'iOS' ] ){
\r
151 // X_AudioSprite_TEMP.tracks[ i ][ '_rawObject' ].load();
\r
153 X_AudioSprite_instance[ 'pause' ]( i );
\r
160 * @param {string} name トラック名
\r
161 * @return {number} uid
\r
163 'play' : function( name ){
\r
164 var bgm = X_AudioSprite_TEMP.bgmTrack,
\r
165 tracks = X_AudioSprite_TEMP.tracks,
\r
166 bgms = X_AudioSprite_TEMP.BGMs,
\r
167 presets = X_AudioSprite_TEMP.presets,
\r
168 preset = presets[ name ],
\r
172 if( bgms[ name ] ){
\r
173 if( name !== X_AudioSprite_TEMP.bgmName ){
\r
175 X_AudioSprite_TEMP.bgmName = name;
\r
176 X_AudioSprite_TEMP.bgmPosition = preset[ 0 ];
\r
177 X_AudioSprite_TEMP.bgmLooped = false;
\r
180 X_AudioSprite_TEMP.bgmPlaying = true;
\r
185 if( 1 < tracks.length ){
\r
186 track = X_AudioSprite_TEMP.bgmTrack = X_AudioSprite_getTrackEnded();
\r
188 track = X_AudioSprite_TEMP.bgmTrack = tracks[ 0 ];
\r
191 if( track[ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_WAITING, X_EVENT_MEDIA_SEEKING, X_EVENT_MEDIA_BEFORE_LOOP ], X_AudioSprite_instance, X_AudioSprite_handleEvent ).playing ){
\r
194 'looped' : X_AudioSprite_TEMP.bgmLooped,
\r
195 'currentTime' : X_AudioSprite_TEMP.bgmPosition,
\r
196 'startTime' : preset[ 0 ],
\r
197 'endTime' : preset[ 1 ],
\r
198 'loopStartTime' : preset[ 3 ],
\r
199 'loopEndTime' : preset[ 4 ]
\r
202 track.setState( { 'looped' : X_AudioSprite_TEMP.bgmLooped } );
\r
203 track.play( preset[ 0 ], preset[ 1 ], true, preset[ 3 ], preset[ 4 ] );
\r
204 track.seek( X_AudioSprite_TEMP.bgmPosition );
\r
208 if( 1 < tracks.length ){
\r
209 track = X_AudioSprite_getTrackEnded( X_AudioSprite_TEMP.bgmPlaying );
\r
211 [ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_WAITING, X_EVENT_MEDIA_SEEKING, X_EVENT_MEDIA_BEFORE_LOOP ], X_AudioSprite_instance, X_AudioSprite_handleEvent )
\r
212 .setState( { 'looped' : false } );
\r
213 track.play( preset[ 0 ], preset[ 1 ], true, 0, X_AudioSprite_lengthSilence );
\r
215 // single track, iOS
\r
217 X_AudioSprite_TEMP.bgmPosition = bgm.currentTime();
\r
218 //console.log( 'bgm position : ' + X_AudioSprite_TEMP.bgmPosition + ' isPlay:' + bgm.playing );
\r
219 X_AudioSprite_TEMP.bgmTrack = null;
\r
221 track = tracks[ 0 ];
\r
223 if( track[ 'listen' ]( [ X_EVENT_MEDIA_PLAYING, X_EVENT_MEDIA_WAITING, X_EVENT_MEDIA_SEEKING, X_EVENT_MEDIA_BEFORE_LOOP ], X_AudioSprite_instance, X_AudioSprite_handleEvent ).playing ){
\r
227 //'currentTime' : preset[ 0 ],
\r
228 'startTime' : preset[ 0 ],
\r
229 'endTime' : preset[ 1 ],
\r
230 'loopStartTime' : 0,
\r
231 'loopEndTime' : X_AudioSprite_lengthSilence
\r
235 track.play( preset[ 0 ], preset[ 1 ], true, 0, X_AudioSprite_lengthSilence );
\r
239 return tracks.indexOf( track );
\r
246 * @param {number} uid トラックID
\r
247 * @return {number} uid
\r
249 'pause' : function( uid ){
\r
250 var track = X_AudioSprite_TEMP.tracks[ uid ];
\r
252 if( X_AudioSprite_TEMP.bgmTrack === track ){
\r
253 X_AudioSprite_TEMP.bgmPosition = track.currentTime();
\r
254 X_AudioSprite_TEMP.bgmPlaying = false;
\r
255 X_AudioSprite_TEMP.bgmTrack = null;
\r
257 track && track.play( 0, X_AudioSprite_lengthSilence, true, 0, X_AudioSprite_lengthSilence );
\r
258 track && track.seek( 0 );
\r
259 X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_MEDIA_PAUSED );
\r
260 return X_AudioSprite_instance;
\r
265 * @param {number} uid トラックID
\r
266 * @param {number} position ms
\r
267 * @return {AudioSprite}
\r
269 'seek' : function( uid, position ){
\r
270 var track = X_AudioSprite_TEMP.tracks[ uid ],
\r
273 delete track.seekTime;
\r
274 end = X_Audio_getEndTime( track );
\r
275 start = X_Audio_getStartTime( track, end );
\r
276 0 <= position && position <= ( end - start ) && track.seek( start + position );
\r
278 return X_AudioSprite_instance;
\r
283 * @param {number} uid トラックID
\r
284 * @param {number} opt_volume= ボリューム
\r
285 * @return {AudioSprite|number}
\r
287 'volume' : function( uid, opt_volume ){
\r
291 if( opt_volume === undefined ){
\r
292 return X_AudioSprite_TEMP.volume;
\r
294 for( i = X_AudioSprite_TEMP.tracks.length; i; ){
\r
295 X_AudioSprite_TEMP.tracks[ --i ].volume( opt_volume );
\r
297 return X_AudioSprite_instance;
\r
299 track = X_AudioSprite_TEMP.tracks[ uid ];
\r
300 if( opt_volume === undefined ){
\r
301 return track ? track.gain : -1;
\r
303 track && track.volume( opt_volume );
\r
304 return X_AudioSprite_instance;
\r
309 * @param {number} uid トラックID
\r
310 * @param {object} opt_obj= 上書きする状態を書き込んだオブジェクト
\r
311 * @return {AudioSprite|object}
\r
313 'state' : function( uid, opt_obj ){
\r
314 var track = X_AudioSprite_TEMP.tracks[ uid ],
\r
317 if( opt_obj === undefined ){
\r
320 state = track.getState();
\r
321 start = state.startTime;
\r
323 'currentTime' : state.currentTime - start,
\r
324 'playing' : start <= state.currentTime && state.currentTime <= state.endTime,
\r
325 'duration' : state.endTime - start,
\r
326 'volume' : X_AudioSprite_TEMP.volume
\r
329 return { 'volume' : X_AudioSprite_TEMP.volume, 'playing' : false };
\r
331 track && track.setState( opt_obj );
\r
332 return X_AudioSprite_instance;
\r
336 function X_AudioSprite_backendHandler( e ){
\r
337 var i, backend, option, src, name, last, _e;
\r
340 case X_EVENT_BACKEND_READY :
\r
342 backend = X_Audio_BACKENDS[ e[ 'backendID' ] ];
\r
343 option = e[ 'option' ];
\r
345 X_AudioSprite_instance[ 'unlisten' ]( X_EVENT_BACKEND_NONE, X_AudioSprite_backendHandler );
\r
346 X_AudioSprite_instance[ 'source' ] = src = e[ 'source' ];
\r
347 X_AudioSprite_instance[ 'backendName' ] = name = backend.backendName;
\r
349 //console.log( i + ' / ' + X_AudioSprite_numTracks );
\r
351 for( i = 0; i < X_AudioSprite_numTracks; ++i ){
\r
352 if( X_AudioSprite_useVideo || ( i === 1 && X_AudioSprite_useVideoForMulti ) ){
\r
353 option[ 'useVideo' ] = true;
\r
354 console.log( 'use video' );
\r
356 // Audiobackend の owner として null を渡すとAudioBackend 自身へ dispatch する
\r
357 X_AudioSprite_TEMP.tracks.push( last = backend.klass( null, e[ 'source' ], option )[ 'listen' ]( X_EVENT_DEBUG, X_AudioSprite_instance, X_AudioSprite_handleEvent ) );
\r
361 'type' : X_EVENT_BACKEND_READY,
\r
363 'backendName' : name
\r
365 // touch 可能で backend ready
\r
367 if( name === 'WebAudio' ){
\r
368 if( _e[ 'needTouchForPlay' ] = X_WebAudio_need1stTouch ){
\r
369 last[ 'listenOnce' ]( X_EVENT_READY, X_AudioSprite_instance, X_AudioSprite_instance[ 'asyncDispatch' ], [ _e ] );
\r
371 X_AudioSprite_instance[ 'asyncDispatch' ]( _e );
\r
373 last[ 'listenOnce' ]( X_EVENT_READY, X_AudioSprite_instance, X_AudioSprite_backendHandler );
\r
375 if( name === 'HTMLAudio' ){
\r
376 if( _e[ 'needTouchForLoad' ] = X_HTMLAudio_need1stTouch ){
\r
377 last[ 'listenOnce' ]( X_EVENT_MEDIA_CAN_TOUCH_LOAD, X_AudioSprite_instance, X_AudioSprite_instance[ 'asyncDispatch' ], [ _e ] );
\r
378 last[ 'listenOnce' ]( !X_HTMLAudio_durationFix ? X_EVENT_MEDIA_PLAYING : X_EVENT_READY, X_AudioSprite_instance, X_AudioSprite_backendHandler );
\r
380 X_AudioSprite_instance[ 'asyncDispatch' ]( _e );
\r
381 last[ 'listenOnce' ]( X_EVENT_READY, X_AudioSprite_instance, X_AudioSprite_backendHandler );
\r
383 // READY, needTouchForPlay, needTouchForLoad
\r
384 if( X_HTMLAudio_durationFix ){
\r
385 for( i = 0; i < X_AudioSprite_TEMP.tracks.length; ++i ){
\r
386 X_AudioSprite_instance[ 'pause' ]( i );
\r
392 X_AudioSprite_instance[ 'asyncDispatch' ]( _e );
\r
394 console.log( 'AudioSprite - X_EVENT_BACKEND_READY' );
\r
396 last[ 'listenOnce' ]( X_EVENT_READY, X_AudioSprite_instance, X_AudioSprite_backendHandler );
\r
398 return X_CALLBACK_STOP_NOW;
\r
400 case X_EVENT_BACKEND_NONE :
\r
401 X_AudioSprite_instance[ 'unlisten' ]( X_EVENT_BACKEND_READY, X_AudioSprite_instance, X_AudioSprite_backendHandler )
\r
402 [ 'asyncDispatch' ]( X_EVENT_BACKEND_NONE );
\r
403 return X_CALLBACK_STOP_NOW;
\r
405 case X_EVENT_READY :
\r
406 case X_EVENT_MEDIA_PLAYING :
\r
407 console.log( 'X.AudioSprite - Ready!' );
\r
408 for( i = 0; i < X_AudioSprite_TEMP.tracks.length; ++i ){
\r
409 X_AudioSprite_instance[ 'pause' ]( i );
\r
411 X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_READY );
\r
417 function X_AudioSprite_handleEvent( e ){
\r
418 var i, tracks, track, _e, k;
\r
421 case X_EVENT_MEDIA_PLAYING :
\r
422 ( e.target === X_AudioSprite_TEMP.bgmTrack || !e.target.looped ) && X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_MEDIA_PLAYING );
\r
424 case X_EVENT_MEDIA_WAITING :
\r
425 case X_EVENT_MEDIA_SEEKING :
\r
426 ( e.target === X_AudioSprite_TEMP.bgmTrack || !e.target.looped ) && X_AudioSprite_instance[ 'asyncDispatch' ]( e.type );
\r
429 case X_EVENT_MEDIA_BEFORE_LOOP :
\r
430 if( e.target === X_AudioSprite_TEMP.bgmTrack ){
\r
431 X_AudioSprite_TEMP.bgmLooped = true;
\r
432 X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_MEDIA_LOOPED ); // TODO uid
\r
434 if( e.target.looped ){
\r
435 // X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_MEDIA_LOOPED ); // TODO uid
\r
437 X_AudioSprite_instance[ 'asyncDispatch' ]( X_EVENT_MEDIA_ENDED ); // TODO uid
\r
440 //console.log( '[AudioSprite] bgmPlaying:' + X_AudioSprite_TEMP.bgmPlaying + ' ' + !X_AudioSprite_TEMP.bgmTrack );
\r
442 // single track | iOS
\r
443 if( X_AudioSprite_TEMP.bgmPlaying && !X_AudioSprite_TEMP.bgmTrack ){
\r
444 X_AudioSprite_TEMP.bgmTrack = e.target;
\r
445 X_AudioSprite_instance.play( X_AudioSprite_TEMP.bgmName );
\r
446 return X_CALLBACK_PREVENT_DEFAULT;
\r
452 case X_EVENT_DEBUG :
\r
453 i = X_AudioSprite_TEMP.tracks.indexOf( e.target );
\r
455 e[ 'trackID' ] = i;
\r
456 X_AudioSprite_instance[ 'dispatch' ]( e );
\r
460 // TODO Android Firefox で アクティブ検出できない!
\r
461 case X_EVENT_VIEW_ACTIVATE :
\r
462 console.log( '■ アクティブ' );
\r
463 // track.play(); or iOS need touch??
\r
464 tracks = X_AudioSprite_TEMP.pauseTracks;
\r
465 while( tracks.length ) tracks.pop().actualPlay();
\r
468 case X_EVENT_VIEW_DEACTIVATE :
\r
469 console.log( '■ デアクティブ' );
\r
471 tracks = X_AudioSprite_TEMP.tracks;
\r
474 track = tracks[ --i ];
\r
475 track.playing && X_AudioSprite_TEMP.pauseTracks.push( track ) && track.pause();
\r
479 case X_EVENT_KILL_INSTANCE :
\r
481 while( X_AudioSprite_TEMP.tracks.length ){
\r
482 X_AudioSprite_TEMP.tracks.pop()[ 'kill' ]();
\r
485 for( k in X_AudioSprite_TEMP.bgms ){
\r
486 delete X_AudioSprite_TEMP.bgms[ k ];
\r
488 for( k in X_AudioSprite_TEMP.presets ){
\r
489 delete X_AudioSprite_TEMP.presets[ k ];
\r
492 X_AudioSprite_TEMP.bgmTrack = null;
\r
493 X_AudioSprite_TEMP.bgmPosition = 0;
\r
494 X_AudioSprite_TEMP.bgmName = '';
\r
495 X_AudioSprite_TEMP.bgmLooped = false;
\r
496 X_AudioSprite_TEMP.bgmPlaying = false;
\r
498 X_ViewPort[ 'unlisten' ]( [ X_EVENT_VIEW_ACTIVATE, X_EVENT_VIEW_DEACTIVATE ], X_AudioSprite_instance, X_AudioSprite_handleEvent );
\r