import { GlCanvas } from './gl-canvas.js'; import { HLS } from './hls.js'; import shader from './greyscale-shader.js'; const ingestion_url_el = document.getElementById('ingestion-url'); ingestion_url_el.value = localStorage.getItem('streamana-example-ingestion-url'); const go_live_el = document.getElementById('go-live'); go_live_el.disabled = false; go_live_el.addEventListener('click', function () { if (this.checked) { start(); } else { stop(); } }); let canvas_el = document.getElementById('canvas'); const canvas_proto = canvas_el.cloneNode(); const waiting_el = document.getElementById('waiting'); const error_alert_el = document.getElementById('error-alert'); const error_alert_el_parent = error_alert_el.parentNode; const error_alert_el_nextSibling = error_alert_el.nextSibling; error_alert_el_parent.removeChild(error_alert_el); const ffmpeg_lib_url_el = document.getElementById('ffmpeg-lib-url'); ffmpeg_lib_url_el.value = localStorage.getItem('streamana-ffmpeg-lib-url'); ffmpeg_lib_url_el.addEventListener('input', function (e) { localStorage.setItem('streamana-ffmpeg-lib-url', this.value); }); let hls; async function start() { const ingestion_url = ingestion_url_el.value.trim(); if (!ingestion_url) { go_live_el.checked = false; return; } localStorage.setItem('streamana-example-ingestion-url', ingestion_url); const ffmpeg_lib_url = ffmpeg_lib_url_el.value.trim() || ffmpeg_lib_url_el.placeholder.trim(); go_live_el.disabled = true; waiting_el.classList.remove('d-none'); const parent_el = canvas_el.parentNode; parent_el.removeChild(canvas_el); canvas_el = canvas_proto.cloneNode(); canvas_el.classList.add('invisible'); parent_el.appendChild(canvas_el); if (error_alert_el.parentNode) { error_alert_el_parent.removeChild(error_alert_el); } let camera_stream, gl_canvas, canvas_stream, done = false; function cleanup(err) { if (err) { console.error(err); } if (done) { return; } done = true; if (err) { error_alert_el_parent.insertBefore(error_alert_el, error_alert_el_nextSibling); error_alert_el.classList.add('show'); } if (camera_stream) { for (let track of camera_stream.getTracks()) { track.stop(); } } if (gl_canvas) { gl_canvas.destroy(); } if (canvas_stream) { for (let track of canvas_stream.getTracks()) { track.stop(); } } if (hls) { hls.end(!!err); } go_live_el.checked = false; go_live_el.disabled = false; waiting_el.classList.add('d-none'); canvas_el.classList.add('d-none'); } try { // create video element which will be used for grabbing the frames to // write to a canvas so we can apply webgl shaders // also used to get the native video dimensions const video_el = document.createElement('video'); video_el.muted = true; video_el.playsInline = true; // Safari on iOS requires us to play() in the click handler and doesn't // track async calls. So we play a blank video first. After that, the video // element is blessed for script-driver playback. video_el.src = 'empty.mp4'; await video_el.play(); // capture video from webcam const video_constraints = { //width: 4096, //height: 2160, width: 1280, height: 720, frameRate: { ideal: 30, max: 30 } }; try { camera_stream = await navigator.mediaDevices.getUserMedia({ audio: true, video: video_constraints }); } catch (ex) { // retry in case audio isn't available console.warn("Failed to get user media, retrying without audio"); camera_stream = await navigator.mediaDevices.getUserMedia({ audio: false, video: video_constraints }); } // use glsl-canvas to make managing webgl stuff easier // because it's not visible, client dimensions are zero so we // need to substitute actual dimensions instead gl_canvas = new GlCanvas(canvas_el, { // as an example, greyscale the stream fragmentString: shader }); gl_canvas.on('error', cleanup); // tell canvas to use frames from video gl_canvas.setTexture('u_texture', video_el); // wait for video to load (must come after gl_canvas.setTexture() since it // registers a loadeddata handler which then registers a play handler) video_el.addEventListener('loadeddata', async function () { try { // make canvas same size as native video dimensions so every pixel is seen if (this.videoWidth > this.videoHeight) { canvas_el.width = this.videoWidth; canvas_el.height = this.videoHeight; } else { canvas_el.width = this.videoHeight; canvas_el.height = this.videoWidth; } // start the camera video this.play(); // capture video from the canvas // Note: Safari on iOS doesn't get any data, might be related to // https://bugs.webkit.org/show_bug.cgi?id=181663 const frame_rate = camera_stream.getVideoTracks()[0].getSettings().frameRate; canvas_stream = canvas_el.captureStream(frame_rate); // add audio if present const audio_tracks = camera_stream.getAudioTracks(); if (audio_tracks.length > 0) { canvas_stream.addTrack(audio_tracks[0]); } // start HLS from the canvas stream to the ingestion URL hls = new HLS(canvas_stream, ingestion_url, ffmpeg_lib_url, frame_rate); hls.addEventListener('run', () => console.log('HLS running')); hls.addEventListener('exit', ev => { const msg = `HLS exited with status ${ev.detail.code}`; if (ev.detail.code === 0) { console.log(msg); cleanup(); } else { cleanup(msg); } }); hls.addEventListener('error', cleanup); hls.addEventListener('start', () => { waiting_el.classList.add('d-none'); canvas_el.classList.remove('invisible'); go_live_el.disabled = false; gl_canvas.onLoop(); }); hls.addEventListener('update', () => { gl_canvas.onLoop(); }); await hls.start(); } catch (ex) { cleanup(ex); } }); // pass the stream from the camera to the video so it can render the frames video_el.srcObject = camera_stream; } catch (ex) { return cleanup(ex); } } function stop() { go_live_el.disabled = true; hls.end(); }