diff --git a/example/index.html b/example/index.html index fda0b32..bf64265 100644 --- a/example/index.html +++ b/example/index.html @@ -118,6 +118,15 @@ } async function main() { + // initialize tfjs + log('FaceAPI Test'); + await faceapi.tf.setBackend('webgl'); + await faceapi.tf.enableProdMode(); + await faceapi.tf.ENV.set('DEBUG', false); + log(`TensorFlow/JS Version: ${faceapi.tf.version_core} Backend ${faceapi.tf.getBackend()}`); + + // load face-api models + log('Loading FaceAPI models'); await faceapi.nets.tinyFaceDetector.load(modelPath); await faceapi.nets.ssdMobilenetv1.load(modelPath); await faceapi.nets.ageGenderNet.load(modelPath); @@ -127,15 +136,11 @@ const optionsTinyFace = new faceapi.TinyFaceDetectorOptions({ inputSize: imgSize, scoreThreshold: minScore }); const optionsSSDMobileNet = new faceapi.SsdMobilenetv1Options({ minConfidence: minScore, maxResults }); - // initialize tfjs - await faceapi.tf.setBackend('webgl'); - await faceapi.tf.enableProdMode(); - faceapi.tf.ENV.set('DEBUG', false); - log('FaceAPI Test'); - log(`TensorFlow/JS Version: ${faceapi.tf.version_core} Backend ${faceapi.tf.getBackend()}`); + // check tf state log(`TF Flags: ${str(faceapi.tf.ENV.flags)}`); const engine = await faceapi.tf.engine(); log(`TF Engine State: ${str(engine.state)}`); + // loop through all images and try to process them log(`Start processing: ${samples.length} images ...
`); for (const img of samples) {