8static const char *
const TAG =
"micro_wake_word";
11namespace micro_wake_word {
16 " Probability cutoff: %.2f\n"
17 " Sliding window size: %d",
18 this->
wake_word_.c_str(), this->probability_cutoff_ / 255.0f, this->sliding_window_size_);
24 " Probability cutoff: %.2f\n"
25 " Sliding window size: %d",
35 ESP_LOGE(TAG,
"Could not allocate the streaming model's variable tensor arena.");
38 this->
ma_ = tflite::MicroAllocator::Create(this->
var_arena_, STREAMING_MODEL_VARIABLE_ARENA_SIZE);
39 this->
mrv_ = tflite::MicroResourceVariables::Create(this->
ma_, 20);
42 const tflite::Model *model = tflite::GetModel(this->
model_start_);
43 if (model->version() != TFLITE_SCHEMA_VERSION) {
44 ESP_LOGE(TAG,
"Streaming model's schema is not supported");
51 if (probed_size > 0) {
52 ESP_LOGD(TAG,
"Probed tensor arena size: %zu bytes", probed_size);
55 ESP_LOGW(TAG,
"Arena size probe failed, using manifest size: %zu bytes", this->
tensor_arena_size_);
63 ESP_LOGE(TAG,
"Could not allocate the streaming model's tensor arena.");
72 if (this->
interpreter_->AllocateTensors() != kTfLiteOk) {
73 ESP_LOGE(TAG,
"Failed to allocate tensors for the streaming model");
80 if ((input->dims->size != 3) || (input->dims->data[0] != 1) ||
81 (input->dims->data[2] != PREPROCESSOR_FEATURE_SIZE)) {
82 ESP_LOGE(TAG,
"Streaming model tensor input dimensions has improper dimensions.");
86 if (input->type != kTfLiteInt8) {
87 ESP_LOGE(TAG,
"Streaming model tensor input is not int8.");
93 if ((output->dims->size != 2) || (output->dims->data[0] != 1) || (output->dims->data[1] != 1)) {
94 ESP_LOGE(TAG,
"Streaming model tensor output dimension is not 1x1.");
98 if (output->type != kTfLiteUInt8) {
99 ESP_LOGE(TAG,
"Streaming model tensor output is not uint8.");
119 for (
size_t attempt_size : attempt_sizes) {
120 uint8_t *probe_arena = arena_allocator.
allocate(attempt_size);
121 if (probe_arena ==
nullptr) {
126 auto probe_interpreter = make_unique<tflite::MicroInterpreter>(
129 if (probe_interpreter->AllocateTensors() != kTfLiteOk) {
130 probe_interpreter.reset();
131 arena_allocator.
deallocate(probe_arena, attempt_size);
132 this->
ma_ = tflite::MicroAllocator::Create(this->
var_arena_, STREAMING_MODEL_VARIABLE_ARENA_SIZE);
133 this->
mrv_ = tflite::MicroResourceVariables::Create(this->
ma_, 20);
139 size_t lower = (probe_interpreter->arena_used_bytes() + 16 + 15) & ~15;
140 probe_interpreter.reset();
141 this->
ma_ = tflite::MicroAllocator::Create(this->
var_arena_, STREAMING_MODEL_VARIABLE_ARENA_SIZE);
142 this->
mrv_ = tflite::MicroResourceVariables::Create(this->
ma_, 20);
144 size_t upper = attempt_size;
146 while (lower < upper) {
147 auto test_interpreter = make_unique<tflite::MicroInterpreter>(
150 bool ok = test_interpreter->AllocateTensors() == kTfLiteOk;
152 test_interpreter.reset();
153 this->
ma_ = tflite::MicroAllocator::Create(this->
var_arena_, STREAMING_MODEL_VARIABLE_ARENA_SIZE);
154 this->
mrv_ = tflite::MicroResourceVariables::Create(this->
ma_, 20);
163 lower = ((lower + upper) / 2 + 15) & ~15;
166 arena_allocator.
deallocate(probe_arena, attempt_size);
208 uint8_t stride = this->
interpreter_->input(0)->dims->data[1];
212 (int8_t *) (tflite::GetTensorData<int8_t>(input)) + PREPROCESSOR_FEATURE_SIZE * this->
current_stride_step_,
213 features, PREPROCESSOR_FEATURE_SIZE);
216 if (this->current_stride_step_ >= stride) {
217 TfLiteStatus invoke_status = this->
interpreter_->Invoke();
218 if (invoke_status != kTfLiteOk) {
219 ESP_LOGW(TAG,
"Streaming interpreter invoke failed");
248 size_t sliding_window_average_size,
const std::string &wake_word,
size_t tensor_arena_size,
249 bool default_enabled,
bool internal_only) {
295 return detection_event;
308 return detection_event;
311VADModel::VADModel(
const uint8_t *model_start, uint8_t default_probability_cutoff,
size_t sliding_window_size,
312 size_t tensor_arena_size) {
330 return detection_event;
342 return detection_event;
346 if (op_resolver.AddCallOnce() != kTfLiteOk)
348 if (op_resolver.AddVarHandle() != kTfLiteOk)
350 if (op_resolver.AddReshape() != kTfLiteOk)
352 if (op_resolver.AddReadVariable() != kTfLiteOk)
354 if (op_resolver.AddStridedSlice() != kTfLiteOk)
356 if (op_resolver.AddConcatenation() != kTfLiteOk)
358 if (op_resolver.AddAssignVariable() != kTfLiteOk)
360 if (op_resolver.AddConv2D() != kTfLiteOk)
362 if (op_resolver.AddMul() != kTfLiteOk)
364 if (op_resolver.AddAdd() != kTfLiteOk)
366 if (op_resolver.AddMean() != kTfLiteOk)
368 if (op_resolver.AddFullyConnected() != kTfLiteOk)
370 if (op_resolver.AddLogistic() != kTfLiteOk)
372 if (op_resolver.AddQuantize() != kTfLiteOk)
374 if (op_resolver.AddDepthwiseConv2D() != kTfLiteOk)
376 if (op_resolver.AddAveragePool2D() != kTfLiteOk)
378 if (op_resolver.AddMaxPool2D() != kTfLiteOk)
380 if (op_resolver.AddPad() != kTfLiteOk)
382 if (op_resolver.AddPack() != kTfLiteOk)
384 if (op_resolver.AddSplitV() != kTfLiteOk)
An STL allocator that uses SPI or internal RAM.
void deallocate(T *p, size_t n)
tflite::MicroAllocator * ma_
bool load_model_()
Allocates tensor and variable arenas and sets up the model interpreter.
uint8_t current_stride_step_
size_t sliding_window_size_
bool tensor_arena_size_probed_
std::unique_ptr< tflite::MicroInterpreter > interpreter_
uint8_t default_probability_cutoff_
tflite::MicroMutableOpResolver< 20 > streaming_op_resolver_
bool register_streaming_ops_(tflite::MicroMutableOpResolver< 20 > &op_resolver)
Returns true if successfully registered the streaming model's TensorFlow operations.
void reset_probabilities()
Sets all recent_streaming_probabilities to 0 and resets the ignore window count.
size_t tensor_arena_size_
std::vector< uint8_t > recent_streaming_probabilities_
size_t probe_arena_size_()
Probes the actual required tensor arena size by trial allocation.
bool unprocessed_probability_status_
tflite::MicroResourceVariables * mrv_
bool perform_streaming_inference(const int8_t features[PREPROCESSOR_FEATURE_SIZE])
uint8_t probability_cutoff_
void unload_model()
Destroys the TFLite interpreter and frees the tensor and variable arenas' memory.
const uint8_t * model_start_
DetectionEvent determine_detected() override
Checks for voice activity by comparing the max probability in the sliding window with the probability...
VADModel(const uint8_t *model_start, uint8_t default_probability_cutoff, size_t sliding_window_size, size_t tensor_arena_size)
void log_model_config() override
void enable() override
Enable the model and save to flash. The next performing_streaming_inference call will load it.
DetectionEvent determine_detected() override
Checks for the wake word by comparing the mean probability in the sliding window with the probability...
void log_model_config() override
ESPPreferenceObject pref_
WakeWordModel(const std::string &id, const uint8_t *model_start, uint8_t default_probability_cutoff, size_t sliding_window_average_size, const std::string &wake_word, size_t tensor_arena_size, bool default_enabled, bool internal_only)
Constructs a wake word model object.
void disable() override
Disable the model and save to flash. The next performing_streaming_inference call will unload it.
Providing packet encoding functions for exchanging data with a remote host.
ESPPreferences * global_preferences
uint32_t fnv1_hash(const char *str)
Calculate a FNV-1 hash of str.
ESPPreferenceObject make_preference(size_t, uint32_t, bool)
uint8_t average_probability