I’m trying to perform quantization-aware training on a customized model that’s not of
tf.keras.Model type. It’s got its own
trainable_variables() functions for the optimizer to apply gradients on. Has anyone had any experience running quantization aware training in such a scenario?