# How I solive the error " ValueError: Dimensions must be equal, but are 124 and 125 for '{{node mean_squared_error/sub}} = Sub[T=DT_FLOAT](MIN2Net/decoder/conv2d_transpose_3/Elu, IteratorGetNext:1)' with input shapes: [?,1,124,16], [?,1,125,16]."?

How I solive the error " ValueError: Dimensions must be equal, but are 124 and 125 for ‘{{node mean_squared_error/sub}} = Sub[T=DT_FLOAT](MIN2Net/decoder/conv2d_transpose_3/Elu, IteratorGetNext:1)’ with input shapes: [?,1,124,16], [?,1,125,16]."?

``````from tensorflow.keras.layers import Concatenate, AveragePooling2D, BatchNormalization, Conv2D, Conv2DTranspose, Dense, Input, Reshape, Flatten
from tensorflow.keras.constraints import max_norm
from tensorflow.keras.models import Model
from tensorflow.keras.callbacks import CSVLogger, ModelCheckpoint, ReduceLROnPlateau
from tensorflow.keras.callbacks import EarlyStopping
import tensorflow as tf
from tensorflow.keras import backend as K

import numpy as np
import os
import time
from sklearn.metrics import classification_report, f1_score
from min2net.loss import mean_squared_error, triplet_loss, SparseCategoricalCrossentropy
from min2net.utils import TimeHistory, compute_class_weight

class MIN2Net:
def __init__(self,
input_shape=(1,125,16),
num_class=2,
loss=[mean_squared_error, triplet_loss(margin=1.0), 'sparse_categorical_crossentropy'],
loss_weights=[1., 1., 1.],
latent_dim = None,
epochs=200,
batch_size=100,
lr=1e-2,
min_lr=1e-3,
factor=0.5,
patience=5,
es_patience=20,
verbose=1,
log_path='logs',
model_name='MIN2Net',
**kwargs):
D, T, C = input_shape
self.latent_dim = latent_dim if latent_dim is not None else C if num_class==2 else 64
self.num_class = num_class
self.input_shape = input_shape
self.loss = loss
self.loss_weights = loss_weights
self.epochs = epochs
self.batch_size = batch_size
self.optimizer = optimizer
self.optimizer.lr = lr
self.lr = lr
self.min_lr = min_lr
self.factor = factor
self.patience = patience
self.es_patience = es_patience
self.verbose = verbose
self.log_path = log_path
self.model_name = model_name
self.weights_dir = log_path+'/'+model_name+'_out_weights.h5'
self.csv_dir = log_path+'/'+model_name+'_out_log.log'
self.time_log = log_path+'/'+model_name+'_time_log.csv'

# use **kwargs to set the new value of below args.
self.f1_average = 'binary' if self.num_class == 2 else 'macro'
self.data_format = 'channels_last'
self.shuffle = False
self.metrics = 'accuracy'
self.monitor = 'val_loss'
self.mode = 'min'
self.save_best_only = True
self.save_weight_only = True
self.seed = 1234
self.class_balancing = False
# 'set params'
self.subsampling_size = 125
self.pool_size_1 = (1,T//self.subsampling_size)
self.pool_size_2 = (1,4)
self.filter_1 = C
self.filter_2 = 10

for k in kwargs.keys():
self.__setattr__(k, kwargs[k])

self.flatten_size = T//self.pool_size_1[1]//self.pool_size_2[1]

np.random.seed(self.seed)
tf.random.set_seed(self.seed)
K.set_image_data_format(self.data_format)
if not os.path.exists(self.log_path):
os.makedirs(self.log_path)

def build(self):
'encoder'
encoder_input  = Input(self.input_shape)
en_conv        = Conv2D(self.filter_1, (1, 64), activation='elu', padding="same",
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(encoder_input)
en_conv        = BatchNormalization(axis=3, epsilon=1e-05, momentum=0.1)(en_conv)
en_conv        = AveragePooling2D(pool_size=self.pool_size_1)(en_conv)
en_conv        = Conv2D(self.filter_2, (1, 32), activation='elu', padding="same",
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(en_conv)
en_conv        = BatchNormalization(axis=3, epsilon=1e-05, momentum=0.1)(en_conv)
en_conv        = AveragePooling2D(pool_size=self.pool_size_2)(en_conv)
en_conv        = Flatten()(en_conv)
encoder_output = Dense(self.latent_dim, kernel_constraint=max_norm(0.5))(en_conv)
encoder        = Model(inputs=encoder_input, outputs=encoder_output, name='encoder')
encoder.summary()

'decoder'
decoder_input  = Input(shape=(self.latent_dim,), name='decoder_input')
de_conv        = Dense(1*self.flatten_size*self.filter_2, activation='elu',
kernel_constraint=max_norm(0.5))(decoder_input)
de_conv        = Reshape((1, self.flatten_size, self.filter_2))(de_conv)
de_conv        = Conv2DTranspose(filters=self.filter_2, kernel_size=(1, 64),
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(de_conv)
decoder_output = Conv2DTranspose(filters=self.filter_1, kernel_size=(1, 32),
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(de_conv)
decoder        = Model(inputs=decoder_input, outputs=decoder_output, name='decoder')
decoder.summary()

'Build the computation graph for training'
latent         = encoder(encoder_input)
train_xr       = decoder(latent)
z              = Dense(self.num_class, activation='softmax', kernel_constraint=max_norm(0.5),
name='classifier')(latent)

return Model(inputs=encoder_input, outputs=[train_xr, latent, z],
name='MIN2Net')

def fit(self, X_train, y_train, X_val, y_val):

if X_train.ndim != 4:
raise Exception('ValueError: `X_train` is incompatible: expected ndim=4, found ndim='+str(X_train.ndim))
elif X_val.ndim != 4:
raise Exception('ValueError: `X_val` is incompatible: expected ndim=4, found ndim='+str(X_val.ndim))

csv_logger    = CSVLogger(self.csv_dir)
time_callback = TimeHistory(self.time_log)
checkpointer  = ModelCheckpoint(monitor=self.monitor, filepath=self.weights_dir,
verbose=self.verbose, save_best_only=self.save_best_only,
save_weight_only=self.save_weight_only)
reduce_lr     = ReduceLROnPlateau(monitor=self.monitor, patience=self.patience,
factor=self.factor, mode=self.mode, verbose=self.verbose,
min_lr=self.min_lr)
es            = EarlyStopping(monitor=self.monitor, mode=self.mode, verbose=self.verbose,
patience=self.es_patience)
model = self.build()
model.summary()

if self.class_balancing: # compute_class_weight if class_balancing is True
class_weight  = compute_class_weight(y_train)
self.loss[-1] = SparseCategoricalCrossentropy(class_weight=class_weight)

model.compile(optimizer=self.optimizer, loss=self.loss, metrics=self.metrics, loss_weights=self.loss_weights)

model.fit(x=X_train, y=[X_train,y_train,y_train],
batch_size=self.batch_size, shuffle=self.shuffle,
epochs=self.epochs, validation_data=(X_val, [X_val,y_val,y_val]),
callbacks=[checkpointer,csv_logger,reduce_lr,es, time_callback])

def predict(self, X_test, y_test):

if X_test.ndim != 4:
raise Exception('ValueError: `X_test` is incompatible: expected ndim=4, found ndim='+str(X_test.ndim))

model = self.build()
model.summary()
model.compile(optimizer=self.optimizer, loss=self.loss, metrics=self.metrics, loss_weights=self.loss_weights)

start = time.time()
y_pred_decoder, y_pred_trip, y_pred_clf = model.predict(X_test)
end = time.time()
loss, decoder_loss, trip_loss, classifier_loss, decoder_acc, trip_acc, classifier_acc  = model.evaluate(x=X_test,
y=[X_test,y_test,y_test],
batch_size=self.batch_size,
verbose=self.verbose)
y_pred_argm = np.argmax(y_pred_clf, axis=1)
print("F1-score is computed based on {}".format(self.f1_average))
f1 = f1_score(y_test, y_pred_argm, average=self.f1_average)
print('(loss: {}, accuracy: {})'.format(loss, classifier_acc))
print(classification_report(y_test, y_pred_argm))
evaluation = {'loss': loss,
'decoder_loss': decoder_loss,
'triplet_loss':trip_loss,
'classifier_loss': classifier_loss,
'accuracy': classifier_acc,
'f1-score': f1 ,
'prediction_time': end-start}
Y = {'y_true': y_test,
'y_pred': y_pred_argm,
'y_pred_decoder': y_pred_decoder}

return Y, evaluation
model = MIN2Net(input_shape=(1, 125, 16), num_class=3, monitor='val_loss', shuffle=True)
model.fit(X_train, y_train, X_val, y_val)
``````

Hi @ruorch ,

Thanks,
Laxma.

Hi @Laxma_Reddy_Patlolla ,
This is reproducible code.

``````!pip install min2net
from tensorflow.keras.layers import Concatenate, AveragePooling2D, BatchNormalization, Conv2D, Conv2DTranspose, Dense, Input, Reshape, Flatten
from tensorflow.keras.constraints import max_norm
from tensorflow.keras.models import Model
from tensorflow.keras.callbacks import CSVLogger, ModelCheckpoint, ReduceLROnPlateau
from tensorflow.keras.callbacks import EarlyStopping
import tensorflow as tf
from tensorflow.keras import backend as K

import numpy as np
import os
import time
from sklearn.metrics import classification_report, f1_score
from min2net.loss import mean_squared_error, triplet_loss, SparseCategoricalCrossentropy
from min2net.utils import TimeHistory, compute_class_weight

class MIN2Net:
def __init__(self,
input_shape=(1,125,16),
num_class=2,
loss=[mean_squared_error, triplet_loss(margin=1.0), 'sparse_categorical_crossentropy'],
loss_weights=[1., 1., 1.],
latent_dim = None,
epochs=200,
batch_size=100,
lr=1e-2,
min_lr=1e-3,
factor=0.5,
patience=5,
es_patience=20,
verbose=1,
log_path='logs',
model_name='MIN2Net',
**kwargs):
D, T, C = input_shape
self.latent_dim = latent_dim if latent_dim is not None else C if num_class==2 else 64
self.num_class = num_class
self.input_shape = input_shape
self.loss = loss
self.loss_weights = loss_weights
self.epochs = epochs
self.batch_size = batch_size
self.optimizer = optimizer
self.optimizer.lr = lr
self.lr = lr
self.min_lr = min_lr
self.factor = factor
self.patience = patience
self.es_patience = es_patience
self.verbose = verbose
self.log_path = log_path
self.model_name = model_name
self.weights_dir = log_path+'/'+model_name+'_out_weights.h5'
self.csv_dir = log_path+'/'+model_name+'_out_log.log'
self.time_log = log_path+'/'+model_name+'_time_log.csv'

# use **kwargs to set the new value of below args.
self.f1_average = 'binary' if self.num_class == 2 else 'macro'
self.data_format = 'channels_last'
self.shuffle = False
self.metrics = 'accuracy'
self.monitor = 'val_loss'
self.mode = 'min'
self.save_best_only = True
self.save_weight_only = True
self.seed = 1234
self.class_balancing = False
# 'set params'
self.subsampling_size = 125
self.pool_size_1 = (1,T//self.subsampling_size)
self.pool_size_2 = (1,4)
self.filter_1 = C
self.filter_2 = 10

for k in kwargs.keys():
self.__setattr__(k, kwargs[k])

self.flatten_size = T//self.pool_size_1[1]//self.pool_size_2[1]

np.random.seed(self.seed)
tf.random.set_seed(self.seed)
K.set_image_data_format(self.data_format)
if not os.path.exists(self.log_path):
os.makedirs(self.log_path)

def build(self):
'encoder'
encoder_input  = Input(self.input_shape)
en_conv        = Conv2D(self.filter_1, (1, 64), activation='elu', padding="same",
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(encoder_input)
en_conv        = BatchNormalization(axis=3, epsilon=1e-05, momentum=0.1)(en_conv)
en_conv        = AveragePooling2D(pool_size=self.pool_size_1)(en_conv)
en_conv        = Conv2D(self.filter_2, (1, 32), activation='elu', padding="same",
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(en_conv)
en_conv        = BatchNormalization(axis=3, epsilon=1e-05, momentum=0.1)(en_conv)
en_conv        = AveragePooling2D(pool_size=self.pool_size_2)(en_conv)
en_conv        = Flatten()(en_conv)
encoder_output = Dense(self.latent_dim, kernel_constraint=max_norm(0.5))(en_conv)
encoder        = Model(inputs=encoder_input, outputs=encoder_output, name='encoder')
encoder.summary()

'decoder'
decoder_input  = Input(shape=(self.latent_dim,), name='decoder_input')
de_conv        = Dense(1*self.flatten_size*self.filter_2, activation='elu',
kernel_constraint=max_norm(0.5))(decoder_input)
de_conv        = Reshape((1, self.flatten_size, self.filter_2))(de_conv)
de_conv        = Conv2DTranspose(filters=self.filter_2, kernel_size=(1, 64),
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(de_conv)
decoder_output = Conv2DTranspose(filters=self.filter_1, kernel_size=(1, 32),
kernel_constraint=max_norm(2., axis=(0, 1, 2)))(de_conv)
decoder        = Model(inputs=decoder_input, outputs=decoder_output, name='decoder')
decoder.summary()

'Build the computation graph for training'
latent         = encoder(encoder_input)
train_xr       = decoder(latent)
z              = Dense(self.num_class, activation='softmax', kernel_constraint=max_norm(0.5),
name='classifier')(latent)

return Model(inputs=encoder_input, outputs=[train_xr, latent, z],
name='MIN2Net')

def fit(self, X_train, y_train, X_val, y_val):

if X_train.ndim != 4:
raise Exception('ValueError: `X_train` is incompatible: expected ndim=4, found ndim='+str(X_train.ndim))
elif X_val.ndim != 4:
raise Exception('ValueError: `X_val` is incompatible: expected ndim=4, found ndim='+str(X_val.ndim))

csv_logger    = CSVLogger(self.csv_dir)
time_callback = TimeHistory(self.time_log)
checkpointer  = ModelCheckpoint(monitor=self.monitor, filepath=self.weights_dir,
verbose=self.verbose, save_best_only=self.save_best_only,
save_weight_only=self.save_weight_only)
reduce_lr     = ReduceLROnPlateau(monitor=self.monitor, patience=self.patience,
factor=self.factor, mode=self.mode, verbose=self.verbose,
min_lr=self.min_lr)
es            = EarlyStopping(monitor=self.monitor, mode=self.mode, verbose=self.verbose,
patience=self.es_patience)
model = self.build()
model.summary()

if self.class_balancing: # compute_class_weight if class_balancing is True
class_weight  = compute_class_weight(y_train)
self.loss[-1] = SparseCategoricalCrossentropy(class_weight=class_weight)

model.compile(optimizer=self.optimizer, loss=self.loss, metrics=self.metrics, loss_weights=self.loss_weights)

model.fit(x=X_train, y=[X_train,y_train,y_train],
batch_size=self.batch_size, shuffle=self.shuffle,
epochs=self.epochs, validation_data=(X_val, [X_val,y_val,y_val]),
callbacks=[checkpointer,csv_logger,reduce_lr,es, time_callback])

def predict(self, X_test, y_test):

if X_test.ndim != 4:
raise Exception('ValueError: `X_test` is incompatible: expected ndim=4, found ndim='+str(X_test.ndim))

model = self.build()
model.summary()
model.compile(optimizer=self.optimizer, loss=self.loss, metrics=self.metrics, loss_weights=self.loss_weights)

start = time.time()
y_pred_decoder, y_pred_trip, y_pred_clf = model.predict(X_test)
end = time.time()
loss, decoder_loss, trip_loss, classifier_loss, decoder_acc, trip_acc, classifier_acc  = model.evaluate(x=X_test,
y=[X_test,y_test,y_test],
batch_size=self.batch_size,
verbose=self.verbose)
y_pred_argm = np.argmax(y_pred_clf, axis=1)
print("F1-score is computed based on {}".format(self.f1_average))
f1 = f1_score(y_test, y_pred_argm, average=self.f1_average)
print('(loss: {}, accuracy: {})'.format(loss, classifier_acc))
print(classification_report(y_test, y_pred_argm))
evaluation = {'loss': loss,
'decoder_loss': decoder_loss,
'triplet_loss':trip_loss,
'classifier_loss': classifier_loss,
'accuracy': classifier_acc,
'f1-score': f1 ,
'prediction_time': end-start}
Y = {'y_true': y_test,
'y_pred': y_pred_argm,
'y_pred_decoder': y_pred_decoder}

return Y, evaluation
X_train = np.random.random_sample((10, 1, 125, 16))
X_val = np.random.random_sample((10, 1, 125, 16))
y_train = np.random.randint(0, 3, (10, 1))
y_val = np.random.randint(0, 3, (10, 1))
model = MIN2Net(input_shape=(1, 125, 16), num_class=3, monitor='val_loss', shuffle=True)
model.fit(X_train, y_train, X_val, y_val)
``````