Issue facing in multihead attention

Hi I am facing some issues with tensorflow multihead attention.

Even though the input shape is consistent. It’s still throwing an error saying incorrect input dimension.

Can any one help please.

I am creating a local transformer that also single layer