I'm building a Recurrent Auto-encoder to make feature extraction on some time series sequences. All this sequences have different lengths and looking at different tutorials it seems that LSTM input layer accept only data in a format like (nb_sequence, nb_timestep, nb_feature) with the same timestep for all the sequences.
My model look like this:
encoder=Sequential([
LSTM(128,activation="relu",input_shape=(timesteps,n_features),return_sequences=True),
LSTM(64,activation="relu",return_sequences=False)
],name="Encoder")
decoder=Sequential([
RepeatVector(timesteps,input_shape=[64]),
LSTM(64,activation="relu",return_sequences=True),
LSTM(128,activation="relu",return_sequences=True),
TimeDistributed(Dense(n_features))
],name="Decoder")
autoencoder=Sequential([encoder,decoder])
Is it possible to train my model with different sequences with different lengths ? If yes, how have to proceed ?