Skip to content

Commit d8523f1

Browse files
committed
positional embedding not needed given latents should have them already
1 parent 398c0dc commit d8523f1

File tree

1 file changed

+2
-3
lines changed

1 file changed

+2
-3
lines changed

train_latent_with_text.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -81,16 +81,15 @@ def encode_tokens(str: str) -> Tensor:
8181
num_text_tokens = 256,
8282
dim_latent = 4,
8383
channel_first_latent = True,
84-
modality_default_shape = (4, 4),
84+
modality_default_shape = (8, 8),
8585
modality_encoder = Encoder(vae),
8686
modality_decoder = Decoder(vae),
8787
pre_post_transformer_enc_dec = (
8888
nn.Conv2d(4, 128, 3, 2, 1),
8989
nn.ConvTranspose2d(128, 4, 3, 2, 1, output_padding = 1),
9090
),
91-
add_pos_emb = True,
91+
add_pos_emb = False,
9292
modality_num_dim = 2,
93-
velocity_consistency_loss_weight = 0.1,
9493
reconstruction_loss_weight = 0.1,
9594
transformer = dict(
9695
dim = 128,

0 commit comments

Comments
 (0)