-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Various parts of the spacetimeformer embedding #11
Comments
As there are
|
Now, there are For the whole batch, this gives a tensor of shape |
For example, if there are 14 target variables, these can be indexed from 0 to 13. Then, what Or, in this competition, In general, there are Now, the target variable, the name itself, is independent of time, so each of the For the batch, this gives a tensor of embeddings of shape |
Here, "available or not" does not concern missing values in the original data, but rather whether the target value is meant to be predicted (not available) or to be used for prediction (available). Being available is mapped to an embedding vector, while being unavailable is mapped to another. For the entire batch, the resulting tensor is of shape (N, L * d_y, d_model). This embedding tensor is actually summed together with the embedding tensor returned by |
Overall, the spacetimeformer embedding,
spacetimeformer_model.nn.embed.SpacetimeformerEmbedding
takes inputs of shape(N, L, d_x)
and(N, L, d_y)
, and it outputs several embeddings of shape(N, L * d_y, d_model)
.N
is the batch size.L
is the sequence length. For the encoder, this iscontext_points
. For the decoder, it's -target_points
.d_x
is the number of input features.d_y
is the number of output features.The overall embedding consists of several embeddings:
x_emb
,y_emb
,var_emb
, andgiven_emb
.The text was updated successfully, but these errors were encountered: