View RSS Feed

mql5

  1. Neural networks made easy (Part 37): Sparse Attention

    by , 12-07-2023 at 01:03 AM
    In the previous article, we discussed relational models which use attention mechanisms in their architecture. We used this model to create an Expert Advisor, and the resulting EA showed good results. However, we noticed that the model's learning rate was lower compared to our earlier experiments. This is due to the fact that the transformer block used in the model is a rather complex architectural solution performing a large number of operations. The number of these operations grows in a quadratic
    ...
    Categories
    Uncategorized