Skip to content

Switch from BiLSTM to the modern attention architecture #32

@vmarkovtsev

Description

@vmarkovtsev

Our current NN splitter is based on BiLSTM, which has problems with performance. We should leverage the recent advancements in deep learning and implement the new attention-based (seq2seq-like?) architecture of the model.

Stage 1 - research

Follow the paper, take the same dataset, and design the model. Calculate the metrics.

Stage 2 - production

Package the model, publish it on Modelforge.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions