Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Switch from BiLSTM to the modern attention architecture #32

Open
vmarkovtsev opened this issue Jul 25, 2019 · 3 comments
Open

Switch from BiLSTM to the modern attention architecture #32

vmarkovtsev opened this issue Jul 25, 2019 · 3 comments

Comments

@vmarkovtsev
Copy link
Collaborator

Our current NN splitter is based on BiLSTM, which has problems with performance. We should leverage the recent advancements in deep learning and implement the new attention-based (seq2seq-like?) architecture of the model.

Stage 1 - research

Follow the paper, take the same dataset, and design the model. Calculate the metrics.

Stage 2 - production

Package the model, publish it on Modelforge.

@vmarkovtsev
Copy link
Collaborator Author

Assigning to you @zurk because you worked for solutions and missed interesting tasks.

@Guillemdb
Copy link
Contributor

@vmarkovtsev I think it's time to close this issue 😉, for some reason I cannot do it myself.

@vmarkovtsev
Copy link
Collaborator Author

I'd rather leave these to indicate what was lacking in the project when we stopped. Thanks for pinging anyway!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants
@zurk @vmarkovtsev @Guillemdb and others