Skip to content
This repository has been archived by the owner on Sep 24, 2021. It is now read-only.

Tensorflow implementation of "A Structured Self-Attentive Sentence Embedding"

Notifications You must be signed in to change notification settings

flrngel/Self-Attentive-tensorflow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Self-Attentive-Tensorflow

model image of Self Attentive

Tensorflow implementation of A Structured Self-Attentive Sentence Embedding

You can read more about concept from this paper

Key Concept

Frobenius norm with attention

Usage

Download ag news dataset as below

$ tree ./data
./data
└── ag_news_csv
    ├── classes.txt
    ├── readme.txt
    ├── test.csv
    ├── train.csv
    └── train_mini.csv

and then

$ python train.py

Result

Accuracy 0.895

visualize without penalization

visualize with penalization

To-do list

  • support multiple dataset

Notes

This implementation does not use pretrained GloVe or Word2vec.

About

Tensorflow implementation of "A Structured Self-Attentive Sentence Embedding"

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages