The code repository for the Nature Scientific Reports (2023) paper Interpretable Sentiment Analysis of COVID-19 Tweets using Attention-based BiLSTM and Twitter-RoBERTa
Created and maintained by Md Abrar Jahin <abrar.jahin.2652@gmail.com, mdabrar.jahin@oist.jp>
.
Extended Datasets
In order to address the research gaps identified by Qi and Shabrina (2023), we have expanded the existing COVID-19 Twitter dataset. Our datasets overcome the limitations highlighted in their paper, specifically the short timeline and geographical constraints of the tweets. Each dataset includes a column of cleaned tweets, which have undergone preprocessing of the raw tweets and comments. Additionally, the datasets are accompanied by sentiment labels categorizing the tweets as negative (-1), neutral (0), or positive (1).
External Datasets
To assess the robustness and generalizability of our proposed model, we employed external datasets for benchmarking purposes. These additional datasets were utilized to evaluate how well our model performs beyond the confines of the original dataset used for training and testing. By incorporating these external datasets, we aimed to obtain a more comprehensive understanding of our model's capabilities and its ability to handle diverse and unseen data. The inclusion of these benchmark datasets allowed us to gauge the model's performance under varying conditions and validate its effectiveness in real-world scenarios.
Datasets | Description |
---|---|
UK Twitter COVID-19 Dataset | This dataset was developed by collecting COVID-19 tweets from only the major cities in the UK (Qi and Shabrina, 2023) |
Global Twitter COVID-19 Dataset | We extended the existing UK COVID-19 dataset by scraping additional 411885 tweets from 32 English-speaking countries |
USA Twitter COVID-19 Dataset | We extended the existing UK COVID-19 dataset by scraping additional 7500 tweets from only the USA |
External Reddit Dataset | 36801 comments |
External Twitter Dataset | 162980 tweets |
External Apple Twitter Dataset | 1630 tweets |
External US Airline Twitter Dataset | 14640 tweets |
Qi and Shabrina (2023) benchmarked their UK COVID-19 Twitter dataset's 3000 observations using Random Forest, Multinomial NB, and SVM. We additionally benchmarked the same portion of the dataset using the existing tree-based gradient boosting models (LGBM, CatBoost, XGboost, GBM), RandomForest+KNN+MLP stacking, RandomForestBagging, and RandomForest+GBM voting. The evaluation of these traditional models was performed individually using CountVectorizer, TF-IDF, and word2vec tokenizers as the tokenization methods.
We also showed how classical models and ensemble work on the pretrained transformer-based tokeizers: BERT (classical and ensemble), roBERTA (classical and ensemble), Sentence Transformer (classical and ensemble)
[1] twitter-roberta-base-sentiment-latest
[3] all-MiniLM-L6-v2
All the implemented DL model architectures with their associated codes and outputs can be found in Twitter-RoBERTa+LSTM. Our proposed model Attention-based biLSTM was trained on Twitter-RoBERTa tokenized inputs.
You can find the relevant files in XAI
The installation requirements for the Python packages are already included within the Notebooks, which are not discussed here.
The Jupyter Notebook can be executed on CPU using Google Colab or Kaggle, but it may take a significant amount of time to obtain the desired outputs.
Some of the notebooks were executed using Kaggle's GPU T4x2 and GPU P100. Kaggle provides a GPU quota of 30 hours per week, while Colab has a restricted usage limit.
Some of the notebooks were executed on Kaggle's TPU VM v3-8, which proved to be much faster than GPU. Kaggle provides a quota of 20 hours per week for TPU usage. However, the following additional code needs to be added before constructing the neural network model:
# detect and init the TPU
tpu = tf.distribute.cluster_resolver.TPUClusterResolver.connect()
# instantiate a distribution strategy
tpu_strategy = tf.distribute.experimental.TPUStrategy(tpu)
# instantiating the model in the strategy scope creates the model on the TPU
with tpu_strategy.scope():
model = tf.keras.Sequential( β¦ ) # define your model normally
model.compile( β¦ )
# train model normally
model.fit(training_dataset, epochs=EPOCHS, steps_per_epoch=β¦)
Below is an illustration of the directory structure of nlp-roBERTa-biLSTM-attention.
π nlp-roBERTa-biLSTM-attention
βββ π BERT
π nlp-roBERTa-biLSTM-attention\BERT
βββ π all_models1.png
βββ π all_models2.png
βββ π all_models3.png
βββ π all_models4.png
βββ π lgb_knn_mlp.png
βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\BERT\model1_keras_1_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\BERT\model2_keras_3_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π rf_knn_mlp.png
βββ π rf_stacking_voting.png
βββ π BoW
π nlp-roBERTa-biLSTM-attention\BoW
βββ π all_models_1.png
βββ π all_models_2.png
βββ π all_models_3.png
βββ π all_models_4.png
βββ π lgb_knn_mlp.png
βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\BoW\model1_keras_1_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\BoW\model2_keras_3_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π rf_knn_mlp.png
βββ π rf_stacking_voting.png
βββ π Data_scraping
π nlp-roBERTa-biLSTM-attention\Data_scraping
βββ π Twint-data collection.ipynb
βββ π Twitter academic api.ipynb
βββ π Extended_datasets
π nlp-roBERTa-biLSTM-attention\Extended_datasets
βββ π Global_covid_twitter_data
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data
β βββ π BiLSTM+CNN
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\BiLSTM+CNN
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π Global.csv
β βββ π Global_twitter_data_preprocessing.ipynb
β βββ π global-tweets_4_baseline_models.ipynb
β βββ π model1_keras_1_dense_layers
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\model1_keras_1_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\model2_keras_3_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\model3_BiLSTM
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\model4_BiLSTM+attention
β β βββ π best-model-global.ipynb
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π preprocessed_dataset
β β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Global_covid_twitter_data\preprocessed_dataset
β β βββ π sample_data_global_0.csv
β β βββ π sample_data_global_1.csv
β β βββ π sample_data_global_10.csv
β β βββ π sample_data_global_11.csv
β β βββ π sample_data_global_12.csv
β β βββ π sample_data_global_13.csv
β β βββ π sample_data_global_14.csv
β β βββ π sample_data_global_15.csv
β β βββ π sample_data_global_16.csv
β β βββ π sample_data_global_17.csv
β β βββ π sample_data_global_18.csv
β β βββ π sample_data_global_19.csv
β β βββ π sample_data_global_2.csv
β β βββ π sample_data_global_20.csv
β β βββ π sample_data_global_21.csv
β β βββ π sample_data_global_22.csv
β β βββ π sample_data_global_23.csv
β β βββ π sample_data_global_24.csv
β β βββ π sample_data_global_25.csv
β β βββ π sample_data_global_26.csv
β β βββ π sample_data_global_27.csv
β β βββ π sample_data_global_28.csv
β β βββ π sample_data_global_29.csv
β β βββ π sample_data_global_3.csv
β β βββ π sample_data_global_30.csv
β β βββ π sample_data_global_31.csv
β β βββ π sample_data_global_32.csv
β β βββ π sample_data_global_33.csv
β β βββ π sample_data_global_34.csv
β β βββ π sample_data_global_35.csv
β β βββ π sample_data_global_36.csv
β β βββ π sample_data_global_37.csv
β β βββ π sample_data_global_38.csv
β β βββ π sample_data_global_39.csv
β β βββ π sample_data_global_4.csv
β β βββ π sample_data_global_40.csv
β β βββ π sample_data_global_5.csv
β β βββ π sample_data_global_6.csv
β β βββ π sample_data_global_7.csv
β β βββ π sample_data_global_8.csv
β β βββ π sample_data_global_9.csv
β βββ π tweets_distribution_global.png
β βββ π word_cloud_global.png
β βββ π word_freq.png
βββ π Only_USA_covid_twitter_data
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data
β βββ π BiLSTM+CNN
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data\BiLSTM+CNN
β βββ π classification_report1.png
β βββ π classification_reports.png
β βββ π confusion_matrix.png
β βββ π Only_USA.csv
β βββ π frequency.png
β βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data\model1_keras_1_dense_layers
β βββ π classification_report1.png
β βββ π classification_reports.png
β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data\model2_keras_3_dense_layers
β βββ π classification_report1.png
β βββ π classification_reports.png
β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data\model3_BiLSTM
β βββ π classification_report1.png
β βββ π classification_reports.png
β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β π nlp-roBERTa-biLSTM-attention\Extended_datasets\Only_USA_covid_twitter_data\model4_BiLSTM+attention
β βββ π accuracy.png
β βββ π best-model-only-usa.ipynb
β βββ π classification_report1.png
β βββ π classification_reports.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π model_architecture.png
β βββ π only_USA-tweets-4_baseline_models.ipynb
β βββ π only_USA_twitter_data_preprocessing.ipynb
β βββ π sample_data_only_USA.csv
β βββ π uk_covid_twitter_sentiment.ipynb
β βββ π word_cloud.png
βββ π External_datasets
π nlp-roBERTa-biLSTM-attention\External_datasets
βββ π Apple_twitter_sentiments
β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments
β βββ π BiLSTM+CNN
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments\BiLSTM+CNN
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π apple-tweets.ipynb
β βββ π model1_keras_1_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments\model1_keras_1_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments\model2_keras_3_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments\model3_BiLSTM
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Apple_twitter_sentiments\model4_BiLSTM+attention
β β βββ π accuracy.png
β β βββ π best-model-apple-twitter.ipynb
β β βββ π classification_reports1.png
β β βββ π classification_reports2.png
β β βββ π confusion_matrix.png
β β βββ π loss.png
βββ π Reddit
β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit
β βββ π BiLSTM+CNN
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit\BiLSTM+CNN
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π Reddit_Data.csv
β βββ π model1_keras_1_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit\model1_keras_1_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit\model2_keras_3_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit\model3_BiLSTM
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Reddit\model4_BiLSTM+attention
β β βββ π LIME.png
β β βββ π SHAP_bar.png
β β βββ π SHAP_bar_ascending.png
β β βββ π SHAP_bar_descending.png
β β βββ π SHAP_explain.png
β β βββ π best-model-reddit.ipynb
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β β βββ π target_predictions.png
β βββ π reddit-tweets-1.ipynb
β βββ π reddit-tweets-2.ipynb
βββ π Twitter
β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter
β βββ π BiLSTM+CNN
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter\BiLSTM+CNN
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π Twitter_Data.csv
β βββ π model1_keras_1_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter\model1_keras_1_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter\model2_keras_3_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter\model3_BiLSTM
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β β π nlp-roBERTa-biLSTM-attention\External_datasets\Twitter\model4_BiLSTM+attention
β β βββ π best-model-twitter-external.ipynb
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π twitter-tweets-2.ipynb
β βββ π twitter_tweets_1.ipynb
βββ π US_airlines_twitter_sentiments
β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments
β βββ π BiLSTM+CNN
β β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments\BiLSTM+CNN
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model1_keras_1_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments\model1_keras_1_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model2_keras_3_dense_layers
β β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments\model2_keras_3_dense_layers
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model3_BiLSTM
β β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments\model3_BiLSTM
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
β βββ π model4_BiLSTM+attention
β β π nlp-roBERTa-biLSTM-attention\External_datasets\US_airlines_twitter_sentiments\model4_BiLSTM+attention
β β βββ π best-model-us-airlines.ipynb
β β βββ π classification_report1.png
β β βββ π classification_report2.png
β β βββ π confusion_matrix.png
βββ π token.txt
βββ π LICENSE
βββ π Previous_research
π nlp-roBERTa-biLSTM-attention\Previous_research
βββ π 1.png
βββ π 2.png
βββ π Vaibhav 2022.pdf
βββ π Yuxing 2023.pdf
βββ π README.md
βββ π RoBERTa
π nlp-roBERTa-biLSTM-attention\RoBERTa
βββ π cardiff_all_models_1.png
βββ π cardiff_all_models_2.png
βββ π cardiff_all_models_3.png
βββ π cardiff_all_models_4.png
βββ π lgb+knn+mlp.png
βββ π rf_stacking_voting.png
βββ π roberta_base_rf+knn+mlp.png
βββ π SBERT
π nlp-roBERTa-biLSTM-attention\SBERT
βββ π all_models_1.png
βββ π all_models_2.png
βββ π all_models_3.png
βββ π all_models_4.png
βββ π all_models_5.png
βββ π lgb_knn_mlp.png
βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\SBERT\model1_keras_1_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\SBERT\model2_keras_3_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π rf_knn_mlp.png
βββ π rf_stacking_voting.png
βββ π TF-IDF
π nlp-roBERTa-biLSTM-attention\TF-IDF
βββ π all_models_1.png
βββ π all_models_2.png
βββ π all_models_3.png
βββ π all_models_4.png
βββ π lgbm_knn_mlp.png
βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\TF-IDF\model1_keras_1_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\TF-IDF\model2_keras_3_dense_layers
β βββ π classification_report.png
β βββ π confusion_matrix.png
βββ π rf_bagging.png
βββ π rf_knn_mlp.png
βββ π rf_stacking_voting.png
βββ π Target_lexicon_selection
π nlp-roBERTa-biLSTM-attention\Target_lexicon_selection
βββ π target_lexicon_selection.ipynb
βββ π textblob1.png
βββ π textblob2.png
βββ π textblob3.png
βββ π textblob4.png
βββ π vader1.png
βββ π vader2.png
βββ π vader3.png
βββ π vader4.png
βββ π wordnet1.png
βββ π wordnet2.png
βββ π wordnet3.png
βββ π wordnet4.png
βββ π Twitter-RoBERTa+LSTM
π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM
βββ π BiLSTM+CNN
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\BiLSTM+CNN
β βββ π accuracy.png
β βββ π biLSTM+CNN.ipynb
β βββ π classification_report.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π model_architecture.png
βββ π model1_keras_1_dense_layers
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model1_keras_1_dense_layers
β βββ π accuracy.png
β βββ π classification_report1.png
β βββ π classification_report2.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π model1.ipynb
β βββ π model_architecture.png
β βββ π summary.png
βββ π model2_keras_3_dense_layers
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model2_keras_3_dense_layers
β βββ π accuracy.png
β βββ π classification_report.png
β βββ π classification_report1.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π model2.ipynb
β βββ π model_architecture.png
β βββ π model_summary.png
βββ π model3_BiLSTM
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model3_BiLSTM
β βββ π accuracy.png
β βββ π classification_report1.png
β βββ π classification_report2.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π model_architecture.png
β βββ π summary.png
β βββ π target_val_counts.png
β βββ π train_val.png
βββ π model4_BiLSTM+attention
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention
β βββ π XAI
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention\XAI
β βββ π Lime
β β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention\XAI\Lime
β β βββ π lime1.png
β β βββ π lime2.png
β β βββ π lime3.png
β β βββ π lime4.png
β β βββ π lime5.png
β β βββ π lime6.png
β β βββ π lime7.png
β βββ π SHAP
β β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention\XAI\SHAP
β β βββ π shap_neg1.png
β β βββ π shap_neg2.png
β β βββ π shap_neg_bar_ascending.png
β β βββ π shap_neg_bar_descending.png
β β βββ π shap_neu1.png
β β βββ π shap_neu2.png
β β βββ π shap_neu_bar.png
β β βββ π shap_neu_bar_ascending.png
β β βββ π shap_neu_bar_descending.png
β β βββ π shap_pos1.png
β β βββ π shap_pos2.png
β β βββ π shap_pos_bar_ascending.png
β β βββ π shap_pos_bar_descending.png
β βββ π learning_rates.png
β βββ π model_architecture.png
β βββ π summary.png
β βββ π uk_twitter_data_3k
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention\uk_twitter_data_3k
β βββ π accuracy.png
β βββ π best-model_uk-tweet_3k.ipynb
β βββ π classification_report.png
β βββ π classification_report2.png
β βββ π confusion_matrix.png
β βββ π loss.png
β βββ π train_val_loss.png
β βββ π uk_twitter_data_all
β π nlp-roBERTa-biLSTM-attention\Twitter-RoBERTa+LSTM\model4_BiLSTM+attention\uk_twitter_data_all
β βββ π accuracy.png
β βββ π best-model-uk-twitter-all.ipynb
β βββ π classification_report1.png
β βββ π classification_report2.png
β βββ π collage.png
β βββ π confusion_matrix.png
β βββ π loss.png
βββ π UK_covid_twitter_data
π nlp-roBERTa-biLSTM-attention\UK_covid_twitter_data
βββ π all_cities.csv
βββ π sample_data_3000.csv
βββ π sample_data_all.csv
βββ π stacked bar graph.png
βββ π tweets distribution.png
βββ π uk_twitter_data_preprocessing.ipynb
βββ π list.md
βββ π sentiment_distribution_barchart.png
βββ π sentiment_distribution_pie_chart.png
βββ π uk-twitter-3k-classical-modelling.ipynb
βββ π word2vec
π nlp-roBERTa-biLSTM-attention\word2vec
βββ π all_models_1.png
βββ π all_models_2.png
βββ π all_models_3.png
βββ π all_models_4.png
βββ π lgb_knn_mlp.png
βββ π model1_keras_1_dense_layers
π nlp-roBERTa-biLSTM-attention\word2vec\model1_keras_1_dense_layers
βββ π classification_report.png
βββ π confusion_matrix.png
βββ π model2_keras_3_dense_layers
π nlp-roBERTa-biLSTM-attention\word2vec\model2_keras_3_dense_layers
βββ π classification_report.png
βββ π confusion_matrix.png
βββ π rf_knn_mlp.png
βββ π rf_stacking_voting.png
β
@misc{md abrar jahin_2023,
title={Extended Covid Twitter Datasets},
url={https://www.kaggle.com/ds/3205649},
DOI={10.34740/KAGGLE/DS/3205649},
publisher={Kaggle},
author={Md Abrar Jahin},
year={2023}
}
Jahin, Md Abrar (2023), βExtended Covid Twitter Datasetsβ, Mendeley Data, V1, doi: 10.17632/2ynwykrfgf.1
@software{md_abrar_jahin_2024_13840678,
author = {Md Abrar Jahin},
title = {Abrar2652/nlp-roBERTa-biLSTM-attention: v1.0.0},
month = sep,
year = 2024,
publisher = {Zenodo},
version = {v1.0.0},
doi = {10.5281/zenodo.13840678},
url = {https://doi.org/10.5281/zenodo.13840678}
}
MIT licensed, except where otherwise stated.
See LICENSE.txt
file.