Loading...
Works

NLP Model Evaluation
NLP Machine Learning Neural Nets Deep-Learning

cm
  • SkillsPyTorch, Keras, Scikit-learn, TensorFlow, Seaborn, NLTK
  • CodeGitHub Link
  • ProgressCompleted

This project was completed under the supervision of NLP experts at Northeastern University. The goal was to compare the performance of three different models on a Kaggle dataset that provided text data and emotional labels.

The models used in the experiment were Logistic Regression, LSTM, and a DistilBERT transformer model. I prepared the data and created training, testing, and validation dataset. The three models performed well, resulting in accuracies of 87%, 92%, and 93% respectively, with the DistilBERT model showing the most promise (likely due to its self-attention mechanism).

banner

This website is based on Takuya Matsuyama's website.