Hello Everyone,
Here is the list of packages needed for our second Data Mining Lab Session.
- Python 3 (coding will be done strictly using Python 3)
- Anaconda Environment (recommended but not mandatory) (https://www.continuum.io/downloads)
- Jupyter (http://jupyter.org/)
- Google's word2vec (Download the file... warning! it is really huge)(https://drive.google.com/file/d/0B7XkCwpI5KDYNlNUTTlSS21pQmM/edit?usp=sharing)
- Gensim (https://radimrehurek.com/gensim/)
- Scikit Learn (http://scikit-learn.org/stable/) (get the latest version)
- Pandas (http://pandas.pydata.org/)
- Matplotlib (https://matplotlib.org/)
- NLTK (for stopwords) (http://www.nltk.org/)
- Operating System: Preferably Linux or MacOS (Windows break but you can try it out)
- RAM: 4GB
- Disk Space: 8GB (mostly to store word embeddings)
Once you have installed all the necessary packages, you can test to see if everything is working by running the following python code:
import logging
logging.root.handlers = [] # Jupyter messes up logging so needs a reset
logging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO)
from smart_open import smart_open
import pandas as pd
import numpy as np
from numpy import random
import gensim
import nltk
from sklearn.cross_validation import train_test_split
from sklearn import linear_model
from sklearn.feature_extraction.text import CountVectorizer, TfidfVectorizer
from sklearn.metrics import accuracy_score, confusion_matrix
import matplotlib.pyplot as plt
from gensim.models import Word2Vec
from sklearn.neighbors import KNeighborsClassifier
from sklearn import linear_model
from nltk.corpus import stopwords
%matplotlib inline
If you have any further questions please feel free to contact me at ellfae@gmail.com
Have Fun,
Elvis Saravia (Data Mining TA)