Skip to content

Latest commit

 

History

History
87 lines (60 loc) · 1.8 KB

README.md

File metadata and controls

87 lines (60 loc) · 1.8 KB

PascalSentenceDataset

This program is utility to download pascal sentence dataset.

Installation

You can install by "git clone" command.

git clone https://github.com/rupy/PascalSentenceDataset.git

Dependency

You must install some python libraries. Use pip command.

pyquery 1.2.9
requests 2.6.0

Usage

To download dataset, just run program as follow:

python pascal_sentence_dataset.py

You can also write code like this:

# import
from pascal_sentence_dataset import PascalSentenceDataSet

# create instance
dataset = PascalSentenceDataSet()
# download images
dataset.download_images()
# download sentences
dataset.download_sentences()
# create correspondence data by dataset
dataset.create_correspondence_data()

That's it!

Correspondence data is the csv data to correspond data id to image data.

Additional information for Nakayama lab members

Our lab created Japanese translation of Pascal Sentence Dataset. Translation class is the utility to use parallel translation data, "pascal_sentence_numbers.csv". You can get text files of two languages by the class. To use the class, you have to install depencent libraries as follow:

mecab-python 0.996

To use mecab-python, you have to install MeCab in addition.

Usage

To create Japanese & English parallel translation data, just run program as follow:

python pascal_sentence_dataset.py

You can also write code like this:

# import
from translation import Translation

# put parallel translation data somewhere in advance
csv_file = 'translations/pascal_sentence_numbers.csv'
# initialize instance
ps = Translation(csv_file)
# create text data from csv file
ps.read_csv_and_save_as_txt()
# create wakati-gaki text data (Japanese text data separated by space between each word)
ps.wakati()