Hello, dear friend, you can consult us at any time if you have any questions, add WeChat: daixieit

COMP4650/6490 Document Analysis – Semester 2 / 2022

Assignment 2

Overview

In this assignment you will:

1. Develop a better understanding of how machine learning models are trained in practice, including partitioning of datasets and evaluation.

2. Become familiar with the scikit-learn1package for machine learning with text.

3. Become familiar with the PyTorch2 framework for implementing neural network-based machine learning models.

Throughout this assignment you will make changes to the provided code to improve or complete existing models. In some cases, you will write your own code from scratch after reviewing an example.

Submission

● The answers to this assignment (including your code files) have to be submitted online in Wattle.

● You will produce an answers le with your responses to each question. Your answers file must be a PDF file named u1234567 .pdf where u1234567 should be replaced with your Uni ID.

● You should submit a ZIP file containing all of the code files and your answers PDF file, BUT NO DATA.

Marking

This assignment will be marked out of 15, and it will contribute 15% of your final course mark.

Your answers to coding questions (or coding parts of each question) will be marked based on the quality of your code (is it efficient, is it readable, is it extendable, is it correct).

Your answers to discussion questions (or discussion parts of each question) will be marked based on how convincing your explanations are (are they sufficiently detailed, are they well-reasoned, are they backed by appropriate evidence, are they clear, do they use appropriate visual aids such as tables, charts, or diagrams).

Question 1: Movie Review Sentiment Classification (4 marks)

For this question you have been provided with a movie review dataset.  The dataset consists of 50,000 review articles written for movies on IMDb, each labelled with the sentiment of the review  either positive or negative. Your task is to apply logistic regression with dense word vectors to the movie review dataset to predict the sentiment label from the review text.

A simple approach to building a sentiment classifier is to train a logistic regression model that uses aggregated pre-trained word embeddings. While this approach, with simple aggregation, normally works best with short sequences, you will try it out on the movie reviews.

You have been provided with a Python file dense_linear_classifier .py which reads in the dataset and splits it into training, testing, and validation sets; and then loads the pre-trained word embeddings. These embeddings were extracted from the spacy-2 .3 . 5 Python package’s en_core_web_md model and, to save disk space, were filtered to only include words that occur in the movie reviews.

Your task is to use a logistic regression classifier with aggregated word embedding features to determine the sentiment labels of documents from their text.  First implement the document_to_vector function which converts a document into a vector by rst tokenising it (the TreebankWordTokenizer in the nltk package would be an excellent choice) and then aggregating the word embeddings of those words that exist in the dense word embedding dictionary. You will have to work out how to handle words that are missing from the dictionary. For aggregation, the mean is recommended but you could also try other functions such as max. Next, implement the fit_model and test_model functions using your document_to_vector function and LogisticRegression from the scikit-learn package. Using fit_model, test_model, and your training and validation sets you should then try several values for the regularisation parameter C and select the best based on accuracy.  To try regularisation parameters, you should use an automatic hyperparameter search method. Next, re-train your classifier using the training set concatenated with the

validation set and your best C value. Evaluate the performance of your model on the test set. Answer the following questions in your answer PDF:

1. What range of values for C did you try? Explain, why this range is reasonable. Also explain what search technique you used and why it is appropriate here.

2. What was the best performing C value?

3. What was your final accuracy?

Also make sure you submit your code.

Hint: If you do the practical exercise in Lab 3 this question will be much easier.

Tip: If you use TreebankWordTokenizer then for efficiency you should instantiate the class as a global variable. The TreebankWordTokenizer compiles many regular expressions when it is initialised; doing this every time you want to tokenise a sentence is very inefficient. For more details see the documentation3 for TreebankWordTokenizer.

Question 2: Genre Classification (Kaggle competition: 4 marks, Write-up: 3 marks)

For this task you will design and implement a classification algorithm that identifies the genre of a piece of text. This task will be run as a competition on Kaggle. Your marks for this question will be partially based on your results in this competition, but your mark will not be affected by other students’ scores, instead you will be graded against several benchmark solutions. The other part of your mark will come from your code and write-up.

The dataset consists of text sequences from English language books in the genres:

● horror (class id 0),

●  science  fiction (class id 1),

● humour (class id 2), and

● crime  fiction (class id 3).

Each text sequence is 10 contiguous sentences. Your task is to build the best classifier when evaluated with macro averaged F1 score.

Note: the training data and the test data come from different sets of books. You have been provided with docids (examples with the same docid come from the same book) for the training data but not the test data.

You have been provided with an example solution in genre_classifier_0pc .py which shows you how to read in the training data (genre_train .json), test data (genre_test .json) and output a CSV le that the judging system can read. This solution is provided only as an example (it is the 0% benchmark for this problem), you will want to build your own solution from scratch.

Setup

Please register for Kaggle4 using any email account you like.  You do not have to use your full or real name if you would prefer not to.

Submit solutions to https://www.kaggle.com/t/5daca656a1e14e159cfd31131ae1c1e0

To ensure student submissions are anonymous to other students but not the examiner, a unique Kaggle ID has been provided on Wattle as text feedback to a dummy assignment called “Kaggle Unique ID” . This dummy assignment does not accept submissions, and its only purpose is to distribute unique ids.  Set

your team name” to your unique id (e.g. pid123456789). This will allow us to match your submission to your assignment for marking purposes. Note that “team name” is a Kaggle term, this is an individual assignment, and so you should not work with others to complete it.

The Kaggle contest deadline is the same as the assignment deadline. For late submissions see below.

Rules

These rules are designed to ensure a degree of fairness between students with different access to compu- tational resources and to ensure that the task is not trivial. Breaching the contest rules will likely result in 0 marks for the competition part of this assignment.

● Do not use additional supervised training data. That is, you are not allowed to collect a new genre classification dataset to use for training. Pre-training on other tasks, such as language modelling, is permitted.

● Pre-trained non-contextual word vectors (such as word2vec, GloVe, fasttext) may be used even if they require an additional download (e.g.  you may use word vectors from spacy, genism, or fasttext).

● You may use the pre-trained transformers distilbert-base-cased or distilbert-base-uncased from the transformers5library, but other pre-trained transformers are not permitted.

● You can use the following libraries (in addition to Python standard libraries): numpy, scipy, pandas, torch, transformers, tensorflow, nltk, sklearn, xgboost, gensim, spacy, imblearn, torchnlp.

If you would like to use other libraries, please ask on the Piazza forum well in advance of the assignment deadline.

● This is an individual task, do not collude with other individuals.   Copying code from other peoples models or models available on the internet is not permitted.

Judging system

● You will upload your predictions for the test set as a CSV file for judging.

● You are allowed to submit up to 5 times per UTC Day. Since you get immediate feedback from every submission it is best to start submitting early and plan ahead.

● The results shown on the public scoreboard prior to the conclusion of the contest only include 50% of the test data.  Your solution will be judged on the other 50% of the test data when computing

final rankings and marks.

● The judging system allows you to choose which of all your submissions you want to be your final

one.

Competition marking

Marks will be assigned based on which judge baselines you beat on the hidden 50% of the test data. If, for example, you beat the 80% baseline but do not beat the 90% baseline you will be awarded a mark on a linear scale between these two based on your macro averaged F1 score.  Exceeding the score of the 100% baseline will give you a mark of 100% for the competition component of this question. (The

100% baseline and all other baselines can be trained in less than 24 hours on a laptop PC without GPU support your solution may make use of any compute resources available to you). Using Google Colab6 is recommended if you want to do GPU training but do not have access to a dedicated GPU.

Write-up

A fraction of your marks will be based on a write-up that a minimum describes:

● How your nal solution works.

● How you trained and tested your model (e.g. validation split(s), hyperparameter search etc.).

● What models you tested, which worked, and which didn’t.

● Why you think these other models didn’t work.

Aim for 1 page or slightly less. Only the rst 2 pages will be marked. Bullet points are acceptable if they are understandable.

What to submit

● Submit the code of your best solution (including training pipeline) to Wattle. Also make sure your team name is set to your unique Kaggle ID. Do not submit stored parameters or data.

● Submit your write-up in your answer PDF file.

● In one of the three cases below, you should also submit a CSV file with your model’s output in the correct judging format. The CSV file should be named with your Uni ID, e.g. your_uid .csv.

(a) You have been granted an extension, OR

(b) You could not use Kaggle and only if you could not use Kaggle, OR

(c) You decide to submit within 24 hours after the assignment deadline (i.e.  late submission, with 5% penalty).

Question 3: RNN Name Generator (4 marks)

Your task is to develop an autoregressive RNN model which can generate people’s names. The RNN will generate each character of a person’s name given all previous characters.  Your model should look like the following when training:

 

Note that the input is shown here as a sequence of characters but in practice the input will be a sequence of character ids. There is also a softmax non-linearity after the linear layer but this is not shown in the diagram.  The output (after the softmax) is a categorical probability distribution over the vocabulary, what is shown as the output here is the ground truth label. Notice that the input to the model is just the expected output shifted to the right one step with the <bos> (beginning of sentence token) prepended. The three dots to the right of the diagram indicate that the RNN is to be rolled out to some maximum length. When generating sequences, rather than training, the model should look like the following:

 

Specifically, we choose a character from the probability distribution output by the network and feed it as input to the next step. Choosing a character can be done by sampling from the probability distribution or

by choosing the most likely character (otherwise known as argmax decoding).

The character vocabulary consists of the following:

“”  The null token padding string.

<bos>  The beginning of sequence token.

.  The end of sequence token.

a-z  All lowercase characters.

A-Z  All uppercase characters.

0-9  All digits.

“    The space character.

Starter code is provided in rnn_name_generator.py, and the list of names to use as training and validation sets are provided in names_small .json.

To complete this question you will need to complete three functions and one class method:  the func- tion seqs_to_ids, the forward method of class RNNLM, the function train_model, and the function gen_string. In each case you should read the description provided in the starter code.

seqs_to_ids: Takes as input a list of names.  Returns a 2D numpy matrix containing the names rep- resented using token ids.  All output rows (each row corresponds to a name) should have the same length of max_length, achieved by either truncating the name or padding it with zeros. For example, an input of:  [“Bec .”, “Hannah .”, “Siqi .”]  with a max_length set to 6 should return (normally we will use max_length  =  20 but for this example we use 6)

[[30  7  5  2  0  0]

[36  3  16  16  3  10]

[47  11  19  11  2  0]]

Where the rst row represents Bec .”  and two padding characters, the second row represents “Hannah”, the third row represents Siqi .” with one padding character.

forward: A method of class RNNLM. In this function you need to implement the GRU model shown in the diagram above. The layers have all been provided for you in the class initialiser.

train_model: In this method you need to train the model by mini-batch stochastic gradient decent. The optimiser and loss function are provided to you. Note that the loss function takes logits (output of the linear layer before softmax is applied) as input. At the end of every epoch you should print the validation loss using the provided calc_val_loss function.

gen_string: In this method you will need to generate a new name, one character at a time. You will also need to implement both sampling and argmax decoding.

For this question, please include in your answers PDF the most likely name your code generates using argmax decoding as well as 10 different names generated using sampling.  Also remember to submit your code. Your code should all be in the original rnn_name_generator .py file, other les will not be marked. For this question you should not import additional libraries, use only those provided in the starter code (you may uncomment the import tqdm statement if you want to use it as a progress bar).

For example, one of the names sampled from the model solution was: Dasbie  Miohmazie