Starting from:
$30

$24

Assignment 5: Cancer Classification using Machine Learning Solution

Overview






The goal of this project is to gain more practice with using functions, lists and dictionaries and gain some intuition for Machine Learning, the eld of computer science concerned with writing algorithms that allow computes to \learn" from data.




The problem we’ll be solving is as follows: Given a data le containing hundreds of patient records with values describing measurements of cancer tumors and whether or not each tumor is malignant or benign, develop a simple rule-based classi er that can be used to predict whether an as-yet-unseen tumor is malignant or benign.




The general idea is that malignant tumors are di erent than benign tumors. Malignant tumors tend to have larger radii, to be more smooth, to be more symmetric, etc. Measurements have been taken on many tumors whose class (malignant or benign) is known. The code you are going to write will get the average score across all the malignant tumors for an attribute (e.g. ‘area’) as well as the average score for that attribute for benign tumors. Let’s say that the average area for malignant tumors is 100, and for benign tumors is 50. We can then use that information to try to predict whether a given tumor is malignant or benign.




Imagine you are presented with a new tumor and told the area was 99. All else being equal, we would have reason to think this tumor is more likely to be malignant than had its area been 51. Based on this intuition, we are going to create a simple classi cation scheme. We will calculate the midpoint between the malignant average and the benign average (75 in our hypothetical example), and simply say that for each new tumor, if its value for that attribute is greater than or equal to the midpoint value for that attribute, that is one vote for the tumor being malignant. Each attribute that we are using produces a vote, and at the end of counting votes for each attribute, if the malignant votes are greater than or equal to the benign votes, we predict that the tumor is malignant.







Machine Learning Framework






\Machine learning" is a popular buzzword that might evoke computer brain simulations, or robots walking among humans. In reality (for now, anyway), machine learning refers to some-thing less fanciful: algorithms that use previously observed data to make predictions about new data. It may sound less glamorous than fully sentient robots, but that’s exactly what was described above! You can get more sophisticated about the speci cs of how you go about this, but that’s the core of what machine learning really means.




If using data to make predictions on new data is our goal, you might think it makes sense to use all the data we have to learn from. But in fact, if we truly don’t know the labels (e.g., malignant or benign) of the data we’re testing our algorithm on, we won’t have any idea whether it’s doing a good job! For this reason, it makes sense to split the data we have labels for into a training




set, which we’ll use to \learn" from, and a test set, which we’ll use to evaluate how well the algorithm does on new data (i.e., data it wasn’t trained on). We will take about 80% of the data as our training set, and use the remaining 20% as our test set.







2.1 Training Phase




Here’s how our classi er will work: In the training phase, we will \learn" (read: compute) the average value each attribute (e.g. area, smoothness, etc.) among the malignant tumors. We will also \learn" (again: compute) the average value of each attribute among benign tumors. Then we’ll compute the midpoint for each attribute. This collection of midpoints, one for each attribute, is our classi er.







2.2 Testing Phase




Having trained our classi er, we can now use it to make an educated guess about the label of a new tumor if we have the measurements of all of its attributes. Our educated guess will be pretty simple:







If the tumor’s value for an attribute is greater than or equal to the midpoint value for that attribute, cast one vote for the tumor being malignant.




If the tumor’s attribute value is less than the midpoint, cast one vote for the tumor being benign.




Tally up the votes cast according to these rules for each of the ten attributes. If the malignant votes are greater than or equal to the benign votes, we predict that the tumor is malignant.







If we want to use this classi er to diagnose people, we have an important question to answer: how good are our guesses? To answer this question, we’ll run test our algorithm on the 20% of our data that we held out as the test set, which we didn’t use to train the classi er, but we do know the correct labels. Our rate of accuracy on these data should be indicative of how well our classi er will do on new, unlabeled tumors.







Dataset Description






You have been provided with cancertTrainingData.txt, a text le containing the 80% of the data that we’ll use as our training set.




The le has many numbers per patient record, some of which refer to attributes of the tumor. The skeleton code includes the function make_training_set(), which reads in the important information from this le and produces a list of dictionaries. Each dictionary contains attributes for a single tumor as follows:




0. ID



radius



texture



perimeter



area



smoothness



compactness



concavity



concave



symmetry



fractal



class






The middle 10 attributes (numbered 1 through 10) are the numbers that describe the tumor. The rst attribute is just the patient ID number, and the last attribute is the actual real life state of the tumor, namely, malignant (represented by \M") or benign (represented by \B").




We don’t need to know what these attributes mean: all we need to know is that they are measurements of the tumors, and that benign and malignant tumors tend to have di erent attribute values. For these 10 tumor attributes when comparing to the midpoint values, higher numbers indicate malignancy. Pictorially, the list of dictionaries looks like this (two are shown, but the list contains many more than that):







training_set




list




. . .







dict


ID
897880




radius
10.05




texture
17.53




perimeter
64.41




area
310.8




smoothness
0.1007




compactness
0.07326




concavity
0.02511




concave
0.01775




symmetry
0.189




fractal
0.06331




class
B













dict


ID
89812




radius
23.51




texture
24.27




perimeter
155.1




area
1747




smoothness
0.1069




compactness
0.1283




concavity
0.2308




concave
0.141




symmetry
0.1797




fractal
0.05506




class
M







Figure 1: Illustration of the data layout of the training set returned by make training set










The dictionary stored in the 0th spot in the list gives the attributes for the 0th tumor: training_set[0]["class"] gives the true class label (in this case, "B" for benign) of the 0th tumor.
Getting Started






Download the skeleton code (cancer_classifier.py), training set (cancerTrainingData.txt), and the test set (cancerTestingData.txt). Make sure all three les are in the same directory, or the main program will not be able to load the data from the les.







Tasks






5.0 Overview




Training and evaluating our classi er involves several steps. The rst task, which has been done for you, is to write code to load the training and test data sets from text les into lists of dictionaries representing patient records, as described in the previous section. The functions make_training_set and make_test_set are included in the skeleton code to complete these steps.




You will complete the following four tasks:







TODO 1: Train the classi er




TODO 2: Apply the classi er to the test set




TODO 3: Calculate and report accuracy on the test set




TODO 4: Provide classi er details on user-speci ed patients







The main program has been provided to you: you will be implementing functions that are called from the main program at the bottom of the skeleton code le. Take a moment to read through and understand the main program (notice that the parts of the program that use TODOs 1{4 are commented out).




Each of the above steps is described in detail in the remainder of this section. After you nish each TODO (2 and 3 are completed together), uncomment the corresponding block in the main program and run your code to make sure your output matches the sample output provided below.







5.1 TODO 1: Train the classi er




A classi er is simply some model of a problem that allows us to make predictions about new records. We use the training set to build up a simple model, as described in Section 2:







For all malignant records, calculate the average value of each attribute. For all benign records, calculate the average value of each attribute.




Calculate the midpoint between these averages for each attribute.




Our classi er is a single dictionary that stores this midpoint value for each attribute.




Implement this functionality in train_classifier. My solution for this part totals roughly 30 lines of code. As always, you may nd it useful to write helper methods that perform smaller tasks: for example, you could create a helper function to initialize a dictionary with each of the attributes as keys and 0 as values.




When done, uncomment the block of code in the main program that calls train_classifier and debug your code until your attribute midpoints match the sample output.







5.2 TODO 2: Apply the classi er




After computing the classi er (namely, the dictionary of attribute midpoints), we can use these values to make predictions given the attribute values of a new patient. A record is classi ed as follows:




For each attribute, determine whether the record’s value is less than or equal to the classi er’s midpoint value. If so, cast one vote for Benign; otherwise, cast one vote for Malignant. If the votes for Malignant are greater than or equal to the votes for Benign, the record is classi ed as Malignant; otherwise, it is classi ed as Benign.




Implement this classi cation scheme in the classify function, applying it to each record in the test set. Notice that the prediction for a record is to be stored in the "prediction" eld of the dictionary for that record.







5.3 TODO 3: Report accuracy




For each record in the test set, compare the predicted class to the actual class. Print out the percentage of records that were labeled correctly (i.e., the predicted class is the same as the true class).







5.4 TODO 4: Provide patient details




The nal task is to provide a user the opportunity to examine the details of the predictions made for individual patients. Implement check_patients, which contains commented pseudocode describing its the exact behavior. You are strongly encouraged to write helper functions that are called from within this function: if a pseudocode step requires more than a few lines of code, consider making a helper function to accomplish that step.




If the user-speci ed patient ID is found in the test set, print a table with four columns:




Attribute: the name of the attribute




Patient: the patient’s value for that attribute




Classi er: the classi er’s threshold (midpoint) for that attribute




Vote: the vote cast by the classi er on for that attribute




See the sample output for speci cs of what the table should look like.




Printing a table of results with nice even columns and uniformly formatted decimal numbers requires delving into the details of string formatting in Python. There are multiple ways to do this, but the following tips should be su cient1:




String objects have rjust and ljust methods, which return a copy of the string padded to the given width, justi ed either right or left. For example, "abc".rjust(5) returns " abc".







Floating point numbers can be formatted nicely using the format method, which is called on strings containing special formatting speci ers. For example, "{.2f}".format(8.632) formats the argument (8.632) as a oat with 2 decimal places, resulting in the string "8.63".







Your table does not need to match the sample output character for character, but your columns should be lined up, right justi ed, and oating-point values should be printed with the decimals aligned and a consistent number of digits following the decimal point.







Sample Output






A sample run of my solution program is shown below. User input is bolded.







Reading in training data...




Done reading training data.




Reading in test data...




Done reading test data.




Training classifier...




Classifier cutoffs:




radius: 14.545393772893773




texture: 19.279093406593404




perimeter: 94.91928571428579




area: 693.337728937729




smoothness: 0.09783294871794869




compactness: 0.1104729532967033




concavity: 0.09963735815018318




concave: 0.054678068681318664




symmetry: 0.18456510989010982




fractal: 0.06286657967032966




Done training classifier.




Making predictions and reporting accuracy Classifier accuracy: 92.20779220779221







For much more detail on string formatting, see the Python Tutorial entry: https://docs.python.org/3/ tutorial/inputoutput.html



Done classifying.






Enter a patient ID to see classification
details: 897880
Attribute
Patient
Classifier
Vote
radius
10.0500
14.5454
Benign
texture
17.5300
19.2791
Benign
perimeter
64.4100
94.9193
Benign
area
310.8000
693.3377
Benign
smoothness
0.1007
0.0978
Malignant
compactness
0.0733
0.1105
Benign
concavity
0.0251
0.0996
Benign
concave
0.0177
0.0547
Benign
symmetry
0.1890
0.1846
Malignant
fractal
0.0633
0.0629
Malignant
Classifier’s diagnosis: Benign




Enter a patient ID to see classification
details: 89812
Attribute
Patient
Classifier
Vote
radius
23.5100
14.5454
Malignant
texture
24.2700
19.2791
Malignant
perimeter
155.1000
94.9193
Malignant
area
1747.0000
693.3377
Malignant
smoothness
0.1069
0.0978
Malignant
compactness
0.1283
0.1105
Malignant
concavity
0.2308
0.0996
Malignant
concave
0.1410
0.0547
Malignant
symmetry
0.1797
0.1846
Benign
fractal
0.0551
0.0629
Benign
Classifier’s diagnosis: Malignant


Enter a patient ID to see classification
details: quit






Hints and Guidelines



Start by reading through the skeleton code, and making sure you know what the main program does and how the functions you are tasked with implementing t into the overall program.




If your understanding of lists and dictionaries is shaky, you will have great di culty making progress. Visit my o ce hours, TA o ce hours, or mentor hours early so you don’t spend too much time struggling.




The top of the skeleton le has a global variable called ATTRS, which is a list of the attribute names each patient record has. Using global variables with all-caps names is a common convention when you have variables that need to be referenced all over your program and (crucially) never change value. You may refer to ATTRS from anywhere in your program, including inside function de nitions, without passing it in as a parameter.




As in A4, all variables (other than ATTRS) referenced from within functions must be local variables - if you need access to information from outside the function, it must be passed into the function as a parameter.




When iterating over patient record dictionaries, use loops over the keys stored in ATTRS rather than looping directly over the dictionary’s keys. An example of this appears in the main program where the classi er cuto s are printed.




The functions provided in the skeleton code include headers and speci cations. Make sure you follow the given speci cations (and don’t modify them!).




Keep the length of each function short: if you’re writing a function that takes more than about 30 lines of code (not including comments and whitespace), consider how you might break the task into smaller pieces and implement each piece using a helper function.




All helper functions you write must have docstrings with precise, clearly written speci - cations.




Test each function after you’ve written it by running the main program with the corre-sponding code block uncommented. Don’t move on until the corresponding portion of the output matches the sample.







Submission







Upload cancer_classifier.py to Canvas and ll in the A5 Hours quiz with an estimate of how many hours you spent working on this assignment.







Rubric







Submission Mechanics (2 points)






File called cancer_classifier.py is submitted to Canvas
2












Code Style and Clarity (28 points)






Comment at the top with author/date/description
3












Comments throughout code clarify any nontrivial code sections
5












Variable and function names are descriptive
5












Helper functions are used to keep functions no longer than about 30 lines of
5




code (not counting comments and blank lines)














ATTRS is used to iterate over dictionary attributes
5












No global variables except ATTRS are referenced from within functions
5












Correctness (70 points)






The trained classi er has the correct midpoint values for each attribute
30












Prediction is performed as described using the midpoints computed in training
5












Accuracy is computed and reported correctly as shown in the demo output
10












User is repeatedly prompted for Patient ID
5












Message is printed if given ID is not in the test set.
5












If ID is in the test set, table is printed with all four columns and rows for all 10
10




attributes














Table columns are right-justi ed and aligned
3












Floating-point values in the table are lined up on the decimal point and have a
2




xed number of digits after the decimal.






















Total
100 points










Acknowledgements







This assignment was adapted from a version used by Perry Fizzano, who adapted it from an original assignment developed at Michigan State University for their CSE 231 course.







Challenge Problem






The following challenge problem is worth up to 10 points of extra credit. As usual, this can be done using only material we’ve learned in class, but it’s much more open-ended. If you are trying to tackle this, feel free to come talk to me about it in o ce hours.




In this assignment, you trained a simple classi er that used means over the entire training set to classify unseen examples. This simple classi er does quite well (92% accuracy) on the test set. There are many more sophisticated methods for learning classi ers from training data, some of which depend on some pretty heavy-duty mathematical derivations.




One type of classi er that doesn’t require a lot of math but nonetheless performs pretty well on a lot of real-world problems is called the Nearest Neighbor Classi er or its more general cousin, the K Nearest Neighbors Classi er. The idea behind KNN is that records with similar attributes should have similar labels. So a reasonable way to guess the label of a previously-unseen record is to nd the record from the training set that is most similar to it and guess that record’s label.




To implement a nearest neighbor classi er, we need some de nition of what it means to be "near". One of the the simplest choices for numerical data like ours is the Sum of Squared Di erences metric. Given two records, compute the di erence between the two records’ values, square the di erence, and add up all the squared di erences over all 10 attributes. The smaller the SSD metric, the more similar the two records are.




(up to 5 points) Implement a nearest neighbor classi er using the SSD metric in a le called KNN.py. Feel free to copy and re-use the data loading functions, and any other functions that remain relevant, from cancer_classifier.py. Evaluate your classi er’s accuracy like we did in the base assignment. Write a comment in KNN.py reporting your classi er’s performance.




You might notice an issue with the SSD metric applied to our dataset: some attributes have huge values (e.g., in the hundreds) and others have tiny values. When computing SSD, the large-valued attributes will dominate the SSD score, even if they aren’t the most important attributes. Come up with a way to modify the distance metric so that it weights attributes evenly. Describe your approach and compare the performance of your new metric with SSD in a comment in KNN.py.




(up to 5 points) The nearest neighbor classi er can be a bit ddly, because the guess depends on a single datapoint in your training set, so a single unusual (or mislabeled!) training record could cause a wrong prediction. A more robust classi er looks not just at the single nearest neighbor, but each of K nearest neighbors, for some choice of K. Generalize your nearest neighbor classi er to a KNN classi er. Try out di erent values of K and include a comment discussing the classi cation accuracy for values of K. Do any of them beat the base assignment classi er’s performance?

More products