Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
27 changes: 20 additions & 7 deletions learning_curve.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,17 +8,30 @@

data = load_digits()
print data.DESCR
num_trials = 10
train_percentages = range(5,95,5)
num_trials = 100
train_percentages = range(5, 95, 5)
test_accuracies = numpy.zeros(len(train_percentages))

# train a model with training percentages between 5 and 90 (see train_percentages) and evaluate
# the resultant accuracy.
# You should repeat each training percentage num_trials times to smooth out variability
# for consistency with the previous example use model = LogisticRegression(C=10**-10) for your learner
# train a model with training percentages between 5 and 90 (see
# train_percentages) and evaluate the resultant accuracy.
# You should repeat each training percentage num_trials times to smooth out
# variability for consistency with the previous example use
# model = LogisticRegression(C=10**-10) for your learner

# TODO: your code here
# create a model
model = LogisticRegression(C=10**-100)

# for loop so the things happen for the correct values enough times
for n in range(len(train_percentages)):
for i in range(num_trials):
# Split the data and train on some of it, store data correctly
X_train, X_test, y_train, y_test = train_test_split(data.data,
data.target,
train_size=train_percentages[n])
model.fit(X_train, y_train)
test_accuracies[n] = test_accuracies[n] + model.score(X_test, y_test)

# create plot and put the correct things on it
fig = plt.figure()
plt.plot(train_percentages, test_accuracies)
plt.xlabel('Percentage of Data Used for Training')
Expand Down
7 changes: 7 additions & 0 deletions questions.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
1. The general trend in the curve is positive.

2. At only 10 trials, almost the entire curve seems to be noisy. Typically, accuracy from training on 5% on the test is lower than on 90%, but everything between is pretty inconsistent.

3. 100 trials makes a fairly smooth curve, but it's still pretty piecewise.

4. The scale of accuracy on the test set gets smaller with a smaller C value and the entire curve becomes less smooth.