-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
group#5: lin reg script added #9
base: main
Are you sure you want to change the base?
Changes from 1 commit
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,77 @@ | ||
import numpy as np | ||
|
||
def compCostFunction(estim_y, true_y): | ||
E = estim_y - true_y | ||
C = (1 / 2 * m) * np.sum(E ** 2) | ||
return C | ||
|
||
def test_dimensions(x, y): | ||
# this checks whether the x and y have the same number of samples | ||
assert isinstance(x, np.ndarray), "Only works for arrays" | ||
assert isinstance(y, np.ndarray), "Only works for arrays" | ||
return x.shape[0] == y.shape[0] | ||
|
||
# To be deleted later | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. are these comments obsolete? if yes, please remove |
||
# feature_1 = np.linspace(0, 2, num=100) | ||
|
||
X = np.random.randn(100,3) # feature matrix | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. could the variables be named with more informative names? |
||
y = 1 + np.dot(X, [3.5, 4., -4]) # target vector | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I'd write this as y = 1 + X @ [3.5, 4., -4]) # target vector
# z = 2 + y @ feature_matrix @ feature_matrix.T |
||
|
||
m = np.shape(X)[0] # nr of samples | ||
n = np.shape(X)[1] # nr of features | ||
|
||
def iterativeLinearRegression(X, y, alpha=0.01): | ||
""" | ||
This makes iterative LR via gradient descent and returns estimated parameters and history list. | ||
""" | ||
steps=500 | ||
X = np.concatenate((np.ones((m, 1)), X), axis=1) | ||
|
||
W = np.random.randn(n + 1, ) | ||
|
||
# stores the updates on the cost function | ||
cost_history = [] | ||
# iterate until the maximum number of steps | ||
for i in np.arange(steps): # begin the process | ||
|
||
y_estimated = X.dot(W) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
|
||
|
||
cost = compCostFunction(y_estimated, y) | ||
# Update gradient descent | ||
E = y_estimated - y | ||
gradient = (1 / m) * X.T.dot(E) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more.
|
||
|
||
W = W - alpha * gradient | ||
if i % 10 == 0: | ||
print(f"step: {i}\tcost: {cost}") | ||
|
||
cost_history.append(cost) | ||
|
||
return W, cost_history | ||
|
||
params, history = iterativeLinearRegression(X, y) | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The code should be restructured so that the module can be imported and does nothing. |
||
|
||
# test 1 | ||
print(params) | ||
print(history) | ||
|
||
import matplotlib.pyplot as plt | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This should be moved to the header |
||
plt.plot(history) | ||
plt.xlabel("steps") | ||
plt.show() | ||
|
||
# test 2 | ||
|
||
X = np.random.randn(500,2) # feature matrix | ||
y = np.dot(X, [5, -1]) # target vector | ||
|
||
m = np.shape(X)[0] # nr of samples | ||
n = np.shape(X)[1] # nr of features | ||
|
||
params, history = iterativeLinearRegression(X, y) | ||
print(params) | ||
|
||
import matplotlib.pyplot as plt | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This duplicates line above… |
||
plt.plot(history) | ||
plt.xlabel("steps") | ||
plt.show() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
where does the variable m come from here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, global variables are terrible. Please move all the code defining variables into a function.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
... and
m
should be a parameter that is passed into the function.