Statistics:
Exam 2 will take place on Monday, June 3. It will take place in CSE2 G10 from 2:30 pm - 3:20 pm.
You are allowed to bring ONE 8.5x11 INCHES 🙂 piece of paper for notes. You may use both front and back sides and it may be handwritten or typed.
There will also be a reference sheet provided with every exam with a list of methods for the classes and libraries we have learned so far. This list shown at the end of this page.
You are responsible for understanding the following topics for Exam 2. You will still be expected to know the essential things from Exam 1 (e.g. how to write a method, how to write a class, how to tell the efficiency of a method), but the emphasis will be on material we covered that was not on Exam 1.
set
to make access \(\mathcal{O}(1)\)geopandas
, matplotlib
):
GeoDataFrame
using the various methods we have discussed in class or used on the assignmentsGeoSeries
or GeoDataFrame
(e.g. max
, count
, mean
) as well as how to aggregate data using groupby
and dissolve
numpy
):
The following topics will NOT be covered on Exam 2:
np.dot
)geopandas.read_file
, imageio
)scikit-image
libraryAs described above, a reference sheet will be provided on the exam that has a list of functions and their parameters that we learned so far. This reference sheet will only contain method names and parameters and will not have descriptions of the methods or examples; if you feel like you would like descriptions for the methods, you may use space on your cheat sheet to do write them down.
The reference sheet for Exam 2 will build off of the one for Exam 1 with new functions added in. We removed things that will be guaranteed that you will not need to write for Exam 2.
We wil adopt a new convention that puts optional parameters with a question mark after their name.
print(*strings)
range(end)
range(start, end, step?)
abs(v)
min(v1, v2)
max(v1, v2)
sum(v1, v2)
open(fname)
zip(l1, l2)
int(v)
, float(v)
, str(v)
, bool(v)
upper()
lower()
find(s)
strip()
split()
list()
or []
append(val)
extend(lst)
insert(idx, val)
remove(val)
pop(idx)
index(val)
reverse()
sort(key?)
set()
add(val)
remove(val)
dict()
or {}
keys()
values()
items()
__init__
__repr__
__eq__
__hash__
mean()
min() / max()
idxmin() / idxmax()
count()
unique()
groupby(col)
apply(fun)
isnull()
notnull()
dropna()
fillna(val)
sort_values(col)
sort_index()
nlargest(n, col)
merge(other, left_on, right_on, how)
index
loc[row, col]
plot(column?, legend?, ax?, color?, vmin?, vmax?)
dissolve(by, aggfunc)
geopandas.sjoin(left, right, op, how)
plt.subplots(nrows, ncols)
plt.show()
plt.savefig(f_name)
np.array(vals?)
np.arange(end)
np.arange(start, end, step?)
np.ones(shape)
, np.zeros(shape)
np.dot(a1, a2)
np.sum(a)
, np.min(a)
, np.max(a)
, np.mean(a)
reshape(shape)
sum()
, min()
, max()
, mean()
copy()
shape
Remember that exams are only worth 25% of your total grade, with exam 1 being about half of that. If exam 1 did not go as well as you wanted it to, you still have time to improve for exam 2. This also means much more of your grade is determined by things outside of exams so there is a lot of other things you can also focus on.
Statistics:
Exam 1 will take place on Friday, May 10. It will take place in CSE2 G10 from 2:30 pm - 3:20 pm.
You are allowed to bring one 8.5x11 INCHES 🙂 piece of paper for notes. You may use both front and back sides and it may be handwritten or typed.
There will also be a reference sheet provided with every exam with a list of methods for the classes and libraries we have learned so far. This list shown at the end of this page.
You are responsible for understanding the following topics:
Series
or DataFrame
using the various methods we’ve discussed in class or used on the assignments.DataFrame
(e.g. max
, count
, mean
) as well as how to aggregate data using groupby
.seaborn
.sklearn
to train a machine learning model using best practices.The following topics will NOT be covered on the Exam 1:
6:0:-2
, but we won’t test your knowledge of negative indices like -4:-1
.resample
.matplotlib
to customize visualizations.As described above, a reference sheet will be provided on the exam that has a list of functions and their parameters that we learned so far. This reference sheet will only contain method names and parameters and will not have descriptions of the methods or examples; if you feel like you would like descriptions for the methods, you may use space on your cheat sheet to do write them down.
print(*strings)
range(end)
range(start, end[, step])
abs(v)
min(v1, v2)
max(v1, v2)
sum(v1, v2)
open(fname)
int(v)
, float(v)
, str(v)
, bool(v)
upper()
lower()
find(s)
strip()
split()
list()
or []
append(val)
extend(lst)
insert(idx, val)
remove(val)
pop(idx)
index(val)
reverse()
sort(key=None)
set()
add(val)
remove(val)
dict()
or {}
keys()
values()
items()
readlines()
read()
__init__
__repr__
__eq__
pd.read_csv
mean()
min() / max()
idxmin() / idxmax()
count()
unique()
groupby(col)
apply(fun)
isnull()
notnull()
dropna()
fillna(val)
sort_values(col)
sort_index()
nlargest(n, col)
index
loc[row, col]
sns.catplot(x, y, data, kind[, hue])
["count", "bar", "violin"]
sns.relplot(x, y, data, kind[, hue[, size]])
["scatter", "line"]
sns.regplot(x, y, data)
sklearn.metrics.accuracy_score(y_true, y_pred)
sklearn.metrics.mean_square_error(y_true, y_pred)
sklearn.model_selection.train_test_split(X, y, test_size)
sklearn.tree.DecisionTreeClassifier()
sklearn.tree.DecisionTreeRegressor()
fit(X, y)
predict(X)