Complete visualizations for Milestone 4

This commit is contained in:
Nat 2023-03-25 15:54:16 -07:00
parent 4566009cae
commit db02382154
Signed by: nat
GPG Key ID: B53AB05285D710D6
20 changed files with 5922 additions and 11134 deletions

File diff suppressed because one or more lines are too long

View File

@ -1,3 +1,5 @@
import numpy as np
from math import pi
import pandas as pd
def load_and_process():
@ -30,11 +32,12 @@ def load_and_process():
pol = (
pd.read_csv("../data/raw/countypres_2000-2020.csv")
.query("`year` == 2012")
.query("`year` == 2012 & `party` == 'DEMOCRAT'")
.reset_index()
.drop([
"year", "state", "county_fips", "office",
"candidate", "version", "mode", "index",
"party"
], axis="columns")
.rename({
"county_name": "county",
@ -42,28 +45,102 @@ def load_and_process():
"candidatevotes": "votes",
"totalvotes": "total"
}, axis="columns")
.apply(lambda x: x.str.capitalize() if x.name == "county" or x.name == "party" else x)
.apply(lambda x: x.str.capitalize() if x.name == "county" else x)
.apply(combine_name_state, axis="columns")
.merge(counties, left_on="county", right_on="name")
.drop(["state", "name"], axis="columns")
.assign(percent=lambda x: x.votes/x.total)
.drop(["votes", "total"], axis="columns")
)
## gb - the gaybourhoods dataset
gb = (
pd.read_csv("../data/raw/gaybourhoods.csv")
.merge(cords, left_on="GEOID10", right_on="ZIP") \
.merge(cords, left_on="GEOID10", right_on="ZIP")
.drop([
"Tax_Mjoint", "TaxRate_SS", "TaxRate_FF", "TaxRate_MM",
"Cns_RateSS", "Cns_RateFF", "Cns_RateMM", "CountBars",
"Mjoint_MF", "Mjoint_SS", "Mjoint_FF", "Mjoint_MM",
"Cns_TotHH", "Cns_UPSS", "Cns_UPFF", "Cns_UPMM",
"ParadeFlag", "FF_Tax", "FF_Cns", "MM_Tax", "MM_Cns",
"SS_Index_Weight", "Parade_Weight", "Bars_Weight",
"GEOID10", "ZIP",
], axis="columns") \
"GEOID10", "ZIP", "FF_Index", "MM_Index",
], axis="columns")
.rename({
"LAT": "lat",
"LNG": "long",
}, axis="columns")
)
def kinsify(index, **kwargs):
max_index = 25
if index < max_index/7:
return 0
elif index < max_index*2/7:
return 1
elif index < max_index*3/7:
return 2
elif index < max_index*4/7:
return 3
elif index < max_index*5/7:
return 4
elif index < max_index*6/7:
return 5
else:
return 6
gb["kinsey"] = gb.SS_Index.apply(kinsify, axis="columns")
percent_democrat = np.empty(len(gb.index))
neighbourhood_kinsey = np.empty(len(gb.index))
for i, row in gb.iterrows():
percent_democrat[i] = nearest_neighbour(pol, (row.long, row.lat), interval=.1).percent
neighbourhood_kinsey[i] = select_smallest_neighbourhood(gb, (row.long, row.lat), interval=.1).kinsey.mean()
gb["percent_democrat"] = pd.Series(data=percent_democrat)
gb["neighbourhood_kinsey"] = pd.Series(data=neighbourhood_kinsey)
return (gb, pol, counties, cords)
def select_region(df, left, right, bottom, top):
"""
Takes a dataframe with columns `long` and `lat` corresponding to
coordinates and returns a subset of the dataframe containing only entries
between the given boundaries
"""
return df[(df["long"] > left) & (df["long"] < right) & (df["lat"] > bottom) & (df["lat"] < top)]
def select_smallest_neighbourhood(df, pos, interval=1, multiplier=1.5, expansion_limit=10):
subset = select_region(df, pos[0]-interval, pos[0]+interval, pos[1]-interval, pos[1]+interval)
cinterval = interval
while subset.count().lat == 0:
cinterval += interval
#interval *= multiplier
subset = select_region(df, pos[0]-cinterval, pos[0]+cinterval, pos[1]-cinterval, pos[1]+cinterval)
return subset
def nearest_neighbour(df, pos, interval=1, multiplier=1.5, expansion_limit=10):
"""
Given a dataframe with columns `long` and `lat` corresponding to
coordinates and a `pos` pair of long/lat coordinates, determine the
coordinates of the nearest observation in the dataset by running the
following algorithm:
1. Find all points within (long+-interval, lat+-interval)
2. If there are no other points within the range, start from step 1 and
set interval *= multiplier
3. Calculate the distance between pos and each point in the interval
3. Return the point with the lowest distance that isn't pos
"""
subset = select_smallest_neighbourhood(df, pos, interval, multiplier, expansion_limit)
subset = subset.assign(distance=distance(*pos, subset["lat"], subset["long"]))
return subset.sort_values("distance").reset_index().iloc[0]
# Efficient implementation of the haversine formula
# Source: https://stackoverflow.com/a/21623206
def distance(lat1, lon1, lat2, lon2):
p = pi/180
a = 0.5 - np.cos((lat2-lat1)*p)/2 + np.cos(lat1*p) * np.cos(lat2*p) * (1-np.cos((lon2-lon1)*p))/2
return 12742 * np.arcsin(np.sqrt(a))

BIN
analysis/images/atlanta.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 556 KiB

BIN
analysis/images/austin.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 204 KiB

BIN
analysis/images/boston.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 381 KiB

BIN
analysis/images/chicago.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 413 KiB

BIN
analysis/images/denver.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 427 KiB

BIN
analysis/images/houston.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 205 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 523 KiB

BIN
analysis/images/miami.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 285 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 481 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 240 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 405 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 452 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 230 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 269 KiB

BIN
analysis/images/seattle.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 168 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 538 KiB

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff