---
title: "City Compare"
output:
html_document:
toc: true
toc_depth: 5
toc_float:
collapsed: false
smooth_scroll: true
editor_options:
chunk_output_type: console
---
# Input Data & Configuration
## Libraries
```{r libs, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
date()
rm(list=ls())
library(tidyverse)
library(tidycensus)
library(sf)
library(openmeteo)
library(maps)
library(scales)
```
## API keys
```{r api_keys, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
# load census api key
census_api_key(key = substr(read_file(file = "api_keys/census_api_key"), 1, 40))
```
## Date ranges
```{r date_range, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
date_start <- "2014-01-01"
date_end <- "2024-12-31"
```
## Cities to compare
```{r cities, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
cities <- read_csv(file = "cities.csv")
cities <- cities |>
mutate(city_name = paste0(City, " ", Type))
```
# Data
## Census data
```{r census, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
populations <- list(NULL)
for(city in cities$city_name){
state <- cities |> filter(city_name == city) |> pull(State)
populations[[city]] <- get_acs(
geography = "place",
variables = "B01003_001",
state = state,
year = 2023,
geometry = TRUE
) |>
filter(str_detect(NAME, city))
}
populations <- bind_rows(populations)
city_center <- populations |>
st_centroid() |>
st_transform(4326) %>%
mutate(
lon = st_coordinates(.)[,1],
lat = st_coordinates(.)[,2]
) |>
st_drop_geometry() |>
select(lat, lon)
cities <- bind_cols(cities, populations, city_center)
cities <- cities |> mutate(
density = estimate/as.double(st_area(geometry))*1000000
)
ggplot(cities) +
geom_col(aes(x = City,
y = estimate)) +
labs(title = "City Population",
x = "City",
y = NULL) +
scale_y_continuous(label = unit_format(unit = "K", scale = 1e-3, sep = ""))
ggplot(cities) +
geom_col(aes(x = City,
y = density)) +
labs(title = "City Density",
x = "City",
y = "people/km^2") +
scale_y_continuous()
```
## Map cities
```{r cities_map, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
ggplot(data = cities) +
geom_polygon(data = map_data(map = "state"),
aes(long, lat, group = group),
fill = "white", colour = "grey50") +
geom_point(aes(x = lon,
y = lat),
shape = 21,
fill = "lightgreen",
color = "black",
size = 4)
```
## weather
```{r weather, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
weather <- list(NULL)
for(city in cities$City){
city_info <- cities |> filter(City == city)
city_run <- weather_history(
location = c(city_info |> pull(lat), city_info |> pull(lon)),
start = date_start,
end = date_end,
daily = c("apparent_temperature_max", "apparent_temperature_min", "precipitation_hours"),
response_units = list(temperature_unit = "fahrenheit")
)
city_run$city <- city
weather[[city]] <- city_run
}
weather <- bind_rows(weather)
weather |>
mutate(year = year(ymd(date)),
month = month(ymd(date))) |>
group_by(year, city) |>
summarise(days_above_80 = sum(daily_apparent_temperature_max > 80)) |>
group_by(city) |>
summarise(median_days_above_80 = median(days_above_80)) |>
ggplot() +
geom_col(aes(x = city,
y = median_days_above_80)) +
labs(title = "Days above 80°F (apparent temperature)",
y = "Median days per year",
x = "City",
fill = NULL)
weather |>
pivot_longer(cols = starts_with("daily"),
names_to = "max_min",
values_to = "temperature") |>
filter(max_min %in% c("daily_apparent_temperature_min", "daily_apparent_temperature_max")) |>
ggplot() +
geom_violin(aes(x = city,
y = temperature,
fill = max_min)) +
scale_fill_manual(labels = c("daily max", "daily min"),
values = c("firebrick", "dodgerblue")) +
labs(title = "Apparent Temperature",
y = "°F",
x = "City",
fill = NULL)
weather |>
mutate(year = year(ymd(date)),
month = month(ymd(date))) |>
group_by(year, city) |>
summarise(days_above_4hr = sum(daily_precipitation_hours > 4)) |>
group_by(city) |>
summarise(median_days_above_4hr = median(days_above_4hr)) |>
ggplot() +
geom_col(aes(x = city,
y = median_days_above_4hr)) +
labs(title = "Days with more than 4 hrs of rain",
y = "Median days per year",
x = "City",
fill = NULL)
weather |>
mutate(year = year(ymd(date)),
month = month(ymd(date))) |>
group_by(year, month, city) |>
summarise(days_above_4hr = sum(daily_precipitation_hours > 4)) |>
group_by(city, month) |>
summarise(median_days_above_4hr = median(days_above_4hr)) |>
ggplot() +
geom_line(aes(x = month,
y = median_days_above_4hr,
color = city)) +
labs(title = "Days with more than 4 hrs of rain",
y = "Median days per month",
x = "Month",
color = "City")
```
## Air Quality
```{r air_quality, eval = TRUE, echo = TRUE, results = "show", warning = FALSE, error = TRUE, message = FALSE}
# download data
# Create data directory if it doesn't exist
if (!dir.exists("data")) {
dir.create("data")
}
# Define years
years <- seq(year(ymd(date_start)), year(ymd(date_end)), 1)
# Initialize empty list to store dataframes
aqi_list <- list()
# Download and process files for each year
for (year in years) {
# Construct URL
url <- paste0("https://aqs.epa.gov/aqsweb/airdata/daily_aqi_by_cbsa_", year, ".zip")
# Define local file paths
zip_file <- file.path("data", paste0("daily_aqi_by_cbsa_", year, ".zip"))
csv_file <- file.path("data", paste0("daily_aqi_by_cbsa_", year, ".csv"))
# Download zip if it doesn't exist
if (file.exists(zip_file)) {
cat(paste0("Zip file for year ", year, " already exists. Skipping download.\n"))
} else {
cat(paste0("Downloading data for year ", year, "...\n"))
tryCatch({
download.file(url, zip_file, mode = "wb")
cat(paste0("Successfully downloaded data for year ", year, "\n"))
}, error = function(e) {
cat(paste0("Error downloading data for year ", year, ": ", e$message, "\n"))
next
})
}
# Extract zip if CSV doesn't exist
if (file.exists(zip_file)) {
if (file.exists(csv_file)) {
cat(paste0("CSV for year ", year, " already extracted. Skipping extraction.\n"))
} else {
cat(paste0("Extracting zip for year ", year, "...\n"))
tryCatch({
unzip(zip_file, exdir = "data")
cat(paste0("Successfully extracted data for year ", year, "\n"))
}, error = function(e) {
cat(paste0("Error extracting data for year ", year, ": ", e$message, "\n"))
next
})
}
}
# Read CSV if it exists
if (file.exists(csv_file)) {
cat(paste0("Reading CSV for year ", year, "...\n"))
tryCatch({
aqi_year <- read.csv(csv_file)
aqi_list[[as.character(year)]] <- aqi_year
cat(paste0("Successfully read ", nrow(aqi_year), " rows for year ", year, "\n"))
}, error = function(e) {
cat(paste0("Error reading CSV for year ", year, ": ", e$message, "\n"))
})
}
}
# Combine all dataframes
aqi_data <- bind_rows(aqi_list)
aqi_data <- aqi_data |>
filter(CBSA.Code %in% (cities |> pull(`CBSA-Code`))) |>
mutate(year = year(ymd(Date)),
month = month(ymd(Date)))
aqi_threshhold <- 50
aqi_data |>
group_by(year, CBSA) |>
summarise(hours_above = sum(AQI >= aqi_threshhold, na.rm = TRUE)) |>
group_by(CBSA) |>
summarise(mean_hours_above = mean(hours_above, na.rm = TRUE)) |>
ggplot() +
geom_col(aes(x = CBSA,
y = mean_hours_above)) +
labs(title = "Hours above 50 AQI",
y = "Mean hours per year",
x = "City",
fill = NULL)
aqi_data |>
group_by(year, CBSA) |>
summarise(hours_above = sum(AQI >= 2*aqi_threshhold, na.rm = TRUE)) |>
group_by(CBSA) |>
summarise(mean_hours_above = mean(hours_above, na.rm = TRUE)) |>
ggplot() +
geom_col(aes(x = CBSA,
y = mean_hours_above)) +
labs(title = paste0("Hours above ", aqi_threshhold * 2," AQI"),
y = "Mean hours per year",
x = "City",
fill = NULL)
aqi_data |>
ggplot() +
geom_violin(aes(x = CBSA,
y = AQI)) +
labs(title = "AQI",
y = "AQI",
x = "City",
fill = NULL) +
coord_cartesian(ylim = (c(0,500)))
aqi_data |>
group_by(year, month, CBSA) |>
summarise(hours_above = sum(AQI >= aqi_threshhold, na.rm = TRUE)) |>
group_by(CBSA, month) |>
summarise(mean_hours_above = mean(hours_above, na.rm = TRUE)) |>
ggplot() +
geom_line(aes(x = month,
y = mean_hours_above,
color = CBSA)) +
scale_x_continuous(breaks = seq(1,12,1)) +
scale_color_brewer(type = "qual") +
labs(title = paste0("Hours with an AQI of greater than or equal to ", aqi_threshhold),
y = "Mean days per month",
x = "Month",
color = "City")
```