Healthy Cities GIS Assignment

Author

Cody Paulay-Simmons

Load the libraries and set the working directory

library(tidyverse)
library(tidyr)
cities500 <- read_csv("500CitiesLocalHealthIndicators.cdc.csv")
data(cities500)

The GeoLocation variable has (lat, long) format

Split GeoLocation (lat, long) into two columns: lat and long

latlong <- cities500|>
  mutate(GeoLocation = str_replace_all(GeoLocation, "[()]", ""))|>
  separate(GeoLocation, into = c("lat",
                                 "long"), 
           sep = ",",
           convert = TRUE)
head(latlong)
# A tibble: 6 × 25
   Year StateAbbr StateDesc  CityName  GeographicLevel DataSource Category      
  <dbl> <chr>     <chr>      <chr>     <chr>           <chr>      <chr>         
1  2017 CA        California Hawthorne Census Tract    BRFSS      Health Outcom…
2  2017 CA        California Hawthorne City            BRFSS      Unhealthy Beh…
3  2017 CA        California Hayward   City            BRFSS      Health Outcom…
4  2017 CA        California Hayward   City            BRFSS      Unhealthy Beh…
5  2017 CA        California Hemet     City            BRFSS      Prevention    
6  2017 CA        California Indio     Census Tract    BRFSS      Health Outcom…
# ℹ 18 more variables: UniqueID <chr>, Measure <chr>, Data_Value_Unit <chr>,
#   DataValueTypeID <chr>, Data_Value_Type <chr>, Data_Value <dbl>,
#   Low_Confidence_Limit <dbl>, High_Confidence_Limit <dbl>,
#   Data_Value_Footnote_Symbol <chr>, Data_Value_Footnote <chr>,
#   PopulationCount <dbl>, lat <dbl>, long <dbl>, CategoryID <chr>,
#   MeasureId <chr>, CityFIPS <dbl>, TractFIPS <dbl>, Short_Question_Text <chr>

Filter the dataset

Remove the StateDesc that includes the United Sates, select Prevention as the category (of interest), filter for only measuring crude prevalence and select only 2017.

latlong_clean <- latlong |>
  filter(StateDesc != "United States") |>
  filter(Data_Value_Type == "Crude prevalence") |>
  filter(Year == 2017)
head(latlong_clean)
# A tibble: 6 × 25
   Year StateAbbr StateDesc  CityName  GeographicLevel DataSource Category      
  <dbl> <chr>     <chr>      <chr>     <chr>           <chr>      <chr>         
1  2017 CA        California Hawthorne Census Tract    BRFSS      Health Outcom…
2  2017 CA        California Hawthorne City            BRFSS      Unhealthy Beh…
3  2017 CA        California Hayward   City            BRFSS      Unhealthy Beh…
4  2017 CA        California Indio     Census Tract    BRFSS      Health Outcom…
5  2017 CA        California Inglewood Census Tract    BRFSS      Health Outcom…
6  2017 CA        California Lakewood  City            BRFSS      Unhealthy Beh…
# ℹ 18 more variables: UniqueID <chr>, Measure <chr>, Data_Value_Unit <chr>,
#   DataValueTypeID <chr>, Data_Value_Type <chr>, Data_Value <dbl>,
#   Low_Confidence_Limit <dbl>, High_Confidence_Limit <dbl>,
#   Data_Value_Footnote_Symbol <chr>, Data_Value_Footnote <chr>,
#   PopulationCount <dbl>, lat <dbl>, long <dbl>, CategoryID <chr>,
#   MeasureId <chr>, CityFIPS <dbl>, TractFIPS <dbl>, Short_Question_Text <chr>

What variables are included? (can any of them be removed?)

names(latlong_clean)
 [1] "Year"                       "StateAbbr"                 
 [3] "StateDesc"                  "CityName"                  
 [5] "GeographicLevel"            "DataSource"                
 [7] "Category"                   "UniqueID"                  
 [9] "Measure"                    "Data_Value_Unit"           
[11] "DataValueTypeID"            "Data_Value_Type"           
[13] "Data_Value"                 "Low_Confidence_Limit"      
[15] "High_Confidence_Limit"      "Data_Value_Footnote_Symbol"
[17] "Data_Value_Footnote"        "PopulationCount"           
[19] "lat"                        "long"                      
[21] "CategoryID"                 "MeasureId"                 
[23] "CityFIPS"                   "TractFIPS"                 
[25] "Short_Question_Text"       

Remove the variables that will not be used in the assignment

latlong_clean2 <- latlong_clean |>
  select(-DataSource,
         -Data_Value_Unit, 
         -DataValueTypeID, 
         -Low_Confidence_Limit, 
         -High_Confidence_Limit,
         -Data_Value_Footnote_Symbol,
         -Data_Value_Footnote)
head(latlong_clean2)
# A tibble: 6 × 18
   Year StateAbbr StateDesc  CityName  GeographicLevel Category UniqueID Measure
  <dbl> <chr>     <chr>      <chr>     <chr>           <chr>    <chr>    <chr>  
1  2017 CA        California Hawthorne Census Tract    Health … 0632548… Arthri…
2  2017 CA        California Hawthorne City            Unhealt… 632548   Curren…
3  2017 CA        California Hayward   City            Unhealt… 633000   Obesit…
4  2017 CA        California Indio     Census Tract    Health … 0636448… Arthri…
5  2017 CA        California Inglewood Census Tract    Health … 0636546… Diagno…
6  2017 CA        California Lakewood  City            Unhealt… 639892   Obesit…
# ℹ 10 more variables: Data_Value_Type <chr>, Data_Value <dbl>,
#   PopulationCount <dbl>, lat <dbl>, long <dbl>, CategoryID <chr>,
#   MeasureId <chr>, CityFIPS <dbl>, TractFIPS <dbl>, Short_Question_Text <chr>

The new dataset “Prevention” is a manageable dataset now.

For your assignment, work with a cleaned dataset.

1. Once you run the above code and learn how to filter in this format, filter this dataset however you choose so that you have a subset with no more than 900 observations.

Filter in Physical Inactivity from updated prevention dataset

lpa <- latlong_clean2 |>
  filter(MeasureId == "LPA")

Down to 28.5k observations

Filter in Maryland from lpa datset

lpa_md <- lpa |>
  filter(StateAbbr == "MD")

And here we are with just 201 observations of physical inactivity cases in Maryland.

2. Based on the GIS tutorial (Japan earthquakes), create one plot about something in your subsetted dataset.

quakes <- read_csv("japan_quakes_01-18.csv")
Rows: 14092 Columns: 22
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr   (8): magType, net, id, place, type, status, locationSource, magSource
dbl  (12): latitude, longitude, depth, mag, nst, gap, dmin, rms, horizontalE...
dttm  (2): time, updated

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.
gaprmsmwr <- quakes |>
  filter(magType == "mwr") |>
  filter(!is.na(mag), 
         !is.na(rms), 
         !is.na(gap))

ggplot(gaprmsmwr,
       aes(x = rms,
           y = gap,
           size = mag)) +
  geom_point(color = "red",
             alpha = 0.3) +
  scale_color_hue()+
  theme_bw() +
  labs(title = "Reliability in predictions (RMS) and Informations (GAP) of MWR type Japanese Earthquakes",
       caption = "Source: USGS")

3. Now create a map of your subsetted dataset.

library(leaflet)

japan_lon <- 138.129731
japan_lat <- 36.2615855

leaflet() |>
  setView(lng = japan_lon, 
          lat = japan_lat, 
          zoom = 6) |>
  addProviderTiles("Esri.NatGeoWorldMap") |>
  addCircles(data = gaprmsmwr,
             radius = 5000,
             color = "white",
             fillColor = "red",
             fillOpacity = 0.5)
Assuming "longitude" and "latitude" are longitude and latitude, respectively

4. Refine your map to include a mouse-click tooltip

popuprmsgap <- paste0("<b>RMS: </b>", gaprmsmwr$rms, "<br>",
                      "<b>GAP: </b>", gaprmsmwr$gap, "<br>",
                      "<b>Magnitude: </b>", gaprmsmwr$mag, "<br>",
                      "<b>Place: </b>", gaprmsmwr$place, "<br>")

leaflet() |>
  setView(lng = japan_lon, 
          lat = japan_lat, 
          zoom = 6) |>
  addProviderTiles("Esri.NatGeoWorldMap") |>
  addCircles(data = gaprmsmwr,
             radius = 5000,
             color = "white",
             fillColor = "red",
             fillOpacity = 0.5,
             popup = popuprmsgap)
Assuming "longitude" and "latitude" are longitude and latitude, respectively

5. Write a paragraph

In a paragraph, describe the plots you created and what they show.

For this assignment, I focused on understanding how reliable earthquake data can be by looking into the RMS and GAP variables in the dataset. After looking up their definitions, I learned that RMS (Root-Mean-Square) represents the margin of error in the recorded event, while GAP (Azimuthal Gap) tells us how close the surrounding recording stations were. The lower either variable are, the better the event was covered… even more if both are. I filtered the dataset to focus only on MWR magnitude types and removed any missing values. I created a scatterplot to visualize the relationship between RMS and GAP. I sized the dots by magnitude to give extra context, which worked well visually. Originally, I was going to size the points based on RMS or GAP in the leaflet map too, but I wasn’t able to get that part working and that’s something I’d love to get advice on, since I followed the class notes step by step and still couldn’t figure it out. Then I created an interactive leaflet map showing the earthquake locations around Japan. The map includes tooltips with details like RMS, GAP, magnitude, and place. Seeing this all together helped me better understand the quality and reliability of real world data… not just what it says, but how well supported it is. It was also a solid opportunity to apply the mapping and filtering techniques we’ve been learning. I could definitely see myself doing more geography-related data analysis in a future job.