Jeffrey Harding, Ph.D.
Jeffrey L. Harding, Ph.D.
jeffrey.l.harding@gmail.com
PH: 615.713.0473
LinkedIn
A public-school classroom teacher turned researcher and creator. Accountability Research Specialist at Georgia Department of Education analyzing and reporting on millions of administrative student- and school-level data points. Former Senior Director of Cognia Innovation Lab. Seasoned presenter, facilitator, and trainer of internal and external constituents. Published author of empirical research.
Hello, and thank you for taking the time to review these application materials. Given the nature of this particular position, I felt it might be helpful to present (as an alternative) all of my applications in the dynamic HTML I created using the R Markdown languages. Of course, a standard CV and Cover Letter have been provided as well if you prefer. It is my hope, though, that this document communicates that same passion and demonstrates my qualifications for this position at least as well as the traditional application materials and, if all goes well, elevates those arguments through this interactive format. Please note that all data used in this report come from either publicly available datasets or those included (and available free of charge) in the R and R studio applications.
Please use the tabs below to navigate to the relevant parts of my résumé. Again, you will find all of the same information here that you would in the PDF version of my CV in my application. However, you may find this format easier to navigate.
Ph.D., Higher Education (Policy), University of Georgia, Athens, GA
Additional Concentrations in Quantitative Methods and K-12 Policy
Dissertation: Timely Information and STEM Pipeline Outcomes
Committee:
M.Ed., Higher Education Administration, Vanderbilt University
Additional Concentration in Institutional Advancement
B.A., English Middle Tennessee State University
Minors in Secondary Education and Spanish
Senior Research and Data Scientist (Nov. 2021 - Present)
Georgia Department of Education
Adjunct Instructor (2021 - Current)
Logan University, Doctor of Education Program
Quantitative Research Methods Course
Accountability Research Specialist (2019 - 2021)
Georgia Department of Education
Senior Director, Innovation Lab (July 2019 - December 2019)
Cognia (Formerly AdvancED|Measured Progress)
Director, Research & Development (2018 - 2019)
Cognia
Senior Specialist, Research & Development (2017 - 2018)
Cognia
Specialist, Research & Development (2015 - 2017)
Cognia
Graduate Research Assistant (2012 - 2015)
UGA Institute of Higher Education, University System of Georgia
Middle School Classroom Teacher (2004 - 2011)
Rutherford County (TN) School System
Peer-Reviewed:
González Canché,M.S., Lee, J.C., Harding, J.L., Turk, J.M., Bae, J.Y., & Zhang, C. (2023). Post-baccalaureate federal loans de-subsidization: Impacts on compositional attributes, extensive and intensive borrowing margins, and anticipatory effects. The Journal of Higher Education. DOI: 10.1080/00221546.2023.2187176
Harding, J., Parker, M., Toutkhoushian, R. (2017). Deciding about college: How soon is soon enough. Teachers College Record, 119,(4), 1-40.
Non-Peer Reviewed:
Harding, J. (2019). The bullying problem in our schools. Retrieved from https://go.cognia.org/the-bullying-problem-in-our-schools
Harding, J. (2019). So, did they learn anything? Strengthening performance-based assessments in the classroom. The Source, Winter. Retrieved from https://source.cognia.org/issue-article/so-did-they-learn-anything/
Harding, J. (2018). Did we do that? Evaluating the impact of turnaround efforts. The Source, Spring/Summer. Retrieved from https://source.cognia.org/issue-article/did-we-do-evaluating-impact-turnaround-efforts/
Keynote Addresses:
Harding, J. (2018). Building and growing STEM education models: Lessons learned from the field. Keynote address given to attendees at the 2018 AdvancED Midwest Regional Conference in Kansas City, Missouri to ~200 attendees comprised of teachers, school and district leaders.
Harding, J. (2018). The who, what, when, where, why and how of Student Engagement. Keynote address given to attendees at the 2018 AdvancED North Dakota State Conference in Bismarck, North Dakota to ~400 attendees comprised of teachers, school and district leaders, and representatives from the North Dakota Department of Education.
Practitioner Conference Presentations
Harding, J. (2018). Program evaluation in K-12 educational settings. Conference session created for and presented at AdvancED North Dakota State Conference in Bismarck, North Dakota to ~50 attendees comprised of teachers, school and district leaders, and representatives from the North Dakota Department of Education.
Harding, J. (2015-2018). Numerous AdvancED conference training sessions covering a range of topics related to AdvancED-specific and related education content presented in Georgia, North Carolina, Tennessee, Missouri, Wyoming, New York, New Jersey, Illinois, Arkansas, Cairo (Egypt) and Dubai (United Arab Emirates.
Workshop Sessions
Harding, J. (2016). Everyday data: Utilizing your school’s data to its full potential. Full-day training session created for and presented to the principals and leaders of the Diocese of Trenton Schools in Point Pleasant Beach, NJ.
Harding, J. (2016). Leveraging Instructional Design for STEM Educational Models. Full-day training session created for and presented to the principals and leaders of the Academy of Greatness and Excellence in Teaneck, NJ.
Stollberg, R., Harding, J., & Gonzalez Canché, M. (2016). Virtuous or Vicious Circles in Student Loan Debt: Measuring the Effects of Debt Level on Early Career Salary and Early Career Salary on Debt Repayment. Paper presented at the Association for Education Finance and Policy annual conference, Denver, CO.
Harding, J., Parker, M.C., & Toutkoushian, R.K. (2014). Deciding about college: How soon is soon enough? Paper presented at the Southern Economic Association annual conference, Atlanta, GA.
Stollberg, R., Harding, J., & Gonzalez Canché, M. (2014). Buy now, pay later: Post-graduate earnings as a function of loan debt. Paper presented at the annual conference for the Association for the Study of Higher education, Washington, D.C.
Harding, J. , Parker, M., & Toutkousian, R. (2014). Deciding about college: How soon is soon enough? Paper presented at the annual conference for the Association for Education Finance and Policy, San Antonio, TX.
Harding, J., Gonzalez Canché, M., Stollberg, R., Quarles, D., & Brajkovic, L. (2014). Fitting the mold: Using text network analysis to model university peer groupings. Paper presented at the annual International Sunbelt Social Network Conference, St. Pete Beach, FL.
Harding, J. and Gonzalez Canché, M. (2013). A wolf in sheep’s clothing? The influence of proprietary institutions on student loan outcomes. Paper presented at the annual conference for the Association for the Study of Higher Education, St. Louis, MO.
Harding, J. and Julseth D. (2012). Social media and service-learning: Using technology to enhance reflection and communication. Paper presented at the annual Gulf South Summit on Service-Learning in Higher Education, Hattiesburg, MS.
Wages By Degree Type
This visualization, while basic in appearance, is meant to demonstrate the ways in which we can leverage publicly available datasets to explore areas of interest. Data for this particular visual come from the website of the Governor’s Office for Student Achievement in Georgia. By using public data (when possible) and by providing the code used in performing analyses, we enable others to review and critique (or confirm!) our work. It’s also good practice to annotate code to let others know what steps you’ve taken, especially since must of us have our own coding conventions. Of course, having both annotated and plain code allows for transparency and readability.
The code provided here can be copy/pasted into your own R terminal session and then executed as written to render the figure on the first tab. The ‘pacman’ package may ask to be installed (if it is not found on your machine). Once you have installed the pacman package, it will automatically check for (and install and load) any other packages you may need to produce the figure.
if (!require("pacman")) install.packages("pacman")
pacman::p_load(dplyr, readr, ggplot2,RColorBrewer )
url <- "https://gosa.georgia.gov/sites/gosa.georgia.gov/files/related_files/site_page/Higher%20Learning%20and%20Earnings.zip"
download.file(url, 'wagedata.zip')
unzip('wagedata.zip')
wage_data <- read.csv("GHLE/Earnings by Degree Type.csv")
wage_data %>%
filter(percentile == '50') %>%
mutate(`Degree Type` = factor(degree,
levels = c("Certificate", "Associate's", "Bachelor's","Master's",
"Education Specialist", "PhD", "Professional")),
Year = as.factor(year)) %>%
ggplot(aes(x = `Degree Type`, y = wages , fill = Year)) +
geom_bar(stat = 'identity', position = 'dodge') +
scale_fill_brewer('Years After Graduation', palette = 'Paired') +
theme(axis.text.x = element_text(angle = 45, vjust = .5),
plot.title = element_text(hjust = 0.5),
plot.subtitle = element_text(hjust = 0.5)) +
scale_x_discrete(name = '') +
labs(title = 'Median Wages in Georgia by Degree Type',
subtitle = 'One and five years after graduation',
y = 'Median Wages')
### Checking to see whether this R environment has the necessary packages installed and loaded. The pacman package, if not already installed, should ask the user permission to install and run. From there, pacman will take care of the other packages needed for the plot below.
if (!require("pacman")) install.packages("pacman")
pacman::p_load(dplyr, readr, ggplot2,RColorBrewer )
### Pointing R to the URL where the publicly available data are stored. R then downloads and unzips the compressed file (into the working directory)/. The final line reads in the .csv file necessary for creating this figure and assigns it to the object "wage_data".
url <- "https://gosa.georgia.gov/sites/gosa.georgia.gov/files/related_files/site_page/Higher%20Learning%20and%20Earnings.zip"
download.file(url, 'wagedata.zip')
unzip('wagedata.zip')
wage_data <- read.csv("GHLE/Earnings by Degree Type.csv")
wage_data %>% # Intializng the object
filter(percentile == '50') %>% # Filtering for the median values
mutate(`Degree Type` = factor(degree, #Renaming 'degree' and converting to a factor
levels = c("Certificate", "Associate's", "Bachelor's","Master's",
"Education Specialist", "PhD", "Professional")),
Year = as.factor(year)) %>% # Assigning year to a factor variable for plotting
ggplot(aes(x = `Degree Type`, y = wages , fill = Year)) + # Mapping the Data layer
geom_bar(stat = 'identity', position = 'dodge') + #Building the geom layer
scale_fill_brewer('Years After Graduation', palette = 'Paired') + # Changing color and Legend name
theme(axis.text.x = element_text(angle = 45, vjust = .5), # Tilting Labels for readability
plot.title = element_text(hjust = 0.5), # Ajdusting title for readability
plot.subtitle = element_text(hjust = 0.5)) + # Ajdusting subtitle for readability
scale_x_discrete(name = '') + #Removing the label from the x axis
labs(title = 'Median Wages in Georgia by Degree Type', #Adding labels for (sub)title and y axis
subtitle = 'One and five years after graduation',
y = 'Median Wages')
Student Assessments During a Pandemic
Many models are used for predicting results. Predictions are important, especially when we can later test the accuracy of those predictions. Sometimes, though, we are missing data that we may never be able to capture. For instance, consider a state where a large percent of students who would be expected to take the end of grade test in a normal year did not do so in the spring of 2021 due to circumstances created by the Covid-19 Pandemic. It is interesting, and possibly instructive, to simulate what scores might look like given some different assumptions. For this example, we will hold constant the idea that all students in a given subgroup (e.g., race, gender) in a given setting (system/county and school) that did not take the test would have behaved similarly and according to a normal distribution. We then use information on the number of participants and the mean achievement of tested (4th grade) students in each content area (found at this link here on the Georgia Department of Education Website) to build a simulated dataset of test scores for the number of students who took the test. We then want to simulate the scores of students who did not take the test. We can add some variation in to the mix by allowing different mean scale scores and standard deviations according to our priors. For instance, we may believe that students who did not take the test participated in a 100% virtual mode of instruction during the 2020 - 2021 school year. We may further believe that given the abrupt change to this model (as opposed to traditional face-to-face), these students would have been less prepared for the test and would have performed worse, (lower mean scores), and with much more variation (larger standard deviation). Alternative, we might believe that these students were from families that chose to keep their children home rather than send them into school for a test. These families may have had more resources for childcare or more opportunities to stay home and work closely with their children due to parents having the flexibility to work remotely. Either case might lead to varying results. The shiny app on the next tab allows for the user to explore any number of possibilities.
Mean Adjustment Scalar
By default, this slider is set to 1. This means that you expect that the non-participants, had they taken the test, would have received the same scale scores on average as those who actually took the test. The more representative you believe the participants were of non-participants, the closer you should set this to 1. If you know something about the population of non-participants that leads you to believe they would have performed worse, on average, than those who actually took the test, move the slider toward .5, at which point you are simulating scores based on the assumption that non-participants would have earned a score on average that was half the scale score of participants (an unlikely outcome, but theoretically possible). Alternatively, if you believe non-participants would have performed better on average, then adjust the slider to the right. At the far end, 1.5 means these students would have performed 1.5 times (150%) as well as those who took the test.
Standard Deviation Adjustment Scalar
By default, this slider is set to 1. This means that you expect that the non-participants, had they taken the test, would have had the same dispersion/variation in their scores as those who took the test. The more representative you believe the participants were of non-participants, the closer you should set this to 1 (the default standard deviation for this simulation is an arbitrary 37.5 scale score points). If you believe that the non-participants were a different group than those who took the test (e.g., more high achieving) and would have had less variation in their scores, move the slider toward the left, where a value of .5 simulates scores based on a standard deviation for non-participants that is half what we ‘observed’ for participants. If you know something about the population of non-participants that leads you to believe they would have performed in ways that were not as predictable or as consistent as those who took the test, move the slider to the right toward 3, which simulates scores assuming the standard deviation would have been 3 times as large for non-participants.
Number of Bins Slider
This slider only exists to adjust the visualized shape of the distribution. Increasing the number of bins adds more detail. Decreasing the number of bins places more students in each bar of the histogram, possibly at the expense of visualizing nuance.
It is important to note that this particular histogram is filled (color) based on Performance Category designation. Because these categories shift around a cut-score, certain places on the histogram will display as two bars side-by-side. In reality, the count in that bin should be interpreted as if those two bars were “stacked” on top of one another.