辅导代写案例-Group exercise 3: The spread of COVID-19 in the US DATA5207 欢迎使用51辅导,51作业君孵化低价透明的学长辅导平台,服务保持优质,平均费用压低50%以上! 51fudao.top Group exercise 3: The spread of COVID- 19 in the US DATA5207: Data Analysis in the Social Sciences In the final two labs of the unit, we will be working on your last assessable group exercise. This task, which counts towards your group work grade, should take you no more than approximately four hours for each group member to complete over the next two sessions. It is worth 10 per cent of your total grade for the unit and will be conducted in your groups for the assessable group projects. You should submit your final written document as a Markdown file through a link that will be provided in this module. You do not need to knit this, just submit the RMD with the files needed to make it run. Only one member of your group needs to submit your work. As long as the groups are properly registered on canvas, all members of the group will receive the full grades awarded to the group. The project We will finish the unit by working on what I think is an interesting and contemporary problem. I am sharing with you some data I have been working with to predict the extent of COVID- 19 confirmed cases in the United States . Using these data and the methods we have covered so far, you will write a short article in the style of a data journalist that explains the spread of COVID- 19 in the United States . This should be written in a clear and accessible manner and the explanations should include high quality visualisations. In preparation for this, you should explore your data and conduct descriptive analysis. You can run a regression model, although this is optional. Remember, the key goal is to convey your findings in a clear and accessible way. The teaching team will be available during the labs and to talk during their office hours, to provide any help you need. You will do this using three datasets, which you need to combine. One contains a time series of COVID- 19 cases (which you will use to create your dependent variable). The other two contain data that can be used for potential predictors: Google mobility data by county (also a time series) and demographic and other information on individual counties (static across time). I have provided you with a cleaned and combined version of these data. This is the file covid.data_cleaned. Rdata. This does not include all of the variables from the original data, though. The code used to download and clean the data is shown below. Start with the cleaned file, but add variables from these datasets as needed. You will need to really get into the data to understand it and then think about how you are going to use it to answer your question. The county data (which contains most of your potential predictors) is quite large. If you dig into this for additional predictors, you might want to use the select() function to only pull out those predictors you intend to use before combining the data. Your data You have three datasets for this task. Two time-series datasets: the first is the daily number of new cases of COVID- 19 in each county in the United States from January 2020 to May 2022, and the other tracks daily changes in mobility in each county from February 2020 to May 2022. You also have a set of stationary data on other features for each county, including demographic, economic and climate characteristics. The first of these datasets is the file containing information on new confirmed COVID- 19 cases in the US, by county. These data are sourced from the Johns Hopkins dataset, available I have downloaded and cleaned these data so you do not have to. I have used them to create a file that contains the number of confirmed cases and a 14-day rolling average for each county in the US. The file is called covid.RData and it is saved in the data folder on the canvas page for this session. So you can see how I did this, the code is included here: # load data confirmed.cases.data <- read.csv(“https://raw.githubusercontent.com/ CSSEGISandData/COVID-19/master/csse_covid_19_data/ csse_covid_19_time_series/ time_series_covid19_confirmed_US.csv”) # load packages library(tidyr) library(DataCombine) # clean data confirmed.cases.data2 <- confirmed.cases.data %>% dplyr::rename(state = Province_State, county = Admin2) %>% dplyr::select(-UID, -iso2, -iso3, -code3, -Lat, -Long_, -Combined_Key, -Country_Region) %>% gather(date, confirmed.cases,-state,-county,-FIPS) %>% dplyr::mutate(date = gsub( ‘X’, ”, date), date = as.Date(as.character(date), “%m.%d.%y”), county_state = paste0(county, ‘, ‘, state)) %>% arrange(county_state, date) %>% dplyr::mutate(lag = slide(., Var = ‘confirmed.cases’, NewVar = ‘new’, GroupVar = ‘county_state’, slideBy = -1)[, ‘new’], new.cases = confirmed.cases – lag) ## calculate rolling averages and remove non-states covid.smooth.data_county <- confirmed.cases.data2 %>% dplyr::group_by(county_state, date) %>% dplyr::summarise(new.cases = sum(new.cases)) %>% dplyr::group_by(county_state) %>% dplyr::mutate(cases_14days = zoo::rollmean(new.cases, k = 14, fill = 0)) %>% dplyr::ungroup() %>% mutate() %>% merge(confirmed.cases.data2 %>% dplyr::select(county_state, date, county, state)) %>% This provides us with our dependent variable, the confirmed cases column (and also the rolling 14-day average). The next dataset is Google mobility data. This shows how mobility patterns have changed in different US counties, and can be obtained with the code: google.mobility <- read.csv(‘https://www.gstatic.com/covid19/mobility/ Global_Mobility_Report.csv’) %>% filter(country_region == ‘United States’) This is a large fil and many people struggle to download it, so I have saved the csv file in the Data folder in the canvas module for this lab. The documentation for this can be found here. The county-level variable is called sub_region_2 in this dataset. The rest of your predictors can be found in this file here: county.data <- read.csv(‘https://raw.githubusercontent.com/JieYingWu/ COVID-19_US_County-level_Summaries/master/data/counties.csv’) The codebook for these data can be found here. Additional information on these data are available here. The cleaned ile I conducted some additional cleaning and rolled these into the file covid.data_cleaned. Rdata. There is a lot here and you can just run your analysis on these data if you want. Or you can use the syntax below to add additional variables to the dataset and your analysis. The cleaned file contains: COVID- 19 data You can use any of these variables as your dependent variable in your analysis, or transform them as you wish. • new.cases – The number of new daily cases • cases_14days – The rolling 14 day average of new cases • new.cases.lag – A one-day lag of new cases Date and location data • date – The date of the observation • county – The county in which the observation was taken • state – The state in which the observation was taken • county_state – The combined county and state in which the observation was taken Demographic data (this is stationary and does not change by date) • total.population – The total population of the county • pop.density – The population density of the county (per square mile) • housing.density – The density (per square mile) of housing units • age.65.plus – The share of the population that is aged 65 and older • pcnt.university – The share of the population with a university education • pcnt.less.than.high.school – The share of the population that did not complete high school or obtain further education • avg.hhold.size – The mean size of households in the county • transit.scores – How well a county is served by public transit • median.income – Estimated median household income in 2018 Climate data (this is stationary and does not change by date) I think these are self explanatory. • avg.dec.tem • avg.jan.temp • avg.feb.temp • avg.winter.temp Healthcare resource data (this is stationary and does not change by date) • icu.beds – Number of ICU beds per county • active.physicians – Active Physicians per 100,000 Population, 2018 • active.patient.care.physicians – Total Active Patient Care Physicians per 100,000 Population, 2018 Mobility data • mobility.transit – Mobility trends for places like public transport hubs such as subway, bus, and train stations. Counties with no public transport have been scored zero • mobility.workplaces – Mobility trends for places of work • workplaces.lag14 – A 14 day lag for this • mobility.grocery – Mobility trends for places like grocery markets, food warehouses, farmers mar- kets, specialty food shops, drug stores, and pharmacies How these data were cleaned and combined Here is the code used to clean and combine these data: # patterns for cleaning county names patterns <- c(‘ Borough| Census Area| County| Parish| city’) # covid case data covid.data <- covid.smooth.data_county %>% # county data left_join(county.data %>% # grab file with full state names left_join(read.csv(‘Data/fips concordance.csv’) %>% dplyr::select(State = Alpha.code, state_full = Name)) %>% dplyr::mutate(county = gsub(patterns, ”, Area_Name), county_state = paste0(county, ‘, ‘, state_full)) %>% dplyr::select(county_state, pop.density = Density.per.square.mile.of.land.area…Population, age.65.plus = Total_age65plus, total.population = POP_ESTIMATE_2018, median.income = MEDHHINC_2018, avg.hhold.size = Total.households..Average.household.size, avg.dec.temp = Dec.Temp.AVG…F, avg.jan.temp = Jan.Temp.AVG…F, avg.feb.temp = Feb.Temp.AVG…F, icu.beds = ICU.Beds, active.physicians = Active.Physicians.per.100000