Main content

Home

Menu

Loading wiki pages...

View
Wiki Version:
This is the page for the project **No effect of different types of media on well-being**. Link to the preprint: <> Link to rendered analysis report: <https://digital-wellbeing.github.io/cultural-consumption/> ## Components The `Github` component holds the source code, hosted on Github. The code details all steps from raw data to final analysis. The `data` component holds the raw data. These data were processed before uploading to remove other variables that we aren't allowed to share. You can see those processing steps in the processing section of the rendered [analysis report](https://digital-wellbeing.github.io/cultural-consumption/). For an explanation of the data structure and processing, see the report and the `codebook.xlsx` file in the `data` component. The component contains the following objects: - `wave1.csv`: The (preprocessed) raw data for the first wave. - `waves_2_to_6.csv`: The (preprocessed) raw data for waves two through six. - `example_id.csv`: Merely a (made up) example of how the ID variables are used across all waves. See [this section](https://digital-wellbeing.github.io/cultural-consumption/data-processing.html#read-waves-2-6) for an explanation. This data file wasn't used in the analysis. - `processed_data.rds`: The final data set after raw data processing, on which summary stats and analyses are based. The file gets written at the end of [this section](https://digital-wellbeing.github.io/cultural-consumption/analysis.html#data-preparation). If you just want to do your own checks/analyses, it's probably easiest to load the `processed_data.rds` file straight into R. Another component holds brms model objects (`models`). We ran 28 total models, and each one is about one GB large. We only ran those models once and stored them as external files. You’ll see in the analysis section that we downloaded those models and loaded them into R for model inspection and diagnostics. The `figures` component holds the figures we report in the paper. You can reproduce them with the source code. The `materials` component holds the questionnaires used at each wave for all items. ## How to reproduce processing and analyses All processing, analysis, description, and modelling steps organized into separate [R Markdown](https://rmarkdown.rstudio.com/) files. The source code is on GitHub: <https://github.com/digital-wellbeing/cultural-consumption>. Those source files are organized as an R [bookdown](https://bookdown.org/yihui/bookdown/) project. That project was knitted to an online book, whose results are at <https://digital-wellbeing.github.io/cultural-consumption/>. The project used a local library to make sure package versions are stable for anyone who wants reproduce the analysis. We used [renv](https://rstudio.github.io/renv/articles/renv.html) for that. If you want to reproduce the entire book, it's best to download the entire R project from the Github repo: <https://github.com/digital-wellbeing/cultural-consumption>. As a first step, open the project and make sure you have all packages installed in the right version by calling `renv::restore()`. For an explanation of `renv`, see here: <https://rstudio.github.io/renv/articles/renv.html>. That'll install all packages in the version that are saved in the lockfile of the project (`renv.lock`). For that, it's best if you are on R 4.0.1 or higher. Once all packages are installed, you should be able to build the book and reproduce all files that are displayed at <https://digital-wellbeing.github.io/cultural-consumption/>. To do that, either run `bookdown::render_book("index.Rmd")` or click the "Build Book" button under the `Build` tab in RStudio. Each `.Rmd` source file will be knitted and stored in a `docs` folder. You can double click on `docs/index.html` and view the results in a web browser. Note that before running the entire book, that you either need to set chunk options or download the data and the models (see below). You'll see that the source code doesn't run the models anymore. In total, running each chunk would take the good part of a week on a modern 4-core laptop. I ran them once and then stored them on the OSF. In the source code, you can download those files (note: downloading the models will take some time, they're 34GB in total). To download the models, make sure the chunk option for chunk `download-models` in the analysis chapter is set to `eval=TRUE`. If you want to run the models rather than downloading them and loading them into R, you can either set chunk options for each model to `eval=TRUE` or run each of the five `.Rmd` files separately. I set this book up so that it runs in one go, meaning you need to run the source files in their order, and can't run source files independently. The same goes for the data: You'll need to set the chunk `download-data` to `eval=TRUE`. If that doesn't work (but it should, even with an anonymized link), you'll need to manually download the data files from the OSF into a `data/` folder in the main directory of the project.