Loading wiki pages...

Wiki Version:
<p>Data/code dictionary for project:</p> <ul> <li>/Countries<ul> <li>crosscultural.R - primary analysis script for the country-level analyses. 3d Mind Model testing, individual dimension analyses, and model comparison with Circumplex, as well as permutation/bootstraping for inference on each. Visualization code for figures 1 and S1 (note: these also require files from <a href="http://naturalearthdata.com" rel="nofollow">naturalearthdata.com</a> for the choropleth). This version of the analyses is more heavily commented than the historical and language versions - refer here for clarification if portion of those appear opaque.</li> <li>Twitter/ - contains code for retrieving, geolocating, cleaning data from <a href="http://Twitter.com" rel="nofollow">Twitter.com</a>. Primary webscraping was carried out using <a href="https://github.com/DominicBurkart/SocialNetworkAnalysis" rel="nofollow">this</a> package. Auxiliary files are included to support geolocation, including shape files and cross-reference tables of geographical units. FastText code is also included (<a href="https://github.com/xeb/fastText-docker" rel="nofollow">this</a> docker container was used to process these data).</li> <li>Vectors/ - Contains the raw word vectors extracted from each country's fastText embedding</li> </ul> </li> <li>/Dimensions<ul> <li>Contains two csv files with averaged human ratings of where each of 166 mental states falls on the dimensions of either the 3d Mind Model, or the Circumplex Model. The data were collected as part of another investigation (Tamir, Thornton, Contreras & Mitchell, 2016). The OSF repository for that project is linked.</li> <li>Also contains 'filt.csv' which contains ratings of the pairwise similarity between the 166 mental states (used in the noise ceiling analysis at each level of analysis).</li> </ul> </li> <li>/Historical<ul> <li>combined_historic.R - primary analysis script for historical data</li> <li>Vectors/ - Contains the raw word vectors extracted from each society's fastText embedding</li> <li>Chinese/ - code and translations for analyzing the Classical Chinese corpus from the Chinese Text Project (<a href="https://ctext.org/" rel="nofollow">https://ctext.org/</a>).</li> <li>Gutenberg/ - code and translations for analyzing the French and English portions of the Standardize Project Gutenberg Corpus (<a href="https://zenodo.org/record/2422561" rel="nofollow">https://zenodo.org/record/2422561</a>)</li> </ul> </li> <li>/Languages<ul> <li>crosslinguistic.R - primary analysis script for language level data.</li> <li><a href="http://proc_translated.sh" rel="nofollow">proc_translated.sh</a> - code for extracting translated terms from pretrained language vectors (downloaded from <a href="https://fasttext.cc/docs/en/crawl-vectors.html" rel="nofollow">https://fasttext.cc/docs/en/crawl-vectors.html</a>).</li> <li><a href="http://split_translations_to_txt.py" rel="nofollow">split_translations_to_txt.py</a> - divides translated.csv into txt files for fastText</li> <li>translated.csv - translated mental state terms</li> <li>Vectors/ - Contains the raw word vectors extracted from each language's fastText embedding</li> </ul> </li> </ul>
OSF does not support the use of Internet Explorer. For optimal performance, please switch to another browser.
This website relies on cookies to help provide a better user experience. By clicking Accept or continuing to use the site, you agree. For more information, see our Privacy Policy and information on cookie use.

Start managing your projects on the OSF today.

Free and easy to use, the Open Science Framework supports the entire research lifecycle: planning, execution, reporting, archiving, and discovery.