<p>Data/code dictionary for project:</p>
<ul>
<li>/Countries<ul>
<li>crosscultural.R - primary analysis script for the country-level analyses. 3d Mind Model testing, individual dimension analyses, and model comparison with Circumplex, as well as permutation/bootstraping for inference on each. Visualization code for figures 1 and S1 (note: these also require files from <a href="http://naturalearthdata.com" rel="nofollow">naturalearthdata.com</a> for the choropleth). This version of the analyses is more heavily commented than the historical and language versions - refer here for clarification if portion of those appear opaque.</li>
<li>Twitter/ - contains code for retrieving, geolocating, cleaning data from <a href="http://Twitter.com" rel="nofollow">Twitter.com</a>. Primary webscraping was carried out using <a href="https://github.com/DominicBurkart/SocialNetworkAnalysis" rel="nofollow">this</a> package. Auxiliary files are included to support geolocation, including shape files and cross-reference tables of geographical units. FastText code is also included (<a href="https://github.com/xeb/fastText-docker" rel="nofollow">this</a> docker container was used to process these data).</li>
<li>Vectors/ - Contains the raw word vectors extracted from each country's fastText embedding</li>
</ul>
</li>
<li>/Dimensions<ul>
<li>Contains two csv files with averaged human ratings of where each of 166 mental states falls on the dimensions of either the 3d Mind Model, or the Circumplex Model. The data were collected as part of another investigation (Tamir, Thornton, Contreras & Mitchell, 2016). The OSF repository for that project is linked.</li>
<li>Also contains 'filt.csv' which contains ratings of the pairwise similarity between the 166 mental states (used in the noise ceiling analysis at each level of analysis).</li>
</ul>
</li>
<li>/Historical<ul>
<li>combined_historic.R - primary analysis script for historical data</li>
<li>Vectors/ - Contains the raw word vectors extracted from each society's fastText embedding</li>
<li>Chinese/ - code and translations for analyzing the Classical Chinese corpus from the Chinese Text Project (<a href="https://ctext.org/" rel="nofollow">https://ctext.org/</a>).</li>
<li>Gutenberg/ - code and translations for analyzing the French and English portions of the Standardize Project Gutenberg Corpus (<a href="https://zenodo.org/record/2422561" rel="nofollow">https://zenodo.org/record/2422561</a>)</li>
</ul>
</li>
<li>/Languages<ul>
<li>crosslinguistic.R - primary analysis script for language level data.</li>
<li><a href="http://proc_translated.sh" rel="nofollow">proc_translated.sh</a> - code for extracting translated terms from pretrained language vectors (downloaded from <a href="https://fasttext.cc/docs/en/crawl-vectors.html" rel="nofollow">https://fasttext.cc/docs/en/crawl-vectors.html</a>).</li>
<li><a href="http://split_translations_to_txt.py" rel="nofollow">split_translations_to_txt.py</a> - divides translated.csv into txt files for fastText</li>
<li>translated.csv - translated mental state terms</li>
<li>Vectors/ - Contains the raw word vectors extracted from each language's fastText embedding</li>
</ul>
</li>
</ul>