Editing Wiki language research
From CommunityData
Warning: You are not logged in. Your IP address will be publicly visible if you make any edits. If you log in or create an account, your edits will be attributed to your username, along with other benefits.
The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then publish the changes below to finish undoing the edit.
Latest revision | Your text | ||
Line 3: | Line 3: | ||
== Action Items == | == Action Items == | ||
* | * write rough draft of preliminary findings for 6/21 | ||
=== Undergrads === | === Undergrads === | ||
Line 15: | Line 14: | ||
* Week 1: Write initial analysis, get google doc - latex pipeline setup | * Week 1: Write initial analysis, get google doc - latex pipeline setup | ||
* Week 2-3: Flag bot edits, pull new samples for coding based on updated percentiles, write new draft of analysis | * Week 2-3: Flag bot edits, pull new samples for coding based on updated percentiles, write new draft of analysis | ||
* | * week 3-6: Develop hypotheses and run analysis | ||
** Cross cultural deliberative practices | ** Cross cultural deliberative practices | ||
** Discussion structure | ** Discussion structure | ||
Line 23: | Line 22: | ||
== meeting logs & notes == | == meeting logs & notes == | ||
=== 05-22-16 === | === 05-22-16 === | ||
DG: I spent some time looking at the data distributions and ran a bunch of models on the simple EN models overnight. The data for len_1 are reallllly long-tailed with very low frequencies -- this is causing the convergence issues. Below is a table of the simple model (len_1 ~ num_editors_1), run through a series of truncated data sets. The models will converge all the way up to removing the final data point out of the 4,077,819 data points we have. In other words, I was able to get convergence by dropping a single data point. Here's a quick table of the results from running the models: | DG: I spent some time looking at the data distributions and ran a bunch of models on the simple EN models overnight. The data for len_1 are reallllly long-tailed with very low frequencies -- this is causing the convergence issues. Below is a table of the simple model (len_1 ~ num_editors_1), run through a series of truncated data sets. The models will converge all the way up to removing the final data point out of the 4,077,819 data points we have. In other words, I was able to get convergence by dropping a single data point. Here's a quick table of the results from running the models: | ||
Line 140: | Line 114: | ||
== project resources & links == | == project resources & links == | ||
'''05-16-16''' | '''05-16-16''' | ||