Abstract | ||
---|---|---|
When social networking sites give users granular control over their privacy settings, the result is that some content across the site is public and some is not. How might this content--or characteristics of users who post publicly versus to a limited audience--be different? If these differences exist, research studies of public content could potentially be introducing systematic bias. Via Mechanical Turk, we asked 1,815 Facebook users to share recent posts. Using qualitative coding and quantitative measures, we characterize and categorize the nature of the content. Using machine learning techniques, we analyze patterns of choices for privacy settings. Contrary to expectations, we find that content type is not a significant predictor of privacy setting; however, some demographics such as gender and age are predictive. Additionally, with consent of participants, we provide a dataset of nearly 9,000 public and non-public Facebook posts. |
Year | DOI | Venue |
---|---|---|
2017 | 10.1145/2998181.2998223 | CSCW |
Keywords | Field | DocType |
privacy, content analysis, Facebook, dataset, machine learning, Mechanical Turk, mixed methods, prediction, research methods, social media | Content type,Categorization,Content sharing,Content analysis,Internet privacy,Social media,Social network,Computer science,Coding (social sciences),Demographics | Conference |
Citations | PageRank | References |
6 | 0.41 | 31 |
Authors | ||
11 |
Name | Order | Citations | PageRank |
---|---|---|---|
Casey Fiesler | 1 | 190 | 32.32 |
Michaelanne Dye | 2 | 75 | 6.84 |
Jessica L. Feuston | 3 | 22 | 3.91 |
Chaya Hiruncharoenvate | 4 | 38 | 2.64 |
Clayton J. Hutto | 5 | 14 | 1.25 |
Shannon Morrison | 6 | 25 | 1.74 |
Parisa Khanipour Roshan | 7 | 10 | 0.90 |
Umashanthi Pavalanathan | 8 | 55 | 5.10 |
Amy S. Bruckman | 9 | 22 | 4.78 |
Munmun De Choudhury | 10 | 1864 | 123.30 |
Eric Gilbert | 11 | 25 | 2.34 |