Federating the Social Web

Prospectus

Author
Affiliation

Carl Colglazier

Northwestern University

Published

February 28, 2025

Abstract

Decentralized online social networks (DOSNs) represent a new way of organizing online communities and present both challenges and opportunities. The decentralized design of spaces like the Fediverse provide more indepdence, but this increased autonomy comes at the expense of collective action problems which must be solved in new ways. This work considers how the collective actions problems are addressed in practice and designs potential new ways to solve them.

Introduction

The social web has fundamentally changed how people connect, share information, and form communities. Like all social communities, the Web must deal with social challenges such as integrating newcomers, enforcing norms, and coordinating collective action ().

Corporate platforms’ values and approaches follow the interests of their key stakeholders (); however, these approaches cannot satisfy all groups. Recommendation systems inherently must uprank some posts at the expense of others. Content moderation has inherent trade-offs which make it difficult to satisfy all possible parties. The limitations of corporate platforms have spurred people to produce alternates, including decentralized online social networks (DOSNs) like the Fediverse—–a collection of websites that pass messages between themselves using shared, open protocols.

In the Fediverse, control is distributed among various independent server operators. While this distributed design allows for more autonomy and independence, it also creates new coordination challenges and operational nuances: in a distributed system, location matters. The kind of content which is viewable and easily seen is different on each server based on their position within the network. Content moderation approaches vary from server to server. While some Fediverse software like Mastodon allows people to move accounts, attracting and retaining newcomers can be more challenging because many newcomers may lack a mental model for why they might want to join one server rather than other.

Taken as a complete body of work, this dissertation considers how federation affects the function of the Social Web. It frames the challenges of the Federated Social Web as social coordination problems. Across the studies, I use a mixed methods approach starting first with an analysis-motivated design for server recommendations, an interview study with Fediverse administrators and moderators, and a quantitative study on the effects of de-federation events (server-to-server blocks).

Background

The World Wide Web has reshaped our communication landscape with pervasive economic and cultural effects (; ; ; ). The Internet, which forms the technological underpinning for the Web, solved the problem of connecting disparate networks, computers, and operators around the world by creating a flexible system based around interoperability: protocols like TCP/IP and HTTP allowed enough standardization for meaningful communication while allowing individual systems to still experiment and retain control over their data, access, and code. One key to its success is that the largely decentralized nature of the TCP/IP protocol allowed new nodes to join the network with limited friction ().

This approach has facilitated a strong network of person-to-person communication, developing various systems to facilitate communication between their stakeholders. Early Internet developers used email to coordinate their efforts (). Later, early online communities developed on Bulletin Board Systems, USENET, and discussion forums (; ). The blogosphere of the 2000s became a significant media power in its own right and many alumni from this era successfully transitioned into roles in traditional mass media, which itself adopted many of the norms and practices from the bloggers (). Further, the rise of social networking sites brought online communication to a larger audience (see ) and marked a shift in the organization of communities from being around topics to being around people ().

The history of the design and use of these systems each reflect the needs and values of the people who built and use them. Early email, for instance, was closed off to academics and government employees who quite literally met in person to plan the early email programs (). As operators could easily be mapped to their real name and identities, there was less need for security and privacy protections when the protocols were first developed.

Figure 1: Reported use of a selection of major social media platforms by adults in the U.S. over time according the surveys conducted by the Pew Research Center (PRC). Note that in 2021 and before, PRC administered the survey over the phone, while after 2021 they administered the survey online and via mail.

Online communities have not come without problems; connecting so much together creates a number of benefits, but it also comes with challenges. Significant resources are spent by the companies who run them to try to keep them free from spam, harassment, and illegal content (). Differences in international law and norms can make this a challenge and norms can vary by culture and location (). The largest platforms have becoming increasingly closed off and less transparent to researchers (). Concerns persist over the dominant economic model for the commercialized Web, which relies heavily on advertising and attention ().

While the commercial Web has received significant media and lawmaker focus, it does not represent the totality of Web. Much of the early online communities were run by hobbyists and non-profit organizations (). Projects like IndieWeb have sought to create a more decentralized Web, where people own their own data and can interact with others on their own terms (). Today’s Fediverse is an extension of the original spirit of the Web: powered by the ActivityPub standard, the Fediverse is a collection of interoperable and interconnected websites that retain their independence and autonomy. With millions of active accounts, the Fediverse has established itself as a viable alternative to the commercial Web.

While significant work has gone into technical interoperability on the Fediverse (e.g. how do we pass messages between servers?), work on social interoperability is still emerging (e.g. how do we determine which servers we want to get messages from?). Fediverse servers must handle many of the challenges faced by the commercial Web in addition to some of the new challenges imposed by their decentralized design. This prospectus outlines a research agenda to understand how the Fediverse handles these challenges.

The Emergence of the Fediverse

Interoperability

A system supports interoperability when information can be exchanged between parts of the system. Interoperability can have tremendous benefits because it guarantees parts of the system can work together while at the same time supporting components that are developed or operated independently. For example, the Internet is a highly interoperable system because it allows computers and networks to communicate with each other using shared protocols. Similarly, Email allows servers and accounts to pass messages to each other using shared protocols, but run their own software of choice.

What makes federated systems different from simple linking in practice is how they handle data. Individual nodes in a federated system do not simply link to data from other nodes in practice, but often store and replicate a copy of the data. In practice, this means such systems are less vulnerable to censorship, but it also introduces new complications regarding privacy. Deleting data from such a system is non-trivial.

Early Examples of Federated Social Websites

The Fediverse’s cultural roots are in the free software movement, which emphasizes permissive licensing and open source software. Early Fediverse projects largely attempted to create libre alternatives to corporate social media platforms (). For example, the GNU Social project (formerly known as StatusNet) was created as a free software approach to microblogging. Similarly, the Diaspora project launched in 2010, inspired by its founders’ shared concerns over the consolidation of information on the cloud (). The network claimed over two hundred thousand users by November 2011 () and was designed to be decentralized, with data stored and managed on independent servers known as “pods”.

More recently, the IndieWeb created a method for interlinking websites using shared standards such as Microformats 2 to mark semantic data (). Culturally, the IndieWeb often encouraged people post on their own websites and then syndicate to other websites using a system called POSSE (Post on your Own Site, Syndicate Elsewhere). Protocols like WebMentions allow IndieWeb sites to interact with each other.

ActivityPub

The ActivityPub (AP) protocol’s story starts from a place of fragmentation. While projects like GNU Social and StatusNet had a small but dedicated following, it was difficult to pass messages between servers with incompatible protocols like OStatus and Pump.io (). The ActivityPub project sought to bridge the collective of early Fediverse projects under a unifying standard and was recommended by the World Wide Web Consortium in 2018.

Although the protocols underlying the internet remains invisible for most of its users, a tremendous amount of time and effort go into their development. The development of a protocol can be challenging because it is impossible to anticipate all possible use-cases and there are trade-offs to most design decisions. ActivityPub, for instance, has been criticized for its lack of optimization and vulnerability to accidental distributed-denial-of-service attacks (). It is important to remember that the ActivityPub standard represents the specific needs of their stakeholders at the time it was designed and adopted: namely, bridging the prior Fediverse protocols under a single, unified standard.

Fragmentation remains a challenge for DOSNs. It is hard to balance keeping wide compatibility with the creation of new and innovative features. While ActivityPub has been largely successful at bridging the fledgling communities it was designed for, further development on DOSNs may well leave AP behind. For instance, Bluesky has opted to produce its own protocol instead of adopting AP citing issues with data portability and scalability ().

Mastodon

Mastodon represents the most important Fediverse project to date. But despite its millions of registered users and coverage in major media publications, the project had humble beginnings. Eugen Rochko released the first public edition of Mastodon in October 2016. In a comment on the Hacker News thread on launch, Rochko wrote about the project’s ethos: “This isn’t a startup, it’s an open-source project. Most likely the Twitters and Facebooks will win, but people should have a viable choice… Plus this is an incredibly fun project to be working on, to be quite honest” (). The software soon found an audience and eclipsed the user base of other Fediverse software.

Early reporting on Mastodon often described it an alternative to other platforms like Twitter, a framing which Zulli, Liu, and Gehl () criticized.

Figure 2: Histogram of Mastodon servers by number of users on October 2023. Note that the bins and y-axis both use a log 10 scale. The majority of Mastodon servers are small.

Most Mastodon servers are small. The vast majority of Mastodon servers have fewer than 10 accounts. Many of these have only a single account. The distribution of accounts on servers, however, is highly skewed: the median server has 3 accounts, while the mean has 595 accounts.

Challenges for Online Communities

Collective Action Problems

Kollock () defines social dilemmas as situations where individually rational behavior leaves the collective worse off. These social dilemmas have a deficient equilibrium where there is an outcome that leaves everyone better off, but no individual incentive to move toward that outcome ().

Collective action problems are social dilemmas that occur when self-interested individuals have no incentive to work toward a public good (). Hardin () described one such collective action problem, where people individually overexploit shared resources, as the tragedy of the commons in a Malthusian argument against population growth. More recent scholarship, however, has shown that the tragedy of the commons is not an inevitable outcome from shared resources (). Instead, people can and do work together to manage shared resources.

Kollock () argues that online communities produce public goods, often in the form of knowledge with a nearly limitless potential audience. Access to free and accurate information leaves everyone better off. This idea has been the foundation of knowledge production projects like Wikipedia and open source software projects like the Linux kernel, both of which function as public goods (non-excludable and non-rivalrous).

Due to the differences in the means of production within the Fediverse, many of the challenges faced by commercial websites are trasformed into collective action problems. These can create challenegs for the sytem as a whole. For instance, several major Fediverse servers have shut down over the years due to hosting costs. The vast majority of people with accounts on the Fediverse do not directly financially contribute to their servers—though many do.

Content Moderation

All websites which rely on third-party, user-generated content must perform some form of content moderation to remain viable (). Without it, online spaces would become dominated by spam, pornography, or other unwanted content (). Much of this work remains largely invisible by design ().

While all social websites must perform content moderation, approaches vary. Large, well-resourced websites like Facebook hire teams of contractors which do the bulk of cleaning up their website. Other websites built around named subcommunities like Reddit hand off moderation duties to unpaid volunteer community members.

The small size of the average Mastodon server affects content moderation. Despite the small size of most Mastodon servers, the average Mastodon account is on a large server. Raman et al. () found the top 5% of Mastodon servers host 90.6% of Mastodon accounts and send 94.8% of the posts. This means while the bulk of the moderation work concentrates on a few large servers, vulnerabilities and problems can come from a large set of smaller, less resourced servers.

Nicholson, Keegan, and Fiesler () characterized the written rules on a number of Mastodon servers.

Discovery

Recommender systems help people filter information to find resources relevant to some need (). The development of these systems as an area of formal study harkens back to information retrieval (e.g. Salton and McGill ()) and foundational works imagining the role of computing in human decision-making (e.g. Bush ()). Early work on these systems produced more effective ways of filtering and sorting documents in searches such as the probabilistic models that motivated the creation of the okapi (BM25) relevance function (). Many contemporary recommendation systems use collaborative filtering, a technique which produces new recommendations for items based on the preferences of a collection of similar users ().

Collaborative filtering systems build on top of a user-item-rating (UIr) model where there is a set of users who each provide ratings for a set of items. The system then uses the ratings from other users to predict the ratings of a user for an item they have not yet rated and uses these predictions to create a ordered list of the best recommendations for the user’s needs (). Collaborative filtering recommender systems typically produce better results as the number of users and items in the system increases; however, they must also deal with the “cold start” problem, where limited data makes recommendations unviable (). The cold start problem has three possible facets: boostrapping new communities, dealing with new items, and handling new users (). In each case, limited data on the entity makes it impossible to find similar entities without some way of building a profile. Further, uncorrected collaborative filtering techniques often also produce a bias where more broadly popular items receive more recommendations than more obscure but possibly more relevant items (). Research on collaborative filtering has also shown that the quality of recommendations can be improved by using a combination of user-based and item-based collaborative filtering ().

Although all forms of collaborative filtering use some combination of users and items, there are two main approaches to collaborative filtering: memory-based and model-based. Memory-based approaches use the entire user-item matrix to make recommendations, while model-based approaches use a reduced form of the matrix to make recommendations. This is particularly useful because the matrix of items and users tends to be extremely sparse, e.g. in a movie recommendor system, most people have not seen most of the movies in the database. Singular value decomposition (SVD) is one such dimension reduction technique which transforms a m×n matrix M into the form M=UΣVT (). SVD is particularly useful for recommendation systems because it can be used to find the latent factors which underlie the user-item matrix and use these factors to make recommendations.

While researchers in the recommendation system space often focus on ways to design the system to produce good results mathematically, human-computer interaction researchers also consider various human factors which contribute to the overall system. Crucially, McNee et al. argued “being accurate is not enough”: user-centric evaluations, which consider multiple aspects of the user experience, are necessary to evaluate the full system. HCI researchers have also contributed pioneering recommender systems in practice. For example, GroupLens researchers Resnick et al. () created a collaborative filtering systems for Usenet and later produced advancements toward system evaluation and explaination of movie recommendations (; ). Cosley et al. () created a system to match people with tasks on Wikipedia to encourage more editing. This prior work shows that recommender systems can be used to help users find relevant information in a variety of contexts.

Mastodon and other decentralized online social networks are particularily vulnerable to discovery problems. As information and accounts are spread out across many different servers, location matters in a way that is not relevant on centralized social networks. At the same time, any recommendation system run on a particular server is limited to the information on that server unless some system is in place to spread recommendations across servers, e.g. using federated machine learning.

Studies

De-federation

Note

This study has been peer-reviewed and published as:

Colglazier, Carl, Nathan TeBlunthuis, and Aaron Shaw. 2024. “The Effects of Group Sanctions on Participation and Toxicity: Quasi-Experimental Evidence from the Fediverse”. Proceedings of the International AAAI Conference on Web and Social Media 18 (1):315-28. https://doi.org/10.1609/icwsm.v18i1.31316.

Content moderation in response to toxic and anti-social behavior is pervasive in social media. In general, moderation interventions strive to balance the value of wide and active user bases with the threats posed by conflicts and hate speech (). Websites that host user-generated content and sub-communities apply many kinds of policies and interventions. However, when norms diverge across interconnected, independent communities in the absence of a single (corporate or not) parent or owner, governance and moderation pose acute challenges.

A growing empirical literature has investigated social media content moderation and governance. Moderation actions most frequently target individual posts and accounts, but other group-level sanctions affect entire communities or websites. For example, Reddit has banned subreddits and Discord has blocked servers, reducing the prevalence of unwanted behavior within and sometimes beyond the targeted groups (, ; ; ; ; ). Most prior work on group-level sanctions focuses on sanctions applied by central actors such as commercial social media platform staff. However, autonomous community administrators can also enact group-level sanctions such as when a sub-community restricts contributions from members of another sub-community. These decentralized group-level sanctions are distinct in that the targeted sub-community remains part of the larger network. To our knowledge, the effects of such sanctions remain unexplored empirically.

Figure 3: Illustration of how defederation disconnects two servers and thereby disconnects the subnetworks of people using each server. The top row shows the network of servers before (left) and after (right) defederation. The bottom row shows the corresponding networks of users. On the left, an edge connects a user on one server with a user on a different server. Defederation (right) disconnects them so they can no longer exchange messages.

To investigate the effects of decentralized group-level sanctions, we analyze defederation events in the Fediverse, a decentralized social media system which consists of independently managed servers that host individual accounts and pass messages using shared protocols. Communication between servers can happen only when the administrators of both servers permit it. Server administrators can revoke such permission by “defederating” from (blocking all interactions with) specific servers. Defederation is one of the few tools administrators in a decentralized system have to protect against bad actors or enforce norms from beyond their own servers. While many defederation events on the Fediverse occur between servers with no known interaction history, many also come in response to norm violations and toxic interactions across server boundaries with a history of previous interactions. Defederations that cutoff cross-server interactions provide an opportunity to identify the effects of these group sanctions on accounts most likely to be directly affected.

We collect data from 214 defederation events between January 1, 2021 to August 31, 2022 that involved 275 servers and 661 accounts which had previously communicated across subsequently defederated inter-server connections. Using a combination of non-parametric and parametric methods, we estimate the effects of defederation on two outcomes: posting activity and toxic posting behavior among affected accounts. We find an asymmetric impact on posting activity: Accounts on blocked servers reduce their activity, but not accounts on blocking servers. By contrast, we find that defederation has no effects on post toxicity on either the blocked or blocking servers.

These findings suggest that defederation, although a common group-level sanction on the Fediverse, has mixed effectiveness: Despite the risks of severing communication channels, communities implementing group-level sanctions do not lose activity. This implies that defederation may avoid some of the costs associated with other moderation techniques such as account requirements or group sanctions like geographic blocks (; ). Although defederation reduces activity by blocked accounts, we did not find evidence that it made their posts less toxic. This suggests that defederation may not improve adherence to broadly held norms. Our study contributes to knowledge of content moderation on social media in that it (1) describes defederation, a novel form of group-level sanction as instantiated on the Fediverse, (2) derives hypotheses regarding the effects of defederation from prior literature, (3) creates a novel dataset of defederation events, (4) conducts a quasi-experimental analyses to quantify effects of defederation on parties affected on the blocked and blocking servers and (5) finds that defederation has asymmetric effects on activity and no measurable effect on toxicity.

Research Questions

  • S1 RQ1: How does defederation impact the activity levels for affected accounts on (a) the defederated instance (blocked server); and (b) the defederating instance (blocking server)?
  • S1 RQ2: How does defederation impact toxic posting behavior among the affected accounts on either (a) the blocked or (b) blocking servers?

Data

We pursue an observational, quasi-experimental research design to identify effects of defederation on the activity and toxic posting behavior in the Fediverse. We collected longitudinal trace data from 7,445 publicly listed defederation events and about 104 million public posts that occurred in the Fediverse on either the Mastodon or Pleroma networks between April 2, 2021 and May 31, 2022. Using this data, we analyze activity of user accounts (for RQ1) and the toxicity of their messages (for RQ2) on the blocking and blocked servers impacted by these events in comparison to matched control accounts. We apply a difference-in-differences approach and present both non-parametric and parametric estimates of the effects of defederation.

Figure 4: The y-axis shows the cumulative number of blocked and blocking accounts included in our analysis over our study period.
Figure 5: Box and whisker plots visualize the distributions of our dependent variables within the blocked and blocking groups of user accounts and their matched controls before and after defederation. The lines correspond to the median, the boxes to the inter-quartile range (IQR), the whiskers to the range of the data within 1.5 * IQR, and the dots to data points outside the range of the whiskers.

Methods

Figure 6: A covariate balance plot shows the standardized mean difference between treatment and control groups for each measure used in our matching procedure before (unadjusted) and after (adjusted) matching. Our procedure effectively found a group of matched controls similar to the treated accounts along these measures.

We used one-to-one matching, selecting the closest match according to Mahalanobis distance and discarded accounts for which there was not a sufficiently good match. shows the distribution of variables in the treatment and control groups, while shows the effectiveness of the matching process for the selection variables.

For both activity and toxicity, we present the median counts per account by week as well as the results of the non-parametric Wilcoxon signed-rank test and a difference-in-differences estimate.

Findings

Figure 7: Visualization of activity among blocked and blocking user accounts shows an asymmetric change in activity following defederation. An account with a median post count on the blocked server declines in activity much more rapidly following defederation compared to matched controls while an account with a median post count on the blocking server declines similarly to matched controls.
Table 1: Non-parametric tests for differences in activity before and after defederation events (summed across all weeks) find a measurable decrease in posting activity for the accounts on blocked servers compared to matched controls but no such change for accounts on blocking servers.
Group median W p
$U_0$ -135.5 41197.5 0.000
$C_0$ -18.0 35762.0 0.143
$U_1$ -54.5 12413.0 0.122
$C_1$ -53.5 12520.0 0.091
$\Delta_0$ -39.0 39927.0 0.000
$\Delta_1$ 3.0 10645.5 0.421
Table 2: Difference-in-differences analysis of activity level for user accounts whose server was defederated (blocked group) or whose server defederated another (blocking group). The 95% credible interval negative coefficient for membership in the blocked group post-defederation (β5) is less than 0, indicating that activity by accounts in this group decreased more than accounts in the matched control group. We do not draw such a conclusion about members of the blocking server because the corresponding credible interval contains 0.
Blocked Blocking
Term Estimate Std. err. Low High Estimate Std. err. Low High
β0 (Intercept) 3.216 0.086 3.049 3.380 2.901 0.118 2.684 3.121
β1 Group -0.024 0.116 -0.255 0.198 0.023 0.177 -0.290 0.367
β2 Treatment -0.089 0.045 -0.178 0.002 -0.080 0.060 -0.201 0.034
β3 Time 0.006 0.004 -0.002 0.015 -0.002 0.006 -0.014 0.009
β4 Treatment : Time -0.026 0.006 -0.038 -0.014 -0.027 0.008 -0.042 -0.010
β5 Group : Treatment -0.241 0.065 -0.367 -0.116 -0.084 0.082 -0.247 0.072
β6 Group : Time -0.007 0.006 -0.020 0.004 0.006 0.008 -0.010 0.021
β7 Group : Treatment : Time -0.015 0.009 -0.031 0.002 0.009 0.011 -0.013 0.032
Figure 8: Median toxicity among accounts which posted each week for blocked and blocking user accounts. The median toxicity remained flat for all groups.
Table 3: Non-parametric difference-in-differences for median post toxicity before and after de-federation events. The W test statistic represents the sum of the ranks of the positive differences between paired observations while the p-value compares to the alternative hypothesis that the changes are zero.
Group median W p
U0 -0.006 17746 0.538
C0 0.004 14000 0.950
U1 -0.008 6514 0.619
C1 0.001 5546 0.873
Δ0 -0.005 17161 0.072
Δ1 0.000 6414 0.305
Table 4: Beta regression coefficients drawn from the posterior of the parametric toxicity DiD model for user accounts whose server was defederated (blocked group) or whose server defederated another (blocking group). For all groups, the 95% credible intervals for a change in toxicity levels after treatment (β1, β5) contain 0.
Blocked Blocking
Group Term Estimate Std. err. Low High Estimate Std. err. Low High
Treatment β0 (Intercept) -0.143 0.438 -1.197 0.763 -0.042 0.284 -0.754 0.518
Treatment β1 Treatment 0.002 0.006 -0.009 0.018 -0.001 0.010 -0.020 0.017
Treatment β2 Time 0.001 0.001 -0.000 0.002 0.001 0.001 -0.001 0.003
Treatment β3 Treatment : Time -0.002 0.001 -0.004 0.000 -0.001 0.002 -0.004 0.003
Control β4 (Intercept) 0.136 0.436 -0.907 1.044 0.048 0.283 -0.613 0.649
Control β5 Treatment -0.000 0.006 -0.016 0.011 0.004 0.009 -0.014 0.024
Control β6 Time 0.001 0.001 -0.000 0.002 0.003 0.001 0.001 0.005
Control β7 Treatment : Time -0.005 0.001 -0.007 -0.003 -0.006 0.002 -0.009 -0.002

We find that the effects of defederation are asymmetric: accounts on the blocked servers decreased their posting activity (, , ) while accounts on the blocking servers did not decrease their posting activity compared to the matched controls (, , ). We found no evidence for a change in toxicity for accounts on either the blocked or blocking servers.

Conclusion

In this study, we investigated the effects of defederation events on the activity levels and toxic posting behavior of accounts in the Fediverse. The results indicate that such events produce asymmetric effects on activity for affected accounts on blocked servers versus those on blocking servers with no increase in toxicity for any groups. The results also highlight the potential of decentralized social networks and their unique mechanisms, such as defederation, in providing communities with tools to manage content moderation and other aspects of online interactions. Future research could explore the causes or reasons behind defederation events, the long-term consequences of defederation, as well as the mechanisms by which defederation produces (asymmetric) effects. By continuing to study the Fediverse and its affordances, we can better understand how to foster healthy online communities and effective content moderation strategies in a decentralized environment.

Newcomers and Server Recommendations

Note

An earlier version of this study was presented at the 1st International Workshop on Decentralizing the Web and the 10th International Conference on Computational Social Science.

Colglazier, Carl. “Do Servers Matter on Mastodon? Data-driven Design for Decentralized Social Media.” 1st International Workshop on Decentralizing the Web, Buffalo, 2024.

Colglazier, Carl. “Do Servers Matter on Mastodon? Data-driven Design for Decentralized Social Media.” 10th International Conference on Computational Social Science, Philadelphia, 2024.

Following Twitter’s 2022 acquisition, Mastodon saw an increase in activity and attention as a potential Twitter alternative (; ). While millions of people set up new accounts and significantly increased the size of the network, many newcomers found the process confusing. Many accounts did not remain active.

Unlike centralized social media platforms, Mastodon is a network of independent servers with their own rules and norms (). While each server can communicate with each other using the shared ActivityPub protocols and accounts can move between Mastodon servers, the local experience can vary widely from server to server.

Although attracting and retaining newcomers is a key challenge for online communities (), Mastodon’s onboarding process has not always been straightforward. Variation among servers can also present a challenge for newcomers who may not even be aware of the specific rules, norms, or general topics of interest on the server they are joining (). Various guides and resources for people trying to join Mastodon offered mixed advice on choosing a server. Some downplayed the importance of server choice and suggested that the most important thing is to simply join any server and work from there (; ); others created tools and guides to help people find potential servers of interest by size and location (; ).

Mastodon’s decentralized design has long been in tension with the disproportionate popularity of a small set of large, general-topic servers within the system (). Analysing the activity of new accounts that join the network, we find that users who sign up on such servers are less likely to remain active after 91 days. We also find that many users who move accounts tend to gravitate toward smaller, more niche servers over time, suggesting that established users may also find additional utility from such servers.

In response to these findings, we propose a potential way to create server and tag recommendations on Mastodon. This recommendation system could both help newcomers find servers that match their interests and help established accounts discover “neighborhoods” of related servers to enable further discovery.

Research Questions

  1. S2 RQ1: What kinds of Mastodon servers do a better job of retaining newcomers?
  2. S2 RQ2: How can we build an opt-in, low-resource recommendation system for finding Fediverse servers?

Data

Mastodon has an extensive API which allows for the collection of public posts and account information. We collected data from the public timelines of Mastodon servers using the Mastodon API with a crawler that runs once per day. We also collected account information from the opt-in public profile directories on these servers.

Figure 9: Mastodon accounts in the dataset created between January 2022 and March 2023. The top panels shows the proportion of accounts still active 45 days after creation, the proportion of accounts that have moved, and the proportion of accounts that have been suspended. The bottom panel shows the count of accounts created each week. The dashed vertical lines in the bottom panel represent the annoucement day of the Elon Musk Twitter acquisition, the acquisition closing day, a day where Twitter suspended a number of prominent journalist, and a day when Twitter experienced an outage and started rate limiting accounts.

Initial Findings

Account Survival

Our initial findings suggest that servers do matter for newcomers on Mastodon. uses a Kaplan–Meier estimator to show that accounts on the largest Mastodon servers featured on the Join Mastodon website are less likely to remain active compared to accounts on smaller servesr featured on Join Mastodon. Further, uses a Cox Proportional Hazard Model with mixed effects to suggest that small servers are significantly better at retaining new accounts and that general servers are less likely to retain new accounts; being featured on the Join Mastodon website appears to have no significant effect.

Figure 10: Survival probabilities for accounts created during May 2023. Using N accounts created from May 1 to June 30, 2023, we create a Kaplan–Meier estimator for the probability that an account will remain active based on whether the account is on one of the largest general instances featured at the top of the Join Mastodon webpage or otherwise if it is on a server in the Join Mastodon list. Accounts are considered active if they have made at least one post after the censorship period M days after account creation.
Table 5: Coefficients for the Cox Proportional Hazard Model with Mixed Effects. The model includes a random effect for the server.
Term Estimate Low High p-value
Join Mastodon 0.115 0.972 1.296 0.117
General Servers 0.385 1.071 2.015 0.017
Small Server -0.245 0.664 0.922 0.003

Moved Accounts

Table 6: Exponential family random graph models for account movement between Mastodon servers. Accounts in Model A were created in May 2022 and moved to another account at some later point. Accounts in Model B were created at some earlier point and moved after October 2023.
Model A Model B
Coef. Std.Error Coef. Std.Error
(Sum) -9.529 ***0.188 -10.268 ***0.718
nonzero -3.577 ***0.083 -2.861 ***0.254
Smaller server 0.709 ***0.032 0.629 ***0.082
Server size (outgoing) 0.686 ***0.013 0.655 ***0.042
Open registrations (incoming) 0.168 ***0.046 -0.250 0.186
Languages match 0.044 0.065 0.589 0.392

To corroborate these findings, we also looked at accounts that moved from one server to another. We find that accounts are more likely to move from larger servers to smaller servers.

Recommender System

Figure 11: Screenshot of the prototype Mastodon server recommendation system. In this example, the system is recommending servers based on the inputs of “academia”, “programming”, and “research”.

Based on the empirical findings, we create a recommendation system that can help people find Mastodon servers based on their shared interests. shows the system in practice. We use Okapi BM25 to construct a term frequency-inverse document frequency (TF-IDF) model to associate the top tags with each server using counts of tag-account pairs from each server for the term frequency and the number of servers that use each tag for the inverse document frequency. We then L2 normalize the vectors for each tag and calculate the cosine similarity between the tag vectors for each server.

tf=ft,s(k1+1)ft,s+k1(1b+b|s|avgstl)

where ft,s is the number of accounts using the tag t on server d, k1 and b are tuning parameters, and avgstl is the average sum of account-tag pairs. For the inverse document frequency, we use the following formula:

idf=logNn+0.5n+0.5

where N is the total number of servers and n is the number of servers where the tag appears as one of the top tags. We then apply L2 normalization:

tfidf=tfidftfidf2

We then used the normalized TF-IDF matrix to produce recommendations using SVD where the relationship between tags and servers can be presented as A=UΣVT. We then use the similarity matrix to find the top servers which match the user’s selected tags. We can also suggest related tags to users based on the similarity between tags, UΣ.

Model Evaluation

Evaluating recommender systems can be tricky because a measure of good performance must take into account various dimensions (). A measure of accuracy must be paired with a question of “accuracy toward what?” Explainability requires a transparent way to show the user why a certain item was recommended.

It is often important to both start with an end goal in mind and to keep evaluation integrated throughout the entire process of creating a recommender systems, from conceptualization to optimization. There are several considerations to keep in mind such as the trade-off between optimizing suggestions and the risks of over-fitting. For example, a system designed to prioritize user interest may struggle with reduced diversity.

Recommender systems can be evaluated using three board categories of techniques: offline, online, and user studies. Offline evaluation uses pre-collected data and a measure to describe the performance of the system, assuming there is insufficient relevance to the difference in time between when the data was collected and the present moment. Online evaluation uses a deployed, live system, e.g. A/B testing. In this case, the user is often unaware of the experiment. In contrast, user studies involve subjects which are aware they are being studied.

To evaluate my system, I plan to use a mix of offline evaluation and user studies. For the offline evalulation, I plan to leverage known accounts which moved from one server to another. Based on their posting history (e.g. hashtags), can the recommendation system predict where they will move to? If the system ranks their destination server highly, this suggests it can do a good job recommending servers for the population of accounts which care enough about servers to move from one to another.

As my recommender system operates under the assumption that smaller, more topic-focused servers are better, it follows that a diverse set of niche results which only match a few tags are more helpful than a set of results which match a larger and more broad set of tags. The system therefore presents results sorted in a manner which encourages a higher diversity of results.

The inital results suggest the system does a decent job of matching moved accounts to their destination servers. On average, the system suggests the destination server as the 8th most likely option given an account’s post history.

Rules, Norms, and Content Moderation

While the Fediverse promises more autonomy in social media operations, most servers on the Fediverse have actually adopted similar rules, even directly copying them in many cases. Why, despite the opportunity to innovate, do Fediverse servers tend to adopt the same rules?

We consider this puzzle through the persepctive of institutional isomorphism. Through interviews with 17 server moderators and administrators combined with longitudinal records from thousands of servers, we find evidence of isomorphism between Fediverse servers and describe some of the mechanisms through which this occurs.

  • RQ1: How do rules originate and spread on the Fediverse?

  • RQ2: Why do Fediverse servers have similar rules?

Our study suggests that rules on the Fediverse perform an isometric role in addition to a practical role: as a signal to other servers and to people on the server. Rules encode not only what behaviors are permissible, but also community values.

Background

Institutional Isomorphism

Meyer and Rowan () consider rules as myths which insitutions incorporate to bolster their own legitimatacy, sometimes at the expense of internal efficiency. Under this neo-institutionalist view, individual actions are constrained by societal expectations, which in turn lead to the development of formal and informal rules within insitutions.

P. J. DiMaggio and Powell () consider the problem of why orginizations tend to look similar to each other and theorize that orginizations tend to adopt similar norms through a process of insitutional isomorphism, which has three types: normative, coersive, and mimetic. Normative isomorphic change is a byproduct of professionalization. Coersive isomporpohic change comes from pressures from other organizations. Mimetic isomorphism describes how orgainizations tend to imitate each other.

Rules and Content Moderation

Significant scholarship has investigated content moderation in the context of commercial, centralized social media platforms (e.g. ; ). The function of content moderation in this context is shaped by economic factors: platforms often use contract labor and the people making case-by-case decisions are often divorced from the context of the communities they moderate (). However, a significant amount of the Web also relies on volunteer moderation: service work from unpaid people who are usually active members of their own communities. These volunteers might have different motivations for their actions.

While some commercial websites like Reddit allow volunteer moderators for sub-communities to create their own rules, moderators must also consider the site-wide rules (). The Fediverse, by contrast, requires no base layer of rules for each community and account: there are no site-wide rules ().

Content Moderation on the Fediverse

Due to its technical design, the Fediverse faces unique content moderation challenges. Administrators may be responsible not only for posts on their own servers, but they may also have to deal with posts and conflicts from external servers. Like the servers themselves, the costs of content moderation is distributed among many servers. Server administrators may choose to block or silence accounts on other servers, but only the server operators have the powers to delete or directly sanction accounts.

Servers must also consider their relationship to other servers from which they may receive posts. Hassan et al. () explored the use of server-to-server federation policies and suggested that some server-level blocks may have collatoral damage on some users. Colglazier, TeBlunthuis, and Shaw () measured the effects of server-to-server de-federation events and found asymmetric effects on activity for affected accounts on the blocked servers.

Data

Figure 12: Cumulative counts of servers as their metadata appears in the quantitative data.

I collected metadata from daily snapshots from thousands of Fediverse servers between 2020 and 2025.

The period of this study came at a somewhat transitional point on Mastodon. Originally Mastodon rules were posted on the “about” page for the server. These rules often were long lists, often clearly copied from other servers. Mastodon v3.4 added an API field that allowed Mastodon administrators to add rules as server metadata.

Raw Longitudinal Data

Filter Servers

Vectorize Rules with SBERT

Cluster Similar Rules

Figure 13: Diagram for the processing of rules data.

Initial Findings

Table 7: Most common rules on Fediverse servers.
Rule Count
No racism, sexism, homophobia, transphobia, xenophobia, or casteism 1925
No incitement of violence or promotion of violent ideologies 1629
No harassment, dogpiling or doxxing of other users 1581
No illegal content. 1273
Sexually explicit or violent media must be marked as sensitive when posting 1220
Do not share intentionally false or misleading information 1074
Be nice. 510
No spam or advertising. 410
Don't be a dick. 405
Be excellent to each other. 264

Interviews

We conducted interviews with 17 moderators and admins of Fediverse communities in 2022.

Table 8: Descriptions of the seventeen interviews in the study.
Descriptions of the seventeen interviews in the study.
ID Software Size Topic
FV1 Mastodon [100–1K) Regional
FV2 Mastodon [1K–10K) Language
FV3 Mastodon [10–100) Interest/Language
FV4 Mastodon [1K–10K) Interest(?)
FV5 Mastodon [10-100) Regional/Interst
FV6 Pleroma ?
FV7 Mastodon [100–1K)
FV8 Mastodon [100–1K)
FV9 Mastodon [1K–10K)
FV10 Mastodon [100–1K)
FV11 Mastodon [100–1K)
FV12 Mastodon [10–100)
FV13 Mastodon [100–1K) Interest
FV14 Pleroma ?
FV15 Pleroma ? Religion
FV16 Misskey ?
FV17 Mastodon [100–1K)

Identifying Communities

To identify communities of interest, we first considered the written rules on Mastodon and Pleroma servers. Starting with Mastodon Social, we used iterative sampling to discover new servers from the set of peer connections on previously known servers. We then downloaded the “/about/more” HTML data on Mastodon servers and the “/static/terms-of-service.html” HTML data on Pleroma servers as they appeared on September 25, 2021. I chose to go with a snapshot from September 25, 2021 as this was before the rules API feature got added to Mastodon and so we get the richest text content. Using these HTML data, we filtered to only servers which had texts in English, had over 100 tokens of text on these pages (after removing stop words, articles, and non-letters). From this matrix, we calculate the pairwise similarity between all M documents by counting the number of n-grams they share in common using the following formula:

count(X1X2)min(count(X1),count(X2))

I then created clusters using the HDBSCAN algorithm to construct clusters of similar rules. We recruited participants with the goal of representing servers in each main cluster.

Qualitative Analysis Methods

I plan to use directed qualitative content analysis, starting under the framework of institutional isomorphism. Several practices that lead to rule creation suggest strong institutional isomorphism amoung Fediverse servers ().

Figure 14: Diagram of three kinds of institutional isomorphism and how they shape rules in the Fediverse.

Rules Memos

Applying local rules to external posts

Some servers attempt to apply their own local rules to posts on other servers. This may create significant difficulties when federating with other servers.

For example, FV15 is a religious community whose members hold strong beliefs about appropriate content. For this community, part of the service they provide to their members is not only ensuring that the posts on the local timeline follow the community values, but also that all posts in their known network are compliant. A further motivation toward their approach comes from how the Fediverse works: they do not want to host copies of posts which violate their policies.

Yeah. Honestly, most moderation is manageable if your users stick to the same rules that everybody else that the server has. If they don’t, that’s when you have most of your issues with moderation.

Proactive and reactive moderation

There are two common approaches toward moderation. One tries to be reactive, only taking action after harm has occured; the other tries to be proactive to reduce the chance of harm in the future.

The proactive approach can be represented in the #FediBlock hashtag, which Mastodon moderators use to share and discuss harmful users and servers with each other.

Rules as signposts

Some Fediverse servers use their rules and community metadata as signals to other communities.

When dealing with external servers, staff have limited information to get a picture of how the server works. The rules can thus be a strong signal of a server’s values. If a post is clearly in violation of its host server’s rules, this can signal that the server staff simply have yet to get around to removing it rather than the alternative narrative that it represents typical content found on the server.

Rules to solicit reports

On Mastodon, the rules API is tied to the reporting feature. People can report posts that violate the rules to the staff on their local server and optionally choose to forward that report on to the post’s originating server.

If server staff deal with posts commonly on large servers, it may make sense to adopt similar rules so they can forward those reports on to the larger server’s staff.

Network integrity

Some staff value upholding the network integrity of the Fediverse. They believe that servers blocking each other breaks the user experience. Instead, they prefer to take actions against people, not entire communities.

Plan of Study

I’m interested in using these data to answer some questions about how these rules developed. Are rules more likely to change over time on larger servers when new issues arise such as AI-generated content? Are certain servers more likely to change and adopt those rules than others? Do servers tend to adopt more rules if they talk more with servers that have those types of rules? These are angles that the quantitative data can help answer.

On the qualitative side of things, I have data collected from 17 interviews with Fediverse administrators and moderators. Those interviews can tell us not only what is happening, but also why this is happening. They have the potential to generate a lot of insights into how server administrators think about the rules and the relationship between rules and Fediverse communities. In preparation for the qualitative data anlysis, I read through Sarah J. Tracy’s Qualitative Research Methods book.

Status and Timeline

My current goal is to complete my defense by the end of the current academic year. I believe this should allow for a realistic timeline barring any unexpected circumstances.

Thankfully, I am not starting from scratch on any of the three studies:

  • Study 1 (de-federation) has been published as Colglazier, TeBlunthuis, and Shaw () at ICWSM

  • Study 2 (newcomers) has been presented as Colglazier () at a workshop and at IC2S2

  • Study 3 (rules) has completed the data collection phase and some of the analysis has been completed

Significant work is still needed on analysis of the qualitative interviews for Study 1, the rules data for Study 1, and refining and evaluating the recommender system in Study 2.

Figure 15: Gantt chart for the timeline and major milestones remaining.

References

AT Protocol FAQ.” n.d. AT Protocol. https://atproto.com/guides/faq. Accessed December 4, 2024.
Bielenberg, Ames, Lara Helm, Anthony Gentilucci, Dan Stefanescu, and Honggang Zhang. 2012. “The Growth of Diaspora - A Decentralized Online Social Network in the Wild.” In 2012 Proceedings IEEE INFOCOM Workshops, 13–18. https://doi.org/10.1109/INFCOMW.2012.6193476.
boyd, danah m., and Nicole B. Ellison. 2007. “Social Network Sites: Definition, History, and Scholarship.” Journal of Computer-Mediated Communication 13 (1): 210–30. https://doi.org/10.1111/j.1083-6101.2007.00393.x.
Bush, Vannevar. 1945. “As We May Think.” The Atlantic 176 (1): 101–8.
Campbell-Kelly, Martin, and Daniel D Garcia-Swartz. 2013. “The History of the Internet: The Missing Narratives.” Journal of Information Technology 28 (1): 18–33. https://doi.org/10.1057/jit.2013.4.
Chandrasekharan, Eshwar, Shagun Jhaver, Amy Bruckman, and Eric Gilbert. 2022. “Quarantined! Examining the Effects of a Community-Wide Moderation Intervention on Reddit.” ACM Transactions on Computer-Human Interaction 29 (4): 29:1–26. https://doi.org/10.1145/3490499.
Chandrasekharan, Eshwar, Umashanthi Pavalanathan, Anirudh Srinivasan, Adam Glynn, Jacob Eisenstein, and Eric Gilbert. 2017. “You Can’t Stay Here: The Efficacy of Reddit’s 2015 Ban Examined Through Hate Speech.” Proc. ACM Hum.-Comput. Interact. 1 (CSCW): 31:1–22. https://doi.org/10.1145/3134666.
Colglazier, Carl. 2024. “Do Servers Matter on Mastodon? Data-driven Design for Decentralized Social Media.” International Workshop on Decentralizing the Web 1.
Colglazier, Carl, Nathan TeBlunthuis, and Aaron Shaw. 2024. “The Effects of Group Sanctions on Participation and Toxicity: Quasi-experimental Evidence from the Fediverse.” Proceedings of the International AAAI Conference on Web and Social Media 18 (May): 315–28. https://doi.org/10.1609/icwsm.v18i1.31316.
Cosley, Dan, Dan Frankowski, Loren Terveen, and John Riedl. 2007. SuggestBot: Using Intelligent Task Routing to Help People Find Work in Wikipedia.” In Proceedings of the 12th International Conference on Intelligent User Interfaces, 32–41. IUI ’07. New York, NY, USA: ACM. https://doi.org/10.1145/1216295.1216309.
Crandall, Robert, William Lehr, and Robert Litan. 2007. “The Effects of Broadband Deployment on Output and Employment: A Cross-sectional Analysis of U.S. Data.” 6. Brookings Institution.
Das, Ankush. 2024. “Please Don’t Share Our Links on Mastodon: Here’s Why!” It’s FOSS News. https://news.itsfoss.com/mastodon-link-problem/.
Davenport, Thomas H., and John C. Beck. 2001. The Attention Economy: Understanding the New Currency of Business. Boston, Mass: Harvard Business School Press.
Diaz, Jesus. 2022. “Using Mastodon Is Way Too Complicated to Ever Topple Twitter.” Fast Company. https://www.fastcompany.com/90808984/using-mastodon-is-way-too-complicated-to-ever-topple-twitter.
DiMaggio, Paul J., and Walter W. Powell. 1983. “The Iron Cage Revisited: Institutional Isomorphism and Collective Rationality in Organizational Fields.” American Sociological Review 48 (2): 147–60. https://doi.org/10.2307/2095101.
DiMaggio, Paul, Eszter Hargittai, W. Russell Neuman, and John P. Robinson. 2001. “Social Implications of the Internet.” Annual Review of Sociology 27 (1): 307–36. https://doi.org/10.1146/annurev.soc.27.1.307.
Drezner, Daniel W., and Henry Farrell. 2008. “Introduction: Blogs, Politics and Power: A Special Issue of Public Choice.” Public Choice 134 (1/2): 1–13. https://www.jstor.org/stable/27698207.
Driscoll, Kevin. 2022. The Modem World: A Prehistory of Social Media. Yale University Press.
Ekstrand, Michael D., John T. Riedl, and Joseph A. Konstan. 2011. “Collaborative Filtering Recommender Systems.” Foundations and Trends in Human–Computer Interaction 4 (2): 81–173. https://doi.org/10.1561/1100000009.
Fiesler, Casey, Jialun" Aaron" Jiang, Joshua McCann, Kyle Frye, and Jed R. Brubaker. 2018. “Reddit Rules! Characterizing an Ecosystem of Governance.” In Proceedings of the International AAAI Conference on Web and Social Media, 72–81. Stanford, CA: AAAI.
Freelon, Deen. 2018. “Computational Research in the Post-API Age.” Political Communication 35 (4): 665–68. https://doi.org/10.1080/10584609.2018.1477506.
Gillespie, Tarleton. 2018. Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions That Shape Social Media. New Haven: Yale University Press.
———. 2020. “Platforms Throw Content Moderation at Every Problem.” In Fake News: Understanding Media and Misinformation in the Digital Age. The MIT Press. https://doi.org/10.7551/mitpress/11807.001.0001.
Göndör, Sebastian, and Axel Küpper. 2017. “The Current State of Interoperability in Decentralized Online Social Networking Services.” In 2017 International Conference on Computational Science and Computational Intelligence (CSCI), 852–57. https://doi.org/10.1109/CSCI.2017.148.
Gray, Mary, and Siddharth Suri. 2019. Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. Boston, MA: Houghton Mifflin Harcourt.
Hardin, Garrett. 1968. “The Tragedy of the Commons.” Science 162 (3859): 1243–48. https://doi.org/10.1126/science.162.3859.1243.
Hassan, Anaobi Ishaku, Aravindh Raman, Ignacio Castro, Haris Bin Zia, Emiliano De Cristofaro, Nishanth Sastry, and Gareth Tyson. 2021. “Exploring Content Moderation in the Decentralised Web: The Pleroma Case.” In Proceedings of the 17th International Conference on Emerging Networking EXperiments and Technologies, 328–35. CoNEXT ’21. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3485983.3494838.
Hauben, Michael, and Ronda Hauben. 1997. Netizens: On the History and Impact of Usenet and the Internet. Washington: IEEE computer.
He, Jiahui, Haris Bin Zia, Ignacio Castro, Aravindh Raman, Nishanth Sastry, and Gareth Tyson. 2023. “Flocking to Mastodon: Tracking the Great Twitter Migration.” In Proceedings of the 2023 ACM on Internet Measurement Conference, 111–23. IMC ’23. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3618257.3624819.
Herlocker, Jonathan L., Joseph A. Konstan, and John Riedl. 2000. “Explaining Collaborative Filtering Recommendations.” In Proceedings of the 2000 ACM Conference on Computer Supported Cooperative Work, 241–50. CSCW ’00. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/358916.358995.
Herlocker, Jonathan L., Joseph A. Konstan, Loren G. Terveen, and John T. Riedl. 2004. “Evaluating Collaborative Filtering Recommender Systems.” ACM Transactions on Information Systems 22 (1): 5–53. https://doi.org/10.1145/963770.963772.
Hill, Benjamin Mako. 2014. “Google Has Most of My Email Because It Has All of Yours.” Copyrighteous.
Hill, Benjamin Mako, and Aaron Shaw. 2021. “The Hidden Costs of Requiring Accounts: Quasi-experimental Evidence from Peer Production.” Communication Research 48 (6): 771–95. https://doi.org/10.1177/0093650220910345.
Jamieson, Jack, Naomi Yamashita, and Rhonda McEwen. 2022. “Bridging the Open Web and APIs: Alternative Social Media Alongside the Corporate Web.” Social Media + Society 8 (1): 20563051221077032. https://doi.org/10.1177/20563051221077032.
Kaye, David. 2019. Speech Police: The Global Struggle to Govern the Internet. New York: Columbia Global Reports.
King, Jaz-Michael. 2024. “Mastodon Near Me.” Jaz-Michael King.
Kollock, Peter. 1998. “Social Dilemmas: The Anatomy of Cooperation.” Annual Review of Sociology 24 (January): 183–214. https://www.jstor.org/stable/223479.
———. 1999. “The Economies of Online Cooperation: Gifts and Public Goods in Cyberspace.” In Communities in Cyberspace, edited by Marc Smith and Peter Kollock, 220–39. London, UK: Routledge.
Koren, Yehuda, Steffen Rendle, and Robert Bell. 2022. “Advances in Collaborative Filtering.” In Recommender Systems Handbook, edited by Francesco Ricci, Liʾor Roḳaḥ, and Bracha Shapira, Third edition, 91–142. New York, NY: Springer.
Krasnoff, Barbara. 2022. “Mastodon 101: How to Follow (and Unfollow) Other Accounts.” The Verge. https://www.theverge.com/23519279/mastodon-instance-follow-friend.
Kraut, Robert E., Paul Resnick, and Sara Kiesler. 2011. Building Successful Online Communities: Evidence-Based Social Design. Cambridge, Mass: MIT Press.
La Cava, Lucio, Luca Maria Aiello, and Andrea Tagarelli. 2023. “Drivers of Social Influence in the Twitter Migration to Mastodon.” Scientific Reports 13 (1): 21626. https://doi.org/10.1038/s41598-023-48200-7.
Lam, Xuan Nhat, Thuc Vu, Trong Duc Le, and Anh Duc Duong. 2008. “Addressing Cold-Start Problem in Recommendation Systems.” In Proceedings of the 2nd International Conference on Ubiquitous Information Management and Communication, 208–11. ICUIMC ’08. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/1352793.1352837.
Leiner, Barry M., Vinton G. Cerf, David D. Clark, Robert E. Kahn, Leonard Kleinrock, Daniel C. Lynch, Jon Postel, Larry G. Roberts, and Stephen Wolff. 2009. “A Brief History of the Internet.” ACM SIGCOMM Computer Communication Review 39 (5): 22–31. https://doi.org/10.1145/1629607.1629613.
Litan, Robert E., and Alice M. Rivlin. 2001. “Projecting the Economic Impact of the Internet.” The American Economic Review 91 (2): 313–17. https://www.jstor.org/stable/2677780.
Mansoux, Aymeric, and Roel Rocsam Abbing. 2020. “Seven Theses on the Fediverse and the Becoming of Floss.” In The Eternal Network: The Ends and Becomings of Network Culture, edited by Kristoffer Gansing and Inga Luchs. Institute of Network Cultures and transmediale e.V.
Meyer, John W., and Brian Rowan. 1977. “Institutionalized Organizations: Formal Structure as Myth and Ceremony.” American Journal of Sociology 83 (2): 340–63. https://www.jstor.org/stable/2778293.
Najarzadeh, Reza, Farzad Rahimzadeh, and Michael Reed. 2014. “Does the Internet Increase Labor Productivity? Evidence from a Cross-Country Dynamic Panel.” Journal of Policy Modeling 36 (6): 986–93. https://doi.org/10.1016/j.jpolmod.2014.10.003.
Nicholson, Matthew N., Brian C Keegan, and Casey Fiesler. 2023. “Mastodon Rules: Characterizing Formal Rules on Popular Mastodon Instances.” In Companion Publication of the 2023 Conference on Computer Supported Cooperative Work and Social Computing, 86–90. CSCW ’23 Companion. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3584931.3606970.
Nussbaum, Emily. 2010. “How Diaspora Is a Very Different Kind of Social Network.” New York Magazine, September.
Olson, Mancur. 1965. The Logic of Collective Action: Public Goods and the Theory of Groups. Cambridge, MA: Harvard University Press.
Ostrom, Elinor. 1990. Governing the Commons: The Evolution of Institutions for Collective Action. The Political Economy of Institutions and Decisions. Cambridge ; New York: Cambridge University Press.
Partridge, Craig. 2008. “The Technical Development of Internet Email.” IEEE Annals of the History of Computing 30 (2): 3–29. https://doi.org/10.1109/MAHC.2008.32.
Paterek, Arkadiusz. 2007. “Improving Regularized Singular Value Decomposition for Collaborative Filtering.” Proceedings of KDD Cup and Workshop, August.
Pinch, Trevor J., and Wiebe E. Bijker. 1984. “The Social Construction of Facts and Artefacts: Or How the Sociology of Science and the Sociology of Technology Might Benefit Each Other.” Social Studies of Science 14 (3): 399–441. https://doi.org/10.1177/030631284014003004.
Rafaeli, Sheizaf. 1984. “The Electronic Bulletin Board: A Computer-Driven Mass Medium.” Social Science Micro Review 2 (3): 123–36. https://doi.org/10.1177/089443938600200302.
Raman, Aravindh, Sagar Joglekar, Emiliano De Cristofaro, Nishanth Sastry, and Gareth Tyson. 2019. “Challenges in the Decentralised Web: The Mastodon Case.” In Proceedings of the Internet Measurement Conference, 217–29. IMC ’19. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3355369.3355572.
Resnick, Paul, Neophytos Iacovou, Mitesh Suchak, Peter Bergstrom, and John Riedl. 1994. “Grouplens: An Open Architecture for Collaborative Filtering of Netnews.” In Proceedings of the 1994 ACM Conference on Computer Supported Cooperative Work, 175–86. CSCW ’94. New York, NY, USA: ACM. https://doi.org/10.1145/192844.192905.
Ribeiro, Manoel Horta, Homa Hosseinmardi, Robert West, and Duncan J Watts. 2023. “Deplatforming Did Not Decrease Parler Users’ Activity on Fringe Social Media.” PNAS Nexus 2 (3): pgad035. https://doi.org/10.1093/pnasnexus/pgad035.
Ribeiro, Manoel Horta, Shagun Jhaver, Savvas Zannettou, Jeremy Blackburn, Gianluca Stringhini, Emiliano De Cristofaro, and Robert West. 2021. “Do Platform Migrations Compromise Content Moderation? Evidence from r/The_Donald and r/Incels.” Proceedings of the ACM on Human-Computer Interaction 5 (CSCW2): 316:1–24. https://doi.org/10.1145/3476057.
Ricci, Francesco, Liʾor Roḳaḥ, and Bracha Shapira, eds. 2022. Recommender Systems Handbook. Third edition. New York, NY: Springer.
Roberts, Sarah T. 2019. Behind the Screen: Content Moderation in the Shadows of Social Media. New Haven: Yale University Press.
Robertson, Stephen, and Hugo Zaragoza. 2009. “The Probabilistic Relevance Framework: BM25 and Beyond.” Foundations and Trends in Information Retrieval 3 (4): 333–89. https://doi.org/10.1561/1500000019.
Rochko, Eugen. 2016. “Show HN: A New Decentralized Microblogging Platform.”
Rousseau, Amaury. 2017. “Mastodon Instances.” Instances.social. https://instances.social/.
Russo, Giuseppe, Luca Verginer, Manoel Horta Ribeiro, and Giona Casiraghi. 2023. “Spillover of Antisocial Behavior from Fringe Platforms: The Unintended Consequences of Community Banning.” Proceedings of the International AAAI Conference on Web and Social Media 17 (June): 742–53. https://doi.org/10.1609/icwsm.v17i1.22184.
Salton, Gerard, and Michael J. McGill. 1987. Introduction to Modern Information Retrieval. 3. pr. McGraw-Hill International Editions. New York: McGraw-Hill Book Comp.
Sarwar, Badrul, George Karypis, Joseph Konstan, and John Riedl. 2001. “Item-Based Collaborative Filtering Recommendation Algorithms.” In Proceedings of the 10th International Conference on World Wide Web, 285–95. WWW ’01. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/371920.372071.
Schafer, J. Ben, Dan Frankowski, Jon Herlocker, and Shilad Sen. 2007. “Collaborative Filtering Recommender Systems.” In The Adaptive Web: Methods and Strategies of Web Personalization, edited by Peter Brusilovsky, Alfred Kobsa, and Wolfgang Nejdl, 291–324. Lecture Notes in Computer Science. Berlin, Heidelberg: Springer. https://doi.org/10.1007/978-3-540-72079-9_9.
Silberling, Amanda. 2023. “A Beginner’s Guide to Mastodon, the Open Source Twitter Alternative TechCrunch.” TechCrunch. https://techcrunch.com/2023/07/24/what-is-mastodon/.
Zangerle, Eva, and Christine Bauer. 2022. “Evaluating Recommender Systems: Survey and Framework.” ACM Computing Surveys 55 (8): 170:1–38. https://doi.org/10.1145/3556536.
Zhang, Xiaoquan Michael, and Feng Zhu. 2011. “Group Size and Incentives to Contribute: A Natural Experiment at Chinese Wikipedia.” American Economic Review 101 (4): 1601–15. https://doi.org/10.1257/aer.101.4.1601.
Zhu, Ziwei, Yun He, Xing Zhao, Yin Zhang, Jianling Wang, and James Caverlee. 2021. “Popularity-Opportunity Bias in Collaborative Filtering.” In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 85–93. WSDM ’21. New York, NY, USA: Association for Computing Machinery. https://doi.org/10.1145/3437963.3441820.
Zulli, Diana, Miao Liu, and Robert Gehl. 2020. “Rethinking the ‘Social’ in ‘Social Media’: Insights into Topology, Abstraction, and Scale on the Mastodon Social Network.” New Media & Society 22 (7): 1188–1205. https://doi.org/10.1177/1461444820912533.

Footnotes

  1. While email is a decentralized system by design, in practice it has become more centralized over time. For instance, Google’s Gmail service is a dominant provider of email services. Smaller email providers may struggle to get through Gmail’s spam filters, increasing the incentive to simply use Gmail. This means that in practice, Google still has a great amount of data even on people who opt not to use Google’s services ().↩︎

  2. And certainly before my wedding in September!↩︎

  3. Illnesses, injuries, etc.↩︎