1. Introduction
In recent years, there has been an increasing interest in the relationship between social media platforms and civic participation. The scandals which followed the 2016 US elections, along with the impact of algorithmic processes in social media on public opinion during the Brexit referendum, are indicative examples of this tremendous relationship. The exact way in which Facebook in general, and its algorithms, predetermine the news media content displayed to users is uncertain; however, we do know that the content made visible to a user is determined by a series of actions and set by multiple actors: the user, their friends, the designers, the advertisers, and the publishers [
1]. Users ‘like’ and ‘follow’ their friends’ content, publishers and politicians target users with paid content, and algorithms rank and classify the content that is visible on users’ Facebook News Feeds [
2].
Civic participation has long been synonymous to democratic rights, liberties, and claims of citizens, which support political engagement and reinforce democratic representation [
3]. In recent scholarship, many authors have argued that one of the major components of participation—being ‘informed’ and able to acquire political knowledge through and within the media—is highly dependent on the algorithmic functions imposed on users by social media platforms like Facebook [
1]. It is argued that this is due to partial conversion of social media into a source of political media which provides and predicts the political tendencies of users [
4] and aims to attract of the attention of individuals by increasing their personal interest in the content. This is due in part to Facebook’s business model. As Thorson et al. argue, the objective of Facebook’s business model “is to create attention that can be sold to advertisers”, and not specifically to create political content [
4] (p. 12). In the same vein, Bode [
5] for example has demonstrated that online political content is associated with, and provided to, users with similar ideological orientations—which raises questions about those who interact with such content without actively seeking it, and are thus exposed to content which does not correspond to their ideological beliefs (this is known as ‘incidental exposure’) [
6]. The results of studies on the relation between social media and civic participation have been generally mixed, finding little or no effect of social media on civic participation [
7]. Other more optimistic approaches, such as that of Boulianne [
8], argue for the primacy of positive rather than negative effects, claiming that “the metadata demonstrate a positive relationship between social media and participation” [
9] (p. 524). Furthermore, other empirical studies have also demonstrated that exposure to political content on Facebook may in fact constitute an interpretive tool for the elucidation of offline political participation [
10]. Likewise, Kümpel [
11] has argued that accidental exposure to news content by friends may be positively correlated with information gain and can strengthen offline political participation.
As regards of Facebook, scholars seem to agree that the platform possesses embedded features which facilitate users’ exposure to different types of news necessary to acquire political information around an issue [
7]. This view has also been confirmed by previous studies which have illustrated how citizens’ consumption (whether intentional or unintentional) of political content tends to increase their political capital [
12,
13] and strengthen their political activity offline. This is related to the concepts of ‘selective exposure’ and ‘incidental exposure’, both of which play a central role in civic participation. Selective exposure emphasizes the role of individual choice in predicting political content exposure, while incidental exposure focuses on the role of friends and ‘unintentional’ encounters with content. Both concepts, however, tend to be insufficient to predict the way that algorithmic curation works [
14,
15].
Individuals’ pre-existing attitudes are a significant variable which must be determined when studying empirically the influence of social media on civic attitudes. Political efficacy can be considered as a variable when examining pre-existing understandings of politics and subjects’ real-life political participation. Real-life participation includes political activities such as active political campaigning, attendance of rallies and protests, and voting in elections. Online activities include sharing posts, liking pictures, and commenting on videos on digital platforms [
16]. Pre-existing attitudes can in fact be strengthened and reinforced through social media [
17,
18] rather than changed. Experimental studies have indicated that individuals’ attitudes can be affected by the exposure or discussion of a political matter on social media, depending on the individuals’ level of engagement, their sociopolitical background, and the degree of politicization of the media message. The degree to which users are affected may depend on their level of awareness, their degree of partisanship, and the directionality of media politicization. Users’ civic attitudes may encourage them to prefer consistent information and to defend any motivation to amplify this preference [
16,
19]. In addition, a significant strand of research has confirmed that social media offers a way to reach users with political information, and that this potentially allows these users to ‘catch up’ in terms of civic knowledge and civic interest, and may facilitate political dialogue [
20]. It is evident that within datafied environments where political content or information is arranged by algorithmic curation, the prediction of media exposure is relatively complex [
21].
However, media exposure is also associated with users’ civic interest in a specific form of content. Several empirical studies examining the impact of Facebook News Feed on civic participation have reported that civic interest can be a valuable source for the prediction of political use of the internet [
22], providing evidence for the interrelation of personal interest and attention on social media. While political interests may correlate with users’ specific interests and preferences in political news choices—thus being easier to predict—a question remains to be answered: what happens when users see content that they did not seek out and which may not correspond to their interests [
23]? Civic information is often curated by algorithms which select and display content to users and attempt to form attention patterns [
1]. While users may employ certain filtering options to pre-select the kind of content they wish to be exposed to, the efficacy of these filters remains debatable. The realm of social media can therefore be conceptualized as a space in which the mediation of political interests occurs regularly, thus rendering such space a means, rather than a cause, of social and political action [
24].
This theoretical discussion posits that the relationship between Facebook use and civic participation may be mediated by the algorithmic curation process, which largely determines the selection of news and political content shown on users’ Facebook News Feeds. Therefore, to understand the communicative process hidden within the platform, we need to move “constantly between the technical and the social […] the inside and the outside of technical objects” [
25]. The objective of this article is thus twofold: firstly, to determine the extent to which Facebook News Feed results related to news media articles are customized to reinforce the existing civic attitudes of users; and secondly, to identify the roles which users come to embody, as these relate to the selection of and exposure to different types of political content. These considerations and emerging questions invite further research in this field, and an investigation of whether users’ previous political interests, preferences and beliefs defined as civic attitudes are encouraged by the Facebook News Feed (thus influencing the classification of news and political content online).
4. Results
Τhe highest participation rate was among those aged between 18 and 30 years old (
n = 50.9%), while 57.8% of the sample were male, and 42.6% were female. As shown in
Table 1, most of the sample considered themselves to be ‘very interested’ in offline political life (
n = 30.6%), while the rest of the sample varied slightly between ‘moderately interested’ and ‘extremely interested’ in politics.
Although a high percentage of participants were interested in politics, as
Table 2 shows the vast majority of users did not belong to a political party (
n = 75%).
There seemed to be much more active participation in online rather than offline engagement in civic activities (see
Table 3). In the last twelve months, most of the participants had sent a political message via Facebook (
n = 71.4%), and most had signed up as volunteers for a campaign or a political cause (
n = 54.3%). Minorities of the sample had contributed to a campaign (
n = 31.4%), written to a politician (
n = 28.6%), subscribed to a political list service (
n = 24.3%) or written messages to the editor of a newspaper using the official Facebook page of the newspaper (
n = 22.9%). This finding demonstrates that the participants are actively engaging in political activities within the social media.
Continuing the analysis of participants’ online civic attitudes and political interests, 71% of the respondents answered that they ‘always’ came across political content on Facebook, whereas 56% encountered such content ‘often’. A total 71% of those who ‘always’ encountered political content on Facebook also sent political messages, while half of these individuals had also signed up as volunteers for a political campaign or cause. In addition, about one in three participants in that group had made a campaign contribution (
Table 3 and
Table 4). This finding evokes that our participants are mostly engaged in low-cost political engagement activities online.
A significant positive correlation was found between all the actions (
Table 5). The strongest correlation, however, was found between responses to the questions
“How often do you create a political content on Facebook?” and
“How often do you share political content on Facebook?” (r(108) = 0.82,
p < 0.01), followed by responses to the questions
“How often do you read political content on Facebook?” and
“How often did you see content on Facebook news media pages related to politics or political issues during the last week?” (r(108) = 0.72,
p < 0.01). This finding indicates that the most active and engaged citizens are often the one who create, share, see and read political content on Facebook. This demonstrates that social media contribute to high-level engagement in various political activities online.
A Chi-square test for association between responses to the questions “Do you have particular Facebook news media pages from which you get informed?” and “Did you join, like, or follow the Facebook media pages of your preferable news media sources?” was conducted (
Table 6). All expected cell frequencies were greater than fine. A statistically significant association was found between responses to these two questions (x2(1) = 17.726,
p < 0.001). Another Chi-square test was conducted to test for association between responses to the questions “Which of the following have you done in the last year?” and the “Did you join, like, or follow the Facebook media pages of your preferable news media sources?”. All expected cell frequencies were greater than fine.
There were statistically significant associations between responses to the questions:
“Added, followed or became friends with a user or organization because of news items they had posted or shared”, x2(1) = 13.325, p < 0.001.
“Deleted or blocked another user or organization because of news they had posted or shared”, x2(1) = 6.130, p = 0.013.
“Changed my settings so that I would see more news from a user or organization”, x2(1) = 5.683, p = 0.017.
An additional Chi-square analysis was conducted to test for association between responses to the questions “Which of the following have you done in the last year?” and “Did you find yourself trying to avoid news stories from a particular Facebook source?” (
Table 7). All expected cell frequencies were greater than fine, proving that there was significant correlation among the elements. There was a statistically significant association for “Deleted or blocked another user or organization because of news they had posted or shared” (x2(1) = 7.691,
p = 0.006), see
Table 8.
The same analysis procedure (Chi-square) was used to test for association between responses to the questions “Did you find yourself trying to avoid news stories from a particular Facebook source?” and “ Have you been exposed to an oppositional ideological form of content through news media articles?” (see
Table 9). All expected cell frequencies were greater than fine. There was a statistically significant association between the two actions: x2(1) = 3.864,
p = 0.049. This finding indicates that participants who answered positive related to their exposure with oppositional ideological form of content through news stories, they find themselves trying to avoid news stories from a particular Facebook source. This evokes a tendency towards what has been called ‘filter bubble’ effect, where users prefer to read, share, and comment news stories which are similar to their own political beliefs. It is still uncertain whether algorithmic curation can effectively amplify individual preferences or enclose users in their own filter bubble. Users’ interests, preferences and habits form the foundation of the platform’s optimal algorithmic categorisation, which in turn promotes content deemed likely to attract users
A Chi-square test for association between responses to the questions “Did you find yourself trying to avoid news stories from a particular Facebook source?” and “Have you ever been surprised by someone’s views on politics or a political issue, based on something they posted on social media?” was also conducted. All expected cell frequencies were greater than fine. There was a statistically significant association between the two reactions, x2(1) = 3.984,
p = 0.046 (
Table 10).
Only four respondents responded, ‘always’ to the question “
How often do you share political content on Facebook?”. Of the 20 respondents that ‘often’ shared political content, 40% felt ‘inspired’, 35% felt ‘anger’, 30% felt ‘sadness’, and 30% felt ‘disgusted’. Similar percentages appeared in the group that ‘occasionally’ shared content. On the other hand, the vast majority (67%) of those that never shared political content replied that they ‘didn’t care’ that they did not do so (
Table 11).
5. Discussion
Our analysis indicates that Facebook’s News Feed is a novel and influential source of news content for users and plays a central role in conveying and controlling the flow of information in different formats [
29] within the platform. It thus constitutes a key source of news information and encourages people to create and receive news stories from media sources and friends among others [
27]. Facebook’s News Feed, of course, operates by means of algorithmic processes which are comprised of systems of criteria used to sort the pollical content to be included from that which is, so to speak, undesirable or not corresponding to the preferences of the user. As such, algorithmic processes user’s experience within the platform in a unique way and according to many parameters; some of them are entirely technical and some others are depended to user’s online behaviour [
27]. As Thorson et al. [
26] point out, “Facebook uses digital trace data about each user to infer their interests in order to (a) aid the newsfeed ranking algorithm in deciding which stories will be most ‘meaningful’ and ‘relevant to that user, and (b) package that user to be targeted by advertisers”.
Other studies have identified a set of nine News Feed values: friend relationships, explicitly expressed user interests, prior user engagement, implicitly expressed user preferences, postage, platform priorities, page relationships, negatively expressed intentions, and content quality [
30]. A variety of different News Feed values are used in everyday story selection, with the most popular being “novelty or oddity, conflict or controversy, interest, importance, impact or consequence, sensationalism, timeliness, and proximity” [
31]. News Feed values, which are algorithmically driven, lack a deep understanding of information due to technological limitations. Such limitations—known as
technical bias—include technical issues in databases, storage (un)availability, processing power, and potential coding errors [
32]. Algorithmic curation is mostly informed by individual users choices (habits, preferences and interests) but also from the arbitrary internal process of the technological artefact as well. This latter form of determination is often referred to as ‘pre-existing bias’ [
33], and its effect on the formation of News Feed values may or may not be taking place in a conscious manner [
34]; in either case, it can affect the basis of algorithmic curation. For instance, if Facebook News Feed is going to arrange posts from close friends of the user, engineers must decide on the criteria that determine what a ‘close friendship’ is, as opposed to a mere contact or association. Engineers must adapt this approach into a single, operational interpretation, and embed in the design of the algorithmic procedure for determining values around the definition of what a ‘close friend’ is. These value-based decisions are a prerequisite for Facebook as a business medium, because advertisers need access to users’ input, to determine the connections and pre-existing preferences of friend groups to serve the advertisers’ promotional interests [
35,
36,
37].
Facebook designers and engineers need to decide on the degree of relevance and choose the variables to use for evaluation [
38] as well as the data sources the algorithm will draw from [
32]. Furthermore, engineers can prioritize popular values based on individuals’ preferences and their previous interactions (feedback loops and filter bubbles) [
39]. Nicholas Diakopoulos [
40] identifies the following as the major functions of these algorithms:
prioritization,
classification,
association, and
filtering. Value decisions in these areas are evident in Facebook’s technical documentation for the News Feed [
39,
41]. Research has revealed a two-element system: (1) featuring
objects (content), and (2) featuring
edges (relational interactions—tags, comments, etc.). Edges include three components: association between the perceiver and creator, the type of edge (with comments being more important than likes), and ‘time rotting’. Creator and viewer relationships are based on a pathway of interaction through wall posts, likes, private messages, etc.
In general, News Feed story selection and algorithmic decision-making procedures are an ongoing process in which Facebook engineers and users interact, and where new improvements from users’ feedback and market changes drive an increasingly effective personalization of individuals’ Facebook News Feeds [
42]. Such personalization is bound up with the significant correlation between online and offline political and civic participation. In our results, this is more pronounced in the 18–30 age group (the age group most strongly represented by our sample): while most individuals in this group stated they were ‘very interested’ in politics, they nevertheless reported no involvement in any political parties. This should come as no surprise, since mistrust of institutions and generalized suspicion towards mainstream politics is quite common among young people. This partial disengagement from formal political structures and procedures should not be mistaken for political apathy. On the contrary, it is evident (see
Table 3) that online civic participation is quite prevalent among the age group in question due to its flexible low-cost and effortless engagement. At the same time, the question arises whether these users abstain from conventional politics because they genuinely believe that online civic participation is more effective, or because such online action is easier and less demanding which provides a convenient yet inferior alternative to offline political activities.
Furthermore, a significant proportion of the participants reported seeing political content on their news feeds on a regular basis, which indicates that they must have sought it out, and perhaps still do by following pages with content of this nature. This indicates that although our participants do not engage, or they do not belong to any political party are still interested in politics. This observation, of course, coincides with the empirical findings of Karnowski et al. [
43], who demonstrated that individuals who are already interested in political matters are indeed more likely to come across political content in their Facebook News Feed because of incidental exposure. This finding coincides with previous studies arguing that citizens who, “unintentionally or not”, consume politically oriented content can gain in political capital [
44] and thus potentially mobilize offline particularly in collective action. This is often related to what scholars call
selective exposure and
incidental exposure, which can intervene in the making of civic participation. While this is true, this raises the central issue of the ‘filter bubble’ phenomenon, whereby actual exposure (selective or incidental) to a diversity of political news posts and views is hampered by feedback loops created by users’ data traces on the one hand, and algorithmic curation on the other. In other words, social media users are often trapped in “echo chambers” and “filter bubbles”, which expose them only to certain views and opinions, usually ones that they agree with in the first place. Yet, algorithmic curation remains mostly invisible as ‘black boxed’ to the users [
45]. Our findings support this idea (see
Table 6), as statistically significant associations were established between responses to the questions: “
Added, followed or became friends with a user or organization because of news items they had posted or shared”, “
Deleted or blocked another user or organization because of news they had posted or shared” and “
Changed my settings so that I would see more news from a user or organization. All the questions are referring to a rather selective exposure (preferable sources) from the side of users (in terms of page news organization and friends) to news posts and views that are supporting the effect of filter bubble.
The majority of those who constantly viewed political content on their news feed also engaged in such actions as sending political messages and signing up to volunteer for various political causes. A third of this population went so far as to make political contributions. This goes in line with Ekstrom and Shehata [
46] findings where social interaction in social media coincides mostly with engagement in political information and interaction seen as rather an unconventional form of participation especially for young citizens. The fact that citizens are more frequently engaged within social media by sending a political message (71.4%) or by making an online campaign contribution is not surprising. These types of online activities are less demanding and more open in terms of political affiliations. The questions that naturally arise from such observations are concerned with the nature of the political content being viewed, and potential fluctuations (or otherwise) in offline political activity. Thus, a strong correlation between the responses to
“Did you join, like, or follow the Facebook media pages of your preferable news media sources?” and
“Which of the following have you done in the last year?”, perhaps provides an indication of the validity of our first hypothesis. Users reported having added or followed users or pages due to their news posts (stories); conversely, they also reported having deleted or blocked users or pages for precisely the same reason. Furthermore, a correlation was established between responses to the questions
“Did you find yourself trying to avoid news stories from a particular Facebook source?” and
“Deleted or blocked another user or organization because of news they had posted or shared”.
Our results complicate our understanding of the phenomena under investigation, due to the correlation observed between responses to the questions “Did you find yourself trying to avoid news stories from a particular Facebook source?” and “Have you been exposed to an oppositional ideological form of content regarding the news media articles?” When encountering pages that share views contrary to their own, a large proportion of users will most likely avoid these pages/or users altogether instead of critically engaging with said views. Furthermore, many users will go so far as to delete or block pages and individuals because of their sharing of certain news stories. Such observations seem to confirm both the presence of the filter bubble phenomenon and as our first Hypothesis (H1) regarding customization and the reinforcement of pre-existing views and convictions. That being the case, we argue that the customization processes initiated and established by the platform are often reinforced by users themselves, who seem to enjoy the cognitive consonance that personalization provides. In these instances, users are not exclusively the passive objects of algorithmic curation: they may also be active yet conservative subjects who are, for whatever reason, unwilling to engage with opposing views. Therefore, our second Hypothesis (H2) also seems to be validated by our findings. The proactive stance users take towards oppositional pages and views presupposes that they assume a specific role, in which they ‘censor indirectly’, so to speak, what is undesirable. Such censoring, however, is not necessarily conservative per se, since it may very well be applied for purposes related to the obfuscation and/or subversion of algorithmic curation.
The last section of the survey results concerned the influence of political content on users’ emotional interactions, and participants were asked to specify their feelings towards the creation and/or sharing of political content on Facebook. The participants’ responses indicated that they were affected emotionally, albeit in a negative way: their affective states ranged from neutral to negative. This indicates that the sharing or creation of online political content by users aims to express their subjective, opposing viewpoint by filtering and criticizing the content they interact with, which is often contrary to their actual political beliefs. Overall, the often-neutral stance of users indicates that users are indifferent to political content that is not politically consistent with their personal political beliefs. Individuals’ emotional tendencies range from neutral to negative, depending on the political content in question and the current political context [
44]. Therefore, the creation or sharing of political content satisfies a need for the expression of contrary viewpoints or opinion.