This document summarizes and expands discussions from mid-2018 within the W3C Credible Web Community Group. It contains ideas and other content contributed by group members, often in discussion with the wider community.
Comments are welcome and are especially useful if they offer specific improvements which can be incorporated into future versions. Please comment either by raising a github issue or making inline comments on the Google Docs version of this document. If neither of those options work for you, please email your comments to public-credibility-comments@w3.org (archive, subscribe).
Can you tell, when looking at a random web page, whether you should trust it? When scanning a page of reviews or search results, do you know which matches come from legitimate sources and which are scams? When reading a news feed, can you tell which items ought to be believed and which are slanted or manipulative? Can you detect propaganda or outright lies? Perhaps most importantly, what happens when you inevitably guess wrong while making some of these credibility assessments, and you unknowingly share falsehoods with your community, helping to make them viral? What if you are misled into making bad decisions for yourself and the people you care about, with potentially disastrous consequences?
The Credible Web Community Group was formed at W3C, the organization which develops technical standards for the web, to look for technological approaches to this “credibility assessment” problem. It's not that we think technology can solve every problem, especially ones as deeply human and complex as this one, but it seems likely that some technology is making matters worse and that certain designs could probably serve people better. For some of us, creating better approaches to credibility assessment seems like a good way to help.
This report, like the Community Group which developed it, is focused on web-centric technical solutions which require standardization.
The web isn't everything, but it's arguably the knowledge backbone of the world today. Everything can and usually does tie back to the web to some degree, so if the web is a reliable source of accurate information, that should help quite a bit.
With standardization, we can help people (via their computers) work together toward improving credibility assessment and the availability of trustworthy information. As detailed below in Potential New Web Standards, the area we find most suitable for standardization is web data interchange, where websites publish data for other systems to consume. W3C standards already cover the basic mechanics for this, and they are widely implemented, most popularly with vocabularies defined by schema.org (a collaboration among search engine companies). What remains is to standardize vocabularies (also known as schemas or ontologies) for exchanging data which bears directly on credibility assessment, that is, “credibility data.”
There may also be a benefit in standardizing new browser functionality. For example, browsers might manage a collection of independent credibility assessment tools which work together to guide and inform the user. This kind of new feature can be pioneered in browser extensions and then later adopted into browsers to increase the user base.
While this report is forward-looking, some of the technologies discussed are already deployed to some degree in existing systems, particular where they require no user action, like in prioritizing search results and news feed items. The hope here is to make such systems more accurate and effective by providing them with better data from around the web in a standard format.
Changing the technology of the web to empower users to be more accurately informed, while challenging, sounds laudable and even overdue. But there are likely to be unintended consequences resulting from these changes. To use a medical analogy, treatments used to help address this painful condition are likely to have side effects. That doesn’t mean we can afford to do nothing, which itself presents significant risk, but we do need to proceed with caution.
This section lists types of side effects, with some ideas about how to mitigate the risks. A general approach to managing these is discussed later, as a recommendation, in Review Board Process.
In general, this should be understood within the context of W3C’s mission of a Web for All, a Web for Rich Interaction, and a Web of Trust. There may need to be trade-offs among benefits of the web; if so, let’s make them with our eyes open.
The regulation of content, called “censorship” in some contexts, is controversial. Nearly every country in the world considers publication of certain kinds of material (such as child pornography) immoral and illegal. At the same time, many of the same countries express their belief in the value of free speech and a free press, sometimes in very broad language. In practice, there is rarely a sharp line between content that is prohibited and content that is accepted. Additionally, content that falls on one side of the line in one jurisdiction at one point in time might be on the other side of the line in a different jurisdiction or at a different time. This is a challenge because the web reaches across time and jurisdiction.
Historically, and perhaps largely by accident of design, the web has managed this controversy by making it relatively easy for anyone to publish what they want but also making it hard to run a website while remaining anonymous to law enforcement. Given the current technology, much of the world’s population could easily set up a website and host illegal content. But the same technology makes it fairly easy to track such a person down and, subject to applicable laws, take down the site and/or identify the site’s owner for criminal prosecution. There are technologies which allow sites to be more anonymous, but they are vastly more difficult to use, making them prohibitive for most situations. (This web technology compromise should not be confused with content regulation decisions made within a particular platform, which is a separate issue.)
While some people are uncomfortable with this technology-driven compromise, especially when considering content they would place on the other side of the line from where it currently lies, we suggest that extreme care be taken before sweeping this solution aside for something that might be worse.
Some proposed credibility technologies could upset the current balance. For example, one common suggestion is to allow publishers to be certified as entities which adhere to certain standards. Platforms would then be free to use this certification to provide better services. They might also be pressured to only allow certified content, perhaps through liability concerns. Approaches like this could significantly raise the bar on publication, fundamentally changing the balance in the system. It should not be done lightly.
It may make sense to raise the bar in certain limited ways. For example, in professions and industries which are already regulated, like medicine, making existing certifications available to the computing infrastructure could increase the quality of information with relatively limited side effects. Even then, some “escape hatch” mechanism would likely be important, where people who fail to meet that test can still get word out (e.g., as a whistle-blower), perhaps via a trusted third-party or an opt-in system for bypassing the safety mechanism.
Another possible area of side effects is in shifting the web toward having more centralized control. For example, Elon Musk’s “Pravda” proposal to “create a site where the public can rate the core truth of any article & track the credibility score over time of each journalist, editor & publication” may (or may not) have been tongue-in-cheek, but it received massive coverage and support. Such a system would have strong network effects, likely resulting in monopoly power in its niche if it caught on. As such, despite whatever value it might bring to its users in the short term, it would stifle competition and free innovation. It would also introduce critical issues with scaling, censorship, and political bias.
In contrast, the architecture of the web is largely decentralized, minimizing bottlenecks and central points of control. With this in mind, and with the will to do so, something like “Pravda” could potentially be created as a decentralized system, using a variety of techniques including the W3C ActivityPub protocol.
More modestly, and as a general rule, we suggest that whatever systems are implemented to help with credibility, users always retain control over their data and computing platforms. This aligns with our recommended approach of standardizing vocabularies for credibility data interchange, removing the technological barriers to moving data between systems.
We recognize that some degree of centralization, while it has risks, can have tremendous value and may turn out to be worthwhile. For example, the increased centralization of email has likely helped with reducing the problem of spam. Still, it is vital that independent email systems continue to exist and they all interoperate.
Some proposals are likely to increase the degree to which consumers see only the content they want to see or their community wants them to see, regardless of its accuracy. While many view this as a bad thing, experts disagree.
One theory on addressing this is to make clear to users when they are inside their bubble (seeing content algorithmically selected to fit them personally) and to allow them to easily step outside of it, whenever they want. The theory is that people won’t always choose to be sheltered, but will instead sometimes look for adventure, or something, “outside the wall.” One early tool for “stepping outside” is Blue Feed, Red Feed.
It has also been suggested that prediction markets incentivize accurate broad knowledge, leading people out of their bubbles, and that commerce/trade in general motivates people to connect across different ingroups and helps people communicate productively and build trust across group identity barriers.
Some credibility solutions may impact privacy. For example, a browser plug-in which checks every page the user visits using its proprietary cloud-based service could potentially misuse or leak information. If designed without proper safety measures, it could easily leak private URLs the user visits, and in some cases could even leak secret page contents, such as the user’s medical communications or their employer’s proprietary data. Such systems are strongly encouraged to use independent and public review to build in strong privacy protection, instead of any kind of “just trust us” approach.
It can be hard to predict all the ways a technological intervention can go wrong. For example, Facebook reported that adding warnings to newsfeed items indicating when an item had been debunked by fact-checkers resulted in more people reading and spreading false news. Predicting, detecting, and if necessary changing course away from such effects can be even harder in the standardized, multi-vendor, decentralized systems we otherwise prefer.
For more discussion on approaching and mitigating these risks, see Review Board Process.
In this document we have attempted to use terms consistently and precisely, hoping to help promote wider consensus on terminology for the field. We are being both descriptive and a bit prescriptive here, although we recognize that in the future better terms may emerge and supplant these.
In general, the web is an information ecosystem. When thinking about credibility on the web, it helps to be somewhat careful with our terms. Note that as information flows around the web, a single party may play multiple roles.
(Content) Consumer. Person who is receiving and experiencing some content. Similar to: Audience Member, Reader, Viewer, Listener, Receiver, or User.
(Content) Provider or Source. Person or organization who provided the consumer with some content. There may be a supply chain of providers, creating and assembling content before it reaches the consumer. Alternatively, sometimes a chain of friends acts as a provider, passing the content on to each other. Often this is invisible to the consumer, who perceives (and makes credibility assessments about) a single apparent provider. Similar to: Producer, Creator, Author, or Publisher. “Source” is often more ambiguous, but can be used as an alternative when “provider” gets awkward.
(Content) Promoter. Person or organization who intentionally or unintentionally increases the spread of content. In social media platforms, this can be as simple as “liking.” Commenting on content or linking to it, even to refute it, can increase its spread and visibility due to various algorithms.
(Credibility) Facilitator. Person or organization who is helping the consumer decide what to trust. Similar to: Moderator, Fact-Checker, Forum, or Comments Editor, but also includes members of the crowd in crowd-source designs.
(Web-Based Communication) Platform. Technological system, and by extension the person or organization who maintains and controls it, which is providing the above parties with the infrastructure that enables the content to pass among them. Often controls which items are allowed and ranks items to influence which are seen. Includes content management systems (CMSes), web browsers, search engines, news feeds (as in social networks and microblogging services), and media sharing platforms.
The terminology around credibility and trust varies somewhat across different fields of study and non-technical discussion. Here we propose some terms and an underlying model intended to be precise and formalizable enough for technical and scientific work, while natural enough to make sense in general use. In some cases, motivation, explanation, and advice is included as well.
Claim. An assertion that is open to disagreement; equivalently, a meaningful declarative sentence which is logically either true or false (to some degree); equivalently, a proposition in propositional logic.
Fact. A claim that is true; equivalently, a claim that accurately describes the actual state of the world.
Accuracy. Degree of truth, especially for claims involving measurements. Can be a useful concept to avoid pedantic distinctions around truth. The claim, “The radius of the earth is 6400 km,” is only somewhat true, as 6371 km is a more accurate figure. Of course, it is not exactly 6371 km, either. Similarly, most “true” statements are not perfectly and completely true, so considering their accuracy may be more helpful.
Believe (in a claim, as in: “Alice believes the world is round”). To be in the mental state of accepting (and consequently behaving as if) the claim is true, at least within some limited context. Evidence suggests people can sometimes believe contradictory claims and switch among them depending on the context. Trusting a claim is essentially synonymous, although perhaps a bit stronger, like “firmly believing.” One approach to measuring strength of belief is asking how much someone would bet on the claim being true. At the same time, however, people may not be consciously aware of how much they routinely trust people and engineered systems in return for minor gain.
The following definitions are expressed in terms of “some information.” This idea can be applied narrowly, to a small bit of text, or very broadly, to all the information provided over time by some content provider. For example, at small scale, we can consider the credibility of a particular sentence, or, at large scale, the credibility of a particular news organization in a given month. See additional discussion below in Granularity.
Mislead (some people, by some information). To cause people to believe a false claim, due to consuming some information. It would require omniscience to know with certainty if someone has been misled; in reality we can only see how beliefs change and differ, and assess how useful they are, especially in making predictions. Nonetheless, the realization that one has been misled can be quite clear, and we expect the desire to avoid being misled to motivate user adoption of credibility technologies. For simplicity, we include as “misleading” all degrees of severity, including actions that might count as deception, lying, or fraud.
Nonmisleading (of some information, to some consumers). Does not mislead a significant portion of its consumers. Similar to “accurate” and “true” but includes information other than claims and excludes information which is so confusing that people misunderstand it. That is, “nonmisleading” differs from “true” in that poorly-expressed true information can still mislead, and things that are not true, like jokes or fiction, can still be nonmisleading, as long as they don't actually mislead their consumers.
Disinformation (disinfo). Information which intentionally misleads; information which deceives its consumer.
Misinformation (misinfo). Either: (1) misleading information; or (2) unintentionally misleading information. Sense 2 was promoted by the influential Fake News, It's Complicated and CoE Report, which defined the set of misinfo to be disjoint from the set of disinfo, instead of the superset implied by sense 1. Because of this ambiguity, we avoid using the term in this document, preferring either “misleading information” (for sense 1) or “unintentionally misleading information” (for sense 2).
Malinformation (malinfo). Information which is harmful without being misleading, such as leaked private information or upsetting images. This is out of scope for this report.
Believable (of information, for a given set of consumers). Appearing acceptable to believe, such as because it does not contradict any currently held beliefs; appearing nonmisleading. (Without omniscience, one cannot tell with certainty that claims are nonmisleading; rather one typically compares them to one's current beliefs.)
Credible (of information, for a given set of consumers). Believable, but with a connotation of additional confidence and motivation. Saying a statement is “believable” is saying it's possible, even reasonable, to believe it; saying it's “credible” is saying it should be believed, at least tentatively.
Credibility (of information, for a given set of consumers). Degree to which information is credible; degree to which information appears nonmisleading and useful (for the given audience). People are typically misled when falsehoods have high credibility (appear true), and they problematically resist believing facts which have low credibility (appear untrue). The term credibility can also be used broadly to refer to the problem space around trust, as in “credibility researchers” or “credibility software.”
Trustworthy and Trustworthiness can be synonyms of credible and credibility, respectively, but there seems less consensus on their meaning. For example, a counterfeit that fools everyone is misleading, and if it fools everyone we know it appears nonmisleading, so it is “credible.” People seem to disagree about whether such a counterfeit would be considered “trustworthy.”
Credibility Score (of some information, in some information environment). A numeric value for credibility, on some defined scale. At present, there are no standard scales, so these scores cannot be meaningfully compared. For example, a standard scale might be: the statistical probability that the given information will not measurably mislead each member of the population who consumes it in the given environment. More work is needed to understand how credibility scores might be meaningfully interchanged and compared, and thus what kind of standard scale(s) might be useful.
Someone claiming that information is credible or has high credibility is asserting that the information appears sufficiently unlikely to mislead that they think it should be believed. Similarly, saying information is not credible or that it has low credibility is saying that it should not be believed because it appears too likely to mislead, in the speaker’s estimation.
We distinguish two kinds of credibility assessment acts:
Credibility assessment tools are software features or applications which perform credibility assessment or help people do so. That is, they are tools to help people make better credibility decisions.
Credibility signal. A small unit of information used in making a credibility assessment. This can be a measurable feature of the information being assessed for credibility, or information about it (metadata), or information about entities which relate to it in various ways, such as the entity who provided it.
Credibility indicator. Commonly, same as credibility signal. In some communities, “signal” is used for inputs to credibility assessment algorithms and “indicator” is used for the display features added to the output, to communicate results to a human consumer. This distinction is not used reliably, however, so use of terms like “input” and “output” is suggested when one needs to make the distinction. See CredWeb Issue 3.
Techniques for assessing credibility can be grouped into the following four types of strategies. Each of these strategies could be applied in a range of specific ways, and some tools will likely combine elements of several strategies.
People necessarily make credibility assessments about content items and about the people who provided the content. It seems likely they make these assessments across a wide range of granularities, all of which need to be taken into account in understanding and assisting with credibility.
Content providers can each be assessed for credibility; that is, one can attempt to determine the likelihood the information they provide is nonmisleading. This assessment might be done across a wide range from the general (coarse-grained, wide scope) to the very specific (fine-grained, narrow scope):
Reasoning about the credibility of content similarly requires working with a range of granularities. For example, we might consider the credibility of a complete snapshot of Wikipedia, of a particular Wikipedia page, of the text of a paragraph, or of a specific claim.
Content which does not make explicit assertions, including jokes and photographs, can also be seen as implying a variety of claims, although that implication relationship may itself be subject to uncertainty.
Sites which aggregate content from other providers, including social media sites, will need special consideration, as will sites which analyze or review other content, sometimes including false content as an example or evidence. That is, a site should be able to present misleading information for analysis without risking its own credibility due to tools simply seeing it hosting misleading information.
In general, in designing systems which inspect and annotate web content, it often works to initially assume each interesting chunk of content will be on its own web page. Then, one can expand the granularity up to web origins (loosely, domain names) and shrink it down to page fragments (loosely, identified portions of a page or portions of media shown on the page). Note that the Web Annotation framework specifies a way to make essentially any portion of any version of a page be its own addressable fragment, although that is not yet widely implemented.
Credibility ratings for a web page may at first seem to be just about whether the content is nonmisleading, but in practice websites can change their content or even secretly show different content to different observers, so the trustworthiness of the site maintainer can be a factor as well. To sidestep this added credibility complexity, protocols for securely referencing site content versions and noticing changes are likely to be necessary. Fortunately, secure hashes and third-party archives, already standardized and implemented in memento, seem well-suited to help. In addition to securing against some attack vectors like this, such a secure history mechanism can provide useful historical evidence while assessing credibility.
In avoiding harm, including all the different kinds of done by misleading information, it can help to understand the people and systems involved. This section is a brief survey of some of the aspects of misleading information on the web. People developing credibility tools are strongly encouraged to dig much more deeply, within their areas of focus, to reduce their chances of building the wrong defenses.
Providers of misleading information have many motivations, and often their motivations will never be known. Still, it can help to consider some of these types of attackers and their motives:
There are, of course, many other kinds of difficult people online.
For a survey on types of bad information and some of the motives behind them, see Fake News, It's Complicated.
How much effort is an attacker willing and able to devote to a disinformation attack? Here are some options, roughly in order of increasing threat. The higher resource levels are often not worth designing against and may not even be ethical or legal to design against.
There are numerous ways to scam people online and a vast array of weaknesses in computer security systems which can help. A few are particularly relevant in credibility assessment:
Also see landscape spreadsheet.
[ Ideally for each entry we should have some description of how it relates and any active liaison. ]
There are dozens if not hundreds of research groups studying aspects of credibility around the world. Listed here are groups that have expressed interest in helping with technical standards work around credibility. Please add your group if this applies.
Specific vendors/products are named as an example when an employee or representative has joined the group.
Many of the proposed solutions have strong network effects. They offer users minimal value until they have a critical mass of users. This makes it quite challenging to achieve adoption.
This kind of challenge is usually addressed by achieving critical mass in a small community and then expanding outward. (Facebook famously started at Harvard and then gradually expanded to other universities before eventually opening to the public.) Within the target small community, adoption can be promoted or even required by the people managing the community.
Some specific approaches:
In general, funding is an issue. There are not yet any clear business models for credibility systems. Philanthropic funding may be necessary, at least for now.
The user experience (UX) of credibility assessment tools is likely to vary widely, given the range in granularity, subject matter, and computing environments where credibility matters and where software can help. Some of the ways users might experience credibility assessment tools:
We can organize the discussion of possible tool designs using a spectrum of user effort, from no effort at all, where the user benefits without doing anything, up to major effort, where a user might be spending hours or even days trying to determine whether to trust a single claim. The appropriate tools are quite different at different points along this spectrum. We will use this concept to organize the tool discussions below, but here we consider it in three ranges.
Some of these could impinge on the user's autonomy or confuse the user, but done properly it may be possible to “nudge” the user while fully respecting their wishes. Additionally, these concerns can be managed by allowing users to opt-in and configure settings for these features.
Within this level, we could zoom in, expanding these approaches that require no action from the user, perhaps conceptualizing effort-to-ignore as “intrusiveness”. It is unclear whether to consider blocking/demoting content as highly intrusive (it would take a lot of effort to get around) or completely non-intrusive (since the user won’t usually notice it).
With some of these options there is a risk triggering a backfire effect, as users may cling to false beliefs more tightly when corrected, especially if the correction is not highly credible to that user.
These only help when the user consciously wonders whether to trust the content.
Most of these are perhaps best done at the claim level, where the user indicates a specific questionable claim in the content, rather than a whole page or site.
This section enumerates proposals for technological approaches to reduce the risk of being misled. Items are organized primarily by credibility assessment strategies. Our hope is to promote the development and adoption of effective tools, as well as interoperation between related tools where that would benefit users. In time, this list might evolve to include links to relevant research and available products. In the meantime, we hope it will inspire and guide students, startups, hackathons, incubators, social investment funds, and other people who want to help.
It should be noted that if credibility tools start to become effective at scale, or look like they might do so, they will themselves come under attack from parties that benefit from higher levels of distrust and deception.
This initial summary table lists some of the more promising ideas, described as they might be implemented in a traditional web browser environment. The table is arranged so that features which require more effort from the user are farther down the table. This is per item effort; install and setup effort is not shown.
Effort Level | Per-Item User Action | Inspection | Corroboration | Reputation | Transparency |
0 | None | Credibility score used in feed/search/suggestion ranking | |||
1 | Glance at visible indicator | Score visible on/near item. Highlight when unusually high/low. | Per-claim or whole-article fact-checks give indicators (click for details). | Display verified and recognizable faces and logos. Show third party +/- flags on source. | Score displayed. Key signals (e.g., name, market, item type) displayed. |
2 | Confirm indicator | Very low score results in warning which must be confirmed before proceeding. E.g., pop-up, interstitial page, or confirmation before sharing. | |||
3 | Request additional information, presented immediately | Show scores per content feature and/or per source. | Show all available fact-checks, helpfully arranged. Lead to full content of each fact-check. | Expanded displays with more related info, info from more sources. View social graph. | Show all score elements; leads to showing all disclosures. |
4 | Request additional information, presented later, e.g., after payment | Request (additional) human annotation be done/revealed. | Request (additional) fact-checks be done/revealed. | Request more information from graph, including from friends and certification businesses. | Request the site reveal more about itself. |
5 | Minor tasks (seconds) | Walk user through annotating a few high-value signals. | Identify key claims to assist in search. | Expand user’s trust graph. | Apply other strategies (e.g., corroboration) to verify disclosures. |
6 | Intermediate tasks (minutes) | Walk through all standard signals. | Search and link to evidence; see other uses and arguments surrounding that evidence. | Provide source credibility evaluations to clarify thinking, and as member of gift economy. | Social outreach to site asking for more disclosure. |
7 | Major tasks (hours+) | Help discover and validate new signals. | Do a full fact-check, possibly in adversarial process. | Push others to provide credibility evaluations. | Provide third party versions of the data, to assist and corroborate. |
People in step 1 might be motivated by:
Machines programmed/trained to do step 1 will probably be faster and cheaper, and might soon become more accurate. They will, however, likely have unpredictable failure modes (bugs, unexpected training artifacts, etc.) which might pose a security vulnerability.
If such automated systems are widely deployed, it is likely attackers will try to game the system, producing misleading content engineered to register as highly credible to the algorithms. Defending against this is similar to defending against malicious SEO. Techniques include:
All of these points have an “arms-race” quality, and C and D might quickly devolve into competing AI, with adversarial neural networks being used to help create language which looks perfect to the credibility assessment systems. As such, it seems unlikely inspection alone is a viable long-term strategy.
See issues and discussion at https://github.com/w3c/cred-claims.
Some of these rely on emerging technologies:
Zoom out to many items, across time, and you get a sort of automated reputation, this provider’s fact-check history. This can lead into the public reputation, which is vital to the business model of some providers.
Here, the emphasis shifts toward the provider making an effort to be more visibly trustworthy. Transparency requires the provider reveal information about themselves and their content, which can require considerable effort and often brings significant risk, in the hope of being more credible.
Many specific items about which transparency would be useful are discussed as part of the The Trust Project.
Much of the work here is about how to motivate transparency and avoid it being gamed. Misleading disclosures are a sort of double-or-nothing gamble in trying to mislead consumers.
As detailed above, in First, Do No Harm, it is quite possible for credibility standards work to do significant unintended damage. One technique which might help mitigate this risk would be to adhere to an independent (and potentially adversarial) review process. These review processes are often seen:
These boards can be expensive to operate and their process can be slow and unpleasant. They can, however, provide an antidote to the tendency of people to be remarkably blind to flaws in proposals they support. This antidote works as long as one or more members of the board is careful to remain disinterested. When set up carefully, with board members who have high credibility in various communities, the board itself and the products approved by the board can also have increased credibility.
Alternatively, an adversarial process (as seen in courts) could be set up where some people are tasked with finding flaws and arguing against a proposal. A third option is to simply rely on wide public review (as per normal W3C process) and hope someone will comment strongly on any major issues.
A review board like this, capable of overseeing potential interventions around credibility, might be useful beyond W3C. It could also be useful for vendor-specific interventions, where a vendor seeks advice about the possible impact of a proposed change. In any case, a board like this could bring considerable credibility to the process if staffed with people respected by diverse stakeholders.
Organizations to consider cooperating with in setting up a board like this include the Internet Governance Forum, the Global Network Initiative, and the Association for the Accreditation of Human Research Protection Programs.
It may also be possible to set up experimental situations, to test proposals while limiting negative impact. This might be seen as A/B testing on a large scale. Civil Servant has developed some methodologies which may be helpful.
Each of the potential standards below is intended to allow systems to interoperate by sharing data with common syntax and semantics. The expectation is that data sharing will be done in a manner compatible with that of schema.org using W3C standard data technologies like JSON-LD and Web Annotations. These build on Web Architecture and the underlying general graph model of RDF, with related technologies like the SPARQL query language, and can usually be used as ordinary JSON embedded in web pages. In each case, the purpose is to allow various independent systems to provide data feeds which can be easily and unambiguously understood by other systems across the web.
It remains unclear how much of the formal W3C standards process is necessary or even helpful for establishing practical, working interoperability for this kind of data. It may be that flexible, responsive de facto standard schemas can emerge from an organic incubation-style community process. In any case, people and organizations are encouraged to experiment with using existing standards and new vocabularies as necessary to achieve these goals and push for widespread adoption. Please do share ideas, progress, and results with the Community Group.
A standard for expressing features of web pages and their content which signal credibility and which are relatively expensive to fake. When humans are recognizing the features, this would allow their work to be shared. When software is recognizing the features, it would allow a more standard API between modules. Standardization would also allow people to be trained to work with these features in creating content and in assessing credibility.
A standard for:
Standards for reputation of providers:
Standards for reputation of content:
Some work has already been done in this area by the Trust Project working with schema.org. See their github and Markup for News.
More work is beginning within the Journalism Trust Initiative (JTI), with leadership on framing ethical issues from Ethical Journalism Network, especially with their Ethical Media Audit.
Directions for next steps on transparency, and impact of existing work, are not yet clear.
Define a way for one provider to embed content from another without endorsement. This would be for credibility a bit like nofollow was for SEO. An iframe would be the natural web way to do this, but third-party content on social media doesn’t currently work that way. There is also the question whether the visited site can be entirely blameless for harmful third-party content; perhaps it can embed it with indicators about what editorial/review process it followed, if any.
Define a way for one provider to link to or embed content from another from another as an anti-endorsement. This could interpreted like a “downvote” by a pagerank style algorithm (but would not need to be coupled with a fact-check). This could be used to let high ranked sites not only increase the credibility of other sites, but also decrease the ranking of other sites. Whether or not this is applied by a search engine or other system will depend on how effective it is in practice, just as links are now.
While many approaches involve behavior from browsers (perhaps via extensions), few require any agreement among vendors (i.e., standardization).
One exception is that browsers could manage a collection of independent credibility assessment tools (something like extensions) which could work together to guide and inform the user. This kind of feature would typically be pioneered in browser extensions and then later adopted into browsers to increase the user base.
[Please add more, as long of you’ve read it and honestly recommend it as relevant. Some description is also helpful.]
Other reading lists and literature reviews:
Information Disorder: Toward an interdisciplinary framework for research and policy making. Claire Wardle and Hossein Derakhshan, of First Draft, Oct 2017. Includes specific recommendations for technology companies.
A Structured Response to Misinformation: Defining and Annotating Credibility Indicators in News Articles, April 2018 Web Conference paper from CredCo members experimentally testing 16 signals, 8 inspection-based. Video of paper presentation, by An Xiao Mina.
Credibility Coalition, Our toolkit for people and teams tackling misinformation online (including “MisinfoMap”).
American Views: Trust, Media, and Democracy (Knight Foundation, Jan 2018).
Mistrust, Efficacy and the New Civics (Ethan Zuckerman, MIT Center for Civic Media, Whitepaper for the Knight Foundation, Aug 2017).
The Partisan Brain: Why People Are Attracted To Fake News And What To Do About It
Truth Decay. RAND Corporation report, Jan 2018. Also, 3 minute video and 12 minute video.
Wikipedia articles on machine-readable data and machine-readable documents. In the latter, see especially the four essential characteristics of trustworthy business records set forth in ISO 15489, Information and documentation - Records management.
Propaganda, classic and prescient 1928 book by Edward Bernays, the father of Public Relations.
Web Literacy for Student Fact-Checkers (2017) by Michael A. Caulfield.
Design Solutions for Fake News is a sprawling (200+ page) crowd-sourced Google-Doc, initiated by Eli Pariser after the 2016 US Election. It approaches many of the same ideas as this document, from some different angles.
The Righteous Mind (2012), by Jonathan Haidt.
Factfulness: Ten Reasons We’re Wrong About the World -- and Why Things Are Better Than You Think, (2018) by Hans Rosling.
We are grateful for everyone who participated, asking questions, making suggestions, taking notes, and especially sharing their knowledge and ideas. The following people attended at least one meeting of the group, according to group attendance records, either as a member or an invited guest: (The editor apologizes in advance for any omissions; corrections welcome): Amy Guy, Amy Zhang, An Xiao Mina, Annette Greiner, Aviv Ovadya, Ben Werdmuller, Caroline Burle, Cheryl Langdon-Orr, Christopher Guess, Cong Yu, Connie Moon Sehat, Dan Brickley, Daniel Schwabe, David Karger, Davide Ceolin, Ed Bice, Ed Summers, Farnaz Jahanbakhsh, Giovanni Luca Ciampaglia, Greg McVerry, Jesse Kranzler, John Connuck, Jon Udell, Katie Haritos-Shea, Matt Lee, Meredith Carden, Mevan Babakar, Michael Golebiewski, Michel Weksler, Newton Calegari, Nick Pickles, Owen Ambur, Patrick Hayes, Reto Gmür, Sally Lehrman, Sam Boyer, Sandro Hawke, Sara-Jayne Terp, Scott Lowenstein, Scott Yates, Stuart Myles, Symeon Papadopoulos, Tantek Çelik, Tim Weninger, Ting Cai, Tom Gilbert, Tzviya Siegman, Vagner Diniz, Vinny Green, Zoë Triska.
We are grateful for review comments from these individuals, some of whom are in the group. These people do not necessarily endorse this document. Please add your name/affiliation here, in alphabet order, if submitting review comments: Amy Guy, Amy Zhang, An Xiao Mina, Annette Greiner, Aviv Ovadya, Connie Moon Sehat, Davide Ceolin, Erica Anderson, Georg Rehm, George Linzer, Humphrey Obuobi, Jon Udell, Kelly J. Cooper, Owen Ambur, Scott Yates, Symeon Papadopoulos.