Create a Wikipedia Page: Definitive Guide & Notability

The Definitive Guide to Wikipedia Page Creation: Policies, Processes, and Strategic Compliance

The creation and sustained maintenance of a Wikipedia page is frequently, and erroneously, classified by corporate entities, public relations agencies, and individuals as a standard component of search engine optimization or digital brand management. In reality, the platform operates as a highly regulated, peer-reviewed tertiary encyclopedia governed by a complex matrix of epistemological frameworks, community-enforced policies, evidentiary standards, and systemic checks against promotional bias. Securing a standalone article requires an exhaustive understanding of the platform’s foundational architecture, specifically its rigid definitions of notability, source reliability, conflict of interest mitigation, and deletion protocols. This comprehensive analysis provides an expert-level examination of the exact requirements, technical processes, and strategic compliance measures necessary to successfully navigate the Wikipedia article creation ecosystem.

The Epistemological Framework of Wikipedia

To understand the mechanics of article creation, one must first understand the philosophical evolution of the platform. Wikipedia is not an indiscriminate collection of information, nor is it a web directory, a mirror for links, or a vanity press. The encyclopedia is designed to be a curated repository of human knowledge that has already been analyzed and vetted by the outside world.

The modern framework for inclusion was formalized in 2006 when Wikipedia co-founder Jimmy Wales introduced a core notability criterion in response to growing concerns regarding the proliferation of unverified biographies of living persons. Wales embedded this standard via the “What Wikipedia Is Not” policy, explicitly stating that the platform is not a tabloid newspaper and must make editorial judgments regarding long-term historical notability. Within weeks, this philosophical shift was codified into heuristics for determining long-term inclusion, transforming Wikipedia into the first encyclopedia in history to openly, and collaboratively, define its own criteria for notability.

This paradigm shift fundamentally altered the digital landscape. As the platform grew in authority, a Wikipedia article became recognized as an independent indicator of relevance, eminence, popularity, and reputation for persons, businesses, and historical events. Consequently, the incentive for external actors to manipulate the platform for promotional gain skyrocketed, prompting the community to develop increasingly sophisticated defense mechanisms, policy guidelines, and evidentiary thresholds to protect the integrity of the encyclopedia.

The Architecture of Inclusion: Deciphering Notability Guidelines

The primary gatekeeping mechanism for inclusion in Wikipedia is the concept of notability. Notability is an inherent property of a subject, determined entirely by its reception in the broader world rather than the quality, length, or formatting of the Wikipedia article itself. If a subject has not received sufficient attention from the outside world, no amount of editorial refinement, pristine wikitext formatting, or comprehensive internal linking will render it suitable for inclusion. Conversely, if robust source material exists in the real world, even a poorly written and poorly referenced initial draft will not decrease the underlying notability of the topic.

The Philosophy of Notability

The philosophical underpinning of notability relies on objective, verifiable evidence of significant attention from independent actors. The platform operates on several fundamental axioms regarding this concept:

The first axiom is that there is no inherent notability. A subject is not automatically worthy of a standalone article simply because it exists, regardless of its traffic, sales, or subjective importance. The second axiom is that notability cannot be inherited. A subject does not absorb notability through association with a notable entity. For instance, if a highly notable celebrity or multinational corporation launches a new website or secondary product, that new entity does not automatically qualify for its own article. The website or product must be independently analyzed by the press or academia to warrant inclusion; otherwise, it is simply described as a subsection within the main article of the celebrity or corporation.

The third axiom is that notability is not temporary. Once a topic has been the subject of significant, independent coverage, it generally retains its notable status indefinitely. It does not require ongoing, perpetual news coverage to justify its continued existence on the platform, although sudden bursts of short-term routine news coverage (such as a minor local crime or a standard corporate product launch) are often insufficient to establish the required sustained attention.

The General Notability Guideline (GNG)

The General Notability Guideline (GNG) is the foundational policy governing the vast majority of article inclusions. A topic is presumed to merit a standalone article if it has received significant coverage in reliable secondary sources that are independent of the subject.

The application of the GNG rests on several strict definitions that must be satisfied simultaneously:

GNG Requirement: Operational Definition and Application

Presumed Merit: Meeting the GNG creates an assumption that a subject warrants an article, but it is not an absolute guarantee. A topic might still be excluded if it violates other foundational policies, such as the rule that Wikipedia is not a directory or an indiscriminate collection of data.

Significant Coverage: Source material must address the topic directly and in depth. A passing reference, a brief directory listing, or a single sentence in a broader narrative does not constitute significant coverage. The material must be robust enough to allow an editor to extract comprehensive, verifiable facts without resorting to original research.

Reliable Sources: Sources must possess a recognized reputation for fact-checking, editorial oversight, and accuracy. Publications can encompass all media forms and languages; they do not need to be available online or published in English.

Independence: The coverage must be intellectually and financially independent of the subject. Press releases, self-published blogs, corporate websites, autobiographies, and sponsored content are entirely disqualified from establishing notability, as they lack objective distance.

Multiplicity: The guideline generally expects multiple secondary sources. Multiple publications from the exact same author or the same corporate news syndicate (like a wire service) are usually consolidated and considered a single source for the purpose of establishing notability.

The barometer of notability is ultimately an assessment of human behavior: whether individuals entirely unconnected to the topic (without financial incentive, promotional intent, or other influence) have voluntarily chosen to dedicate significant resources to writing and publishing non-trivial works focusing upon it.

Subject-Specific Notability Guidelines (SNGs)

To address systemic biases, prevent promotional gaming, and provide clarity for edge cases, Wikipedia maintains Subject-Specific Notability Guidelines.

These function in tandem with the GNG; satisfying either a specific SNG or the overarching GNG is typically sufficient to justify an article’s existence.

For organizations and companies, the criteria are explicitly engineered to counteract the sophisticated tactics of marketing and public relations professionals. The depth and quality of coverage are scrutinized far more rigorously than the sheer volume of media mentions. Routine business coverage, local directory listings, syndicated press releases, and trade magazine profiles based on corporate input are routinely discarded during notability assessments. Furthermore, reviews of a company’s products must be evaluated carefully; brief or routine reviews do not qualify, and many ostensible reviews are actually disguised product placements. The source’s audience and editorial rigor must be considered paramount.

For biographical articles about people, the guidelines demand significant coverage independent of the individual, but specific heuristics are applied based on the individual’s profession. Politicians and judges are generally presumed notable if they have held international, national, or state-level office. However, merely being an elected local official or an unelected political candidate does not guarantee notability; such individuals must still satisfy the GNG through widespread, independent press coverage. Athletes are likely notable if they have achieved significant success in major competitions or won significant honors, though their biographies must still include at least one source providing significant, non-routine coverage. Creative professionals (authors, artists, directors) satisfy the criteria if their work has become a significant monument, forms a substantial part of a significant exhibition, or has won major critical attention and permanent representation in notable galleries.

Web content—encompassing podcasts, webcomics, forums, and digital portals—is subject to intense scrutiny to avoid arbitrary biases favoring high-traffic domains. A website is not notable simply due to high bandwidth, user metrics, or commercial success. When evaluating web content, editors must consider whether the platform has had demonstrable, independently verified effects on culture, society, economies, history, literature, or science.

Other domains feature equally tailored SNGs, including those for music, video games, academic figures, films, books, and media outlets, all designed to ensure that only subjects with significant, verifiable external impact are documented.

The Evidentiary Mandate: Sourcing and Verifiability

Wikipedia operates on a strict policy of verifiability, not absolute truth. Every single fact, claim, or data point in an article must be attributable to a reliable, published source. This mandate is absolute; if no reliable sources can be found on a topic, Wikipedia should not have an article on it.

The Hierarchy of Sources

The encyclopedia categorizes evidentiary material into three distinct tiers, dictating exactly how and when they can be utilized in the article creation process:

Primary Sources

Definition: Original materials, historical documents, court transcripts, interviews, corporate press releases, empirical data sets, and a subject’s personal website.

Application in Article Creation: Strongly discouraged for establishing notability. May be used exceedingly sparingly to support basic, uncontroversial facts (e.g., a company’s founding date or an individual’s self-stated birth year). Editors are strictly prohibited from summarizing, analyzing, or drawing novel inferences from primary sources.

Secondary Sources

Definition: Publications that analyze, evaluate, interpret, or synthesize primary sources. Examples include academic journal articles, investigative journalism, and published biographies.

Application in Article Creation: The definitive foundation of Wikipedia. Articles must be built predominantly on reliable secondary sources. These are mandatory for proving notability, establishing context, and ensuring a neutral point of view.

Tertiary Sources

Definition: Compilations or summaries of primary and secondary sources, such as other encyclopedias, almanacs, or general textbooks.

Application in Article Creation: Useful for providing broad overviews or establishing the structural consensus of a highly complex topic, but generally do not count toward establishing the primary notability of a specific subject. Wikipedia itself is a tertiary source and cannot be cited.

Self-published sources are universally prohibited for establishing facts about third parties, particularly concerning living individuals. The singular, narrow exception applies to self-published materials authored by established subject-matter experts whose work in the relevant field has previously been published by reliable, independent academic or journalistic institutions. Even then, such sources must be utilized with extreme caution; if the information is truly suitable for inclusion, it is highly likely that an independent publication will have already covered it.

Evaluating Source Quality and the Risk of Circular Sourcing

The Wikipedia community engages in perpetual, rigorous, and highly nuanced debates regarding the reliability of specific media outlets. Consensus is not static; it is constantly evolving based on publisher behavior, editorial changes, and historical accuracy. These debates culminate in the Perennial Sources list, a centralized database tracking community consensus on thousands of publishers. For example, sources like Amnesty International or Axios are tracked, debated, and assigned caveats; a source might be deemed reliable for general news but unreliable for controversial international politics, or reliable generally but requiring attribution for statements of opinion. Editors must proactively evaluate author credentials, publisher reputation, and historical accuracy before utilizing any source.

A critical systemic risk in the modern information ecosystem is circular sourcing. Because Wikipedia is massively influential, news organizations frequently use Wikipedia articles as uncredited sources for their own journalism. If a Wikipedia editor then cites that newly published news story to support the original fact on Wikipedia, a closed loop of misinformation is created. Editors must vigilantly examine news stories on a case-by-case basis to ensure the publication actually conducted independent fact-checking. Furthermore, Wikipedia is not a repository for gossip, rumors, or unverified leaks; unless a leak is analyzed and reported by a highly reliable secondary source, it cannot be utilized.

Advanced Source Discovery Techniques

Source discovery requires methodical, academic-level research. Relying solely on standard search engine results frequently yields low-quality, search-engine-optimized content, affiliate marketing blogs, and aggregated news rather than encyclopedic material.

Editors are advised to utilize a specific suite of tools for source discovery:

  • The Wikipedia Library: An initiative that provides established editors (those with accounts older than six months and at least 500 edits) free access to premium, paywalled academic databases such as JSTOR.
  • Academic Search Engines: Google Scholar, ScienceOpen, OpenAlex, BASE, and the Internet Archive’s specialized scholar index are essential for locating peer-reviewed secondary sources.
  • Google Books and Bypassing Limits: Google Books indexes millions of texts, which are highly valued on Wikipedia. When encountering preview limits on copyrighted works, editors utilize technical workarounds, such as editing the browser URL and adding specific parameters (e.g., &pg=PA100) to force the engine to jump to a specific page to verify a citation.
  • Free and Open Access Tools: The Wikipedia community maintains lists of free English newspaper sources (WP:FENS) that are text-searchable and do not require subscriptions. The Unpaywall browser extension is also frequently utilized to locate legal, full-text versions of scholarly articles.
  • Citation Chaining: Once a single high-quality scholarly source is identified, editors systematically analyze its bibliography to locate the historical works it cites, and utilize reverse-citation tools to find more recent works that have cited it.

To streamline this exhaustive process, the community utilizes inline templates such as and , which automatically generate targeted search queries across Google, JSTOR, and internal library databases to locate the best possible matches for a proposed topic.

The Prohibition of Artificial Intelligence

The introduction of Large Language Models (LLMs) has profoundly complicated source verification. Wikipedia policy explicitly prohibits the use of LLMs to generate original content for articles. AI models frequently hallucinate non-existent sources, generate synthetic citations that appear legitimate, and fail entirely to adhere to the strict, objective neutral point of view required by the platform. Articles suspected of being generated by an LLM without rigorous human review are classified as a severe threat to the encyclopedia’s integrity and are subject to immediate, speedy deletion. While LLM-powered search engines can theoretically assist in finding concepts, editors are strongly discouraged from relying on them unless every suggested source is manually tracked down and physically verified to exist.

Ethical Paradigms: Conflict of Interest and Paid Editing Disclosures

The most pervasive and insidious threat to the neutral point of view on Wikipedia is the intervention of editors possessing a conflict of interest (COI).

A conflict of interest occurs when an editor attempts to advance the financial, political, or personal interests of their external roles, relationships, clients, or employers.

The Historical Context of Paid Editing Scandals

The platform’s strict rules regarding conflict of interest were not formed in a vacuum; they are the direct result of a decade of high-profile media scandals involving systematic manipulation. In 2006, it was discovered that United States congressional staff were aggressively editing articles about their respective members of Congress. In 2007, Microsoft offered a software engineer financial compensation to edit articles regarding competing code standards. In 2011, the public relations firm Bell Pottinger was caught scrubbing negative information from the articles of its corporate clients.

The scale of manipulation escalated dramatically in 2012 when Wikipedia launched one of its largest sockpuppet investigations, uncovering 250 illicit accounts operated by a firm known as Wiki-PR engaged in covert paid editing. This was followed by the 2015 “Operation Orangemoody” scandal, which uncovered 381 accounts used in a sophisticated extortion scam, charging businesses to create promotional articles and then demanding payment to “protect” them from deletion. Even non-profit entities, such as the Anti-Defamation League in 2020, faced severe backlash for attempting to train staff to edit pages relevant to their organizational goals.

These incidents forced the Wikimedia Foundation (WMF) to legally harden its Terms of Use and forced the volunteer community to adopt a zero-tolerance approach to covert promotional activity.

The Paid Contribution Disclosure Mandate

While editing with a general conflict of interest (such as writing about a friend or a hobbyist organization) is strongly discouraged, editing for compensation without explicit disclosure is a direct violation of the Wikimedia Foundation’s Terms of Use, legally constituting deceptive activity, misrepresentation of affiliation, and fraud.

Any editor receiving or expecting to receive compensation—whether in the form of money, goods, services, or salary—must unequivocally disclose three specific data points for every contribution:

  • The Employer: The specific person or organization remitting payment to the user.
  • The Client: The person or organization on whose behalf the edits are ultimately being made (often the subject of the article itself).
  • Affiliations: Any other relevant connections, such as secondary PR firms, media providers, or signatories to non-disclosure agreements.

Crucially, corporate non-disclosure agreements do not override Wikipedia’s Terms of Use; there is absolutely no recognized confidentiality for paid editing on the platform. If a paid editor cannot legally disclose their client, they are strictly prohibited from editing Wikipedia.

This disclosure must be transparent, persistent, and machine-readable. It must be manifested in at least one of the following venues: on the editor’s primary user page (often utilizing the designated template), on the specific talk page accompanying the paid contributions, or within the metadata of the edit summary. Furthermore, paid editors who advertise their services externally must provide direct hyperlinks to their Wikipedia user accounts on their commercial websites, allowing the community to track their behavior.

Procedural Compliance for COI Editors

Because editors with a conflict of interest are intrinsically incentivized to present their subjects in a positive light, downplay controversies, emphasize minor achievements, and utilize promotional language, they are actively and forcefully discouraged from making direct edits to live articles.

The community-mandated workflow for an editor with a conflict of interest is strictly procedural and relies entirely on peer review:

  • Account Registration: Editors must register with an independent, individual username. Usernames representing a company, a PR agency, or a shared corporate account are prohibited and will be blocked.
  • Propose Changes via Talk Pages: The editor must detail their requested additions, deletions, or new article drafts on the dedicated talk page of the relevant article, rather than altering the article itself.
  • Utilize Review Templates: By deploying specific markup templates, such as or submitting a draft via the Articles for Creation process, the editor flags the request for review by independent, disinterested volunteers.
  • Provide Verifiable Sourcing: All requests must be accompanied by precise citations to independent, reliable secondary sources. Corporate press releases cannot be submitted as justification for content changes.

Attempting to bypass this workflow and unilaterally insert promotional content triggers severe and immediate consequences. Independent administrators actively monitor the platform utilizing advanced heuristic tools to detect COI violations. Violators face public scrutiny, the immediate reversion of their additions, the placement of permanent warning banners (such as or ) on the affected article which damages the brand’s credibility, and the indefinite blocking of their user accounts.

The Creation Lifecycle: From Incubation to Mainspace Publication

Pre-Flight Audits and Incubation

Before initiating any drafting process, prospective authors must execute a series of preparatory checks. The most critical is an exhaustive search to ensure the subject does not already exist on the platform. With millions of existing pages, duplicate creation is a frequent error. This search must encompass alternative spellings, pseudonyms, and potential mentions within broader parent articles. Crucially, the search must examine both the live “Mainspace” (where published articles reside) and the “Draftspace” (where works-in-progress are incubated). If an abandoned draft exists, it is often better to improve it than to start anew.

If the subject is confirmed absent, the creator must immediately perform a rigorous notability assessment, compiling at least three independent, highly reliable secondary sources that address the topic in depth. If these core sources cannot be located prior to drafting, the creation process should be abandoned; an article cannot survive the subsequent review processes without them.

Wikipedia provides dedicated staging areas for manuscript development. The Article Wizard is a highly recommended automated tool that guides new users through the initial stages, verifying notability constraints and citation formatting step-by-step. Drafts should be constructed in either the user’s personal sandbox or the public draftspace. During this phase, editors can utilize the Visual Editor (a standard word-processor interface) or the Source Editor (for direct manipulation of wikitext markup).

Drafting the Manuscript: Avoiding Common Pitfalls

The architectural focus during drafting must remain entirely on neutral, encyclopedic prose. The article must not read like a press release, a persuasive essay, or a defense of the subject.

New creators frequently fail because they commit common sourcing and stylistic errors:

  • Starting with Content, Not Sources: A fatal mistake is writing the article first based on personal knowledge and then attempting to find references to back it up. The text must be a direct summary of what the sources already say; original research and personal knowledge are strictly prohibited.
  • Promotional Tone: Subjective adjectives (e.g., “ground-breaking,” “leading expert,” “visionary,” “revolutionary”) are absolutely prohibited unless they are enclosed in quotation marks and directly attributed to an independent source.
  • Copyright Infringement: Text cannot be copy-pasted from external websites, even if the creator owns the copyright to that external website (like a corporate “About Us” page). All text must be entirely original prose summarizing the factual data.
  • Systemic Bias and Coatracking: Editors must avoid “coatracking,” where an article nominally about one subject is used to disproportionately attack or promote another. The weight of the information in the article must accurately reflect the weight of the coverage in the independent sources.

The Articles for Creation (AfC) Review Matrix

Because newly registered users are technically restricted from creating articles directly in the main encyclopedia space, they must submit their drafts through the Articles for Creation (AfC) process. By appending the markup tag to their draft, the article is placed into a centralized queue where it awaits formal peer review by experienced, vetted volunteer editors.

The AfC process, established in the wake of early platform controversies (such as the Seigenthaler incident), is highly rigorous and historically characterized by massive backlogs, with review wait times frequently extending from several weeks to several months. Reviewers themselves must meet strict criteria, including a minimum of 500 undeleted mainspace edits and a demonstrable understanding of complex deletion policies.

Submissions are routinely declined.

A decline is not a permanent rejection, but rather a mandate for revision. The AfC review matrix utilizes standardized decline rationales:

Insufficient Context

The draft fails to clearly define the subject in the opening sentences. The creator must ensure the lead paragraph identifies the subject in plain English without requiring the reader to guess its significance.

Lack of Demonstrated Notability

The sources provided are primary, trivial, interconnected, or not independent. The creator must return to the research phase and locate in-depth coverage from major, independent publications.

NPOV Violation (Advertisement)

The tone resembles an advertisement or press release. All subjective descriptors must be stripped, and controversies or criticisms mentioned in the external sources must be proportionally included to achieve a neutral point of view.

BLP Policy Violation

Contentious material about a living person lacks strict inline citations to high-quality sources. Immediate removal of the offending text is required; failure to do so will result in the draft’s deletion.

Merge Recommendation

The subject lacks standalone notability, but verifiable information is extracted and combined into a broader, existing parent page. The reviewer will suggest redirecting the effort to an existing page.

The AfC process frequently generates friction. Creators often express frustration regarding perceived “deletionism,” noting that nominating an article for deletion requires minimal effort, whereas creating a properly sourced article demands immense research and labor. Despite proposals to streamline the process and quickly promote promising drafts, the stringent review standards remain a vital defense mechanism for the encyclopedia.

Autoconfirmed Status and the Risks of Direct Movement

Editors who have achieved “autoconfirmed” status—a technical threshold reached when an account is at least four days old and has registered a minimum of ten edits—gain the technical capability to bypass the AfC queue entirely. Autoconfirmed users can move a page directly from the draftspace into the live mainspace by utilizing the page move function to strip the “Draft:” prefix from the title.

However, utilizing this technical privilege is highly risky for inexperienced editors or those harboring a conflict of interest. If an autoconfirmed user publishes an article that violates fundamental inclusion guidelines, it bypasses the safety net of AfC peer review and is immediately exposed to the hostile environment of the rapid deletion processes, often resulting in the permanent erasure of the content.

Architectural Integrity: The Manual of Style and Multimedia Integration

Structural Anatomy of a Standard Article

A properly architected article follows a definitive sequence of elements, designed to optimize data retrieval and maintain visual coherence.

The Lead Section

The article must begin with a concise, standalone overview ranging from one to four paragraphs. The very first sentence must definitively define the subject, placing it in its proper context in plain English. The remainder of the lead must summarize the most important points detailed in the body of the article, acting as an inverted pyramid. The emphasis given to statements in the lead must proportionally reflect their relative importance in the body text. Crucially, the lead should not introduce novel facts that are absent from the subsequent sections.

The Body and Section Headings

The core text is organized via hierarchical section headings. Wikipedia strictly mandates the use of sentence case for headings (capitalizing only the first word and proper nouns, e.g., “Early life and career”) rather than title case (“Early Life and Career”). Sections should progress logically without skipping hierarchical levels (e.g., jumping from a Level 2 heading directly to a Level 4 subheading).

Standard Appendices

The conclusion of the article is highly structured, typically following a strict chronological sequence:

  • See also: A bulleted list of internal links to related Wikipedia articles.
  • Notes and References: The core citation apparatus.
  • Further reading: Relevant publications that provide deeper context but were not utilized as direct sources for the text.
  • External links: Official websites of the subject or highly relevant external databases. This section is heavily restricted to prevent link spam.

Multimedia Logistics: Curation, Uploading, and Licensing

The integration of visual multimedia requires navigation of complex, international copyright frameworks. Images cannot be simply copied from external websites and pasted into an article; hotlinking is technically disabled to prevent copyright infringement and bandwidth theft.

All visual assets must be formally uploaded, and the destination of the upload depends entirely on the rigorous copyright status of the file:

  • Wikimedia Commons: This is the preferred central repository for all media. Only files released under a verified free copyright license (such as Creative Commons) or those undeniably in the public domain may be uploaded here. Images hosted on Commons can be utilized seamlessly across all 300+ language variants of Wikipedia.
  • Local Wikipedia Uploads (Fair Use): If an image is under copyright (e.g., a corporate logo, a movie poster, a book cover, or a specific software screenshot), it cannot be hosted on Wikimedia Commons. It must be uploaded directly to the local English Wikipedia under a strict “Fair Use” rationale. This requires the uploader to provide a detailed, formal justification explaining exactly why the non-free content is essential to the reader’s understanding of the article, and proving that no free alternative could possibly be created to serve the same educational purpose.

Format selection is also critical. Diagrams, logos, and illustrations should ideally be uploaded as Scalable Vector Graphics (SVG) rather than JPEGs, as SVGs are code-based, infinitely scalable without quality loss, and easily translatable. If PNGs are used for diagrams, they must be optimized by reducing the color depth to eliminate anti-aliasing artifacts.

Once uploaded, images are integrated into the text using specific wiki-markup (e.g., []). This markup standardizes the display as a right-aligned thumbnail, preventing visual stack-ups that disrupt the text flow, and mandates the inclusion of descriptive alt-text to maintain strict compliance with web accessibility standards.

Semantic Web Integration: Categorization and Wikidata

Categorization Logic and Ontological Placement

Categorization is the primary mechanism by which articles are indexed, grouped, and discovered by readers and automated bots. Every article must belong to at least one category, but categories must be applied with surgical precision.

Categories are broadly divided into two structural types:

  • Topic Categories: Named after a singular concept (e.g., “Category:France”) containing articles relating broadly to that topic.
  • Set Categories: Named after a plural class (e.g., “Category:Australian astronomers”) containing items that belong to that specific group.

When placing an article into the category tree, editors must select the most specific subcategory available, actively avoiding broad parent categories. For example, an article about a Parisian museum should be placed in “Museums in Paris,” not just the broader “France” category.

The defining principle of categorization is neutrality and verifiability. A category must represent a universally recognized, defining characteristic of the subject, and that characteristic must be explicitly supported by the sourced text within the article. Subjective categories (e.g., “beautiful cities,” “famous inventors,” or “large companies”) are strictly prohibited. For biographical articles, individuals must be categorized by their own defining traits (occupation, nationality, significant achievements) rather than the characteristics of the article itself (e.g., avoiding “Category:Biography”). Technically, category tags are placed at the absolute bottom of the article’s wikicode to ensure they do not disrupt the visual rendering of the text interface.

Wikidata Integration and Interlanguage Connectivity

An often-overlooked but strategically vital final step in article creation is the integration of the new page into Wikidata. Wikidata operates independently of Wikipedia but serves as the centralized knowledge base and infrastructural backbone that powers the semantic web architecture of the entire Wikimedia ecosystem.

When a new English Wikipedia article is published, the creator should immediately query Wikidata to determine if an item (identified by a specific “Q-number”) already exists for the subject—which is highly probable if the subject already possesses an article on a non-English Wikipedia.

  • If the item exists, the new English article is mapped to it via a “sitelink”.
  • If the item does not exist, a new Wikidata item must be generated from scratch, populated with a structural label, a brief description, alternative aliases, and the foundational interlanguage links.

This integration is critical for long-term data management.

It ensures that discrete data points (such as population statistics, mayoral elections, or dates of birth) remain synchronized across the global network. A single value updated in Wikidata can instantly cascade across hundreds of language versions of Wikipedia, drastically reducing maintenance overhead and ensuring global data consistency.

The publication of an article in the mainspace is not the conclusion of the lifecycle; it is merely the beginning of continuous community scrutiny. The Wikipedia community employs aggressive, tiered maintenance protocols to continuously prune content that fails to meet inclusion standards. Understanding these deletion mechanisms is critical for anyone attempting to establish and defend a new page.

Criteria for Speedy Deletion (CSD)

The most immediate and severe threat to a newly published article is the Speedy Deletion process. This mechanism empowers Wikipedia administrators to instantly delete pages without debate, community discussion, or prior warning if they unequivocally violate specific, narrowly defined criteria.

Primary CSD Designations

Triggering Conditions and Mechanisms

Strategic Avoidance Protocols

A7: No Indication of Importance

Applies to articles about people, animals, organizations, or web content that completely fail to assert why the subject is significant. This is the most common cause of rapid deletion for new corporate pages.

Ensure the lead paragraph explicitly states the subject’s major achievements or impact, supported immediately by independent citations.

G11: Unambiguous Promotion

Articles written exclusively to advertise a product, company, or individual, behaving as public relations material and requiring a total rewrite to achieve neutrality.

Strip all marketing jargon, superlative adjectives, and subjective claims. Adhere strictly to the Neutral Point of View protocol.

G12: Copyright Infringement

Text that is demonstrably copy-pasted from external sources, including corporate ‘About Us’ pages or press releases.

Write all prose entirely from scratch. Summarize sources rather than copying them verbatim.

G13: Abandoned Drafts

Unpublished drafts residing in the draftspace or AfC queue that have not been edited for a period of six consecutive months.

Maintain continuous editorial activity on pending drafts, even minor formatting tweaks, to reset the automated deletion timer.

G15: LLM-Generated Content

Pages synthesized by Large Language Models lacking verifiable human review or containing hallucinated sources.

Strictly prohibit the use of AI generation tools during the drafting and research phases.

Crucially, the creator of an article is technically and procedurally prohibited from removing a speedy deletion tag placed by another editor. Attempting to simply delete the warning tag will result in immediate administrative action. Instead, the creator must utilize the “Contest this speedy deletion” button embedded in the tag, which automatically opens a structured dialogue on the article’s talk page, allowing the creator to formally argue why the criteria do not apply based on policy.

Proposed Deletion (PROD) and BLPPROD

If an article survives the immediate threat of speedy deletion but its notability or sourcing remains highly questionable, it enters the secondary deletion tier: Proposed Deletion (PROD).

The PROD process handles uncontroversial deletion candidates. An editor places a PROD tag on the page, outlining the deficiencies. If no other editor objects or removes the tag within seven days, the article is quietly and administratively deleted. Unlike CSD, any editor—including the article’s creator—can remove a standard PROD tag to save the page, which forces the nominator to escalate to a formal debate if they wish to pursue deletion.

However, for Biographies of Living Persons (BLPs) that completely lack sources, a much stronger variant applies: BLPPROD. Deletion via BLPPROD cannot be halted simply by removing the tag; it can only be contested and stopped by the physical insertion of a reliable, independent source into the article text.

Articles for Deletion (AfD)

The ultimate adjudicative process on the platform is Articles for Deletion (AfD). This is a formal, highly structured, community-wide discussion lasting up to two weeks, where editors debate the merits of the article strictly against established policies.

AfD is fundamentally not a democratic vote; it is a consensus-building mechanism. An article will not be saved simply because a large number of affiliated users or coordinated sockpuppets cast “Keep” votes. Closing administrators weigh the arguments based on their logical alignment with the General Notability Guideline and sourcing policies, discarding votes that lack policy-based rationale.

Discussions in AfD can result in a variety of nuanced outcomes:

  • Keep: The consensus determines the article meets notability standards, and the deletion notice is permanently removed.
  • Delete: The article and its entire edit history are erased from the public database.
  • Merge: The subject lacks standalone notability, but verifiable information is extracted and combined into a broader, existing parent page.
  • Draftify / Incubate: The article is removed from the live mainspace and returned to the draftspace, allowing creators more time to improve sourcing without the threat of immediate deletion.
  • No Consensus: If the community debate is fractured and cannot reach a definitive conclusion, the default administrative action is to maintain the status quo and keep the article. However, immediately renominating a page after a “no consensus” result is viewed as highly disruptive behavior.

In rare instances where an article’s survival or deletion is overwhelmingly obvious from the outset—such as a nomination based on blatant vandalism or a nomination against a highly recognized historical figure—administrators may invoke the “Snowball Clause” (or “Speedy Keep”) to close the debate early, avoiding bureaucratic process for its own sake.

Alternatives to Article Creation

Given the extreme friction, strict evidentiary thresholds, and high failure rates of the creation and review processes, prospective authors must critically evaluate whether a standalone article is truly the optimal vehicle for the information.

Often, content that repeatedly fails the General Notability Guideline can be successfully preserved by merging it into an existing, broader article, provided it maintains an appropriate level of detail, avoids self-promotion, and is backed by reliable sources. For example, a minor software product might not deserve its own page, but a paragraph detailing its release is perfectly suited for the page of its parent development company.

The “Requested Articles” mechanism exists for users to suggest topics they believe are notable but lack the technical expertise, time, or conflict-of-interest clearance to write themselves. Users can submit requests organized by profession, nationality, or topic. However, this is emphatically not an on-demand writing service. The vast majority of requests languish indefinitely due to the volunteer nature of the workforce and the stringent evidence required to initiate a new page. Unless a subject possesses overwhelming, obvious notability, volunteer editors rarely take up the arduous task of researching and writing an article on behalf of an external requestor.

Strategic Imperatives for Sustained Inclusion

Securing and maintaining a Wikipedia page is a process of establishing an irrefutable, independently verified evidentiary record rather than a simple exercise in copywriting or digital marketing. The platform’s technical architecture and cultural norms are explicitly designed to resist commercial intervention, filter out ephemeral phenomena, and demand high-quality, independent journalistic or academic validation before granting encyclopedic real estate.

Success in this ecosystem requires total submission to its epistemological rules. Prospective creators must abandon the impulse to control the narrative, replacing promotional intent with rigorous adherence to the Neutral Point of View. They must understand that notability is established by third parties in the real world, and their sole function on Wikipedia is to accurately map that external coverage using the highly structured parameters of the Manual of Style. By comprehensively documenting conflicts of interest, relying exclusively on top-tier secondary sources, utilizing the safety mechanisms of the Articles for Creation process, and preempting the triggers for algorithmic and administrative deletion, creators can successfully navigate the formidable barriers to entry and contribute enduring value to the global knowledge repository.