About NikkiWiki and SITEMAP: Difference between pages

From NikkiWiki
(Difference between pages)
Jump to navigation Jump to search
No edit summary
 
No edit summary
 
Line 1: Line 1:
[[image:MediaWiki.svg|thumb|Image: <cite class="u">NikkiWiki</cite> is powered by the free open-source software of [[wikipedia:Wikimedia MediaWiki|Wikimedia MediaWiki]].  Pictured is a MediaWiki logo with the MediaWiki mascot (a sunflower or ''[[wikipedia:Helianthus annuus|Helianthus annuus]]'') and the MediaWiki slogan of "because ideas want to be free". <ref><code>[[commons:category:MediaWiki]]</code></ref> <ref><code>[[commons:category:MediaWiki logos]]</code></ref> <ref><code>[[commons:category:SVG MediaWiki logos]]</code></ref>]]
Adding a sitemap to your website allows searchbots to find pages much faster and more efficiently, allowing them to be quickly indexed for search engines.&ensp; Sitemaps can be saved as either "<code>/sitemap.txt</code>" or "<code>/sitemap.xml</code>".&ensp; Using plaintext ("<code>/sitemap.txt</code>") is much faster and easier than writing extensible markup language (XML).&ensp; I recommend keeping the sitemap as plaintext, allowing the [https://www.sitemaps.org/ SITEMAP] protocol to join the ranks of the other plaintext website protocols for <u>[[ROBOTS]]</u>, [https://www.securitytxt.org/ SECURITY], and [https://humanstxt.org/ HUMANS].


A wiki for [[Nicole Sharp]] of [https://www.frostburg.edu/about-frostburg/campus-and-community/ Frostburg State University], Maryland, United States of America (USA). <ref><code>[[Nicole Sharp's Website]]</code></ref>
To create a sitemap, you simply make a plaintext list of each URL (uniform resource locator) for the website.&ensp; Only URLs for a single domain should be included — do not add URLs for subdomains or alias domains.&ensp; You should also only list canonical URLs.&ensp; This means that if a particular webpage can be accessed from multiple URLs, only one URL should be listed for that webpage in the sitemap.


<cite class="n">[[NikkiWiki]]</cite> and <cite class="n">[[Nicole Sharp's Website]]</cite> are [https://www.freedomdefined.org/ <strong>Free Cultural Works</strong>]. <ref><code>[[NikkiLicense]]</code></ref>
For example, there are many different ways to access <u>[[Nicole Sharp's Homepage]]</u>:


<cite class="n">NikkiWiki</cite> is powered by [[mw:What is MediaWiki?|<strong>Wikimedia MediaWiki</strong>]], the same [[wikipedia:FOSS|free open-source wikisoftware]] used for [[wikipedia:Main Page|<cite class="n">Wikipedia</cite>]], [[wikibooks:Main Page|<cite class="n">Wikibooks</cite>]], [[wikiversity:Main Page|<cite class="n">Wikiversity</cite>]], [[commons:Main Page|<cite class="n">Wikimedia Commons</cite>]], and the other wikiprojects of the [https://www.wikimedia.org/ Wikimedia Foundation]. <ref><code>[[Wikimedia MediaWiki]]</code></ref> <ref><code>[[mw:Main Page]]</code></ref> <ref><code>[[mw:sites using MediaWiki]]</code></ref> <ref><code>https://www.wikimedia.org/</code></ref>
<code><pre>
https://www.nicolesharp.net/
https://www.nicolesharp.net/index.htm
https://www.nicolesharp.net/index.html
https://www.nicolesharp.net/w/
https://www.nicolesharp.net/w/index.php
https://www.nicolesharp.net/w/index.php?title=NikkiWiki
https://www.nicolesharp.net/w/index.php?title=Main_Page
https://www.nicolesharp.net/w/index.php?title=NikkiWiki:Main_Page
https://www.nicolesharp.net/wiki/
https://www.nicolesharp.net/wiki/NikkiWiki
https://www.nicolesharp.net/wiki/Main_Page
https://www.nicolesharp.net/wiki/index
</pre></code>


[[NikkiLicense|Unless otherwise noted, content on <cite class="n">NikkiWiki</cite> and <cite class="n">Nicole Sharp's Website</cite> is available under the <cite>Creative Commons Attribution-ShareAlike (CC BY-SA) 4.0 International Public License</cite>.]] This is the same license used by <cite class="n">Wikipedia</cite>, <cite class="n">Wikibooks</cite>, <cite class="n">Wikiversity</cite>, and the other wikiprojects of the Wikimedia Foundation, allowing to freely copy and paste wikitext (with attribution) back and forth between NikkiWiki and Wikimedia.
The canonical URL though is
: <code>[[about Nicole Sharp's Homepage|https://www.nicolesharp.net/wiki/NikkiWiki]]</code>
since all of the other URLs redirect to that URL.


[[NikkiAnalytics|<cite class="n">NikkiWiki</cite> and <cite class="n">Nicole Sharp's Website</cite> use webanalytics cookies from Cloudflare Analytics, Matomo Analytics, Google Analytics, Microsoft Clarity, and Yandex Metrica.]] <ref><code>[[NikkiAnalytics]]</code></ref>
Here are even more ways to access Nicole Sharp's Homepage:


<cite class="n">NikkiWiki</cite> and <cite class="n">Nicole Sharp's Website</cite> are hosted, published, and written in the United States of America (USA).
<code><pre>
https://nicolesharp.net/
https://www.nicolesharp.net/
https://web.nicolesharp.net/
https://en.nicolesharp.net/
https://eng.nicolesharp.net/
https://us.nicolesharp.net/
https://usa.nicolesharp.net/
https://wiki.nicolesharp.net/
https://w.nicolesharp.net/
http://www.nicolesharp.net/
http://nicolesharp.net/
http://nicolesharp.altervista.org/
http://nicolesharp.dreamhosters.com/
https://nicolesharp.dreamhosters.com/
</pre></code>
 
With the exception of "<code><nowiki>https://www.nicolesharp.net/</nowiki></code>", none of these other URLs should be included in "<u><code>https://www.nicolesharp.net/sitemap.txt</code></u>".&ensp; All of the URLs should have the same protocol (either all HTTPS [Hypertext Transfer Protocol Secure] or all HTTP [Hypertext Transfer Protocol]) and all of the URLs should be on the same subdomain (for example, either all with "<code>www</code>" or all without "<code>www</code>").
 
The following "<code>/sitemap.txt</code>" example gives a compliant sitemap for "<u><code>[[Nicole Sharp's Website|https://www.nicolesharp.net/]]</code></u>":
 
<code><syntaxhighlight lang="text">
https://www.nicolesharp.net/wiki/NikkiWiki
https://www.nicolesharp.net/wiki/about_NikkiWiki
https://www.nicolesharp.net/wiki/Nicole_Sharp
https://www.nicolesharp.net/wiki/license_for_Nicole_Sharp's_Website
https://www.nicolesharp.net/wiki/analytics_for_Nicole_Sharp's_Website
https://www.nicolesharp.net/wiki/donations
https://www.nicolesharp.net/wiki/security
</syntaxhighlight></code>
 
Only canonical URLs are included, all of the URLs have the same protocol ("<code>https://</code>"), and all of the URLs are on the same subdomain ("<code>www.nicolesharp.net</code>").&ensp; Each new subdomain will need its own sitemap.
 
Once your sitemap is completed, you can add it to the Robots Exclusion Protocol to be indexed by searchbots.&ensp; An example "<code>/robots.txt</code>" with a sitemap is given below.
 
<code><highlight lang="robots">
User-agent: *
Disallow:
Sitemap: https://www.example.net/sitemap.txt
</highlight></code>


== see also ==
== see also ==


* <code>[[NikkiWiki]]</code>
* <u><code>https://www.nicolesharp.net/sitemap.txt</code></u>
* <code>[[NikkiSite]]</code>
* <code>https://www.sitemaps.org/</code>
* <code>[[NikkiLicense]]</code>
* <u><code>[[ROBOTS]]</code></u>
* <code>[[NikkiAnalytics]]</code>
* <code>https://www.securitytxt.org/</code>
* <code>[[Nicole Sharp]]</code>
* <code>https://humanstxt.org/</code>


== references ==
== keywords ==


<references />
<code>bots, development, indexing, ROBOTS, robots.txt, searchbots, SITEMAP, sitemap.txt, TXT, web, webcrawlers, webcrawling, webdevelopment, WWW</code>


{{#seo:|keywords=about, aboutpage, academic, Alleganian, Allegany, Alleghenian, Allegheny, America, American, Appalachia, Appalachian, Attribution-ShareAlike, BY-SA, CC, Chesapeake, Cumberland, Cumberlander, Earth, Earther, Earthling, Earth-Moon, EDU, education, educational, EN, ENG, English, EN-US, homewiki, homewikisite, learning, Maryland, Marylander, MD, MediaWiki, MW, NET, Nicole, nicolesharp.net, Nikki, NikkiSite, NikkiWiki, NLS, nonprofit, NS, Potomac, Ruiz, scholarly, Sharp, Terra, Terran, Terra-Luna, US, USA, US-MD, wiki, wikisite}}
{{#seo:|keywords=bots, development, indexing, ROBOTS, robots.txt, searchbots, SITEMAP, sitemap.txt, TXT, web, webcrawlers, webcrawling, webdevelopment, WWW}}


[[category:NikkiWiki]]
[[category:webdevelopment]]

Revision as of 2023-09-04T22:30:00

Adding a sitemap to your website allows searchbots to find pages much faster and more efficiently, allowing them to be quickly indexed for search engines.  Sitemaps can be saved as either "/sitemap.txt" or "/sitemap.xml".  Using plaintext ("/sitemap.txt") is much faster and easier than writing extensible markup language (XML).  I recommend keeping the sitemap as plaintext, allowing the SITEMAP protocol to join the ranks of the other plaintext website protocols for ROBOTS, SECURITY, and HUMANS.

To create a sitemap, you simply make a plaintext list of each URL (uniform resource locator) for the website.  Only URLs for a single domain should be included — do not add URLs for subdomains or alias domains.  You should also only list canonical URLs.  This means that if a particular webpage can be accessed from multiple URLs, only one URL should be listed for that webpage in the sitemap.

For example, there are many different ways to access Nicole Sharp's Homepage:

https://www.nicolesharp.net/
https://www.nicolesharp.net/index.htm
https://www.nicolesharp.net/index.html
https://www.nicolesharp.net/w/
https://www.nicolesharp.net/w/index.php
https://www.nicolesharp.net/w/index.php?title=NikkiWiki
https://www.nicolesharp.net/w/index.php?title=Main_Page
https://www.nicolesharp.net/w/index.php?title=NikkiWiki:Main_Page
https://www.nicolesharp.net/wiki/
https://www.nicolesharp.net/wiki/NikkiWiki
https://www.nicolesharp.net/wiki/Main_Page
https://www.nicolesharp.net/wiki/index

The canonical URL though is

https://www.nicolesharp.net/wiki/NikkiWiki

since all of the other URLs redirect to that URL.

Here are even more ways to access Nicole Sharp's Homepage:

https://nicolesharp.net/
https://www.nicolesharp.net/
https://web.nicolesharp.net/
https://en.nicolesharp.net/
https://eng.nicolesharp.net/
https://us.nicolesharp.net/
https://usa.nicolesharp.net/
https://wiki.nicolesharp.net/
https://w.nicolesharp.net/
http://www.nicolesharp.net/
http://nicolesharp.net/
http://nicolesharp.altervista.org/
http://nicolesharp.dreamhosters.com/
https://nicolesharp.dreamhosters.com/

With the exception of "https://www.nicolesharp.net/", none of these other URLs should be included in "https://www.nicolesharp.net/sitemap.txt".  All of the URLs should have the same protocol (either all HTTPS [Hypertext Transfer Protocol Secure] or all HTTP [Hypertext Transfer Protocol]) and all of the URLs should be on the same subdomain (for example, either all with "www" or all without "www").

The following "/sitemap.txt" example gives a compliant sitemap for "https://www.nicolesharp.net/":

https://www.nicolesharp.net/wiki/NikkiWiki
https://www.nicolesharp.net/wiki/about_NikkiWiki
https://www.nicolesharp.net/wiki/Nicole_Sharp
https://www.nicolesharp.net/wiki/license_for_Nicole_Sharp's_Website
https://www.nicolesharp.net/wiki/analytics_for_Nicole_Sharp's_Website
https://www.nicolesharp.net/wiki/donations
https://www.nicolesharp.net/wiki/security

Only canonical URLs are included, all of the URLs have the same protocol ("https://"), and all of the URLs are on the same subdomain ("www.nicolesharp.net").  Each new subdomain will need its own sitemap.

Once your sitemap is completed, you can add it to the Robots Exclusion Protocol to be indexed by searchbots.  An example "/robots.txt" with a sitemap is given below.

User-agent: *
Disallow:
Sitemap: https://www.example.net/sitemap.txt

see also

keywords

bots, development, indexing, ROBOTS, robots.txt, searchbots, SITEMAP, sitemap.txt, TXT, web, webcrawlers, webcrawling, webdevelopment, WWW