{"id":10486,"date":"2025-09-13T07:13:59","date_gmt":"2025-09-13T07:13:59","guid":{"rendered":"https:\/\/affinite.io\/cs\/?p=10486"},"modified":"2025-09-13T07:19:31","modified_gmt":"2025-09-13T07:19:31","slug":"jak-spravne-pouzivat-soubor-robots-txt","status":"publish","type":"post","link":"https:\/\/affinite.io\/cs\/jak-spravne-pouzivat-soubor-robots-txt\/","title":{"rendered":"Jak spr\u00e1vn\u011b pou\u017e\u00edvat soubor robots.txt"},"content":{"rendered":"\n

Soubor robots.txt<\/code> je jednoduch\u00fd textov\u00fd soubor um\u00edst\u011bn\u00fd ve ve\u0159ejn\u011b p\u0159\u00edstupn\u00e9m ko\u0159enov\u00e9m adres\u00e1\u0159i webu (nap\u0159. https:\/\/www.example.com\/robots.txt<\/code>). A\u010dkoliv je jeho syntaxe trivi\u00e1ln\u00ed, jeho v\u00fdznam v r\u00e1mci technick\u00e9ho SEO je z\u00e1sadn\u00ed. \u0160patn\u00e1 konfigurace m\u016f\u017ee v\u00e9st k z\u00e1va\u017en\u00fdm d\u016fsledk\u016fm \u2013 od ignorace d\u016fle\u017eit\u00fdch \u010d\u00e1st\u00ed webu roboty a\u017e po \u00fapln\u00e9 vy\u0159azen\u00ed webu z v\u00fdsledk\u016f vyhled\u00e1v\u00e1n\u00ed.<\/p>\n\n\n\n

C\u00edlem tohoto \u010dl\u00e1nku je nab\u00eddnout nejen p\u0159ehled z\u00e1kladn\u00edch pravidel, ale tak\u00e9 poskytnout kontext, pokro\u010dil\u00e9 p\u0159\u00edklady a specifick\u00e1 doporu\u010den\u00ed pro r\u016fzn\u00e9 platformy \u2013 zejm\u00e9na WordPress.<\/p>\n\n\n\n


\n\n\n\n

Co je robots.txt<\/code> a pro\u010d ho pou\u017e\u00edvat<\/h2>\n\n\n\n

Soubor robots.txt<\/code> slou\u017e\u00ed k \u0159\u00edzen\u00ed p\u0159\u00edstupu robot\u016f (tzv. user-agents) vyhled\u00e1va\u010d\u016f, jako je Googlebot, Bingbot, Yandexbot apod., k jednotliv\u00fdm \u010d\u00e1stem va\u0161eho webu. Je sou\u010d\u00e1st\u00ed tzv. Robots Exclusion Protocolu (REP), kter\u00fd byl navr\u017een pro efektivn\u00ed spr\u00e1vu proch\u00e1zen\u00ed web\u016f.<\/p>\n\n\n\n

Prim\u00e1rn\u00ed c\u00edle pou\u017eit\u00ed robots.txt<\/code> jsou:<\/p>\n\n\n\n

    \n
  1. Zamezit indexaci duplicitn\u00edho nebo nevhodn\u00e9ho obsahu.<\/strong><\/li>\n\n\n\n
  2. Omezit p\u0159\u00edstup ke struktur\u00e1m, kter\u00e9 nejsou pro ve\u0159ejnost ur\u010deny.<\/strong><\/li>\n\n\n\n
  3. Optimalizovat tzv. crawl budget<\/strong> \u2013 tedy \u010das a prost\u0159edky, kter\u00e9 robot na webu str\u00e1v\u00ed.<\/li>\n\n\n\n
  4. Zabr\u00e1nit proch\u00e1zen\u00ed technick\u00fdch struktur<\/strong>, kter\u00e9 nemaj\u00ed informa\u010dn\u00ed hodnotu.<\/li>\n\n\n\n
  5. Zamezit prohled\u00e1v\u00e1n\u00ed rozhran\u00ed jako jsou REST API, vyhled\u00e1vac\u00ed dotazy, AJAX skripty apod.<\/strong><\/li>\n<\/ol>\n\n\n\n
    \n\n\n\n

    Struktura souboru robots.txt<\/h2>\n\n\n\n

    Soubor se skl\u00e1d\u00e1 z blok\u016f, z nich\u017e ka\u017ed\u00fd za\u010d\u00edn\u00e1 direktivou User-agent<\/code>, po n\u00ed\u017e n\u00e1sleduj\u00ed Disallow<\/code>, Allow<\/code> a p\u0159\u00edpadn\u011b Sitemap<\/code>.<\/p>\n\n\n\n

    User-agent: *\nDisallow: \/wp-admin\/\nAllow: \/wp-admin\/admin-ajax.php\nSitemap: https:\/\/www.example.com\/sitemap_index.xml\n<\/code><\/pre>\n\n\n\n