rudimentary search engine optimisation (robots.txt, <meta name="robots"> in /b)
This commit is contained in:
parent
107a9a4c39
commit
41fa245150
|
@ -156,6 +156,7 @@ async fn branch(RawQuery(named_id): RawQuery, State(state): State<Arc<Server>>)
|
|||
String::from("<meta property=\"og:description\" content=\"");
|
||||
write!(per_page_metadata, "{}", EscapeHtml(branch_markup)).unwrap();
|
||||
per_page_metadata.push_str("\">");
|
||||
per_page_metadata.push_str(r#"<meta name="robots" content="noindex">"#);
|
||||
|
||||
const PER_PAGE_METADATA_REPLACEMENT_STRING: &str = "<!-- treehouse-ca37057a-cff5-45b3-8415-3b02dbf6c799-per-branch-metadata -->";
|
||||
return Html(content.replacen(
|
||||
|
|
|
@ -511,6 +511,10 @@ where
|
|||
pub fn target(dirs: Arc<Dirs>, sources: Arc<Sources>) -> DynDir {
|
||||
let mut root = MemDir::new();
|
||||
root.add(VPath::new("static"), dirs.static_.clone());
|
||||
root.add(
|
||||
VPath::new("robots.txt"),
|
||||
Cd::new(dirs.static_.clone(), VPathBuf::new("robots.txt")).to_dyn(),
|
||||
);
|
||||
|
||||
let dir_index = DirIndex::new(sources.parsed_trees.keys().map(|x| &**x));
|
||||
let tree_view = TreehouseDir::new(dirs, sources, dir_index);
|
||||
|
|
11
static/robots.txt
Normal file
11
static/robots.txt
Normal file
|
@ -0,0 +1,11 @@
|
|||
# This file is mirrored under /robots.txt and /static/robots.txt.
|
||||
|
||||
User-Agent: *
|
||||
|
||||
# /static does not contain any pages.
|
||||
Disallow: /static
|
||||
# /_treehouse contains system pages such as the 404 page.
|
||||
Disallow: /_treehouse
|
||||
|
||||
# In addition to this, some system pages contain a <meta name="robots" content="noindex"> tag to
|
||||
# prevent them from appearing in search results.
|
Loading…
Reference in a new issue