馃懡 freezr

Dear search engine makers how can I tell your services to do not crawl some specific folders?

I use a folder where I upload the drafts and I don't want that discoverable.

Can you help me with this? 馃檹

Thanks! 馃憤

2 years ago

Actions

馃憢 Join Station

6 Replies

馃懡 acidus

Its pretty easy. As other's have said, putting a "robots.txt" in the root of our capsule is what you need. It contains lines that tell crawlers what URLs to ignore. Solderpunk defined that Gemini uses a really primative version of Robots.txt. Here is my:

gemini://gemi.dev/robots.txt

Any URLs that start with the text listed on a line will be ignored. So if you want to have a crawler ignore anything inside of "gemini://example.com/drafts/" put this in:

user-agent:*

Disallow: /drafts/ 路 2 years ago

gemini://gemi.dev/robots.txt

gemini://example.com/drafts/

馃懡 freezr

@moddedbear thanks! 馃憤馃嵑 路 2 years ago

馃懡 moddedbear

Robots.txt is way simpler than most pages in search results make it out to be I'm just noticing. 路 2 years ago

馃懡 moddedbear

You could start it off with "User-agent: *" to target all crawlers and then if you wanted to disallow a directory called drafts the next line should be "Disallow: /drafts/". 路 2 years ago

馃懡 freezr

@moddedbear I did not understand how to use it...

My level of abstraction doesn't decode information laid off that way... 馃槱馃槱馃槱 路 2 years ago

馃懡 moddedbear

I don't think you typically have to worry as long you're not linking to your drafts anywhere, but you should look into putting a robots.txt in your capsule root. I'll let you look up the specifics but it's a pretty simple text file that tells crawlers where they should or shouldn't go. 路 2 years ago