jtr
jtr

@lizpixels Hmm, how would that work (can it?) in MB? Can we? @manton do you mind if I toss this in your direction? :)

|
Embed
Progress spinner
sod
sod

@jtr robots.txt in a Micro.blog context is sparsely documented in Search engine indexing. An OpenAI-specific discussion with an example is over at the Help Center. I'm thinking about writing a plug-in to make this a little easier…

|
Embed
Progress spinner
manton
manton

@jtr @sod Looks like @jdm made a plug-in that adds the ChatGPT bot to robots.txt. It's in the plug-in directory called "No Index+". It blocks all bots, though, so we probably should have a plug-in that only blocks GPTBot too.

|
Embed
Progress spinner
jtr
jtr

@sod @manton of course it's under my nose... So my robot.txt currently has User-agent: * Disallow: / which means all crawlers are disallowed (I didn't touch this file below, I thought these are on by default?). So from this, it's pretty easy to add what @lizpixels suggested above, which will only disallow GPTBot.

I'm comfortable with messing with my theme to that degree, I can imagine not all folks here do though.

|
Embed
Progress spinner
ericmwalk
ericmwalk

@sod that would be great if you did!

|
Embed
Progress spinner
sod
sod

@ericmwalk @jtr I did! 😊 Custom Robots makes it a copy-n-paste job.

|
Embed
Progress spinner
jtr
jtr

@sod Thanks....! Good stuff. Now, if you can only make a plug in to refill my coffee cup... 😉

|
Embed
Progress spinner
jdm
jdm

@manton oh that's a good point, it's easy enough to fork and make a specific one. If no one else has done it, I'll make one exclusively for cgpt.

|
Embed
Progress spinner
jdm
jdm

@manton OpenAI lists two bots in their documentation one for their crawler and another for GPT user plugins. Should the explicit version block both or just the cgpt crawler?

|
Embed
Progress spinner