mirror of
				https://github.com/superseriousbusiness/gotosocial.git
				synced 2025-10-31 05:02:25 -05:00 
			
		
		
		
	oop, new-line!!
This commit is contained in:
		
					parent
					
						
							
								9d93273450
							
						
					
				
			
			
				commit
				
					
						e4ba68e4e3
					
				
			
		
					 1 changed files with 1 additions and 2 deletions
				
			
		|  | @ -10,7 +10,6 @@ You can allow or disallow crawlers from collecting stats about your instance fro | ||||||
| 
 | 
 | ||||||
| The AI scrapers come from a [community maintained repository][airobots]. It's manually kept in sync for the time being. If you know of any missing robots, please send them a PR! | The AI scrapers come from a [community maintained repository][airobots]. It's manually kept in sync for the time being. If you know of any missing robots, please send them a PR! | ||||||
| 
 | 
 | ||||||
| A number of AI scrapers are known to ignore entries in `robots.txt` even if it explicitly matches their User-Agent. This means the `robots.txt` file is not a foolproof way of ensuring AI scrapers don't grab your content. In addition to | A number of AI scrapers are known to ignore entries in `robots.txt` even if it explicitly matches their User-Agent. This means the `robots.txt` file is not a foolproof way of ensuring AI scrapers don't grab your content. In addition to this you might want to look into blocking User-Agents via [requester header filtering](request_filtering_modes.md), and enabling a proof-of-work [scraper deterrence](scraper_deterrence.md). | ||||||
| this you might want to look into blocking User-Agents via [requester header filtering](request_filtering_modes.md), and enabling a proof-of-work [scraper deterrence](scraper_deterrence.md). |  | ||||||
| 
 | 
 | ||||||
| [airobots]: https://github.com/ai-robots-txt/ai.robots.txt/ | [airobots]: https://github.com/ai-robots-txt/ai.robots.txt/ | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue