[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"$fQbAgWAO6iMAwtmhGtecpdTE6LBqqITSMxa7GD2kmWK0":3},{"slug":4,"name":5,"version":6,"author":7,"author_profile":8,"description":9,"short_description":10,"active_installs":11,"downloaded":12,"rating":13,"num_ratings":14,"last_updated":15,"tested_up_to":16,"requires_at_least":17,"requires_php":18,"tags":19,"homepage":25,"download_link":26,"security_score":27,"vuln_count":11,"unpatched_count":11,"last_vuln_date":28,"fetched_at":29,"vulnerabilities":30,"developer":31,"crawl_stats":28,"alternatives":36,"analysis":137,"fingerprints":267},"ai-rights-protection-by-simplefeed","AI Rights Protection by SimpleFeed","1.0.22","Alex Khomichenko","https:\u002F\u002Fprofiles.wordpress.org\u002Fakhomichenko\u002F","\u003Cp>The leader in content syndication for WordPress publishers now helps protect you from Artificial Intelligence companies using your content without compensation. If you are not blocking the AI companies access, they have no incentive to license your content.\u003C\u002Fp>\n\u003Cp>Features include:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Block the bots of leading AI companies from scraping your content\u003C\u002Fli>\n\u003Cli>See the bots name, IP address and last seek date\u003C\u002Fli>\n\u003Cli>Add\u002Fremove bots from the block list\u003C\u002Fli>\n\u003Cli>Opt in to create a common AI bot list and give blocked bots instructions on how to license your content.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>SimpleFeed is the leading content syndication solution for WordPress Publishers. We work with over 500 publishers to syndicate their content to Apple News, MSN, SmartNews, Newsbreak, Opera News, Samsung News, Yahoo and just about anywhere your users consume content. We are working with these publishers, and publishers who use this plugin and want to participate, to license their content to leading AI companies.\u003C\u002Fp>\n\u003Cp>SimpleFeed has a non-public plug-in which also enables publishers to be more successful in their syndication efforts. SimpleFeed customers get the complete plug in, best practices for successful syndication, and technical support.\u003C\u002Fp>\n\u003Cp>Please note that blocking the wrong bot and\u002For presenting the wrong Robots.txt can lead to traffic declines. This plug in is free and offered “as is” without warranty or support.\u003C\u002Fp>\n\u003Cp>Install now to preserve your rights to potential future compensation from the AI companies.\u003C\u002Fp>\n\u003Ch3>Use of 3rd Party Services\u003C\u002Fh3>\n\u003Cp>The AI Rights Protection by SimpleFeed plugin communicates with an external service hosted at \u003Ca href=\"https:\u002F\u002Fwww.simplefeed.com\u002Fwp-content\u002Fsimplefeed\u002Fwp-simplefeed-bots-protection-admin-rest.php\" rel=\"nofollow ugc\">SimpleFeed\u003C\u002Fa> to perform the following functions:\u003C\u002Fp>\n\u003Col>\n\u003Cli>\n\u003Cp>\u003Cstrong>Storage of Plugin Settings\u003C\u002Fstrong>: The plugin sends and retrieves plugin settings to\u002Ffrom the SimpleFeed service. This ensures that your settings are securely stored and synchronized with the external service.\u003C\u002Fp>\n\u003C\u002Fli>\n\u003Cli>\n\u003Cp>\u003Cstrong>Bot Request Statistics\u003C\u002Fstrong>: The plugin collects and stores statistics about bot requests on your WordPress site. These statistics are sent to SimpleFeed for analysis and reporting purposes.\u003C\u002Fp>\n\u003C\u002Fli>\n\u003Cli>\n\u003Cp>\u003Cstrong>Premium Access Validation\u003C\u002Fstrong>: The plugin checks with the SimpleFeed service to verify if your WordPress user account has premium access to the plugin features.\u003C\u002Fp>\n\u003C\u002Fli>\n\u003C\u002Fol>\n\u003Cp>For more information, please review the \u003Ca href=\"https:\u002F\u002Fwww.simplefeed.com\u002Fother\u002Fprivacy-policy\u002F\" rel=\"nofollow ugc\">SimpleFeed Privacy Policy\u003C\u002Fa> and \u003Ca href=\"https:\u002F\u002Fwww.simplefeed.com\u002Fother\u002Ftos\u002F\" rel=\"nofollow ugc\">Terms of Use\u003C\u002Fa>.\u003C\u002Fp>\n\u003Cp>By using this plugin, you agree to the data transmissions to SimpleFeed’s external services for the purposes outlined above.\u003C\u002Fp>\n","Protect your rights by reviewing and blocking AI bots from using your content without compensation.",0,641,100,1,"2025-04-01T09:10:00.000Z","6.7.5","3.0","5.3",[20,21,22,23,24],"ai","analytics","bots","crawlers","robots-txt","http:\u002F\u002Fwww.simplefeed.com\u002F","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Fai-rights-protection-by-simplefeed.1.0.22.zip",92,null,"2026-03-15T15:16:48.613Z",[],{"slug":32,"display_name":7,"profile_url":8,"plugin_count":14,"total_installs":11,"avg_security_score":27,"avg_patch_time_days":33,"trust_score":34,"computed_at":35},"akhomichenko",30,88,"2026-04-04T00:44:54.605Z",[37,63,81,97,117],{"slug":38,"name":39,"version":40,"author":41,"author_profile":42,"description":43,"short_description":44,"active_installs":45,"downloaded":46,"rating":47,"num_ratings":48,"last_updated":49,"tested_up_to":50,"requires_at_least":51,"requires_php":52,"tags":53,"homepage":58,"download_link":59,"security_score":60,"vuln_count":61,"unpatched_count":11,"last_vuln_date":62,"fetched_at":29},"better-robots-txt","Better Robots.txt – AI-Ready Crawl Control & Bot Governance","3.0.0","Pagup","https:\u002F\u002Fprofiles.wordpress.org\u002Fpagup\u002F","\u003Cp>Better Robots.txt replaces the default WordPress robots.txt workflow with a smarter, structured version you can configure and preview before publishing.\u003C\u002Fp>\n\u003Cp>Instead of a blank textarea, you get a guided wizard with presets, plain-language explanations, and a final Review & Save step so you can inspect the generated robots.txt before it goes live.\u003C\u002Fp>\n\u003Cp>Built for beginners and advanced users alike, Better Robots.txt helps you control how search engines, AI crawlers, SEO tools, archive bots, bad bots, social preview bots, and other automated agents interact with your site.\u003C\u002Fp>\n\u003Cp>Trusted by thousands of WordPress sites, Better Robots.txt is designed for the AI era without resorting to hype, vague promises, or hidden rules.\u003C\u002Fp>\n\u003Cp>Better Robots.txt is available in Free, Pro, and Premium editions. The free plugin covers the guided workflow and essential crawl control features, while Pro and Premium unlock additional governance, protection, and AI-ready modules. Some screenshots on the plugin page show features from all three editions.\u003C\u002Fp>\n\u003Ch3>A quick overview\u003C\u002Fh3>\n\u003Cp>\u003Ciframe loading=\"lazy\" title=\"Better robots.txt Video — AI-Ready Crawl Control for WordPress\" src=\"https:\u002F\u002Fplayer.vimeo.com\u002Fvideo\u002F1169756981?dnt=1&app_id=122963\" width=\"750\" height=\"372\" frameborder=\"0\" allow=\"autoplay; fullscreen; picture-in-picture; clipboard-write; encrypted-media; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\">\u003C\u002Fiframe>\u003C\u002Fp>\n\u003Ch3>Why Better Robots.txt is different\u003C\u002Fh3>\n\u003Cp>Most robots.txt plugins fall into one of three categories:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Simple text editor\u003C\u002Fli>\n\u003Cli>Virtual robots.txt manager\u003C\u002Fli>\n\u003Cli>Single-purpose AI or policy add-on\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>Better Robots.txt goes further.\u003C\u002Fp>\n\u003Cp>It gives you a complete, guided crawl control workflow so you can:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Choose a preset that matches your goals\u003C\u002Fli>\n\u003Cli>Control major crawler categories without writing everything by hand\u003C\u002Fli>\n\u003Cli>Keep core WordPress protection rules visible and editable\u003C\u002Fli>\n\u003Cli>Clean up low-value crawl paths that waste crawl budget\u003C\u002Fli>\n\u003Cli>Generate a cleaner robots.txt output\u003C\u002Fli>\n\u003Cli>Preview the final result before saving\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>What you can control\u003C\u002Fh3>\n\u003Cp>Better Robots.txt helps you manage:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Search engine visibility\u003C\u002Fli>\n\u003Cli>AI and LLM crawler behavior\u003C\u002Fli>\n\u003Cli>AI usage signals such as search, ai-input, and ai-train preferences\u003C\u002Fli>\n\u003Cli>SEO tool crawlers\u003C\u002Fli>\n\u003Cli>Bad bots and abusive crawlers\u003C\u002Fli>\n\u003Cli>Archive and Wayback access\u003C\u002Fli>\n\u003Cli>Feed crawlers and crawl traps\u003C\u002Fli>\n\u003Cli>WooCommerce crawl cleanup\u003C\u002Fli>\n\u003Cli>CSS, JavaScript, and image crawling rules\u003C\u002Fli>\n\u003Cli>Social media preview crawlers\u003C\u002Fli>\n\u003Cli>ads.txt and app-ads.txt allowance\u003C\u002Fli>\n\u003Cli>llms.txt generation\u003C\u002Fli>\n\u003Cli>Advanced directives such as crawl-delay and custom rules\u003C\u002Fli>\n\u003Cli>Final review before publishing\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>Editions\u003C\u002Fh3>\n\u003Cp>Better Robots.txt is available in three editions:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Free – Includes the guided setup, the Essential preset, core crawl control features, and the final Review & Save workflow.\u003C\u002Fli>\n\u003Cli>Pro – Adds more advanced governance and protection modules, including additional AI, crawler, and cleanup controls.\u003C\u002Fli>\n\u003Cli>Premium – Unlocks the most restrictive and advanced protection options, including the Fortress preset and additional high-control modules.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>Some options shown in the interface are marked Free, Pro, or Premium so users can immediately understand which modules belong to each edition.\u003C\u002Fp>\n\u003Ch3>Presets\u003C\u002Fh3>\n\u003Cp>Setup starts with four modes:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Essential – A clean, practical configuration for most websites that want a better robots.txt without complexity.\u003C\u002Fli>\n\u003Cli>AI-First – For publishers and content sites that want AI-ready governance without shutting down discovery.\u003C\u002Fli>\n\u003Cli>Fortress – For websites that want stronger protection against scraping, archive capture, and unnecessary crawl activity.\u003C\u002Fli>\n\u003Cli>Custom – For users who prefer to configure each module manually.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>For many sites, one preset plus a quick review is enough.\u003C\u002Fp>\n\u003Ch3>Built for beginners and experts\u003C\u002Fh3>\n\u003Cp>Beginners get:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>A guided setup instead of a raw robots.txt box\u003C\u002Fli>\n\u003Cli>Preset-based configuration\u003C\u002Fli>\n\u003Cli>Plain-language explanations for important choices\u003C\u002Fli>\n\u003Cli>A safer workflow with a final preview step\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>Advanced users get:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Editable core WordPress protection rules\u003C\u002Fli>\n\u003Cli>Fine-grained crawler controls by category\u003C\u002Fli>\n\u003Cli>WooCommerce-oriented cleanup options\u003C\u002Fli>\n\u003Cli>Consolidated output options\u003C\u002Fli>\n\u003Cli>Advanced directives and custom rules\u003C\u002Fli>\n\u003Cli>A final output they can inspect before publishing\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>AI-ready, without hype\u003C\u002Fh3>\n\u003Cp>Better Robots.txt includes features for modern AI-related crawl governance, including:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>AI crawler handling\u003C\u002Fli>\n\u003Cli>Optional llms.txt support\u003C\u002Fli>\n\u003Cli>AI usage signals for compliant systems\u003C\u002Fli>\n\u003Cli>Optional machine-readable governance signals for advanced use cases\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>These features help you express how you want automated systems to use your content.\u003C\u002Fp>\n\u003Cp>However, Better Robots.txt does not claim to control AI by force. Like robots.txt itself, these signals are most useful with compliant systems and good-faith crawlers.\u003C\u002Fp>\n\u003Ch3>What Better Robots.txt is\u003C\u002Fh3>\n\u003Cp>Better Robots.txt is:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>A robots.txt governance plugin for WordPress\u003C\u002Fli>\n\u003Cli>A guided configuration workflow instead of a raw text editor\u003C\u002Fli>\n\u003Cli>A crawl control layer to reduce wasteful crawling\u003C\u002Fli>\n\u003Cli>A practical bridge between SEO, crawl hygiene, and AI-era policy signaling\u003C\u002Fli>\n\u003Cli>A way to keep your crawl policy clearer for humans and machines\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>Technical reference for advanced users: Better Robots.txt also maintains a public \u003Ca href=\"https:\u002F\u002Fgithub.com\u002FGautierDorval\u002Fbetter-robots-txt\" rel=\"nofollow noopener noreferrer ugc\">GitHub repository\u003C\u002Fa> with product definition, governance notes, and machine-readable artefacts.\u003C\u002Fp>\n\u003Ch3>What Better Robots.txt is not\u003C\u002Fh3>\n\u003Cp>Better Robots.txt is not:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>A firewall or Web Application Firewall (WAF)\u003C\u002Fli>\n\u003Cli>An anti-scraping enforcement engine\u003C\u002Fli>\n\u003Cli>A legal compliance engine\u003C\u002Fli>\n\u003Cli>A guarantee that every bot will obey your rules\u003C\u002Fli>\n\u003Cli>A replacement for server-level security or access control\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>It helps you publish a clearer crawl policy.\u003C\u002Fp>\n\u003Cp>It does not replace infrastructure-level protection.\u003C\u002Fp>\n\u003Ch3>Typical use cases\u003C\u002Fh3>\n\u003Cp>Use Better Robots.txt if you want to:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Clean up a weak or noisy default robots.txt\u003C\u002Fli>\n\u003Cli>Reduce crawl waste on WordPress or WooCommerce\u003C\u002Fli>\n\u003Cli>Keep major search engines allowed while restricting other bots\u003C\u002Fli>\n\u003Cli>Control whether archive bots can snapshot your site\u003C\u002Fli>\n\u003Cli>Publish AI usage preferences more clearly\u003C\u002Fli>\n\u003Cli>Keep social preview bots allowed while limiting scrapers\u003C\u002Fli>\n\u003Cli>Review the final file before making it live\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>Key Features\u003C\u002Fh3>\n\u003Cul>\n\u003Cli>Guided step-by-step wizard\u003C\u002Fli>\n\u003Cli>Preset-based setup: Essential, AI-First, Fortress, Custom\u003C\u002Fli>\n\u003Cli>Search engine visibility controls\u003C\u002Fli>\n\u003Cli>AI and LLM crawler governance\u003C\u002Fli>\n\u003Cli>AI usage signals support\u003C\u002Fli>\n\u003Cli>SEO tool crawler controls\u003C\u002Fli>\n\u003Cli>Bad bot and abusive crawler options\u003C\u002Fli>\n\u003Cli>Archive and Wayback access controls\u003C\u002Fli>\n\u003Cli>Spam, feed, and crawl trap cleanup\u003C\u002Fli>\n\u003Cli>WooCommerce crawl cleanup options\u003C\u002Fli>\n\u003Cli>CSS, JavaScript, and image crawling rules\u003C\u002Fli>\n\u003Cli>Social media preview crawler controls\u003C\u002Fli>\n\u003Cli>ads.txt and app-ads.txt allowance\u003C\u002Fli>\n\u003Cli>Optional llms.txt generation\u003C\u002Fli>\n\u003Cli>Consolidated output option\u003C\u002Fli>\n\u003Cli>Core WordPress protection rules remain visible and editable\u003C\u002Fli>\n\u003Cli>Final Review & Save preview screen\u003C\u002Fli>\n\u003C\u002Ful>\n","Replace the default WordPress robots.txt workflow with a smarter, structured version you can preview before publishing, with Free, Pro, and Premium ed &hellip;",6000,305034,90,102,"2026-03-10T18:33:00.000Z","6.9.4","5.0","7.4",[54,55,56,24,57],"ai-crawlers","bot-blocker","llms-txt","seo","","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Fbetter-robots-txt.3.0.0.zip",99,2,"2023-02-14 00:00:00",{"slug":64,"name":65,"version":66,"author":67,"author_profile":68,"description":69,"short_description":70,"active_installs":71,"downloaded":72,"rating":73,"num_ratings":74,"last_updated":75,"tested_up_to":50,"requires_at_least":76,"requires_php":77,"tags":78,"homepage":58,"download_link":80,"security_score":13,"vuln_count":11,"unpatched_count":11,"last_vuln_date":28,"fetched_at":29},"block-ai-crawlers","Block AI Crawlers","1.5.6","lastsplash (a11n)","https:\u002F\u002Fprofiles.wordpress.org\u002Flastsplash\u002F","\u003Cp>Protect Your Content from AI Scraping\u003C\u002Fp>\n\u003Cp>This plugin helps you prevent AI crawlers from using your content as training data for their products. By updating your site’s \u003Ccode>robots.txt\u003C\u002Fcode>, it blocks common AI crawlers and scrapers, aiming to protect your content from being used in the training of Large Language Models (LLMs).\u003C\u002Fp>\n\u003Ch3>Features\u003C\u002Fh3>\n\u003Ch3>Blocks AI Crawlers\u003C\u002Fh3>\n\u003Cp>Includes:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>\u003Cstrong>OpenAI\u003C\u002Fstrong> – Blocks crawlers used for ChatGPT\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Google\u003C\u002Fstrong> – Blocks crawlers used by Google’s Gemini AI products\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Facebook \u002F Meta\u003C\u002Fstrong> – Used for Facebook’s AI training\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Anthropic AI\u003C\u002Fstrong> – Blocks crawlers used by Anthropic  \u003C\u002Fli>\n\u003Cli>\u003Cstrong>Perplexity\u003C\u002Fstrong> – Block crawlers used by Perplexity\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Applebot\u003C\u002Fstrong> – Blocks crawlers used by Apple\u003C\u002Fli>\n\u003Cli>… and more!\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>Experimental Meta Tags\u003C\u002Fh3>\n\u003Cp>The plugin adds the “noai, noimageai” directive to your site’s meta tags, instructing AI bots not to use your content in their datasets. Please note that these tags are experimental and have not been standardized.\u003C\u002Fp>\n\u003Ch3>Custom robots.txt Rules\u003C\u002Fh3>\n\u003Cp>Have custom entries for your robots.txt file? You can now add them directly through the plugin!\u003C\u002Fp>\n\u003Ch3>Usage\u003C\u002Fh3>\n\u003Cp>After activation, the plugin will automatically update your \u003Ccode>robots.txt\u003C\u002Fcode> and add the necessary meta tags. No further configuration is required, but you can check the settings page for a full list of blocked crawlers.\u003C\u002Fp>\n\u003Ch3>Limitations\u003C\u002Fh3>\n\u003Cp>While this plugin aims to block specified crawlers, it cannot guarantee complete protection against all forms of scraping, as some bots may disregard \u003Ccode>robots.txt\u003C\u002Fcode> directives.\u003C\u002Fp>\n\u003Ch3>Support\u003C\u002Fh3>\n\u003Cp>For questions or support, \u003Ca href=\"https:\u002F\u002Fwordpress.org\u002Fsupport\u002Fplugin\u002Fblock-ai-crawlers\u002F\" rel=\"ugc\">please post on the forums\u003C\u002Fa> or \u003Ca href=\"https:\u002F\u002Fgithub.com\u002Fbobmatyas\u002Fwp-block-ai-crawlers\u002Fissues\" rel=\"nofollow ugc\">on GitHub\u003C\u002Fa>.\u003C\u002Fp>\n","Tell AI (Artificial Intelligence) companies not to scrape your site for their AI products.",1000,13412,96,5,"2026-02-15T13:47:00.000Z","6.8","8.2",[20,79,23,24],"chatgpt","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Fblock-ai-crawlers.1.5.6.zip",{"slug":82,"name":83,"version":84,"author":85,"author_profile":86,"description":87,"short_description":88,"active_installs":89,"downloaded":90,"rating":13,"num_ratings":91,"last_updated":92,"tested_up_to":50,"requires_at_least":51,"requires_php":52,"tags":93,"homepage":95,"download_link":96,"security_score":13,"vuln_count":11,"unpatched_count":11,"last_vuln_date":28,"fetched_at":29},"ai-content-signals","AI Content Signals","1.0.1","Fernando Tellado","https:\u002F\u002Fprofiles.wordpress.org\u002Ffernandot\u002F","\u003Cp>AI Content Signals allows you to easily implement the Content Signals Policy in your WordPress site’s robots.txt file. This gives you more control over how AI crawlers and large language models (LLMs) can use your content.\u003C\u002Fp>\n\u003Cp>\u003Cstrong>What are Content Signals?\u003C\u002Fstrong>\u003C\u002Fp>\n\u003Cp>Content Signals is an extension to the robots.txt standard created by Cloudflare that lets you specify three types of permissions for AI crawlers:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>\u003Cstrong>search\u003C\u002Fstrong> – Allow or deny search indexing and traditional search results\u003C\u002Fli>\n\u003Cli>\u003Cstrong>ai-input\u003C\u002Fstrong> – Allow or deny using your content for real-time AI responses (RAG, AI Overviews)\u003C\u002Fli>\n\u003Cli>\u003Cstrong>ai-train\u003C\u002Fstrong> – Allow or deny using your content for training AI models\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>\u003Cstrong>Key Features\u003C\u002Fstrong>\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Easy-to-use settings page in WordPress admin\u003C\u002Fli>\n\u003Cli>Set global defaults for all crawlers\u003C\u002Fli>\n\u003Cli>Configure specific settings for individual AI bots (GPTBot, ClaudeBot, PerplexityBot, etc.)\u003C\u002Fli>\n\u003Cli>Add custom bot User-Agents\u003C\u002Fli>\n\u003Cli>Supports both physical and virtual robots.txt files\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Option to create physical robots.txt with basic WordPress rules\u003C\u002Fstrong>\u003C\u002Fli>\n\u003Cli>Preview generated Content Signals before applying\u003C\u002Fli>\n\u003Cli>Optional legal text with EU Directive reference\u003C\u002Fli>\n\u003Cli>Works with existing robots.txt from SEO plugins\u003C\u002Fli>\n\u003Cli>Automatic sitemap detection and inclusion\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>\u003Cstrong>Supported Bots\u003C\u002Fstrong>\u003C\u002Fp>\n\u003Cp>The plugin includes predefined settings for major AI crawlers:\u003C\u002Fp>\n\u003Cul>\n\u003Cli>OpenAI GPTBot and ChatGPT-User\u003C\u002Fli>\n\u003Cli>Anthropic ClaudeBot and Claude-Web\u003C\u002Fli>\n\u003Cli>Perplexity Bot\u003C\u002Fli>\n\u003Cli>Google Extended (Bard\u002FGemini)\u003C\u002Fli>\n\u003Cli>Common Crawl Bot\u003C\u002Fli>\n\u003Cli>Meta\u002FFacebook Bot\u003C\u002Fli>\n\u003Cli>And many more…\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Cp>\u003Cstrong>Important Notice\u003C\u002Fstrong>\u003C\u002Fp>\n\u003Cp>Content Signals is a declarative standard – it expresses your preferences but does not technically enforce them. AI companies are not legally required to respect these signals, though the plugin includes legal text referencing EU copyright directives.\u003C\u002Fp>\n\u003Cp>This plugin works best when combined with other protection measures like traditional robots.txt rules and server-level bot management.\u003C\u002Fp>\n","Add Content Signals to your robots.txt to control how AI crawlers can use your content.",200,605,3,"2025-12-28T17:52:00.000Z",[20,94,23,24,57],"cloudflare","https:\u002F\u002Fservicios.ayudawp.com","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Fai-content-signals.1.0.1.zip",{"slug":98,"name":99,"version":100,"author":101,"author_profile":102,"description":103,"short_description":104,"active_installs":71,"downloaded":105,"rating":13,"num_ratings":61,"last_updated":106,"tested_up_to":107,"requires_at_least":108,"requires_php":58,"tags":109,"homepage":113,"download_link":114,"security_score":115,"vuln_count":14,"unpatched_count":14,"last_vuln_date":116,"fetched_at":29},"robotstxt-rewrite","Robots.txt rewrite","1.6.1","Eugen Bobrowski","https:\u002F\u002Fprofiles.wordpress.org\u002Feugenbobrowski\u002F","\u003Cp>Plugin provide to help search engines to indexing site correctly.\u003C\u002Fp>\n\u003Cp>A simple plugin to manage your robots.txt. Plugin donn’t create the file or edit it. This plugin edit WordPress output of robots.txt content. And get you a easy and usable interface to manage it.\u003C\u002Fp>\n\u003Cp>\u003Cstrong>Features\u003C\u002Fstrong>\u003C\u002Fp>\n\u003Cul>\n\u003Cli>Drag-n-drop robots.txt paths\u003C\u002Fli>\n\u003Cli>Changing \u003Ccode>blog_public\u003C\u002Fcode> option form plugin settings page\u003C\u002Fli>\n\u003Cli>Site map field for robots.txt\u003C\u002Fli>\n\u003Cli>Robots.txt physical file checking.\u003C\u002Fli>\n\u003C\u002Ful>\n","Provide the easy managment of your robots.txt from admin side. It propose you the advanced then standard robots.txt content too.",35511,"2017-11-28T17:32:00.000Z","4.7.32","4.7",[110,23,111,112,24],"crawler","robot","robots","http:\u002F\u002Fwordpress.org\u002Fplugins\u002Frobotstxt-rewrite\u002F","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Frobotstxt-rewrite.1.6.1.zip",63,"2025-12-31 00:00:00",{"slug":118,"name":119,"version":120,"author":121,"author_profile":122,"description":123,"short_description":124,"active_installs":125,"downloaded":126,"rating":73,"num_ratings":127,"last_updated":128,"tested_up_to":50,"requires_at_least":51,"requires_php":129,"tags":130,"homepage":135,"download_link":136,"security_score":13,"vuln_count":11,"unpatched_count":11,"last_vuln_date":28,"fetched_at":29},"dark-visitors","Known Agents – Track AI Bots and Crawlers, Block Scrapers, Analyze LLM Referral Traffic","1.28.0","gavindarkvisitors","https:\u002F\u002Fprofiles.wordpress.org\u002Fgavindarkvisitors\u002F","\u003Cp>Track crawlers, scrapers, LLM assistants, and AI agents on your website. Generate a robots.txt that blocks AI bots. Formerly Dark Visitors.\u003C\u002Fp>\n\u003Ch3>Track, control, and optimize your website for AI agents and bots\u003C\u002Fh3>\n\u003Cp>You have \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002Finsights\" rel=\"nofollow ugc\">40% more visitors than you think\u003C\u002Fa>. Protect your content from unwanted AI scraping and crawling, and turn the rising wave of AI agents, LLM assistants, and other bots crawling your website into a new growth channel for your business. For full details, visit the \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002F\" rel=\"nofollow ugc\">Known Agents\u003C\u002Fa> website.\u003C\u002Fp>\n\u003Ch3>1. Agent Analytics\u003C\u002Fh3>\n\u003Cp>Get realtime insight into the hidden ecosystem of crawlers, scrapers, and AI agents browsing your website.\u003C\u002Fp>\n\u003Cul>\n\u003Cli>\u003Cstrong>Monitor traffic from crawlers and scrapers\u003C\u002Fstrong>: Track the hidden activity of \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002Fagents\" rel=\"nofollow ugc\">all known artificial agents\u003C\u002Fa> visiting your website in real time.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Analyze how AI agents browse your website\u003C\u002Fstrong>: Review sessions from AI models autonomously navigating and interacting with your website. Optimize their experience to increase conversions and human referrals.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Detect bots spoofing legitimate agents\u003C\u002Fstrong>: Identify bad actors and unauthorized scrapers pretending to be legitimate agents like Googlebot or Bingbot.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Get alerted to bot traffic spikes\u003C\u002Fstrong>: Stay informed when an agent’s traffic starts surging or overloading your server.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>See which agents visit most frequently and which pages they target\u003C\u002Fstrong>: Break down bot traffic by agent and URL to understand where crawlers and scrapers are focused.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Export reports to CSV\u003C\u002Fstrong>: Download agent activity reports for custom analysis.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>2. LLM Referral Tracking\u003C\u002Fh3>\n\u003Cp>Measure human conversions from AI chat and search platforms like ChatGPT, Perplexity, Gemini, Claude, and more.\u003C\u002Fp>\n\u003Cul>\n\u003Cli>\u003Cstrong>Find out which AI platforms are sending you traffic\u003C\u002Fstrong>: See which LLMs are recommending your website and measure how many humans click through to visit.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Export reports to CSV\u003C\u002Fstrong>: Download LLM referral reports for custom analysis.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>3. Automatic Robots.txt\u003C\u002Fh3>\n\u003Cp>Protect IP, reduce server cost, and save time by serving a robots.txt that updates continuously.\u003C\u002Fp>\n\u003Cul>\n\u003Cli>\u003Cstrong>Block crawlers and scrapers by category\u003C\u002Fstrong>: Control access for entire agent categories including AI crawlers, scrapers, SEO bots, and more. One selection covers every current and future agent in that group.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Stop tracking new bots manually\u003C\u002Fstrong>: Your robots.txt is updated as new agents emerge, so you never have to manage them yourself or make manual edits.\u003C\u002Fli>\n\u003Cli>\u003Cstrong>Enforce robots.txt rules with the WordPress plugin\u003C\u002Fstrong>: The WordPress plugin detects and blocks bots that ignore your robots.txt rules, protecting your content from unauthorized access.\u003C\u002Fli>\n\u003C\u002Ful>\n\u003Ch3>Demos\u003C\u002Fh3>\n\u003Cp>Demos can be found on the \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002F\" rel=\"nofollow ugc\">home page\u003C\u002Fa>.\u003C\u002Fp>\n\u003Ch3>Further Details\u003C\u002Fh3>\n\u003Cp>This plugin connects your website to \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002F\" rel=\"nofollow ugc\">Known Agents\u003C\u002Fa>. It uses the Known Agents API to connect to your account and provide functionality.\u003C\u002Fp>\n\u003Cp>By signing up, you agree to the \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002Fterms-of-service\" rel=\"nofollow ugc\">Terms of Service\u003C\u002Fa> and \u003Ca href=\"https:\u002F\u002Fknownagents.com\u002Fprivacy-policy\" rel=\"nofollow ugc\">Privacy Policy\u003C\u002Fa>.\u003C\u002Fp>\n\u003Cp>Are you finding this plugin useful? Please \u003Ca href=\"https:\u002F\u002Fwordpress.org\u002Fsupport\u002Fplugin\u002Fdark-visitors\u002Freviews\u002F#new-post\" rel=\"ugc\">leave us a ★★★★★ review\u003C\u002Fa>! Otherwise, please \u003Ca href=\"https:\u002F\u002Fwordpress.org\u002Fsupport\u002Fplugin\u002Fdark-visitors\u002F\" rel=\"ugc\">ask us for help\u003C\u002Fa>.\u003C\u002Fp>\n","Track crawlers, scrapers, LLM assistants, and AI agents on your website. Generate a robots.txt that blocks AI bots. Formerly Dark Visitors.",900,15728,6,"2026-02-24T00:40:00.000Z","7.0",[131,132,133,24,134],"detect","firewall","monitor","training","https:\u002F\u002Fknownagents.com\u002F","https:\u002F\u002Fdownloads.wordpress.org\u002Fplugin\u002Fdark-visitors.1.28.0.zip",{"attackSurface":138,"codeSignals":212,"taintFlows":227,"riskAssessment":254,"analyzedAt":266},{"hooks":139,"ajaxHandlers":182,"restRoutes":199,"shortcodes":200,"cronEvents":201,"entryPointCount":211,"unprotectedCount":211},[140,147,152,156,159,162,166,169,174,179],{"type":141,"name":142,"callback":143,"priority":144,"file":145,"line":146},"action","robots_txt","closure",10,"includes\\classes\\SFAIRP_SimpleFeedBotsProtection.php",906,{"type":141,"name":148,"callback":149,"priority":150,"file":145,"line":151},"template_redirect","renderSitemap",11,911,{"type":153,"name":154,"callback":143,"file":145,"line":155},"filter","query_vars",912,{"type":141,"name":157,"callback":143,"priority":14,"file":145,"line":158},"do_robotstxt",918,{"type":141,"name":142,"callback":143,"priority":160,"file":145,"line":161},9999,921,{"type":141,"name":163,"callback":164,"priority":144,"file":145,"line":165},"simplefeed_boots_protection_sitemap_sync","syncAsync",936,{"type":141,"name":167,"callback":143,"file":145,"line":168},"admin_menu",942,{"type":141,"name":170,"callback":171,"priority":144,"file":172,"line":173},"admin_enqueue_scripts","enqueueAdminScripts","includes\\classes\\SFAIRP_SimpleFeedBotsProtectionAjax.php",29,{"type":141,"name":175,"callback":176,"file":177,"line":178},"admin_init","registerSettings","includes\\classes\\SFAIRP_SimpleFeedBotsProtectionSettings.php",36,{"type":141,"name":180,"callback":164,"priority":144,"file":177,"line":181},"simplefeed_boots_protection_settings_sync",39,[183,188,192,196],{"action":184,"nopriv":185,"callback":186,"hasNonce":185,"hasCapCheck":185,"file":172,"line":187},"wp_simplefeed_bots_protection_history",false,"handleHistory",26,{"action":189,"nopriv":185,"callback":190,"hasNonce":185,"hasCapCheck":185,"file":172,"line":191},"wp_simplefeed_bots_protection_history_action_enable","handleHistoryActionEnable",27,{"action":193,"nopriv":185,"callback":194,"hasNonce":185,"hasCapCheck":185,"file":172,"line":195},"wp_simplefeed_bots_protection_history_action_disable","handleHistoryActionDisable",28,{"action":197,"nopriv":185,"callback":198,"hasNonce":185,"hasCapCheck":185,"file":172,"line":33},"wp_simplefeed_bots_protection_log","handleLog",[],[],[202,204,206,208,209],{"hook":163,"callback":163,"file":145,"line":203},794,{"hook":163,"callback":163,"file":145,"line":205},851,{"hook":163,"callback":163,"file":145,"line":207},872,{"hook":180,"callback":180,"file":177,"line":27},{"hook":180,"callback":180,"file":177,"line":210},115,4,{"dangerousFunctions":213,"sqlUsage":214,"outputEscaping":223,"fileOperations":11,"externalRequests":91,"nonceChecks":11,"capabilityChecks":11,"bundledLibraries":226},[],{"prepared":150,"raw":91,"locations":215},[216,220,222],{"file":217,"line":218,"context":219},"includes\\classes\\SFAIRP_SimpleFeedBotsProtectionDatabase.php",66,"$wpdb->query() with variable interpolation",{"file":217,"line":221,"context":219},71,{"file":217,"line":47,"context":219},{"escaped":224,"rawEcho":11,"locations":225},60,[],[],[228,246],{"entryPoint":229,"graph":230,"unsanitizedCount":14,"severity":245},"registerSettings (includes\\classes\\SFAIRP_SimpleFeedBotsProtectionSettings.php:86)",{"nodes":231,"edges":243},[232,237],{"id":233,"type":234,"label":235,"file":177,"line":236},"n0","source","$_POST",104,{"id":238,"type":239,"label":240,"file":177,"line":241,"wp_function":242},"n1","sink","update_option() [Settings Manipulation]",106,"update_option",[244],{"from":233,"to":238,"sanitized":185},"low",{"entryPoint":247,"graph":248,"unsanitizedCount":14,"severity":245},"\u003CSFAIRP_SimpleFeedBotsProtectionSettings> (includes\\classes\\SFAIRP_SimpleFeedBotsProtectionSettings.php:0)",{"nodes":249,"edges":252},[250,251],{"id":233,"type":234,"label":235,"file":177,"line":236},{"id":238,"type":239,"label":240,"file":177,"line":241,"wp_function":242},[253],{"from":233,"to":238,"sanitized":185},{"summary":255,"deductions":256},"The plugin exhibits several concerning security practices, particularly regarding its attack surface. A significant number of AJAX handlers (4 out of 4) lack authentication checks, creating a substantial entry point for potential attackers. This is further exacerbated by the absence of any nonce checks or capability checks in the code signals. While the plugin demonstrates good practices in SQL query preparation (79%) and output escaping (100%), the lack of authentication on critical entry points overshadows these strengths. The absence of any recorded vulnerabilities or CVEs in its history might suggest a lack of past exploitation or a relatively new plugin, but it does not negate the inherent risks posed by its current code. The taint analysis indicating unsanitized paths is a red flag, even without critical or high severity findings. Overall, the plugin has a weak security posture due to its exposed AJAX endpoints, requiring immediate attention.",[257,260,262,264],{"reason":258,"points":259},"4 unprotected AJAX handlers",20,{"reason":261,"points":144},"0 nonce checks found",{"reason":263,"points":144},"0 capability checks found",{"reason":265,"points":144},"2 flows with unsanitized paths","2026-03-17T06:11:32.980Z",{"wat":268,"direct":279},{"assetPaths":269,"generatorPatterns":273,"scriptPaths":274,"versionParams":275},[270,271,272],"\u002Fwp-content\u002Fplugins\u002Fai-rights-protection-by-simplefeed\u002Fincludes\u002Fcss\u002Fstyle.css","\u002Fwp-content\u002Fplugins\u002Fai-rights-protection-by-simplefeed\u002Fincludes\u002Fjs\u002Fscript.js","\u002Fwp-content\u002Fplugins\u002Fai-rights-protection-by-simplefeed\u002Fincludes\u002Fjs\u002Face@1.4.12.js",[],[271,272],[276,277,278],"ai-rights-protection-by-simplefeed\u002Fincludes\u002Fcss\u002Fstyle.css?ver=","ai-rights-protection-by-simplefeed\u002Fincludes\u002Fjs\u002Fscript.js?ver=","ai-rights-protection-by-simplefeed\u002Fincludes\u002Fjs\u002Face@1.4.12.js?ver=",{"cssClasses":280,"htmlComments":281,"htmlAttributes":283,"restEndpoints":284,"jsGlobals":289,"shortcodeOutput":291},[],[282],"\u003C!-- This plugin is licensed software for the use of SimpleFeed customers only. Copyright ©2004 - 2024 SimpleFeed, Inc. All Rights Reserved. Protected by Patents 8065383 and 8661001. -->",[],[285,286,287,288],"\u002Fwp-json\u002Fwp_simplefeed_bots_protection_history","\u002Fwp-json\u002Fwp_simplefeed_bots_protection_history_action_enable","\u002Fwp-json\u002Fwp_simplefeed_bots_protection_history_action_disable","\u002Fwp-json\u002Fwp_simplefeed_bots_protection_log",[290],"MyAjax",[]]