Robots.txt Generator API

Category: security

Generate robots.txt files with custom rules for search engine crawlers. Runs entirely in the browser.

Endpoint

N/A Browser API
No Server API

Response

Content-Type: N/A

This tool generates robots.txt content in the browser using form inputs. No server-side API is available.

// Browser-only: generates text from form inputs

cURL Example

# No cURL equivalent — this tool runs in the browser

Code Samples

// Generate robots.txt content
function generateRobotsTxt(rules) {
    let content = '';
    rules.forEach(rule => {
        content += `User-agent: ${rule.userAgent}\n`;
        rule.disallow.forEach(path => content += `Disallow: ${path}\n`);
        rule.allow.forEach(path => content += `Allow: ${path}\n`);
        content += '\n';
    });
    if (sitemapUrl) content += `Sitemap: ${sitemapUrl}\n`;
    return content;
}
# Generate robots.txt in Python:
robots_txt = """User-agent: *
Disallow: /admin/
Disallow: /private/
Allow: /

Sitemap: https://example.com/sitemap.xml"""

with open('robots.txt', 'w') as f:
    f.write(robots_txt)
// Generate robots.txt in Java:
String robotsTxt = "User-agent: *\n"
    + "Disallow: /admin/\n"
    + "Allow: /\n\n"
    + "Sitemap: https://example.com/sitemap.xml\n";
Files.writeString(Path.of("robots.txt"), robotsTxt);