robots.txt or a bot-blocking rule on your server is denying the crawler.
This page covers failures whose error message or details mention that the social network could not download the media, typically referencing
facebookexternalhit, robots.txt, "Restricted by robots.txt", "HTTP error code 403", or Meta error 2207052. For aspect-ratio or format errors on Instagram code 138, see Instagram Media Guidelines or Threads Media Guidelines instead.Symptom
When the crawler is blocked, you’ll see errors like these:Error 440 (primary)
Instagram Error 138 (fallback — less specific upstream response)
Threads Error 379
Code 440 is the dedicated Ayrshare code for this failure and its message explicitly names
facebookexternalhit and robots.txt — if you see 440, you’re on the right page. Code 138 is emitted for the same root cause when the upstream response is less specific; 138 is also used for aspect-ratio / format issues, so the media-fetch variant is identifiable by "Restricted by robots.txt" or "HTTP error code 403" in details. Code 379 does not include a details field — if Threads fails alongside an Instagram 440 or 138, the root cause is typically the same.Why This Happens
When you publish to Instagram or Threads via Ayrshare, Meta’s servers fetch your media from the URL you provide. This server-side fetch uses thefacebookexternalhit User-Agent. If your server’s robots.txt disallows this crawler — or a WAF/bot-protection rule blocks it — Meta cannot download the file and the publish fails.
Facebook Page publishing uses a different ingestion path, which is why the same mediaUrl may work for Facebook but fail for Instagram and Threads.
Fix: Update Your robots.txt
Recommended: Allow Meta explicitly, keep others open
Add these rules to yourrobots.txt file:
robots.txt
Advanced: Lock down to social publishers only
If you want to block most crawlers but allow social media platforms:robots.txt
Verify Meta Can Fetch Your URL
Before retrying your post, verify that Meta’s crawler can now access your media. Run this command, replacing$URL with your full media URL:
- Healthy response: HTTP 200 or 206 with binary data in the body.
- Blocked response: HTTP 403 or an empty/HTML error page.
Per Meta’s documentation,
robots.txt changes may take up to 24 hours to propagate through Meta’s crawler cache. If verification succeeds but your post still fails, wait and retry later.If This Doesn’t Fix It
If you’ve updatedrobots.txt and verified with the curl command but still see failures:
- 24-hour propagation delay — Meta caches
robots.txt. Wait up to 24 hours after making changes before retrying. - WAF or bot-fight rules — Cloudflare Bot Fight Mode, AWS WAF managed bot rule groups, and similar services may block Meta’s crawler IP ranges even if
robots.txtallows it. Check your WAF logs and add an exception forfacebookexternalhit. - Hotlink protection / Referer checks — Some CDNs block requests from data-center IPs or without a valid
Refererheader. Whitelist Meta’s crawler or disable hotlink protection for media paths. - Signed-URL / presigned-URL expiry — If your media URL has an expiration timestamp (common with S3 presigned URLs), ensure it doesn’t expire before Meta’s crawler can fetch it. For scheduled posts, generate URLs that remain valid until well after the scheduled time.
- Managed media hosting — If you use a service like Cloudinary, Imgix, or similar where you cannot edit
robots.txt, check their documentation for a Meta/Facebook crawler allow-list setting.
- The failing
postIdfrom the error response - The output of the
curlverification command above - Your
robots.txtcontents
