robots-txt | Check user agents against robots.txt files | Sitemap library
kandi X-RAY | robots-txt Summary
kandi X-RAY | robots-txt Summary
Check user agents against robots.txt files
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of robots-txt
robots-txt Key Features
robots-txt Examples and Code Snippets
Community Discussions
Trending Discussions on robots-txt
QUESTION
The site loses all functionalities after building it. In develop mode everything works fine, but when I build the website it looks like all scripts are missing. Bootstrap (Carousel DropDowns) are not responding, leflet map and background image not loading and react-multi-carousel do not work. I don't see any errors in the browser console, of course I ran gatsby clean
before building. I uploaded the project to netlify. Below I am enclosing the json package:
ANSWER
Answered 2021-Apr-13 at 20:59There's not much debug in the question however, to me, it seems that you are using some dependencies outside React's scope, which may potentially block React's hydration process, which may explain the problem described. For example, using Leaflet instead of React's Leaflet or (depending on its implementation) Bootstrap instead of React's Boostrap.
You should be able to change all React-unfriendly modules to React's ones without too much effort and that should fix your problems.
Keep in mind that if your project "work in develop
and breaks in build
" doesn't mean that your project work or stops working, it just means that is working under certain and specific circumstances. Basically, and summarizing (to avoid extending the answer), gatsby develop
uses the browser as an interpreter, where there are, among other things, global objects like window
or document
. However, gatsby build
occurs in the Node server, where at the build time, there are not global objects because there are not even created yet, that the main reason why you may face a different behavior between both scenarios but doesn't mean that the project stopped working magically.
You can read further details in the Overview of Gatsby Build Process.
Another option, linked with blocking React's hydration, is that some component may be blocking that process because of its own behavior. Be careful when using global objects like window
or document
(or when importing modules that uses it), they use should be always be wrapped inside the following condition, as you can see from the docs:
When referencing window in a React component.
QUESTION
I have a Gatsby site that has been running smoothly for 3 months online. As of Friday 24th July I have started to receive the below result and users only see a blank screen.
...ANSWER
Answered 2020-Dec-14 at 05:47What did you try so far? As @ksav pointed out, in this GitHub thread there are a several ways to fix a similar issue:
- Removing
node_modules
,.cache
and install again - Removing
node_modules
,.cache
fix Gatsby tov2.23.3
/upgrade up to^2.26.1
where the bug is fixed and install again
It seems related to a loading staticQuery bug that can't be reproduced in a fresh install. The final trial is to remove your package-lock
/yarn-lock.json
and generate it again.
QUESTION
ReadTheDocs auto-generates a robots.txt
and sitemap.xml
for projects. Each time I deploy a new minor version of my project (ex. 4.1.10
), I hide previous minor versions (ex. 4.1.9
). ReadTheDocs adds entries for all versions to sitemap.xml
, but hidden versions are also added to robots.txt
. The result is that submitted sitemaps to Google Search Console, at this point, result in "Submitted URL blocked by robots.txt" errors, since the previous sitemap entry is now blocked by the newly generated robots.txt
.
ReadTheDocs generates a sitemap URL for each version, so we have an entry like this for 4.1.9
, for example:
ANSWER
Answered 2020-Aug-25 at 15:38After playing around with a few ideas, here is the solution I came other with. Since this question is asked frequently and often opened as a bug against ReadTheDocs on GitHub (which it's not, it just appears to be poorly supported and/or documented), I'll share my workaround here for others to find.
As mentioned above and in the docs, while ReadTheDocs allows you to override the auto-generated robots.txt
and publish your own, you can't with sitemap.xml
. Unclear why. Regardless, you can simply publish a different sitemap.xml
, I named mine sitemap-index.xml
, then, tell your robots.txt
to point to your custom sitemap.
For my custom sitemap-index.xml
, I only put the pages I care about rather then ever generated version (since stable
and latest
are really what I want search engines to be crawling, not versioned pages):
QUESTION
I am using gatsby. All was working fine recently until I ran npm update
as I wanted to ensure I was up to date. Since then I am getting white pages I navigate to with this error
I believe this error is only occurring as the page is not loading and is not the root cause. I am looking to correct the root cause of the page not loading.
Looking around it seemed like this could be a service worker problem So I removed the service works as per gatsby guides but no luck.
The error only occurs when navigating to a page, for example
...ANSWER
Answered 2020-Jun-09 at 15:22Smoking Gun was a plugin I was using
"embla-carousel" "embla-carousel-react"
Rolled them back and issues went away, raising an issue on github for the team
QUESTION
I am revamping a blog to Gatsby which is lightning fast, everything seems perfect but I am facing different sort of issue, as the images I have via Netlify CMS aren't appearing properly in the blog, the images are appearing blur. I don't know what is going wrong here.
Here is the example of the problem statement
here is the excerpt of my gatsby-config.js.
...ANSWER
Answered 2020-Mar-24 at 06:16I am more interested in how do you call those images in your components rather than in the package.json
(it doesn't seem a dependencies issue) because inspecting the code, it seems that you've added the /static
path which is not required. As it is shown in the following screenshot:
Regarding the updates coming from the comments below, we've figured out that the issue is related directly to this GitHub issue where apparently images retrieved by a markdown are blurring. What solves the issue is to pass a withWebp
parameter in Gatsby's configuration, so in gatsby-config.js
:
QUESTION
I can’t resolve issue with building gatsby site on netlify. Since a few days I’m getting following error:
...ANSWER
Answered 2020-Jan-03 at 22:31I had the same issue. Updating gatsby-plugin-netlify-cms to latest version helped.
QUESTION
I have a Magento 2 website with two stores. At the moment, I can edit the global website and his content is applied to both stores.
What I want to de is replace that behaviour in order to get one robot.txt file by store.
But I really have no idea how I should do that.
Currently, if I go to the back office Content > design > Configuration > (Store Edit) > Search Engine Robots
All the fields are disabled in the stores and can't be modified
But If I go on the global Content > design > Configuration > (Global Edit) > Search Engine Robots
, of course, I can modify.
I also have 3 robots.txt files on my storage, but none of them seems to be matching the information saved in the global search engine robots configuration
src/robots.txt
src/app/design/frontend/theme1/jeunesse/robots.txt
src/app/design/frontend/theme2/jeunesse/robots.txt
I found these two links...but none of them helped me : https://inchoo.net/online-marketing/editing-robots-txt-in-magento-2-admin/ and https://support.hypernode.com/knowledgebase/create-robots-txt-magento-2/
The first one tells me that If I have a robots.txt on my storage it should override the configurations...but looks like no considering I have robots file and they aren't showing when I go to website/robots.txt
. I only find again the one in the global configuration.
The second one tells that saving the configuration should save the robots.txt file on the storage...but once again...that's not what is happening.
Thanks for your help, let me know if there is pieces of code I can show ? I really don't know which one at this point.
...ANSWER
Answered 2020-Jan-28 at 08:28I'm the author of the first link. It's a 2 years old article, Magento 2 has since then introduced a few improvements to the built in robots.txt functionality.
The robots.txt content you save under Content > Design > Configuration has a "website" scope. Meaning you can edit it on website level and if you need it to vary through this config you can do it if you have multiple websites.
It is unclear from the question itself if you have multiple websites or if you have set-up multiple stores and/or storeviews under the same website.
QUESTION
I saw this but it's obsolete. I tried the following:
Created src/main/resources/static/robots.txt
.
ANSWER
Answered 2018-Aug-27 at 15:19I gave up on WebMvcConfigurerAdapter
. I think controller requests have priority over resource handlers. I just used this in the controller.
QUESTION
I want to read a path to an image file from YAML, and use the gatsby-image
to create responsive images, but it doesn't let me do what I want.
data.yaml
...ANSWER
Answered 2019-Sep-17 at 20:42This is likely occurring because Gatsby is inferring your profile.image
field as a String instead of a File. This can happen if one or more of the provided path strings does not resolve to a file when you boot Gatsby. Note that Gatsby will not rerun type-inference for existing fields after it boots, so you will need to restart the development server to pick up these changes.
QUESTION
I've searched but haven't found quite what I need. What I'm trying to do is noindex a plugin folders. Google for some reason has indexed a bunch of urls like this for example: /wp-content/plugins/LayerSlider/static/public
All that pulls up is a "Index of" page, so I'd like to get that and urls like it noindexed. I don't want to just block them with the robots file though because after reading the following article, it seems like a bad idea: WordPress robots.txt example for great SEO
There's also this to backup why blocking any css and js is a bad idea: Google Panda 4, and blocking your CSS & JS
There's no index file or anything like that, so I'm wondering if there's a way to get these noindexed. Perhaps the htaccess file?
...ANSWER
Answered 2017-Nov-21 at 00:16You can use .htaccess file to add X-Robots-Tag
header.
For example create new .htaccess file inside plugins directory and put this code inside it will add the header to all the files inside the folder.
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install robots-txt
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page