English Google SEO office-hours from December 10, 2021

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
JOHN MUELLER: All right. Welcome, everyone, to today's Google Search Central SEO Office Hours Hangout. My name is John Mueller. I'm a Search Advocate at Google in Switzerland. And part of what I do are these office hour hangouts where people can join in and ask their questions around their websites and web search, and we can try to find some answers. Looks like a lot of stuff was submitted on YouTube as well, but a bunch of people are already jumping in and raising their hands. So let's get started with some live questions. Let's see, Praveen, I think you're first up. PRAVEEN SHARMA: Hi, John. How are you? JOHN MUELLER: Hi. PRAVEEN SHARMA: So you might have answered this question in the past, but I didn't find the reference. So I just want to know, as we know that for Core Web Vitals, Google uses its Chrome User experience data. So I want to know for this Chrome User experience data that Google uses for Core Web Vitals, does Google only consider users that were coming from Google organic search or all the Chrome users, irrespective of their traffic channel or source? JOHN MUELLER: It's not tied to people clicking from Google Search. It's not-- PRAVEEN SHARMA: So it's basically all Chrome users? JOHN MUELLER: Well, it's not all Chrome users because of the way that the Chrome User Experience Report collects the data, but it's not tied to any specific source of traffic for your website. PRAVEEN SHARMA: OK. So what kind of users we will see if this is a user who falls into that category that we should consider for this report? JOHN MUELLER: I don't know it offhand. I think it has to do with the settings in Chrome that you have it set up that you're submitting user feedback through Chrome, those kind of things. But I don't know the exact details offhand. PRAVEEN SHARMA: OK, sure. Thank you. JOHN MUELLER: Cool. All right, Christian. CHRISTIAN KUNZ: Now? Hi, good morning, John. So I've got a question. So I have this website I'm looking for, and we have this situation that this website hasn't been migrated to mobile first indexing and there is a mixed state. We have some pages which have a separate mobile version and some other pages are responsive, and the link structure is kind of unclear. So I suspect that this might be the reason for the website not to be migrated to mobile first yet. And I just wanted to confirm if this is true or if there might be other reasons for that. JOHN MUELLER: I don't know. My guess is maybe at the time where we were migrating all of the websites over, maybe that was causing it to be held back. But as far as I know, there's still a batch of mobile first indexing sites that are just not migrated over. And it's something where we're working through some of the bigger sites there step by step to make sure that they can all be migrated cleanly. And then when we have the bigger issues resolved, then we'll migrate the rest as well. So my guess is the website is in this holding position where it's not necessarily that they're doing anything wrong now, it's just back when we checked it wasn't perfect. CHRISTIAN KUNZ: OK, so you would say it's good to just wait a little bit longer and to see what happens? JOHN MUELLER: Yeah. I mean, I don't know quite what the timeline is. But my guess is like towards the first half of next year we will be migrating all of the remaining sites over. CHRISTIAN KUNZ: Oh, OK. JOHN MUELLER: And then you'd probably see that happen. CHRISTIAN KUNZ: OK. OK. Cool. Thank you. JOHN MUELLER: Sure. Aakash? AAKASH SINGH: Hi, John. Hi. JOHN MUELLER: Hi. AAKASH SINGH: One of my clients' website will be down for a week or for two weeks. OK? And the reason why I'm here is there are a few bugs on the website and that's the reason they said they wanted to take the whole website down for a week or two. And my question is only that although there will be traffic loss and so and so, but how can I tell Google that this is a temporary situation? It will be only for the two weeks and our website will be live. Is there anything that I could display, a page showing the website is temporarily down and will be [INAUDIBLE] I know that the countdown is there on my page. It's been within two weeks or the 10 days, something like that. Can I tell Google that this website is currently down, but it will be just going back to live again within two weeks or a week? But there shouldn't be any ranking loss, or there could be a minimum ranking loss that I could get? Something like that. JOHN MUELLER: I don't think you'll be able to do it for that time, regardless of whatever you set up. So for an outage of maybe a day or so, using a 503 result code is a great way to tell us that we should check back. But after a couple of days, we think this is a permanent result code and we think your pages are just gone, and we will drop them from the index. And when the pages come back, we will crawl them again and we will try to index them again. But it's essentially during that time, we will probably drop a lot of the pages from the website from our index. And there's a pretty good chance that it'll come back in a similar way, but it's not always guaranteed. So any time you have a longer outage, I'm thinking more than a couple of days, I would assume that at least temporarily you will have really strong fluctuations and it's going to take a little bit of time to get back in. It's not impossible, because these things happen sometimes. But if there's anything that you can do to avoid this kind of outage, I will try to do that. And that could be something like setting up a static version of the website somewhere and just showing that to users for the time being. But especially if you're doing this in a planned way, I would try to find ways to reduce the outage to less than a day if at all possible. AAKASH SINGH: OK. Well, thank you so much. That ends my query. Thank you so much. [INAUDIBLE] JOHN MUELLER: Good luck. Hazel? HAZEL WRONG: Hey, John. Good morning. It's me again. So we want to ask some questions about Google AdBots. So do you think will disallow ads URL to the Googlebot in the robot text will affect our Google Ads? And maybe that cause ads won't work normally. JOHN MUELLER: So as far as I know, the AdsBot doesn't follow the normal robot text directives. You have to use the user agent directly for it. But I don't know what the effect would be on the ad side if you block AdsBot. My understanding is we use this as a way to do quality checks for the landing page. I don't know what that would mean from a quality point of view for ads if you don't let AdsBot check these pages. HAZEL WRONG: The reason we talk about this question is that we've recently found that Googlebot crawled our ads URL more than normal page URLs. But we want Googlebot to promote normal product pages instead of these ads. So we just think if it's OK to block ads URL for Googlebot. JOHN MUELLER: Sure. Sure. Sure. That's perfectly fine. Blocking the ads landing pages for Googlebot is fine. Again, if you block the ads pages for AdsBot, then that's something you'd have to check with the ads folks. HAZEL WRONG: Yeah, that's not what we do. JOHN MUELLER: OK. HAZEL WRONG: So also about the Googlebot, does AdsBot share a same crawl budget with Googlebot? JOHN MUELLER: Yes. HAZEL WRONG: Yes, so that means maybe the Chrome requests from AdsBot increase and maybe Googlebots crawl less? JOHN MUELLER: Yes, it can happen. Usually it settles down fairly quickly. Again, I don't know the details from the ad side, but my understanding is when you submit new ads campaigns then the AdsBot checks all of those pages. And once that's checked then they don't need to be checked as often as usual. HAZEL WRONG: OK. I see. If we're blocking the ads URLs in the Googlebot ads, Googlebot maybe still crawl too many ads URLs. What can we do? Maybe some measures to handle this situation to let the Googlebot crawl more product pages? JOHN MUELLER: So I think if you're blocking the ads landing pages for Googlebot specifically, then we shouldn't be crawling the ads landing pages with Googlebot. It would really only be the AdsBot that's crawling it. HAZEL WRONG: Because previous time we blocked some certain pages in the robot text, but we still see that sometimes Googlebots still crawl these pages even though we block them. So we just want to make sure that it's working. JOHN MUELLER: Yeah. It should never be the case that if we recognize the robot's text file that we still crawl the URLs anyway from Googlebot. So that feels like either something we would have to know as soon as possible or maybe some mistake with things like upper and lower case in the URLs, or not exactly the right path being included in the robot's text file. HAZEL WRONG: OK. So normally if we in the robots.txt, it shouldn't be crawled by Googlebot? JOHN MUELLER: Yeah. And if you see situations where the AdsBot crawling severely drowns out everything else, I would use that contact form that we have in the Help Center to report problems with Googlebot. I think you've used it before in the past as well. HAZEL WRONG: Yeah. Yeah. JOHN MUELLER: Because that also goes to the Googlebot team and they can help to distribute the requests a little bit better and contact the ads team and tell them to slow down with this website, those kind of things. HAZEL WRONG: Sure. Sure. If we notice something unusual we definitely would report it to Google. So another question is about the 304 response code. Do you think the 304 response code affects crawling? Because logically if we think if Googlebot checks a URL with the same content and it returns 304 code first time, maybe there's a possibility that Googlebot may reduce the crawling for the same URL because it returns to the 304 code. JOHN MUELLER: So I think there are two things. So the 304 is, I think, in response to the "if modified since" requests where Googlebot tries to see if this page has changed. And my understanding is that a 304 response code would not apply to the crawl budget side of things. So that basically means for us, we can reuse that request and crawl something else on a website. So there's that aspect. And the other aspect with regards to crawling that specific URL less, I don't think that would be the case. But we do try to figure out how often pages change and we try to recrawl pages based on the assumed page frequency or update frequency that we have. So it's not so much that that particular URL would get crawled less frequently, it's more that, oh, we understand a bit better how often these pages change. And then based on that, we can update our refresh crawling a little bit. HAZEL WRONG: So if most of the pages on the site returns to 304, so maybe it's a signal for Googlebot that the site has no new updated content, maybe reduce the crawling rate? JOHN MUELLER: No, I don't think so. No. I don't think we would reduce the crawling rate. We would just try to focus more on maybe the parts where we do see updates happening. So I would definitely not artificially hide the 304s is in the hope that it improves the crawling. HAZEL WRONG: Mhm. Sure. Sure. So also about the crawling issues. Since our crawling rate is back to normal, and we noticed that our crawl requests from smartphone is recovering much more faster than desktop. Could you shed some light on it? Any possible reasons behind this? JOHN MUELLER: I don't know. It sounds like what would be expected with mobile first indexing, that we crawl a bit more with mobile. I don't know if your specific site has already moved to mobile first indexing. HAZEL WRONG: Yeah, it is. JOHN MUELLER: Yeah. But then that would be normal, that we just crawl more with mobile and then you would see any changes faster there. HAZEL WRONG: Yeah. OK. So it's mobile first indexing. So the last one question is a more general one. It's about 301 redirects. We are taking down some useful pages and doing 301 redirects to them and we just think, well, will a huge amount of 301 redirects do harm to a site? JOHN MUELLER: No. No. That's perfectly fine. If you're making changes on your website and redirecting, that's perfectly fine. HAZEL WRONG: So no matter how many pages we make redirects? JOHN MUELLER: Yeah. Yeah. HAZEL WRONG: OK. That's good. That's good to hear. JOHN MUELLER: Cool. HAZEL WRONG: Thanks. This will be only question. JOHN MUELLER: Thanks. Let's see. Reshmi. RESHMI MOHAN: Hi, John. JOHN MUELLER: Hi. RESHMI MOHAN: Hi. This is the first time actually I am attending this session. It's regarding to our e-commerce website. This year only we have launched one e-commerce website and for 70-plus countries we have created a [INAUDIBLE] plan. In our URL structures for different countries we have created-- like we have established in UAE, so in UAE we have en_a. And when it comes to US, our website structure is xyz.com/en_us. Like that we have created for 70-plus countries. And the language is in English and for all the regions, our content is same except for some regions we have mentioned few cities. Like UAE we have mentioned cities like Arab Emirates, like Dubai, Sharjah, when it comes to US, but overall the content is the same. And we have more than 800,000 products on our website. And this year only we have launched. And for other countries also, even the language is not in English, we have created locale like bg_bg for Bulgaria and all. And initially we have submitted all the site map for all these regions as well. And first we have experienced that Google started crawling our website and almost 350,000 URLs got indexed. Then we have started experiencing the indexing of the URLs. So we first thought it can be the speed of the website and then we have increased the speed. And then again we started experiencing, OK, the URL got indexed. Later it changed, then again it stopped. Then what we did, we removed all the submitted URLs, and then we have the URL structure like xyz.com/en. Then we submitted the URL. Now we are experiencing the indexing of the URL, and when we submitted the website structure like /en, we had experienced that all the indexed URLs are going to the excluded list, and we don't know what is the exact reason behind that. But what can be the reason? Is it because of since the language is in English and we have created a different URL structure? We don't know exactly. Now the URL, indexed URL is showing only 160,000. JOHN MUELLER: OK. I don't know your website so it's really hard to say, but overall from what you've mentioned it sounds like you have a lot of URLs, and you have those 70 different country and language versions on top of that, so everything is multiplied by 70. From my offhand assumption is that this is just too much. So it's not that from a technical point of view that we can't handle that. But essentially you're submitting so much content and you have everything multiplied by 70, essentially. And that means for us that we start somewhere and we start indexing some things, but there's almost no chance for us to get through to everything. So usually what I recommend in these kind of situations is start off with a very small number of different countries and language versions and make sure that they're working well, and then expand incrementally from there. And in particular, if you have a lot of English variations, then try to make sure that you just have maybe one English landing or English site, rather than all of the different English country versions, which are essentially the same thing. Because if you have fewer versions of the English content, it's a lot easier for us to focus on that for indexing and to treat that a little bit better when it comes to ranking overall. So that's my general recommendation [INAUDIBLE] So that's the thing [INAUDIBLE] With regards to international versions, it's very easy to take a website and say I will just make all English language countries and create a hundred different versions of website. But it just causes so many problems and makes everything in the whole crawling and indexing and ranking cycle, makes everything so much harder. So that would be kind of the direction I would head, is just start off a lot smaller and then build out from there. RESHMI MOHAN: So that means Google is considering us a duplicate? JOHN MUELLER: Sure. I mean, if these English versions are all essentially the same thing, then for the most part we will treat it as duplicates. And then it's kind of hit and miss which version actually ends up being indexed. So that's, I think, a large part of the issue there is that we just see all of these different copies of the same thing and we don't know what you really want us to do. We have to focus on a small part of it. We can't do everything at once so it ends up causing almost more problems for your website by having all of these different versions. RESHMI MOHAN: That means we are going in the wrong direction. JOHN MUELLER: Probably, or moving too fast. I don't know. It's something where it's very easy to take a really large website that's doing well and say, oh, we will just copy their system and take all of the different English versions and language versions and do the same thing. But if you're not in the situation of that large well known website already, then it's very different for you. And then it's a lot harder to actually get to expanding. So I would try to pick one English version and maybe one other language version to start off with, and then expand from there. RESHMI MOHAN: Yeah. Because since we want to target worldwide-- because it's an e-commerce website and we want to target worldwide, and we want to improve our SEO, we just targeted in this way, creating different URLs and locating them in different countries like that. So I think we can focus only on the main domain instead of creating locally, right? JOHN MUELLER: Yeah. I, mean that's what I would do. Yeah. RESHMI MOHAN: And no need of going for the en-- I mean, the URL structure should not-- it should be like www.xyz.com instead of the locale, right? JOHN MUELLER: You can have the locale in there. It doesn't really matter. But again, I would make it so that you primarily just have one English version rather than all of the other variations of English as well RESHMI MOHAN: OK. OK. Thank you. Thank you, Mr. John. JOHN MUELLER: Sure. Let's see. Gakushi. GAKUSHI NAKAMURA: Hi. JOHN MUELLER: Hi GAKUSHI NAKAMURA: I'm new, and bear with me if my question isn't clear enough. But we have a website with about, say 20,000, 30,000 pages and it's like a directory service, but we have unique content. So we put no index to probably about 2/3 of those pages because we wanted to concentrate what Google looks at. But after a month or so, only less than 1% of those pages are indexed and we're getting identified or discovered but not indexed for almost all of our pages. So kind of scratching our heads what we can do. We made a lot of tweaks and it looks like internal linking is all right. We might be restricting ourselves a little bit by the canonical setting, but they don't look wrong. So in the end, we kind of figured going through your videos and stuff, it might be a site quality issue, because there are predecessors with similar content that are already indexed. And yet we do have very unique content. The problem may be that that unique content is not text. It's data. And it looks like a lot of the quality is assessed by things that's readable by human being. But our service is really focused on a number that is very valuable and we're just wondering what we can do. Should we try to squeeze in text which isn't actually central to what we do, or are there things that we can do to tell Google that it's actually valuable to people who visit our sites? JOHN MUELLER: OK. I think in general having numbers on a page is fine and it's something that we do see as being part of unique content on a page. So just because a lot of the content is numerical wouldn't necessarily mean that we would see the whole page as being duplicated. So it's hard for me to estimate what kind of site it is with regards to-- you mentioned it's a directory type site. From a quality point of view, I do think that matters quite a bit in terms of how much we actually index from your website. So that's something where I would probably still focus on that. If you're saying the site is like 25,000 pages large, it feels like something that we should be able to index if it's something that we would consider to be a reasonable quality website. And when it comes to the quality assessment, it is something where we do look at the content of the page and we try to make sure that it's not duplicate, but we also try to understand the bigger picture of that page within the context of the rest of the web. So that sometimes makes it a little bit trickier. GAKUSHI NAKAMURA: Right. Just as a follow up, we're getting discovered and not indexed rather than crawled and not indexed for those 99%. Should we differentiate between those two? Because our site is not that big and this can't be a crawl budget issue in my view. In that case, are those two designations pretty much the same in that it's just a quality issue? JOHN MUELLER: Again, I don't know your website so it's hard for me to say offhand. But if it's something where you're seeing the clean URLs being listed in the discovered not indexed report, essentially the URLs that you do want to have indexed, then to me that sounds like it's really less a matter of Google can't go off and crawl that many URLs. Because, again, with 25,000 pages, most servers that are reasonably sized can easily allow that kind of crawling on a regular basis. And it's probably really more a matter of our understanding of the overall website quality. And with larger websites or if in the discovered report-- discovered not indexed report, you see that there are lots of different variations of URLs, like with parameters or with upper or lowercase, those kind of things, that can be a sign that the internal linking is kind of messy and that we're having trouble finding the right URLs to crawl. But if we're showing the right URLs in the discovered not indexed report and it's a reasonably small website, then to me that more points in the direction of the overall site quality. GAKUSHI NAKAMURA: Right. So do you think we should try to add text in there? What we show is really a directory of companies, and we show what a stock price means in terms of future growth of that company. So it's a number, but there's not a whole lot of readable text that goes with it. It's just a number. Does this $20 stock mean 10% growth or 15% growth is what's valuable. We can add it. We do have a description, but it's common to all those companies and we've got to figure out what to do if we were to squeeze in unique text per each company. But should we head to that direction, do you think? JOHN MUELLER: I don't think the text will affect how we index the pages. So from that point of view, it's something where if you see the text affecting how users look at your pages and are able to interact with your pages, and then sure. But that's more a matter of trying to figure out what users are actually looking for and where you can provide unique value to your users. But just adding text to pages I don't think would affect how we would crawl and index those pages. If it's something where you're providing numbers, I don't know, like the stock numbers there, that's something where I would also try to figure out what you can do to make sure that what you're providing is unique and provides value to users and do something maybe along the lines of a user study to figure out, what is it that we can do to make our website such that users recommend it to other people as well? And that it builds up almost like, I don't know, trust or something from the user point of view. And a lot of times those are not purely technical things that you change on a website, where you change a design or you convert some of the numbers into text, for example. It's really a matter of the overall setup of the website. GAKUSHI NAKAMURA: Great. Thank you so much. JOHN MUELLER: Sure. Let me run through some of the submitted questions, and then we'll get back to the giant number of people who have their hands raised. Wow, so many today. Let's see. Does Google have any trouble indexing sites that have a mobile version on a subdomain, for example, example.com and m.example.com? And then the question goes into they're seeing a lot of pages indexed without content. So from our point of view, we don't have-- well, at least as far as I know-- we don't have any problems with m dot domains in general, in the sense that this is one of the supported formats that we have for mobile websites. We don't recommend the m dot setup. So if you're setting up a new website, I would try to avoid that as much as possible and instead use a responsive setup, but it is something that can work. So if you're seeing this regularly with your website that we're not able to index your mobile content properly, then to me, that would point more at an issue on your website where when mobile Googlebot is trying to crawl, it's not able to access everything as expected. So that's the direction I would head there to try to clean that up. The one thing that throws people off sometimes with m dot domains is with mobile first indexing, we switch to the m dot version as the canonical URL, and it can happen that we show the m dot version in the desktop search results as well. So you also need to watch out for not only redirecting mobile users from the desktop to the mobile version, but also redirecting desktop users from the mobile to the desktop version. And again, that's something if you have a responsive design setup you don't have to worry about that. So it's another reason to go responsive if possible. Would I be penalized if I don't set the rel sponsored or rel no follow for my affiliate links? Probably not. So from our point of view, affiliate links fall into that category of something financial attached to the links, so we really strongly recommend to use this setup. But for the most part if it doesn't come across as you selling links, then it's not going to be the case that we would manually penalize a website for having affiliate links and not marking them up. It is a best practice and I strongly recommend doing that, but it's not something where I'd say we will go off and manually take action on these sites. A few months ago, we made a CMS migration and our URL structure completely changed. All of our URLs execute 301 redirects to the new addresses and we have improved technical indicators. After that, we realized that our organic traffic dropped a lot immediately after the URLs changed. Is that what we should expect? What time does it take? So when you move from one domain to another, it's very easy for us to just transfer everything from that domain to the new one. And that's something that can be processed within almost like a week or so in many cases. However, if you change the internal URLs within your website, then that's something that does take quite a bit of time because essentially, we can't just transfer the whole website in one direction. We have to almost reprocess the whole website and understand the context of all of the pages on the website first, and that can take a significant amount of time. And during that time, you will almost certainly see fluctuations. So my offhand guess would be you will definitely see at least a month of fluctuations there, perhaps even longer, especially if it's a bigger change within the website itself. The other thing is when you change your CMS, oftentimes things that are associated with the CMS also change, and that includes a lot of things around internal linking, for example, and also the way that your pages are structured in many cases. And changing all of those things can also result in fluctuations. It can be that the final state is higher or better than it was before. It can also be that the final state is less strong than it was before. So that's something where changing a CMS and changing all of the internal linking on a website, changing the internal URLs on a website, changing the design of these pages, these are all individually things which can cause fluctuations and can cause drops and perhaps even rises over time. But doing that all together means that it's going to be messy for a while. So that's something where, on the one hand, waiting to see what happens is something you almost have to do. But at the same time, I would also recommend really from a technical point of view, trying to understand all of the differences between your new version of your website and your old one and try to figure out which of these changes are better and which of these changes might be worse, and which things you might want to clean up. Archive.org is a great way of looking at your website in the past. So I would use that to try to figure out what things looked like before, how the internal linking was set up, what anchor tags were there, what headings, what kind of text was there, all of that. Another thing to watch out for with these kind of migrations is oftentimes there's embedded content that you don't think about directly because it's not an HTML page, and a really common one is images. And if you don't redirect your old image URLs, then we have to reprocess them and find them again new because we don't have that connection between the old images and the new ones. So if your site was getting a lot of image search traffic, that can also have a significant effect. And setting up those redirects probably still makes sense even if you've moved after a month or so. I don't know what the timing here was. Then I have not a technical question, but a career question. What is your advice for someone interested in pursuing a career in SEO? What should they focus on if they already started learning and are familiar with the basics? How to go from there to become great at SEO? I don't know. I don't get these kind of questions often, so I don't have a perfect answer lined up. From my point of view what I like about SEO is that there's so many different angles where people can dive in and understand a little bit more. And oftentimes people have interest in, I don't know, figuring out what to do with regards to content, or trying to understand the audiences that are involved, trying to understand the technical details of a website, maybe even going into the direction of JavaScript and developing things. And all of these are, I think, really neat angles, and they're all important for SEO. It's not that one of these is going to go away and suddenly you'll be stuck. So my recommendation there would kind of be to figure out which of these angles you really enjoy doing and try to figure out ways to do a little bit more of that. One of the things I've recommended in the past which might also be appropriate here is to take part in the different Webmaster and SEO forums. So Google has, I think, a help community as well and there are definitely others out there as well. And the good part there with all of these forums is that you see a lot of different scenarios, and you run into situations where you can dive in and try to understand, well, what is the problem here and how could I solve this. And whether or not you actively take part in the forum and post your suggestions there, at least the opportunity to look at individual situations where you see, well, there's a problem here. How can I try to figure this out myself? I think that is really valuable. And that's really important in the long run because a lot of times you will see the same kind of problem come back over and over again. And if you understand the patterns there, it's a lot easier for you to jump in and say, well, this is the same pattern as I've seen before, and that means I need to double check these three or these five things first. So I think that's always really valuable, to see in practice a little bit more. But again, there are so many different options and so many different directions that you could take, it's hard to just say you need to do this one thing to become better at SEO. Let's see. I have a situation with a site where the main navigational links is visually different from the mobile and desktop versions. The mobile version is slimmer and the desktop version has more links. Looking exclusively at crawling, the desktop is more helpful. For technical reasons we can't hide the desktop links in the source and the site is on mobile first indexing already. Is there a chance that Google might think I'm trying to game the algorithms by keeping this more link-rich navigation even though it's invisible to mobile users and bots? So I think overall, this is a fairly common set up for responsive design, in that you have both variations visible in the HTML. And from that point of view, I don't see it as being super problematic, but it's probably also not super clean because you have to maintain both variations of the UI within the same HTML, rather than having just one variation in the HTML that you adjust depending on the viewport size of the device. So from that point of view, I don't think you would be penalized for anything like this-- well, definitely not penalized, because it's a very common setup. It's probably not optimal, but it's also not something where I would say need to jump in and fix that right away. So that's my assessment there. My question is about dealing with outdated blogs on our platform. We have about 450 blogs, some of which are four to five years old and therefore out of date and have almost no traffic on them. Do you recommend deleting them because they hurt our general search rankings? What's the best way? Delete all without traffic at once and request index deletion at Google, or do you recommend step by step approach? So I think with blogs you probably mean blog posts, so individual pages, not whole sets of pages. Because I think if you have so many different sets of pages it's probably a bigger change. But with 450 pages, say, more or less, where you're saying, well, these don't get a lot of traffic. Should I just delete them or not? From my point of view, probably that's something where you can just make that call on your own. I don't see that as being something where from an SEO point of view, you would see a significant change unless these are really, really terrible blog posts. The main thing, however, I would watch out for is that just because something doesn't have a lot of traffic doesn't mean that it's a bad piece of content. It can mean that it's something that just gets traffic very rarely, maybe once a year. Maybe it's something that is very seasonal that overall when you look at it from a website point of view, it's not very relevant, but it's relevant, I don't know, maybe right before Christmas, for example. So from that point of view, I would say it's fine to go through a website and figure out which parts you want to keep and which parts you want to clean out. But just purely looking at traffic for figuring out which parts you want to clean out, I think that's too simplified. But again, from an SEO point of view removing 450 pages from a larger website, that's tiny change and I wouldn't worry about when you do that and how exactly you do that. Delete them whenever you recognize that they're no longer valuable. Delete them all at once, that's also an option. With regards to submitting them with the removal tool as well in Search Console, that probably wouldn't change anything because the removal tool in Search Console just hides the page in their search results, it doesn't actually remove anything from indexing. So that's one thing you don't have to do. But again, otherwise, I would just think about which pages you want to keep, which ones you want to remove, and go through it like that. Would a query string at the end of an image source URL have a negative effect for SEO? We use it for cache invalidation when an image is edited. So no, it wouldn't cause issues with regards to SEO. But with images in general, we tend to recrawl and reprocess them much less frequently. So that means that if you're regularly changing the image URLs linked in your website, we would have to refind those again and put them in our image index again, and that tends to take a lot longer than with normal HTML pages. So from that point of view, if you can avoid doing that too often I would recommend doing that. If it's something that happens, I don't know, very rarely, and where you're saying, well, it doesn't really matter too much how things are processed in image search. We don't rely on image search for traffic to our website, then from that point of view, that's totally non-issue. The thing I would avoid, especially here with image URLs, is that you embed something that changes very quickly. So something like a session ID or just always today's date, because that would probably change more often than we would reprocess the image URLs, and then we would never be able to index any of the images for image search. Let's see. Maybe I'll just go back to more live questions since we have so many people here. Bruno, over to you. BRUNO MOSCONI: Hi, John. JOHN MUELLER: Hi. BRUNO MOSCONI: Thank you for this opportunity. John, we are from a news website here in Brazil and we have a question about a problem that we're going through in the last few days in our website. We're going through a huge drop in the delivery of our content through Google Discover. And now Google Discover is our main acquisition channel, but in the last 10 days the delivery of our content through Discover has dropped to practically zero. Our normal audience was a thousand active users in real time, and now it's about 90 or a hundred active users. And we didn't change anything technical or editorial. And our Search Console did not report any problem. We have all Core Web Vitals rated very high in our CrUX reports, it's OK too, all above 80%. Our website is 100% AMP, and we don't know what's happening. What do you think could be happening with us? Can we do something to fix this problem in Discover? We don't know what's happening. JOHN MUELLER: Yeah. I think it's always tricky with Discover because it's-- at least what I hear from people, it's very binary, in that either you get a lot of traffic or you don't get a lot of traffic from Discover. And that also means that any changes there tend to be very visible. So my main recommendation is not to rely on Google Discover for traffic, but rather to see it as an additional traffic source and not as the main one. When it comes to Discover there are a few things that play in there. You mentioned some of the technical things that I think are good practices. One of the things that also plays in there is, for example, the core updates also play a role. We recently had a core update, maybe from a timing point of view that matches what you saw there. So that's something where if you do see an effect from the core update, then I would double check the blog post that we have about core updates with the large number of tips and ideas that you could focus on. The other thing is with Discover in particular, we have a set of content guidelines that we try to stick to in an algorithmic way. And depending on the website itself, it might be something where some of these content guidelines, your website is kind of borderline. So for example, I don't know the content guidelines all by heart, but I think there is something around clickbait-y titles or clickbait-y content in general, or adult oriented content, for example. And it might be that your website is kind of borderline there with regards to how we evaluate your website in that regard. And then it can also happen that our algorithms say, oh, well, a large part of this website is just clickbait or one or the other categories that we list in the content guidelines. And then we will be a lot more conservative with regards to how we show the website in Discover. So those are-- without knowing your website, that's the direction I would head. On the one hand, the core updates, think about that. On the other hand, the content guidelines that we have. And then finally, I would still make sure that you don't rely on Discover for your business overall, because it can change fairly quickly. And it's something where often there are not pure technical reasons behind those changes. BRUNO MOSCONI: OK, great. Thank you. JOHN MUELLER: Cool. Brian? BRIAN HARNISH: Hi, John. JOHN MUELLER: Hi. BRIAN HARNISH: Nice to meet you finally. JOHN MUELLER: Yeah. BRIAN HARNISH: OK, I have a couple of questions. First one. On a Reddit post a few weeks ago, I believe, you mentioned that having less than 30 pages would make Google consider the website less authoritative, right? And I know that leaves it open to SEO [? writing. ?] You have to have minimum 30 articles to rank, et cetera. But my question concerns the other extreme. What about one page sites? JOHN MUELLER: I think you can make good one page sites. So from that point of view, I'm not too worried about that. I think the Reddit post, as far as I remember, was something along the lines, I created 30 blog posts and they're really good, and therefore my website should be authoritative. And from my point of view, you going off and creating 30 blog posts does not automatically make your website authoritative. And especially for the higher or the more critical topics, it's something where you can't just create 30 blog posts on a medical topic and then say, I am a doctor. I've written 30 articles. So that was the direction I was headed there. And for a lot of websites, it's not that you need to be seen as an authority. You essentially put your content out there. If you're a small business, you're selling something. You don't need to be an authority. And especially things like one page websites, they're often very focused on this one thing and you don't need to be an authority to do that one thing, to sell, I don't know, an e-book, or to give information about opening hours for a business. It's like, it's just information. So from that point of view, having a one page website I think it's perfectly fine. With regards to starting out with a one page website, I think that's fine. I would just think about, well, where do you want to go from there at some point? Maybe you do want to create more pages, and try to find a way that you don't paint yourself into a corner by saying, well, I have to put everything on one page all the time, but rather expand when you see that it fits. BRIAN HARNISH: OK. The second question is actually more of a hypothetical. So for international SEO, say you have a site that is basically generally well established. They have links going to it, all that stuff. It's generally good from an SEO standpoint. But at one point, they decided to build their international URLs and setup into one domain. Say, for example, it would be something like media news United States as opposed to the International URLs and the international setup. And aside from the obvious, from things like links being combined, maybe not being redirected properly, stuff like that, what else would-- and say Google all of a sudden decides to drop traffic 90% after that implementation. Assuming that everything was done 100% correctly, which is likely not to be the case on a larger site like that, what would happen to cause Google to drop traffic like that in such a scenario? JOHN MUELLER: It's hard to say. So usually what I would do in a case like this is try to figure out if things have settled down or not. Because if it's in this flux period, it's really hard to determine what things should be like. And then I would try to dig into both individual pages and individual queries. So in Search Console, figure out what the top 50 queries were and look at the top 50 queries now and see, are there certain queries which have dropped, or is it more a matter of everything overall dropping slightly? To try to figure out, is this based on individual parts of the site, or is it an overall change? And the same thing on a page level as well. Often what I find when I look at things on a page level is that individual pages that used to get a lot of traffic suddenly don't work anymore. They're 404, or they got redirected incorrectly or something like that. And then it's something where it's suddenly not clear what actually happened there, in that you see, well, some of these top traffic pages are gone now, then of course, the overall traffic would be visibly affected by that as well. But usually going and trying to find ways to figure out what the details were of that change, that helps to understand a little bit better which direction people should start looking. BRIAN HARNISH: Thank you, John. Thank you very much. JOHN MUELLER: Cool. Thanks. All right, David? DAVID RADICKE: Hi, Thanks. Last one in, I guess. So I posted this also in the comments. So I have a small website and it's a couple of hundred URLs only. It's a software company, and its a small team producing content like white papers and help articles and blog posts. And it has been going well for a long time. And suddenly now in November, the posted pieces, the published pieces are not indexed anymore, not all of them. Which is very sad because they're writing nice little pieces for Black Friday, how to write a newsletter for Black Friday or for Christmas, and then we're sitting there and seeing that Google is crawling them or discovering them-- we see that in the Search Console-- but it's not indexed. So I tried everything. I looked at the technical issues, the linking is good. So my question is, is there a paradigm shift that Google is saying, well, thanks for suggesting these new-- publishing these articles, but we don't want it right now? Is this something new that has changed recently? JOHN MUELLER: Not really, at least not that I know of. I mean, I think what I see a lot with regards to a lot of the indexing questions that I get nowadays is from a technical point of view, it's very easy for websites to make websites that just work. You set up WordPress and then essentially all of the SEO is done for you, it just works. And from our point of view, this means that it's less often a case that there is a technical problem with a page that it doesn't get indexed. That means all of the content that we get is essentially technically OK and our systems have to be a lot more critical with regards to the overall quality of the website, the quality of the pieces of content that we get. And then it's something where additionally in Search Console, we give you all of the information on things like discovered and not indexed, or crawling but not indexed. And then suddenly you see all of these problems and it seems like something that people have to fix. So from that point of view, it feels like it's normal for us to get a lot more of these indexing questions just because, well, a lot of content is kind of OK and we still can't index everything on the web, so we have to make a cut somewhere. That said, if you're really looking in situations where you're seeing this is a smaller website and the content is really reasonable, and it's not something where they're just rewriting news articles from other people's sites, I would love to have example URLs. DAVID RADICKE: I pasted them actually in the comments. JOHN MUELLER: OK, perfect. Perfect. DAVID RADICKE: And I would love to hear your-- because it's just not-- you could say it's not a big deal. I understand that Google cannot index everything. And we do rank for some of these topics, for example, Black Friday 2020. We rank for that. And now we wrote a new piece, "Black Friday 2021: What to do With Your Customers," and it's a handwritten piece. So the question would be there, should we just lay off the content team? Because it's really expensive to produce content. I mean, they're writing all those pieces in a couple of days. I mean, those are nice ladies. They're sitting there writing those content and then they aren't getting indexed and we're not getting any traffic from Google for that. So basically the management is asking me, well, why should we do that? There's no sense in producing good content if we don't get that part of the traffic that we get from Google normally, because Google says, well, we just don't want it right now. So it's tricky, isn't it? JOHN MUELLER: Yeah. I mean, that's why these examples are really useful. Because we do bring them back to the indexing teams and say, well, what is wrong here? Why aren't we picking this up? This is something reasonable. We should be doing a better job at that. And especially when things are-- DAVID RADICKE: --not good or there's something technically wrong, but I just don't see anything. And again, it's not a site with millions of pages, but we have 500 pieces, I think, and they're all handwritten. It has always worked, just suddenly in November it stopped, during the time of the core update so I was afraid that Google's busy with other things right now. JOHN MUELLER: It shouldn't be. But, I mean, having these kind of examples is really valuable for me, because going to the team and saying, oh, people are complaining that content is not being indexed and I don't have any examples for you, then they'll be like, well, we can't do anything. DAVID RADICKE: --content I have the OK from the company to use this in public, so it's OK. JOHN MUELLER: OK, cool. I mean, you can also send it to me privately if there's anything that you or anyone else runs into. DAVID RADICKE: Great. What's the best way to send it to you, email or Twitter? JOHN MUELLER: Twitter is probably easiest and for the most part, people can ping me and ask to be followed so they can send me a DM if that makes it easier. DAVID RADICKE: All right. Thanks. JOHN MUELLER: Cool. All right, let me take a break here with regards to the recording. I can be around a little bit longer, but not as long as usual this time because we have more meetings. But thank you for joining. Thank you for all of the questions so far and I hope we can get through some of the remaining questions as well. People still have their hands raised. If you're watching this recording on YouTube, feel free to submit questions for the next sessions which I'll be setting up as well. I don't know how the timing of the next sessions will be with regards to holidays and things like that, but we'll figure something out. Cool. And with that, let me pause the recording here. Just a second.
Info
Channel: Google Search Central
Views: 2,899
Rating: undefined out of 5
Keywords: Webmaster Central, Google, SEO, Search Engines, Websites, Search Console, Webmaster Tools, crawling, indexing, ranking, mobile sites, internationalization, duplicate content, sitemaps, pagination, structured data, rich results, English Webmaster Office Hours, Office Hours English, Search, office-hours, office-hour, SEO office hours, English SEO office hours, Search Central, Google Search Central, GSC, SC
Id: GR-A6kYdqfU
Channel Id: undefined
Length: 60min 42sec (3642 seconds)
Published: Fri Dec 10 2021
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.