The Moz Q&A Forum

    • Forum
    • Questions
    • My Q&A
    • Users
    • Ask the Community

    Welcome to the Q&A Forum

    Browse the forum for helpful insights and fresh discussions about all things SEO.

    1. SEO and Digital Marketing Q&A Forum
    2. Categories
    3. Intermediate & Advanced SEO
    4. Pages are Indexed but not Cached by Google. Why?

    Pages are Indexed but not Cached by Google. Why?

    Intermediate & Advanced SEO
    44 4 19.4k
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as question
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • friendoffood
      friendoffood @RyanPurkey last edited by

      Thanks for your reply Ryan..

      Well, that was easy.  How come for some pages they have that code and others they don't?  It seems like a backward way to find the cached version of your own pages too..

      Thanks much.

      Now, for a twist, suddenly I checked just now for site:http://www.qjamba.com/restaurants-coupons/ferguson/mo/all

      and it is no longer there.  It was there yesterday.  Why would it suddenly disappear?  I again checked webmaster tools--not finding any crawling errors or problems for it.   So, the question suddenly gets reversed: how now can I have a cached version without a site:www... listing?

      RyanPurkey 1 Reply Last reply Reply Quote 0
      • RyanPurkey
        RyanPurkey @friendoffood last edited by

        The "site:" operator isn't as precise as what you'll find in Google Webmaster Tools (GWT) and doesn't return each and every page for a site Google has indexed for a few reasons. One reason, Google is preventing people without private access to the site from seeing each and every page it indexes for a site to keep that data from being scraped publicly.  In your case, that's good if your competitor is running similar searches like you're doing now in the attempt to copy your site. Instead Google gives you that information privately via GWT.

        The same goes for cached pages. The overarching reason is that it's about preventing over exposure publicly both in how Google operates and how a site is constructed. Ultimately you'll have to trust GWT and your own site's server records more than what you can find searching Google as an average user.

        friendoffood 1 Reply Last reply Reply Quote 1
        • friendoffood
          friendoffood @RyanPurkey last edited by

          Thanks.  I will give you a 'good answer soon', but dont want to close this out yet since there are some other issues going on with the other response.  In addition I'm trying to figure out the practical way to use all this info:

          Here's what I was trying to do:  Make a program change, fetch with google, and see in a couple of days if the site has risen in the rankings.  It was complicated (for me) by the absence of the fetched site being cached (which even the one you found is not a recent cache), and the the site:www... sometimes not appearing to have changed at all even if there was a new cache.  How is a person to test the immediate effect on ranking when they make a change if they can't tell what Google is doing?

          I haven't found in GWT a list of my indexed pages, or the dates they were indexed, and certainly not the content that has been indexed, yet your answer above seems to indicate there is more under the hood than what i have found there. Is there really a way in GWT to see if a specific page has been indexed (my site has thousands so I would need to search for it), and when it was last indexed?  I can tell from my dbtables what pages and when they were crawled but not whether Google has added the new info into their indexing algorithm.

          1 Reply Last reply Reply Quote 0
          • RyanPurkey
            RyanPurkey last edited by

            I would caution against creating a tool to do what you're describing as you might end up violating Google's terms of use.  Instead, use a dedicated tool for monitoring rankings--like Moz's pro feature set--around a specific set of keywords that have value for you instead of each and every page on your site.  Chasing after the immediate effect of ranking and changes is akin to trying to precisely unravel Google's ranking algorithm, something Google very much doesn't want you to do.  Instead look at past performance of content (analytics, server logs, etc.) and whether or not it improves after changes.  The improvement is also subjective.  Maybe you get less users and sessions, but much higher conversions...

            Within GWT you're going to want to look at Index Status https://www.google.com/webmasters/tools/index-status? and compare it with the number of pages in your sitemap(s).  This most likely isn't going to be an exact match as Google at times limits the amount it caches and indexes a site based on its own determination of page worthiness (high percentage of the page is duplicate content for example).  So look for a decent percentage of indexation versus exacting numbers.  Also, having pages that perform really well for you indexed and ranking well is more important that 100 that don't.

            Ultimately the more precisely you try to deconstruct Google the more difficult things will be. Take old Ben's advice, "Let go..."

            DOFgFAcGHQc?t=12s

            friendoffood 1 Reply Last reply Reply Quote 1
            • Travis_Bailey
              Travis_Bailey @friendoffood last edited by

              Rest assured, that I don't scrape/hammer so hard that it would knock your site down for a period. I often throttle it back to 1 thread and two URI per second. If I forget to configure it, the default is 5 threads at two URI per second. So yeah, maybe a bit of the moz effect.

              Chrome Incognito Settings:

              Just the typical/vanilla/default incognito settings. It should accept cookies, but they generally wouldn't persist after the session ends.

              I didn't receive a message regarding cookies prior to the block notification.

              On a side note, I don't allow plugins/extensions while using incognito.

              Fun w/ Screaming Frog:

              It's hard to say if the 8.5 hour later instance was my instance of Screaming Frog. The IP address would probably tell you the traffic came out of San Antonio, if it was mine. I didn't record the IP at the time, but I remember that much about it. Otherwise it's back in the pool.

              Normally Screaming Frog would display notifications, but in this instance the connection just timed out for requested URLs. It didn't appear to be a connectivity issue on my end, so... yeah...

              Fun w/ Scraping and/or Spoofing:

              Screaming Frog will crawl CSS and JS links in source code. I found it a little odd that it didn't.

              I also ran the domain through the Google Page Speed tool for giggles, since it would be traffic from Googlebot. It failed to fetch the resources necessary to run the test. Though cached versions of pages seemed to render fine, with the exception of broken images in some cases. Though I think that may have something to do with the lazy load script in indexinit.js, but I didn't do much more than read the code comments there.

              In regard to the settings for the crawler, I had it set to allow cookies. The user agent was googlebot, but it wouldn't have came from the typical IPs. Basically just trying to get around the user agent and cookie problem with an IP that hadn't been blocked. You know, quick - dirty - and likely stupid.

              Fun w/ Meta Robots Directives:

              A few of the pages that had noindex directives appeared to lack genuine content, in line with the purpose of the site. So I left that avenue alone and figured it was intentional. The noarchive directive should prevent a cache link. I was just wondering if one or more somehow made into the mix, for added zest. Apparently not.

              While I'm running off in an almost totally unrelated direction, I thought this was interesting. Apparently Bingbot can be cheeky at times.

              Fun w/ The OP:

              It looks like Ryan had your answer, and now you have an entirely new potential problem which is interesting. I think I'm just going to take up masonry and carpentry. Feel free to come along if you're interested.

              friendoffood 1 Reply Last reply Reply Quote 1
              • friendoffood
                friendoffood @RyanPurkey last edited by

                Thanks Ryan.  I am new to this so I appreciate the cautions. Are you saying it is a bad idea to try and run tests on specific page changes (ie like reducing the size of a menu or adding more content in paragraph form, for example) to see how they affect ranking on given searches?  I'm struggling with why that is a problem for Google..  The SEO experts who run tests--to get a sense of what factors are important -- isn't that exactly what they are doing in a lot of cases?  I plan on using the MOZ tools and have much to learn, but it just seems so logical to test things out like I want to do also.

                The GWT index status page shows an increasing number of pages, but no detail on which ones have been indexed.  That's why I was looking into the site:  command and the cache command in the first place.  It seems those exist for a reason, and I thought it was to help us with specific pages.  If that isn't it, why do they exist?  Other than showing the number of indexed pages I don't see much value to the index status page. Perhaps there is a good tutorial somewhere that helps people know how to use the GWT not just to find crawl errors or see general trends but to actually implement SEO page changes to improve their site?

                I hope this doesn't sound argumentative..I just don't quite understand your perspective on it.

                1 Reply Last reply Reply Quote 0
                • friendoffood
                  friendoffood @Travis_Bailey last edited by

                  Travis,

                  First of all, I absolutely appreciate all the time you are taking to address my issues here.  Second of all, it IS very tempting to join you and any others here to go build houses or do something else, especially given the last few days..:)

                  Ok. I'll try to keep it short:

                  I wasn't thinking you had any bearing on my site going down, but that maybe there was a 'moz effect'.  hope not.

                  Re: Chrome Incognito Settings:
                  I'm really worried now that there is a sessions problem since anyone with cookies allowed should have the session ID saved between pages--in which case they would have only 1 entry in my 'user' table and you had 3 in a short amount of time.  That's why it thought you were a robot.  I don't know how to duplicate the problem though because I've never had it personally, and I use a program that connects to other machines with all kinds of combinations of operating systems and browsers and computers and have never had this problem with those.  It's my problem.  I'll have to figure it out somehow.  I have many session variables and it would be a huge overhaul to not use sessions at this point.  If you have any ideas (I'm using php) I'm all ears.

                  re:Fun w/ Screaming Frog:  
                  The IP for the 8.5 hour later instance was the same as your first one. Yet, if you were spoofing it shouldn't have said screaming frog in the user agent, right?  It was in my 'bot-stopped' file as instantly stopping because it was an unexpected bot.  So, confused unless perhaps you tried it separately from running with the spoof?

                  <<normally screaming="" frog="" would="" display="" notifications,="" but="" in="" this="" instance="" the="" connection="" just="" timed="" out="" for="" requested="" urls.="" it="" didn't="" appear="" to="" be="" a="" connectivity="" issue="" on="" my="" end,="" so...="" yeah...="">>
                  Ok.

                  Fun w/ Scraping and/or Spoofing:
                  <_>
                  I'll have to check into it.  I've run Yslow and Gtmetrix without problems.  I see you tried to run it on the ferguson page and the home page.  I just ran the ferguson page in gtmetrix - which uses Page Speed Test (Google?) and Yslow both - and it ran ok, although not a great grade.

                  <<while i'm="" running="" off="" in="" an="" almost="" totally="" unrelated="" direction,="" i="" thought="" this="" was="" interesting.="" apparently="" bingbot="" can="" be="" cheeky="" at="" times.="">>  That is interesting.

                  I'm worried now most about the session issue, as that may be affecting a lot of my users and I've assumed multiple entries were from robots that generally don't keep sessions between page crawls (actually quite a few of the seo crawlers do -- but google, bing, yahoo don't).  If you are ok with going to my home page without incognito and clicking on a few pages and letting me know the first part of your IP when you do that it might really help me. Your shouldn't be blocked anymore (it lasts 1 day). But, no worries if you're ready to move on.</while>_</normally>

                  Sorry, wasn't so short after all.  Thanks again.  Ted

                  Travis_Bailey 1 Reply Last reply Reply Quote 0
                  • RyanPurkey
                    RyanPurkey last edited by

                    You're welcome Teddy. Something that goes undermentioned when SEOs run very precise tests on specific page side changes is that they're typically doing them on completely brand new domains with non-sense words and phrases because of the chance that their manipulations might get the site blacklisted.There's no loss to them if that happens other than unanswered questions. If the site does survive for a bit maybe they'll learn a few new insights.  This level of granular, on site testing isn't a practical method for legitimate, public facing sites.

                    When it comes to sites that serve a business function aside from testing possible granular ranking changes, you're going to be much better served by measuring your changes against your user interaction instead of your search engine rankings. In that vein, design and test for the users on your site, not the search engines. If your site is getting visits but none of your business goals are being met, go nuts on testing those factors. Split test, iterate, and improve things with the focus of better conversions. Dive deep into services like Optimizely and research by the likes of Conversion Rate Experts.  Use focus groups and usability testing to see how the minutiae of your changes affects interaction. You can go as extreme as you want to in that regard.

                    Most importantly, the bulk of search engine ranking strength comes from external factors: the number and variety of sites linking to your site, the quality of sites linking to your site, the trust and high reputation of sites linking to your site, the semantic agreement of sites linking to your site, etc.  These factors are going to have many times greater influence in your ranking than your onsite tweaks in most cases.  If your site is functional and complies with Google's own guidelines (http://static.googleusercontent.com/media/www.google.com/en//webmasters/docs/search-engine-optimization-starter-guide.pdf) you've covered the bulk of what you need to do on site.  Focus instead on off site factors.

                    The site: search function exists mostly to provide searchers the ability to find a piece of information on a given domain. For example, let's say a reporter wants to cite an NBA stat from NBA.com, they'd use "stat thing" site:nba.com as a search.  For users, that's useful in searching specifics, and for Google that makes them look all the better at "categorizing the world's information."  Cache demonstrates the amount of information Google has archived and how quickly it's available. Back in the day--story time--Google used to advertise how quickly and how broadly they indexed things.  In fact, they still do!  If you look at a search result you'll notice a light gray statistic at the very top that says something like, "About 125,000,000 results (0.50 seconds)" for a search about hot dogs for example.  This is Google saying, "We're BIG and FAST." The precise details of your site are way down the list to Google's own story of being big and fast.

                    If you focus your efforts in off site optimization, linking with other reputable sites, and building your network you'll be way better served because you'll be getting referral traffic as well as lift in Google.  Cheers!

                    friendoffood max.favilli 2 Replies Last reply Reply Quote 2
                    • Travis_Bailey
                      Travis_Bailey @friendoffood last edited by

                      I'll PM my public IP through Moz. I don't really have any issue with that. Oddly enough, I'm still blocked though.

                      I thought an okay, though slightly annoying, middle ground would be to give me a chance to prove that I'm not a bot. It seems cases like mine may be few and far between, but it happened.

                      It turns out that our lovely friends at The Googles just released a new version of reCAPTCHA. It's a one-click-prove-you're-not-a-bot-buddy-okay-i-will-friend-who-you-calling-friend-buddy bot check. (One click - and a user can prove they aren't a bot - without super annoying squiggle interpretation and entry.)

                      I don't speak fluent developer, but there are PHP code snippets hosted on this GitHub repo. From the the documentation, it looks like you can fire the widget when you need to. So if it works like I think it could work, you can have a little breathing room to figure out the possible session problem.

                      I've also rethought the whole carpenter/mason career path. After much searches on the Yahoos, I think they may require me to go outside. That just isn't going to work.

                      friendoffood 1 Reply Last reply Reply Quote 1
                      • friendoffood
                        friendoffood @Travis_Bailey last edited by

                        Sorry about your being blocked.  A day hadn't passed due to the timing of your second visit--sorry.  I just changed the ip address in the table so you aren't blocked now.

                        Ok, I think I figured out what happened.  You first went to the ferguson site.  You may clicked on something but the same page was reloaded.  Then in a different tab you clicked on my home page from a google search results page.  Then in a third tab you went directly to my home page.  Then you ran screaming frog and the program stopped it without a message, seeing the word 'spider' in the useragent.  Then you tried it again and it recognized that as a stopped bot and gave the message about suspicious activity.

                        The program wipes out sessions and cookies when a user goes to the home page (it's not even a link anywhere) since that is just a location-choosing page, so when you opened it in a different tab the sessions were wiped out.  It had nothing to do with you being in incognito or not having cookies allowed.

                        Does this sound like what you may have done, and in sequence?

                        That's what it looks like, and that, if correct, is a huge relief for me since that is not usual user activity.  (Although I may have to reconsider whether its still a poor approach).

                        I don't know about what happened with your second visit and the timeout.  curious that you got some 60 pages crawled or so--I don't suppose you have anything that would tell me the first 3 of those so I can look into why it timed out?  The table isn't keeping the ip on crawling so I can only look those up by the url crawled and the time.

                        max.favilli Travis_Bailey 2 Replies Last reply Reply Quote 0
                        • friendoffood
                          friendoffood @RyanPurkey last edited by

                          What a great answer Ryan!  Thanks.  I'll tell you what my concern is.  As a coupon site I know that users don't want a bunch of wording at the beginning of pages.  They just want to find the coupon and get it, but from what I've read Google probably would reward the site more if there was beefier wording in it instead of a bunch of listings that are closer in some ways to just a bunch of links, resembling a simple link page.  I also have a 'mega-menu' on some of my pages which I think is pretty user friendly but have read that Google might not know for sure if it is part of the page content or not, as some forums I found talk about how their rankings improved when they simplified their menus.  Lastly, I have a listing of location links at the top of the page for users to 'drill down' closer to their neighborhood.  This is just about the first thing Google sees and may again be confusing to Google as to what the page is all about.

                          So IF the lack of 'wording content' and the early location of menu-type content is making my site hard to figure out from Google's perspective, I have alternatives and thought I could test those with Google ranking.  For example, I can enter wording content early on so as to 'beef' up the page so that it isnt just a bunch of coupon offer links.  I also could ajax the stuff that is above the 'coupon content' so that Google doesn't read that and get confused, and then put the actual links for Google to read down at the bottom of the page.  Both of those would be moves soley to satisfy Google and with no effect on the user  Google isn't perfect and i don't want to be penalized on ranking as a result of not addressing Google's 'imperfections', as it seems every edge counts and being on page 2 or 3 just isn't quite good enough.  I view this as reasonable testing rather than devious manipulation, but of course what matters with Google ranking is what Google thinks.

                          So in these cases the user response will be neutral -- they generally won't care if I have wording about what is on the page (especially if most requires clicking on 'more info') or am ajaxing the menu information--they again just want to find coupons.  But, if Google cares, as I have read they do, then it would be nice to be able to verify that with some simple tests.  It may be that my issues are somewhat unique as far as the typical webpage is concerned.

                          Having said all of that I do think your advice makes a ton of sense as the user is really what it is all about ultimately.

                          Thanks very much, and I'm giving you a 'good' answer as soon as I hear back!

                          RyanPurkey 1 Reply Last reply Reply Quote 0
                          • max.favilli
                            max.favilli @RyanPurkey last edited by

                            Ryan, I don't agree. It's true external factors (in other words backlinks) nowadays have the biggest impact, but on-page optimization as far as my little experience tell, still does affect ranking and it's worth working on.

                            And if we don't keep track of changes on pages and change on ranking how can we know what is working and what is not?

                            Especially since there's no gold rule and what works for one site doesn't necessarily work for another.

                            To make some example, I had a page which was ranking in position 1 for a search query with a volume of 50+k and very high competition. I expanded content to improve ranking for some additional queries, and it worked, it climbed from 2nd and 3rd serp page to 1st for a couple of those queries (I use both Moz ranktracker, semrush, and proracktracker to monitor ranking).

                            Unfortunately ranking for the search query with the highest volume moved from position 1 to postion 2, I changed the content a little bit, to add some keyword, which made sense because was re-balancing the keyword density now that the content was bigger. And in 24 hours it got back to position 1, without damaging the other search query improvement.

                            **In many other cases, I improved ranking on pages without any backlink, just improving the content, and I am talking about business critical pages with a high competition.

                            So I would say on-page optimization is still worth spending time on, to test the effect of the changes is a must and to monitor google ranking fluctuation is a must too.

                            Of course I am not saying off-page optimization is not important, is fundamental, I am giving that for granted.**

                            RyanPurkey 1 Reply Last reply Reply Quote 1
                            • max.favilli
                              max.favilli @friendoffood last edited by

                              Let me say straight forward, all that bot blocking is not a good idea.

                              I have been there in the past few times, especially for e-commerce, scraping to compare prices is very common, and I tried blocking scrapers many times, maybe I am not that good, but at the end I gave up because the only thing I was able to do was annoy legitimate users, and legitimate bots.

                              I do scrape other website too for price comparison, tens of websites, since I don't want to be blocked I split the requests among different tasks, I add a random delay between each request, I fake header data like user agent pretending to be Firefox from a windows pc, and I cycle through different proxies to continuously change IP address.

                              So as you can see, it's much harder to block scrapers than it seems.

                              Neither I would use JS to block cut&paste. I have no data to base my judgement on. But it's annoying for users, it doesn't sound compliant with accessibility, it stinks and google usually doesn't like things which stinks. Plus... If someone wants to scrape your content you are not going to block him that way.

                              friendoffood 1 Reply Last reply Reply Quote 1
                              • RyanPurkey
                                RyanPurkey @friendoffood last edited by

                                Thanks. There are two main parts to Google "figuring out" your site. One: indexation. That's been solved.  We know that you're getting indexed.  Two: ranking. Your site being so new and young, it's going to need backlinks and network growth to experience dramatic ranking changes.  If your menu isn't causing your site to be indexed poorly and your pages are being counted as unique enough, then you're ok now there as well.  The next most important step is getting your domain trust and authority up.

                                friendoffood 1 Reply Last reply Reply Quote 0
                                • RyanPurkey
                                  RyanPurkey @max.favilli last edited by

                                  Hi Massimliano.  I would disagree with myself if I was talking about your site too... ;^)  But in this specific case, qjamba.com is a site that needs the fundamental quality of backlinks more so than it needs Teddy to write a bot that is constantly pinging Google in order to try and decipher the incremental on-site changes he's making.  I'm speaking to his need to prioritize that aspect of optimization.  Copying what an SEO does when creating a nonsensical site with gibberish words in order to test on-page optimization as purely as possible with a normal, public facing website is a bad idea.

                                  Obviously on-page optimization is important, but again, in this specific example, Teddy isn't even discussing his keyword rankings, rather he was looking to go down an on-site optimization path that might make him more and more frustrated instead of bringing about much more positive results.  Cheers!

                                  friendoffood max.favilli 2 Replies Last reply Reply Quote 0
                                  • friendoffood
                                    friendoffood @RyanPurkey last edited by

                                    Some of my pages are on Google's page 2, 3 and a few on page 1 for certain search terms that don't have a lot of competition but that I know SOME people are using (they are in my logs). and those pages have virtually no backlinks.  I want to boost those on page 2 or 3 to page 1 as quickly as possible because p1 is 10x or more better than p2. Time/Cost is an issue here:  I can make changes overnight at no cost as opposed to blogging or paying someone to blog.

                                    Because domain authority and usage takes so long, it seems worth tweaking/testing NOW to try to boost certain pages from p2 or 3 to page 1 virtually overnight as opposed to waiting for months on end for usage to kick in. I don't know why Google would penalize me for moving a menu or adding content--basically for performing SEO on page, so it would be nice to be able to figure out what tools (cached pages, site:www. GWT, GA or otherwise) to look at to know if Google has re-indexed the new changes.

                                    Of course, the biggest pages with the most common search terms probably HAVE to have plenty of backlinks and usage to get there, and I know that in the long run that's the way to success overall when there is high competion, but it just seems to me that on page SEO is potentially very valuable when the competition is slimmer.

                                    1 Reply Last reply Reply Quote 0
                                    • friendoffood
                                      friendoffood @RyanPurkey last edited by

                                      I think there's been a misunderstanding.  I'm not writing a bot. I am talking about making programming changes and then submitting them to Google via the fetch tool to see how it affects my ranking as quickly as possible, instead of waiting for the next time Google crawls that page -- which could be weeks.  I think the early reply may have given you a different impression.  I want to speed up the indexing by fetching in Google the pages and then look to see what the effect is.  My whole reason for starting this thread was confusion over knowing how to tell when it was indexed because of unexpected results (by me) with the cache and site:www... on Google.

                                      RyanPurkey 1 Reply Last reply Reply Quote 0
                                      • RyanPurkey
                                        RyanPurkey @friendoffood last edited by

                                        Great! Well you have lots of insights here. It sounds like you're ready to test in the near term, and build up the domain in the long term.  Good luck!

                                        friendoffood 1 Reply Last reply Reply Quote 0
                                        • friendoffood
                                          friendoffood @RyanPurkey last edited by

                                          Well, I'm ready to test -- but still not quite sure how since I don't know how to tell when Google has indexed the new content, since sometimes it doesn't get cache'd and sometimes it disappears from the site:www.. listing.  I've read it only takes a couple of days after Google crawls the page, and can go with that, but was hoping there is a way to actually 'see' the evidence that it has been indexed.

                                          So, while I've gotten some great input, I am somewhat unsatisfied because I'm not sure how to tell WHEN my content has really been put in the index so that the algorithm is updated for the newly crawled page.

                                          RyanPurkey 1 Reply Last reply Reply Quote 0
                                          • RyanPurkey
                                            RyanPurkey @friendoffood last edited by

                                            Ah, that answer really varies per website.  For example, if you're site is a major news site, Google's indexation is extremely fast, measured in seconds not days.  Even if you're not a news site, major sites (high domain authority) get crawled and indexed very rapidly. Since you're going to be testing your own changes you'll learn how long this takes for your particular site.

                                            friendoffood 1 Reply Last reply Reply Quote 0
                                            • 1
                                            • 2
                                            • 3
                                            • 1 / 3
                                            • First post
                                              Last post
                                            • Does Google cache every page that is been indexed?
                                              donsilvernail
                                              donsilvernail
                                              0
                                              2
                                              73

                                            • Google is indexing wrong page for search terms not on that page
                                              katemorris
                                              katemorris
                                              0
                                              6
                                              1.1k

                                            • Our client's web property recently switched over to secure pages (https) however there non secure pages (http) are still being indexed in Google. Should we request in GWMT to have the non secure pages deindexed?
                                              N1ghteyes
                                              N1ghteyes
                                              0
                                              3
                                              128

                                            • What is the tool to check if a page (ex. a dynamic page) will properly be indexed by Google?
                                              ThompsonPaul
                                              ThompsonPaul
                                              0
                                              4
                                              78

                                            • Does Google still don't index Hashtag Links ? No chance to get a Search Result that leads directly to a section of a page? or to one of numeras Hashtag Pages in a single HTML page?
                                              Muhammad_Jabali
                                              Muhammad_Jabali
                                              0
                                              3
                                              748

                                            • Incorrect cached page indexing in Google while correct page indexes intermittently
                                              MikeRoberts
                                              MikeRoberts
                                              0
                                              2
                                              298

                                            • Merging your google places page with google plus page.
                                              junkcars
                                              junkcars
                                              0
                                              3
                                              518

                                            • Page Indexed but not Cached
                                              Matt-Williamson
                                              Matt-Williamson
                                              0
                                              2
                                              774

                                            Get started with Moz Pro!

                                            Unlock the power of advanced SEO tools and data-driven insights.

                                            Start my free trial
                                            Products
                                            • Moz Pro
                                            • Moz Local
                                            • Moz API
                                            • Moz Data
                                            • STAT
                                            • Product Updates
                                            Moz Solutions
                                            • SMB Solutions
                                            • Agency Solutions
                                            • Enterprise Solutions
                                            • Digital Marketers
                                            Free SEO Tools
                                            • Domain Authority Checker
                                            • Link Explorer
                                            • Keyword Explorer
                                            • Competitive Research
                                            • Brand Authority Checker
                                            • Local Citation Checker
                                            • MozBar Extension
                                            • MozCast
                                            Resources
                                            • Blog
                                            • SEO Learning Center
                                            • Help Hub
                                            • Beginner's Guide to SEO
                                            • How-to Guides
                                            • Moz Academy
                                            • API Docs
                                            About Moz
                                            • About
                                            • Team
                                            • Careers
                                            • Contact
                                            Why Moz
                                            • Case Studies
                                            • Testimonials
                                            Get Involved
                                            • Become an Affiliate
                                            • MozCon
                                            • Webinars
                                            • Practical Marketer Series
                                            • MozPod
                                            Connect with us

                                            Contact the Help team

                                            Join our newsletter
                                            Moz logo
                                            © 2021 - 2026 SEOMoz, Inc., a Ziff Davis company. All rights reserved. Moz is a registered trademark of SEOMoz, Inc.
                                            • Accessibility
                                            • Terms of Use
                                            • Privacy