Google Cached 800+ https pages????

Discussion in 'Site & Server Administration' started by Sem-Advance, Sep 22, 2006.

  1. #1
    I am lost here as I do not remember ever seeing this occur.

    Has anyone else???

    I have a new client with 810 cached pages in Google.

    Problem is they are behind a secure URL

    Everyone of the URLs Google returns starts as so:

    [SIZE=-1]https://www

    [/SIZE][SIZE=-1]https:/.../index.php

    Any idea whats going on???

    Why would Google cache https ????

    :confused:


    [/SIZE]
     
    Sem-Advance, Sep 22, 2006 IP
  2. GuyFromChicago

    GuyFromChicago Permanent Peon

    Messages:
    6,728
    Likes Received:
    529
    Best Answers:
    0
    Trophy Points:
    0
  3. Sem-Advance

    Sem-Advance Notable Member

    Messages:
    6,179
    Likes Received:
    296
    Best Answers:
    0
    Trophy Points:
    230
    #3
    I see

    Well the problem is these pages dont exist anymore on the server...

    and the client wants the http:// pages cached

    Can I use the URL removal tool safely do you think??
     
    Sem-Advance, Sep 22, 2006 IP
  4. GuyFromChicago

    GuyFromChicago Permanent Peon

    Messages:
    6,728
    Likes Received:
    529
    Best Answers:
    0
    Trophy Points:
    0
    #4
    I had the same problem awhile back. Someone had linked to an https page from another site and that caused all the https indexing...so I blocked https pages from being indexed. I never used the url removal tool but after a few months the https pages dropped from the index.
     
    GuyFromChicago, Sep 22, 2006 IP
  5. skore

    skore All-Star

    Messages:
    90
    Likes Received:
    1
    Best Answers:
    0
    Trophy Points:
    0
    #5
    Yeah - they should eventually drop - I don't think its neccessary to use the UR tool...
     
    skore, Sep 22, 2006 IP
  6. Sem-Advance

    Sem-Advance Notable Member

    Messages:
    6,179
    Likes Received:
    296
    Best Answers:
    0
    Trophy Points:
    230
    #6
    From what I am learning now in the past history the original designer / dev person submitted the https pages as that is how the entire site was set up.....

    Did you block the https through robots.txt or .htaccess?

    Thanks for the help & info
     
    Sem-Advance, Sep 22, 2006 IP
  7. GuyFromChicago

    GuyFromChicago Permanent Peon

    Messages:
    6,728
    Likes Received:
    529
    Best Answers:
    0
    Trophy Points:
    0
    #7
    If I recall correctly (it's been awhile) it was with robots.txt. Also did a 301 redirect - all https to http. That way if anyone linked to the https again in the future it wouldn't be an issue.
     
    GuyFromChicago, Sep 22, 2006 IP
  8. codeber

    codeber Peon

    Messages:
    578
    Likes Received:
    11
    Best Answers:
    0
    Trophy Points:
    0
    #8
    google search engine is screwed up.

    They havent fixed most of it since BigDaddy. Think yourself lucky it spidered your site. If do alot of google search like me, you'll find it littered with dead links (caching dead links aswell), sites which had 100,000 pages indexed and now have 10, and the supplementary search of death, where you have 1 page listed and then 99,999 others are supplementary's

    Google needs to fix their SE before going onto to other markets and products.
     
    codeber, Sep 22, 2006 IP
  9. GuyFromChicago

    GuyFromChicago Permanent Peon

    Messages:
    6,728
    Likes Received:
    529
    Best Answers:
    0
    Trophy Points:
    0
    #9
    Which pretty much has nothing to do with this thread.

    Thanks-for-nothing!
     
    GuyFromChicago, Sep 22, 2006 IP
  10. codeber

    codeber Peon

    Messages:
    578
    Likes Received:
    11
    Best Answers:
    0
    Trophy Points:
    0
    #10
    ^ i thought it is linked. (well except the last line, which went off in a bit of a random comment/rant.)

    The same problem which is causing the problems i mentioned, may also be the reason for google spidering pages you would not think it would normally.
     
    codeber, Sep 22, 2006 IP
  11. SSX

    SSX Guest

    Messages:
    11
    Likes Received:
    0
    Best Answers:
    0
    Trophy Points:
    0
    #11
    use the robots.txt
     
    SSX, Sep 22, 2006 IP
  12. Abhik

    Abhik ..:: The ONE ::..

    Messages:
    11,337
    Likes Received:
    606
    Best Answers:
    0
    Trophy Points:
    410
    Digital Goods:
    2
    #12
    yes, thts a good option. use the URL Removal Tool and then start afresh
     
    Abhik, Sep 22, 2006 IP
  13. GuyFromChicago

    GuyFromChicago Permanent Peon

    Messages:
    6,728
    Likes Received:
    529
    Best Answers:
    0
    Trophy Points:
    0
    #13
    Google will index any url it finds unless it's specifically being told not too. Take the proper steps on your site to control what's indexed and what's not and you'll be set. I would see it as a big problem if there were not relatively easy steps that webmasters can take to control it.
     
    GuyFromChicago, Sep 23, 2006 IP