• davidgeipel

    (@davidgeipel)


    More of a feature request. You may be doing something special with sitemaps… just don’t know.

    Right now you give users the ability to provide their sitemap.
    LiteSpeed Cache > Settings

    Bottom of the page you allow us to specify our sitemap.
    “Custom Sitemap” and you state:
    “The crawler can use your Google XML Sitemap instead of its own. Enter the full URL to your sitemap here.”

    What I suspect is happening is that if you don’t have a server with crawler enabled, this won’t work then, right? Otherwise my questions were:

    1) What does this do?
    2) Are you caching just that single sitemap?
    3) If there are others in the sitemap, do you cache them too?
    4) If you need crawler, can you atleast cache a few of these pages without the use of the crawler. I did just recommend anoter feature where you have a “lite crawler” that can crawl say up to a few hundred pages. Server resources shouldn’t be as bad as what Wordfence uses…
    5) Would be nice to have a date purge for them too so perhaps daily our 20 or so xml files are recached.

    Since ours updates daily (we add about 300 new pages daily since we’re a real estate company), then can you create the abilty (like homepage) to set a time to purge it daily and cache them so it helps our indexing? We have about 10k links on each sitemap page.

    Feel free to check it out.

    HMSRE Site Index

    Any thoughts are always appreciated. Also be sure to see what I was calling a life-crawler too so it COULD still crawl a limited number of pages – even once a day to prime the cache is better than none.

    Thanks!

    David Geipel
    Dave

Viewing 1 replies (of 1 total)
  • Plugin Support LiteSpeed Lisa

    (@lclarke)

    Hi, Dave.

    1) What does this do?

    You are correct. The sitemap is only used for the crawler. If you have the crawler turned off (or your server doesn’t have it enabled) then the sitemap is not used for anything.

    2) Are you caching just that single sitemap?

    If you mean by the crawler, then yes. Anything that appears on the sitemap is visited by the crawler. If it has expired in the cache, then the crawler refreshes the page. If it’s not on the sitemap, it doesn’t get crawled.

    If you mean in general, then no. The sitemap doesn’t have any bearing on what is cached under normal conditions. To control what pages are cached in general, you would use the Cache and Excludes setting tabs.

    3) If there are others in the sitemap, do you cache them too?

    I’m not sure what you mean by “others.” Do you mean does the crawler read sitemaps within sitemaps? I’m not sure about that.

    4) If you need crawler, can you atleast cache a few of these pages without the use of the crawler. I did just recommend anoter feature where you have a “lite crawler” that can crawl say up to a few hundred pages. Server resources shouldn’t be as bad as what Wordfence uses…

    Unfortunately, without the crawler enabled, the only way to cache a page is to visit it. I saw your recommendation. It’s an interesting idea. I don’t know if it’s feasible – that’s something the development team will need to address.

    5) Would be nice to have a date purge for them too so perhaps daily our 20 or so xml files are recached.

    Is there a reason why you can’t set the TTL to 86400 seconds? That would cause pages to expire in the cache every 24 hours.

    If there are just a few specific pages you are targeting with a daily purge, and you’re not looking to make it site-wide, you can use the Scheduled Purge settings on the Purge tab.

    Neither of these options re-cache, though. They simply purge. The only way to re-cache a page is for someone (either a real person or a crawler) to visit the page.

    You can configure LSCache, though, so that the relevant pages are automatically purged when you post a new listing. All of the category, tag, archive, and index pages that include the listing would be purged, and the next person to visit the site would immediately see the most up-to-date information. With this tag-based purging, it shouldn’t be necessary for you to spend to much effort thinking about when and how often to purge your cache. It’s all being handled behind the scenes.

    The only thing missing is the ability to re-cache the purged pages automatically, and without the crawler, I’m afraid there’s nothing you can do about that.

    Does that help? Let me know if I’ve misunderstood any of your questions!

Viewing 1 replies (of 1 total)
  • The topic ‘LiteSpeed Cache Sitemaps’ is closed to new replies.