Jump to content

Matt

Management
  • Posts

    69,446
  • Joined

  • Last visited

  • Days Won

    559

Reputation Activity

  1. Like
    Matt got a reaction from Felipo Antonoff - Opencart for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  2. Thanks
    Matt got a reaction from Ricsca for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  3. Thanks
    Matt got a reaction from Cannabis Connect for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  4. Like
    Matt got a reaction from DreamOn for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  5. Like
    Matt got a reaction from Jalal arefen for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  6. Like
    Matt got a reaction from lordi for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  7. Like
    Matt got a reaction from levsha for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  8. Like
    Matt got a reaction from stu_m for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  9. Thanks
    Matt got a reaction from InvisionHQ for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  10. Like
    Matt got a reaction from SoloInter for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  11. Thanks
    Matt got a reaction from usmf for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  12. Like
    Matt got a reaction from Silnei L Andrade for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  13. Like
    Matt got a reaction from Darek_Hugo for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  14. Like
    Matt got a reaction from Ocean West for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  15. Like
    Matt got a reaction from 4joys for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  16. Like
    Matt got a reaction from sobrenome for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  17. Like
    Matt got a reaction from kyriazhs1975 for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  18. Like
    Matt got a reaction from Ioannis D for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  19. Agree
    Matt got a reaction from COLONER for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  20. Thanks
    Matt got a reaction from evcom for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  21. Thanks
    Matt got a reaction from steel51 for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  22. Thanks
    Matt got a reaction from Unienc for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  23. Like
    Matt got a reaction from BomAle for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  24. Thanks
    Matt got a reaction from Gabriel Torres for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  25. Like
    Matt got a reaction from 403 - Forbiddeen for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  26. Like
    Matt got a reaction from Markus Jung for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  27. Thanks
    Matt got a reaction from SeNioR- for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  28. Like
    Matt got a reaction from The Old Man for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  29. Like
    Matt got a reaction from Clover13 for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  30. Like
    Matt got a reaction from AlexJ for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  31. Like
    Matt got a reaction from Adlago for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  32. Like
    Matt got a reaction from sudo for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  33. Thanks
    Matt got a reaction from Ibai for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  34. Like
    Matt got a reaction from ASTRAPI for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  35. Thanks
    Matt got a reaction from IPCommerceFan for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  36. Like
    Matt got a reaction from OptimusBain for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  37. Like
    Matt got a reaction from Lance... for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  38. Like
    Matt got a reaction from Miss_B for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  39. Like
    Matt got a reaction from Rikki for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  40. Like
    Matt got a reaction from Jim M for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  41. Like
    Matt got a reaction from DawPi for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  42. Thanks
    Matt got a reaction from Maxxius for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  43. Like
    Matt got a reaction from Charles for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  44. Like
    Matt got a reaction from Marc Stridgen for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  45. Thanks
    Matt got a reaction from Sonya* for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  46. Thanks
    Matt got a reaction from LaCollision for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  47. Like
    Matt got a reaction from PPlanet for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  48. Like
    Matt got a reaction from sound for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  49. Like
    Matt got a reaction from aXenDev for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  50. Like
    Matt got a reaction from Real Hal9000 for a blog entry, SEO: Improving crawling efficiency   
    No matter how good your content is, how accurate your keywords are or how precise your microdata is, inefficient crawling reduces the number of pages Google will read and store from your site.
    Search engines need to look at and store as many pages that exist on the internet as possible. There are currently an estimated 4.5 billion web pages active today. That's a lot of work for Google.
    It cannot look and store every page, so it needs to decide what to keep and how long it will spend on your site indexing pages.
    Right now, Invision Community is not very good at helping Google understand what is important and how to get there quickly. This blog article runs through the changes we've made to improve crawling efficiency dramatically, starting with Invision Community 4.6.8, our November release.

    The short version
    This entry will get a little technical. The short version is that we remove a lot of pages from Google's view, including user profiles and filters that create faceted pages and remove a lot of redirect links to reduce the crawl depth and reduce the volume of thin content of little value. Instead, we want Google to focus wholly on topics, posts and other key user-generated content.
    Let's now take a deep dive into what crawl budget is, the current problem, the solution and finally look at a before and after analysis. Note, I use the terms "Google" and "search engines" interchangeably. I know that there are many wonderful search engines available but most understand what Google is and does.
    Crawl depth and budget
    In terms of crawl efficiency, there are two metrics to think about: crawl depth and crawl budget. The crawl budget is the number of links Google (and other search engines) will spider per day. The time spent on your site and the number of links examined depend on multiple factors, including site age, site freshness and more. For example, Google may choose to look at fewer than 100 links per day from your site, whereas Twitter may see hundreds of thousands of links indexed per day.
    Crawl depth is essentially how many links Google has to follow to index the page. The fewer links to get to a page, is better. Generally speaking, Google will reduce indexing links more than 5 to 6 clicks deep.
    The current problem #1: Crawl depth
    A community generates a lot of linked content. Many of these links, such as permalinks to specific posts and redirects to scroll to new posts in a topic, are very useful for logged in members but less so to spiders. These links are easy to spot; just look for "&do=getNewComment" or "&do=getLastComment" in the URL. Indeed, even guests would struggle to use these convenience links given the lack of unread tracking until logged in.  Although they offer no clear advantage to guests and search engines, they are prolific, and following the links results in a redirect which increases the crawl depth for content such as topics.
    The current problem #2: Crawl budget and faceted content
    A single user profile page can have around 150 redirect links to existing content. User profiles are linked from many pages. A single page of a topic will have around 25 links to user profiles. That's potentially 3,750 links Google has to crawl before deciding if any of it should be stored. Even sites with a healthy crawl budget will see a lot of their budget eaten up by links that add nothing new to the search index. These links are also very deep into the site, adding to the overall average crawl depth, which can signal search engines to reduce your crawl budget.
    Filters are a valuable tool to sort lists of data in particular ways. For example, when viewing a list of topics, you can filter by the number of replies or when the topic was created. Unfortunately, these filters are a problem for search engines as they create faceted navigation, which creates duplicate pages.

    The solution
    There is a straightforward solution to solve all of the problems outlined above.  We can ask that Google avoids indexing certain pages. We can help by using a mix of hints and directives to ensure pages without valuable content are ignored and by reducing the number of links to get to the content. We have used "noindex" in the past, but this still eats up the crawl budget as Google has to crawl the page to learn we do not want it stored in the index.
    Fortunately, Google has a hint directive called "nofollow", which you can apply in the <a href> code that wraps a link. This sends a strong hint that this link should not be read at all. However, Google may wish to follow it anyway, which means that we need to use a special file that contains firm instructions for Google on what to follow and index.
    This file is called robots.txt. We can use this file to write rules to ensure search engines don't waste their valuable time looking at links that do not have valuable content; that create faceted navigational issues and links that lead to a redirect.
    Invision Community will now create a dynamic robots.txt file with rules optimised for your community, or you can create custom rules if you prefer.

    The new robots.txt generator in Invision Community
    Analysis: Before and after
    I took a benchmark crawl using a popular SEO site audit tool of my test community with 50 members and around 20,000 posts, most of which were populated from RSS feeds, so they have actual content, including links, etc. There are approximately 5,000 topics visible to guests.
    Once I had implemented the "nofollow" changes, removed a lot of the redirect links for guests and added an optimised robots.txt file, I completed another crawl.
    Let's compare the data from the before and after.
    First up, the raw numbers show a stark difference.

    Before our changes, the audit tool crawled 176,175 links, of which nearly 23% were redirect links. After, just 6,389 links were crawled, with only 0.4% being redirection links. This is a dramatic reduction in both crawl budget and crawl depth. Simply by guiding Google away from thin content like profiles, leaderboards, online lists and redirect links, we can ask it to focus on content such as topics and posts.

    Note: You may notice a large drop in "Blocked by Robots.txt" in the 'after' crawl despite using a robots.txt for the first time. The calculation here also includes sharer images and other external links which are blocked by those sites robots.txt files. I added nofollow to the external links for the 'after' crawl so they were not fetched and then blocked externally.

    As we can see in this before, the crawl depth has a low peak between 5 and 7 levels deep, with a strong peak at 10+.

    After, the peak crawl depth is just 3. This will send a strong signal to Google that your site is optimised and worth crawling more often.
    Let's look at a crawl visualisation before we made these changes. It's easy to see how most content was found via table filters, which led to a redirect (the red dots), dramatically increasing crawl depth and reducing crawl efficiency.

    Compare that with the after, which shows a much more ordered crawl, with all content discoverable as expected without any red dots indicating redirects.

    Conclusion
    SEO is a multi-faceted discipline. In the past, we have focused on ensuring we send the correct headers, use the correct microdata such as JSON-LD and optimise meta tags. These are all vital parts of ensuring your site is optimised for crawling. However, as we can see in this blog that without focusing on the crawl budget and crawl efficiency, even the most accurately presented content is wasted if it is not discovered and added into the search index.
    These simple changes will offer considerable advantages to how Google and other search engines spider your site.
    The features and changes outlined in this blog will be available in our November release, which will be Invision Community 4.6.8.
  51. Like
    Matt reacted to Jordan Miller for a blog entry, Up your community moderation game with these quick tips   
    There’s no one-size-fits-all solution when it comes to moderating/guiding an online community. 
    Each digital world includes its own series of hurdles to overcome. As a community gains momentum and the registration count climbs higher, community moderators are presented with new challenges. These tasks become inherently more complex as a result, siphoning precious resources and oftentimes paint moderators into a corner.
    Take back the brush and illustrate a new scene with these supportive tips on how to better engage with your community. These pro-moderation tactics aim to foster a greater sense of togetherness and unity while also acknowledging individual members for their contributions. 
     
    Clear and concise community guidelines. 
    These are the building blocks for a solid foundation. Community guidance is less about different means of restriction and more about the ability to express one’s thoughts and feelings in a healthy way. All communities will inevitably run into toxic trolls, but instead of exerting energy on the negative, give praise and accolades to the positive.

    As we mentioned in a previous blog post about community guidance, show your community in the guidelines how you want them to post by contributing that way yourself. That sets a precedent. From there, you'll notice other community members participating in a way that is similar to you

    (TL;DR lead by example).
     
    Curate member content.
    Curating member content is the secret sauce to accelerated growth when it comes to community building. Featuring your members' posts publicly demonstrates your desire to embrace participating members. It’s one thing to comment on members’ topics, it’s another to feature and promote them for all to see. It gives added depth to your community by amplifying differing opinions, builds trust and encourages others to partake. It’ll take extra effort on your end to manually curate content (be selective regarding what topics you want to represent your community); establishing multiple authorities in your community builds momentum.

    Remember, people come for the content, but stay for the community. Give them both!


     
    Moderate on the offense.
    Moderating has a bad rap. Generally speaking, most view it as a prohibitive tool. A member posts profanity? Moderate! A topic goes live in the wrong section? Moderate! A spam bot infiltrates the latest blog entry? Moderate, moderate, moderate! That aspect of community building will always exist, but setting up automations will save you time and money. More importantly, it creates an elevated user experience because your attention is spent engaging with your members. 

    We have some powerful tools at your disposal to help, including a designated ‘automatic moderation’ section in your Admin Control Panel. Here, community leaders can set up rules that will trigger automatic content moderation. For example, you could create a rule that automatically hides content when it receives three or more user-submitted reports. If your members collectively agree another member’s post is problematic, and three or more of them report it, the post is hidden from public view until a moderator has a chance to review it. 

    Moderation doesn’t have to be a dirty word. And if it does, we have an automatic moderation tool for that, too. 
     
    Publicly recognizing members.
    Members who refuse to follow your community guidelines likely won’t fall in line with a public reprimand. Either privately message them your concerns or issue a warning only the moderation team can see.

    Instead, focus your attention on recognizing and rewarding contributing members. There’s a number of ways you can energize morale, including responding to members’ positive posts, manually or automatically awarding badges with our Achievements system, highlighting engaged members in a sidebar block, showcasing Leaderboard champs or even interviewing your MVPs. The idea is to approach community guidance with open arms – not a closed fist.
     

     
    Team work makes the dream work.
    There’s a reason it’s cliché. Organizing then empowering a moderation team is essential. Once on the same page in terms of expectations and processes, a community leader is then faced with the daunting task of relinquishing control as their team gears up to congratulate and moderate. 

    Your moderation team is only one moving part of a well-oiled machine. There are times, especially in the early days of a community where, as the community leader, you’re doing most of the heavy lifting. However, there will come a time when the community supports you in return and shares the weight. It’s important to recognize those instances and publicly acknowledge them!

    The world has enough doom and gloom; be an inspiration to your community and in return they’ll inspire you. 
    Do you have a community management tip to share? Drop it into the comments. We’d love to hear from you! 

    Header photo: Unsplash
  52. Like
    Matt got a reaction from Oshi for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  53. Haha
    Matt got a reaction from Notplus for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  54. Like
    Matt got a reaction from abobader for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  55. Like
    Matt got a reaction from ZLTRGO for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  56. Like
    Matt got a reaction from Alexandr.P for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  57. Like
    Matt got a reaction from BomAle for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  58. Like
    Matt got a reaction from sobrenome for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  59. Thanks
    Matt got a reaction from vahid67 for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  60. Thanks
    Matt got a reaction from Unienc for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  61. Like
    Matt got a reaction from WEBTH for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  62. Like
    Matt got a reaction from OptimusBain for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  63. Like
    Matt got a reaction from Noble~ for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  64. Like
    Matt got a reaction from ASTRAPI for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  65. Like
    Matt got a reaction from GazzaGarratt for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  66. Like
    Matt got a reaction from Maxxius for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  67. Like
    Matt got a reaction from Core4D for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  68. Like
    Matt got a reaction from Claudia999 for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  69. Like
    Matt got a reaction from shahed for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  70. Like
    Matt got a reaction from LiquidFractal for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  71. Like
    Matt got a reaction from nodle for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  72. Like
    Matt got a reaction from Markus Jung for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  73. Like
    Matt got a reaction from Ramsesx for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  74. Like
    Matt got a reaction from DawPi for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  75. Like
    Matt got a reaction from Marc Stridgen for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  76. Like
    Matt got a reaction from SeNioR- for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  77. Like
    Matt got a reaction from Jordan Miller for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  78. Like
    Matt got a reaction from Dprock for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  79. Thanks
    Matt got a reaction from Yamamura for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  80. Like
    Matt got a reaction from SoftwareFactory for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  81. Like
    Matt got a reaction from Real Hal9000 for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  82. Like
    Matt got a reaction from Chris Anderson for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  83. Like
    Matt got a reaction from Jimi Wikman for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  84. Like
    Matt got a reaction from Adriano Faria for a blog entry, Editor Stock Replies   
    Wouldn't it be nice if the Invision Community editor could re-use whole replies, text snippets, and even reply templates?
    As we get ready to welcome more customers into our staffed community support area, this feature idea has become a reality to help form personalized replies.
    Invision Community has a saved actions feature that allows the community team to perform multiple actions on multiple topics. For example, you might want to add a title prefix, move the topic and add a reply. This works great for 'canned' responses and actions, but it is less useful if you want to edit the reply to personalize it.
    Stock replies allow you to set up entire replies, partial replies or even reply templates.

    Stock replies via the editor
    Once you have these set up in the Admin Panel, they are visible on the editor.

    Stock replies are configured in the Admin Panel
    You can choose multiple stock replies to build up a message with handy re-usable reply snippets.
    stockreplies_video.mp4
    Each stock reply has full permission capabilities, meaning you can specify which member groups can use each stock reply. For example, you may wish to create partial replies for your team but encourage members to use a reply template to report bugs, etc.

    Using stock actions as a template
    We hope you like this feature, which is coming to our 4.6.7 October release.
  85. Like
    Matt got a reaction from shahed for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  86. Like
    Matt got a reaction from GTServices for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  87. Like
    Matt got a reaction from crmarks for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  88. Like
    Matt got a reaction from Yamamura for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  89. Like
    Matt got a reaction from AlexJ for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  90. Like
    Matt got a reaction from Pjo for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  91. Agree
    Matt got a reaction from Marco Junior for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  92. Like
    Matt got a reaction from Linux-Is-Best for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  93. Like
    Matt got a reaction from ASTRAPI for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  94. Like
    Matt got a reaction from Maxxius for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  95. Thanks
    Matt got a reaction from The Old Man for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  96. Like
    Matt got a reaction from Clover13 for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  97. Like
    Matt got a reaction from Jordan Miller for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  98. Like
    Matt got a reaction from sobrenome for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  99. Thanks
    Matt got a reaction from Abies for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  100. Like
    Matt got a reaction from SoloInter for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  101. Like
    Matt got a reaction from Kellen Deming for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  102. Like
    Matt got a reaction from aXenDev for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  103. Thanks
    Matt got a reaction from DawPi for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  104. Thanks
    Matt got a reaction from Thomas P for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  105. Agree
    Matt got a reaction from Myr for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  106. Agree
    Matt got a reaction from mcartemon2we23 for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  107. Thanks
    Matt got a reaction from IP-Gamers for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  108. Thanks
    Matt got a reaction from Ilya Hoilik for a blog entry, 3 Improvements to Spam Management in 4.6   
    Spam is as much a part of life on the internet as emoji overuse, serial GIF abuse and regretful tweeting.
    But I'm not here to talk about how I conduct myself online; I wanted to talk about three spam improvements coming to Invision Community 4.6.
    As you may be aware, Invision Community has its own Spam Defense functionality, which uses a mixture of crowdsourced data, publicly available data and our own special sauce to help reduce the number of spam accounts that get through the registration system. Invision Community also has several other tools to mitigate spam post-registration.
    These tools have served us well, but as spammers evolve, so must our systems. Here's what's coming to our next release.
    Spam Defense Scoring
    I can't divulge too much on our Spam Defense system lest we give spammers targeting Invision Community information that can assist them. Still, we have made several changes to our Spam Defense system.  These include rebalancing the score thresholds, checking against known TOR networks and proxies and using other data in the public domain to inform our scoring decisions.
    Spam Defense Blocking
    The current implementation of our Spam Defense only allows options to either prevent registration entirely or put the registration in an approval queue. However, the days when Spam Bots stood out from normal registrations are long gone, and it's hard to know if an account in an approval queue is legitimate or not.
    In 4.6, we've added a new Spam Defense option that you can choose to allow the registration but put the new members into the posting approval queue, meaning their posts will need moderator approval before being published.

    This reduces the decision burden and makes it easier to take a chance on a low score from the Spam Defense system and review their posts before they are made public.
    Word Filters
    We have added a new option to the Word Filters to allow content containing specific words or phrases to be held for moderator approval where the author has less than a set threshold of posts.
    For example, you may notice an increase in spam targeting "CBD Oil" and add it to the word filter list to hold the content for moderator approval. This works great and captures a good number of spam posts; however, your regular members get frustrated when they want to talk about CDB Oil in their posts.
    This new option allows you to set a trust level for allowing these words to be used without capturing them for approval.

    We hope these three changes to our spam controls will reduce the level of spam you get in your community!
    I'd love to know what's the weirdest spam (that is safe for work!) you've seen in your community.
  109. Like
    Matt got a reaction from media for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  110. Like
    Matt got a reaction from Morgin for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  111. Like
    Matt got a reaction from Lindy for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  112. Like
    Matt got a reaction from Miss_B for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  113. Like
    Matt got a reaction from Jordan Miller for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  114. Like
    Matt got a reaction from Ryan Ashbrook for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  115. Thanks
    Matt got a reaction from ipbhero for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  116. Like
    Matt got a reaction from Sonya* for a blog entry, 3 lessons content creators can learn from conspiracy theories   
    Conspiracy theories have roots in the 19th century and have been popular for decades. Until recently, conspiracy theorists have lived in the margins. They are often convinced the earth is flat, Lee Harvey Oswald didn't act alone, and the moon landings were faked in a Hollywood sound stage.
    More recently, with 9/11 and the coronavirus pandemic, these conspiracy theories have become more mainstream, with celebrities and politicians sharing them over their official social media channels. From the evil machinations of Bill Gates, the rise of QAnon, to the conflation that 5G is responsible for spreading coronavirus, it's hard to ignore the impact they have in creating misinformation which undermines attempts at effective communication from governments and public health bodies.
    Despite reams of facts, logic and critical thinking, those that follow conspiracy theories will not be budged from their positions. They trust their sources implicitly, and a mountain of research disproving the argument does not interest them.
    The number of people that succumbs to these narratives grows every day. When you consume the content shared by the primary sources of this misinformation, it's easy to see why.
    Conspiracy theories are created and shared in a way that is engaging and irresistible to many seeking stability in a confusing world. Whatever your position is on these conspiracy theorists, you can leverage these tactics to make your own content more engaging and shareable.
    Lesson 1: Make it emotive
    Human beings have two distinct and independent thinking centres in the brain. One works on emotion (the limbic system) and the other on logic (the neocortex).
    The emotional brain works much faster than the logical brain. It is what has kept us alive as a species. If you hear a loud bang, your emotional brain processes this first and triggers the urge to move before your logical brain kicks in and deduces the bang was from a book expertly pawed from its shelf by your cat.
    The emotional brain is continually processing the world, and even though it's part of you, you do not have much control over it. Your logic brain, however, works on facts, truths and analysis.
    When you watch harrowing whistleblower testimony telling of their suffering in a conspiracy theory video, your emotional brain is powerfully stirred.
    It's why challenging conspiracy theorists who are emotionally committed to the point of view with just logic often fails. The emotional commitment is incredibly powerful, and when you challenge them, the logic brain is short-circuited, and the emotional brain becomes defensive. In fact, the more logic and evidence you provide, the more the emotional brain digs in and refuses the new evidence.
    How can you use this to your advantage?
    Work on creating an emotional response with your content. Don't purely rely on facts and logic to persuade your audience. Try and evoke an emotional reaction through imagery, metaphors and similes.
    President Obama was a powerful orator and used emotion often to create a strong message. When he spoke of investing in education, he invokes emotion by saying "We believe that when she goes to school for the first time, it should be in a place where the rats don't outnumber the computer."
    Lesson 2: Tell a story
    Conspiracy theory videos don't just reel off a list of events and facts, they tell a story. Some of the more complex theories are akin to a sprawling TV series with several characters linked by circumstance.
    Humans have always been curators of stories. From religious texts to morality fables, we learn and process the world through stories. Stories are memorable. Most adults can recite fairytales read to us when we were children.
    Use a story to link together critical points within your content.
    Consider how "Gamification has been proven to make communities more sticky and encourage more engagement" reads compared to "It was 3am, the flicker of the TV set was the only light in the room. My palms, slick with sweat, fought to keep the controller sticks moving. Even though I had a 6am start, I couldn't put the controller down. I had to finish the quest and collect the reward. Your community is no different."
    Take your reader on a journey, and they're more likely to finish your content. Try and make it personal. When we read, we always try and put ourselves in the shoes of the author or the protagonist.
    Stories and emotion go hand in hand. Recently, the Huffington Post ran a story with the headline "One death a minute" which is a very emotive and powerful alternative to the raw fact that 1,461 Americans lost their lives to COVID-19 on the 29th July.

    Lesson 3: Make it easy to consume
    A key strength for any content creator is to know when to create long-form content and snackable content.
    A single meme is more potent than 300 links to PubMed. A single YouTube video can be more persuasive than an expert in her field.
    Conspiracy theory creators use over-simplification to reduce a complex issue into an easily digestible entertaining snack. A meme generally contains a single idea that is easy to grasp and engaging. You don't have to work very hard to understand it, your visual brain processes it in 1/10th of a second, and it triggers a moment of delight.
    Infographics and memes are often smart ways to create an entrance to your content. If an image containing a straightforward idea from a more complex piece of content is digested quickly, it can leave your audience wanting more, and therefore more likely to involve themselves in your more complex work.

    When creating long-form content, consider the use of iconography, infographics and photography. Visuals help us remember and understand content quickly. I could say that 63% of this blog was written on an iPad, but a piechart would make this easier to process and more memorable.
    No tin foil hats required
    Creating compelling content is key to building your community. Your content sets the tone, helps drive re-engagement and positions you as a key expert in your field. Using the techniques many conspiracy theory creators use to spread their narratives will help your content be more memorable and shareable
    A well-created story with emotional cornerstones made more accessible by key points simplified into snackable quotes or images will help your content find a wider audience, whether you believe Neil Armstrong landed on the moon or not.
  117. Like
    Matt got a reaction from Emanoel for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  118. Like
    Matt got a reaction from MacOS86it for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  119. Thanks
    Matt got a reaction from O9C4 for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  120. Like
    Matt got a reaction from The Old Man for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  121. Like
    Matt got a reaction from tonyv for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  122. Like
    Matt got a reaction from Martin A. for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  123. Like
    Matt got a reaction from princeton for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  124. Like
    Matt got a reaction from Mark H for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  125. Like
    Matt got a reaction from media for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  126. Like
    Matt got a reaction from TSP for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  127. Like
    Matt got a reaction from Maxxius for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  128. Like
    Matt got a reaction from DawPi for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  129. Like
    Matt got a reaction from Mandalala for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  130. Like
    Matt got a reaction from Nebthtet for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
  131. Like
    Matt got a reaction from Daniel F for a blog entry, How to keep your community secure   
    Security should never be an afterthought. Don't wait until an attack has compromised your site before you take action.
    All too often, site owners consider increasing their security only when it's too late, and their community has already been compromised.
    Taking some time now to check and improve the security of your community and server will pay dividends.
    In this blog, we run down 8 ways that you can protect your community with Invision Community. We go through the security features you may not know about to best practices all communities should be following.
    1. Set up Two Factor Authentication
    Invision Community supports Two Factor Authentication (2FA for short), and we highly recommend making use of this feature for your users, but especially for your administrative staff.
    2FA is a system that requires both a user's password and a special code (displayed by a phone app) that changes every few seconds. The idea is simple: if a user's password is somehow compromised, a hacker still wouldn't be able to log in to the account without the current code number.
    You may already be familiar with 2FA from other services you use. Apple's iCloud, Facebook and Google all offer it, as do thousands of banks and other security-conscious businesses.
    Invision Community supports 2FA via the Google Authenticator app (available for iOS and Android) or the Authy service, which can send codes to users via text message or phone call. You can also fall back to security questions instead of codes.
    You can configure which members groups can use 2FA, as well as requiring certain groups to use it. 
    Recommendation: Require any staff with access to the Admin Control Panel or moderation functions to use 2FA. This will ensure that no damage will occur should their account passwords be discovered. Allow members to use 2FA at their discretion.
    2. Configure password requirements
    The password strength feature displays a strength meter to users as they type a new password. The meter shows them approximately how secure it is, as well as some tips for choosing a good password.
    While you can leave this feature as a simple recommendation for users, it's also possible to require them to choose a password that reaches a certain strength on the meter. 
    Recommendation: Require users to choose at least a 'Strong' password.

    3. Be selective when adding administrators
    Administrator permissions can be extremely damaging in the wrong hands, and granting administrator powers should only be done with great consideration. Giving access to the AdminCP is like handing someone the keys to your house. Before doing so, be sure you trust the person and that their role requires access to the AdminCP (for example, would moderator permissions be sufficient for the new staff member?).
    Recommendation: Don't forget to remove administrator access promptly when necessary too, such as the member of staff leaving your organization. Always be aware of exactly who has administrator access at any given time, and review regularly. You can list all accounts that have Administrative access by clicking the Administrators button under staff on the Members tab.
    4. Utilize Admin Restrictions
    In many organizations, staff roles within the community reflect real-world roles - designers need access to templates, accounting needs access to billing, and so forth. 
    Invision Community allows you to limit administrator access to particular areas of the AdminCP with the Admin Restrictions feature, and even limit what can is done within those areas.
    This is a great approach for limiting risk to your data; by giving staff members access to only the areas they need to perform their duties, you reduce the potential impact should their account become compromised in future.
    Recommendation: Review the restrictions your admins currently have. 
    5. Choose good passwords
    This seems like an obvious suggestion, but surveys regularly show that people choose passwords that are too easy to guess or brute force. Your password is naturally the most basic protection of your AdminCP there is, so making sure you're using a good password is essential.
    We recommend using a password manager application, such as 1password or LastPass. These applications generate strong, random passwords for each site you use, and store them so that you don't have to remember them.
    Even if you don't use a password manager, make sure the passwords you use for your community are unique and never used for other sites too.
    Recommendation: Reset your password regularly and ensure you do not use the same password elsewhere.

    6. Stay up to date
    It's a fact of software development that from time to time, new security issues are reported and promptly fixed.
    But if you're running several versions behind, once security issues are made public through responsible disclosure, malicious users can exploit those weaknesses in your community.
    When we release new updates - especially if they're marked as a security release in our release notes - be sure to update promptly.
    Invision Community allows you to update to the latest version via the AdminCP. You no longer need to download a thing!
    Recommendation: Update to the latest version whenever possible. Remember, with Invision Community's theme and hook systems, upgrades to minor point releases should be very straight forward.
    7. Restrict your AdminCP to an IP range where possible
    If your organization has a static IP or requires staff members to use a VPN, you can add an additional layer of security to your community by prohibiting access to the AdminCP unless the user's IP matches your whitelist.
    This is a server-level feature, so consult your IT team or host to find out how to set it up in your particular environment.
    Recommendation: Consider IP restriction as an additional security layer when you are not able or willing to use 2FA.
    8. Properly secure your PHP installation
    Many of PHP's built-in functions can leave a server vulnerable to high-impact exploits, and yet many of these functions aren't needed by the vast majority of PHP applications you might run. We, therefore, recommend that you explicitly disable these functions using PHP's disable_functions configuration setting. Here's our recommended configuration, although you or your host may need to tweak the list depending on your exact needs:
    disable_functions = escapeshellarg,escapeshellcmd,exec,ini_alter,parse_ini_file,passthru,pcntl_exec,popen,proc_close,proc_get_status,proc_nice,proc_open,proc_terminate,show_source,shell_exec,symlink,system Another critical PHP configuration setting you need to check is that open_basedir is enabled. Especially if you're hosted on a server that also hosts other websites (known as shared hosting), if another account on the server is comprised and open_basedir is disabled, the attacker can potentially gain access to your files too.
    Naturally, Cloud customers needn't worry about this, we've already ensured our cloud infrastructure is impervious to this kind of attack.
    Recommendation: Review your PHP version and settings, or choose one of our cloud plans where we take care of this for you.
    So there we go - a brief overview of 8 common-sense ways you can better protect your community and its users.
    As software developers, we're constantly working to improve the behind-the-scenes security of our software. As an administrator, there's also a number of steps you should take to keep your community safe on the web.
    If you have any tips related to security, be sure to share them in the comments!
     
×
×
  • Create New...