Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is it better to drip feed content?
-
Hi All,
I've assembled a collection of 5 closely related articles each about 700 words for publishing by linking to them from on one of my pages and would appreciate some advice on the role out of these articles.
Backround: My site is a listings based site and a majority of the content is published on my competitors sites too. This is because advertisers are aiming to spread there adverts wide with the hope of generating more responses. The page I'm targeting ranks 11th but I would like to link it to some new articles and guides to beef it up a bit. My main focus is to rank better for the page that links to these articles and as a result I write up an introduction to the article/guide which serves as my unique content.
Question: Is it better to drip feed the new articles onto the site or would it be best to get as much unique content on as quickly as possible to increase the ratio of unique content vs. external duplicate content on the page that links to these articles**?**
Thank you in advance.
-
Good luck. Like I said, this is just me being silly. I pray to my Google shrine twice a day and this is what it tells me.
All at once or drip feed, either way your content gets up there!
-
Thanks guys for your help. Think I'm going to publish it all at once. Was originally in agreement with Bill but after doing a bit of reading it's probably safe to say that the SE's prioritise good content over content age. I've noticed blogs having slightly inflated PR because of the regular content but it's unlikely I'll be able to keep up regular posts and as a result any benefit derived from drip feeding would fall away when I run out of articles. If it doesn't work I'm calling my lawyer on you guys, hehe kidding :)))))
-
I don't think there is any right or wrong answer to this question. More of a preference.
For me, I like to drip my content.
In my own silly mind, it looks more natural to the search engines rather than dumping a bunch of content on your site.
I also think it keeps the search engines coming back to your site as you posting content through the months and years rather than all at one time.
Mind you. I have no scientific basis for this... just my own anal retentivity. LOL
-
When I have new content I can't wait to get it indexed. So even if I am not promoting it yet on the homepage I will put up links to it on relevant pages just to get spiders into it.
Five articles is no worry.
-
There is no advantage to holding back from a search engine perspective. The only reason I can think of to hold back relates to promotion opportunities for the articles. You could publish one article each week, tweet it and otherwise generate interest around the weekly article. If that is not of interest to you, then I would publish all five articles.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Best practices for publishing sponsored content
Hello, Our website hosts sponsored content from different brands. Should we be listing the sponsor either on the frontend and/or through markup? - Would either way have any sort of an impact? The content itself is already clearly marked as 'sponsored content' but we were more interested in listing the specific sponsor. Also, we’re assuming the outbound links would need to be marked rel="sponsored" but are there any other best practices we should be implementing? Any insight would be appreciated.
On-Page Optimization | | Ben-R
Thank you in advance.
Best,0 -
How to fix duplicate content for homepage and index.html
Hello, I know this probably gets asked quite a lot but I haven't found a recent post about this in 2018 on Moz Q&A, so I thought I would check in and see what the best route/solution for this issue might be. I'm always really worried about making any (potentially bad/wrong) changes to the site, as it's my livelihood, so I'm hoping someone can point me in the right direction. Moz, SEMRush and several other SEO tools are all reporting that I have duplicate content for my homepage and index.html (same identical page). According to Moz, my homepage (without index.html) has PA 29 and index.html has PA 15. They are both showing Status 200. I read that you can either do a 301 redirect or add rel=canonical I currently have a 301 setup for my http to https page and don't have any rel=canonical added to the site/page. What is the best and safest way to get rid of duplicate content and merge the my non index and index.html homepages together these days? I read that both 301 and canonical pass on link juice but I don't know what the best route for me is given what I said above. Thank you for reading, any input is greatly appreciated!
On-Page Optimization | | dreservices0 -
How does Indeed.com make it to the top of every single search despite of having aggregated content or duplicate content
How does Indeed.com make it to the top of every single search despite of having duplicate content. I mean somewhere google says they will prefer original content & will give preference to them who have original content but this statement contradict when I see Indeed.com as they aggregate content from other sites but still rank higher than original content provider side. How does Indeed.com make it to the top of every single search despite of having aggregated content or duplicate content
On-Page Optimization | | vivekrathore0 -
Less Tags better for SEO?
I am currently reviewing my strategy when it comes to categories and tags on my site. Having been no-indexed for some time, and having many tags with just one entry I am thinking that this is not optimal for SEO purposes. This is what I am planning: Categories - Change these to Index, but only after adding a hundred words or so by way of introduction (see this example - https://www.besthostnews.com/news/hosting/a-small-orange-news/). With the categories I am thinking of highlighting key articles as well to improve link juice distribution to older articles that are important. Tags - About half my tags have only 1 entry, with a few more just having 2 entries. I am thinking of deleting all tags with just one entry, and trying to merge those with just two or 3 entries where it makes sense to do so. I will keep these as no-index, but I think this will mean more optimal distribution of link juice within the site. I would appreciate your thoughts \ suggestions on the best practices here.
On-Page Optimization | | TheWebMastercom0 -
Duplicate page titles and Content in Woocommerce
Hi Guys, I'm new to Moz and really liking it so far!
On-Page Optimization | | jeeyer
I run a eCommerce site on Wordpress + WooCommerce and ofcourse use Yoast for SEO optimalisation I've got a question about my first Crawl report which showed over 600 issues! 😐 I've read that this is something that happens more often (http://cloudz.click/blog/setup-wordpress-for-seo-success). Most of them are categorized under:
1. Duplicate Page Titles or;
2. Duplicate Page Content. Duplicate Page Titles:
These are almost only: product category pages and product tags. Is this problem beeing solved by giving them the right SEO SERP? I see that a lot of categories don't have a proper SEO SERP set up in yoast! Do I need to add this to clear this issue, or do I need to change the actual Title? And how about the Product tags? Another point (bit more off-topic) I've read here: http://cloudz.click/community/q/yoast-seo-plugin-to-index-or-not-to-index-categories that it's advised to noindex/follow Categories and Tags but isn't that a wierd idea to do for a eCommerce site?! Duplicate Page Content:
Same goes here almost only Product Categories and product tags that are displayed as duplicate Page content! When I check the results I can click on a blue button for example "+ 17 duplicates" and that shows me (in this case 17 URLS) but they are not related to the fist in any way so not sure where to start here? Thanks for taking the time to help out!
Joost0 -
SEO value of old press releases (as content)?
Howdy Moz Community, I'm working with a client on migrating content to a new site/CMS and am wondering whether anyone has thoughts on the value of old press releases. I'm familiar with the devaluation of press release links from early 2013, but I'm wondering more about their value as content. Does importing old press releases (3-5 years old) create contextual depth of content that has some value for the site as a whole (even though the news contained within is useless)? Or, do these old press releases just create clutter and waste time (in migration). The site has a wealth of additional content (articles and videos), so the press releases wouldn't be covering up for thin content. I'm just wondering whether there's any best practices or a general rule of thumb. Thanks!
On-Page Optimization | | MilesMedia0 -
Does Rel=canonical affect google shopping feed?
I have a client who gets a good portion of their sales (~40%) from Google Product Feeds, and for those they want each (Product X Quantity) to have it’s own SKU, as they often get 3 listings in a given Google shopping query, i.e. 2,4,8 units of a given product. However, we are worried about this creating duplicate content on the search side. Do you know if we could rel=canonical on the site without messing with their google shopping results? The crux of the issue is that they want the products to appear distinct for the product feed, and unified for the web so as not to dilute. Thoughts?
On-Page Optimization | | VISISEEKINC0 -
Solve duplicate content issues by using robots.txt
Hi, I have a primary website and beside that I also have some secondary websites with have same contents with primary website. This lead to duplicate content errors. Because of having many URL duplicate contents, so I want to use the robots.txt file to prevent google index the secondary websites to fix the duplicate content issue. Is it ok? Thank for any help!
On-Page Optimization | | JohnHuynh0