Everything Eli said is spot on. Optimize your on-page and create compelling and valuable content on a regular basis focused on your customers/prospects and the links and PA/DA will follow.
- SEO and Digital Marketing Q&A Forum
- KevinBudzynski
KevinBudzynski
@KevinBudzynski
Job Title: CMO
Company: Monroe
Favorite Thing about SEO
The ever-changing landscape
Latest posts made by KevinBudzynski
-
RE: How to increase website Moz DA/PA?
-
RE: 403 error but page is fine??
This error can occur if you are blocking the bot trying to index your site. What is interesting , it's only for that page. Did you check your robots.txt file or .htaccess file?
-
RE: Quick SEO Audit of my site.
Outside of the ordinary on-page SEO, focus on speed and responsiveness. I would look at putting your site in https://www.webpagetest.org/ and see if you can prioritize scripts and optimize images. Also, run it through Chrome Dev Tools (F12) and play around. One example is look at the mobile view in there. Is the mobile page to long? Does it need different a different CTA? Where can you use media queries to hide and etc.? Good luck!
-
RE: Dropdown content on page being crawled
It depends. If the drop down uses ajax or related technology, it may be difficult for the spider to parse and may not be indexed. In most cases, yes it will be indexed. I guess we need to understand how the click has been implemented.
-
RE: How does EAT work?
The page looks great but is relatively new. Give it some time, get some inbound links and add content on regular basis. Also, query "What Eats A Squirrels" and look at the results. The bioexplorer.net has a great landing page for this query. and look for something that they are and are not doing. Good luck.
-
RE: Should I move my blog from subdomain to subdirectory?
Google is good at understanding the relationship between sub/primaries via linking and etc. and some of the issues may be more of server response time for the sub and etc.
I agree w/Ross--I prefer directories. However, if the structure has been around a while and migration is difficult, I don't see a huge benefit in migrating. Also, I believe there are some ways to use .htaccess that you can rewrite a subdomain to the directory.
See this video from Google's John Mueller for more information.
-
RE: Is DA (Domain Authority) increasing in 20 days gig worth it?
Yeah, second that. Remember DA is not Google. If a gig is offered for this service and seems to be to easy/sleazy or doesn't make sense, not worth doing.
Understand your point--I'm sure MOZ will continually update DA and close up the loopholes.
-
RE: Redirecting 2 established websites to 1 new one.
No, there isn't an issue with Google when you consolidate 2 sites into one. There is a slight erosion in link equity via 301, but it's nominal. Just make sure the old/new landing pages are closely related and you will be fine.
I know its nerve-racking, but make sure you do yourpyour work and fix any issues after implementation.
When you go live, I typically type site:nameofsite.com in google and click on the legacy link and make sure the old domain goes to the correct page on new site.
-
RE: Redirecting 2 established websites to 1 new one.
I agree with Joseph. The key to 301's redirects is to redirect to a landing page that is "closely" related. If there is no closely related page, best 404 as the user (as Joseph says) may feel cheated with the 301 and bounce. Also, a blanket 301 to a homepage is rarely a good idea. Yes, Google has no issue w/a 301 that is closely related. Good luck!
-
RE: Content that's behind CSS..
Yes, still valuable but less valuable than unhidden ("importnat content"). If the hidden content is an extension of your core/important content, no worries. Do what is best for the user even if you have a concern with drilling down for additional info. This is from G's website Guidelines:
"Make your site's important content visible by default. Google is able to crawl HTML content hidden inside navigational elements such as tabs or expanding sections, however we consider this content less accessible to users, and believe that you should make your most important information visible in the default page view."
Best posts made by KevinBudzynski
-
RE: KW density and idiot clients. HELP!!!!
Show them high-ranking content for the KW's and I can almost guarantee that it doesn't follow the criteria that your client is suggesting.
-
RE: Where is my Hug from Roger?
The rumor I heard was that Roger couldn't give hugs until his chiropractor fixed his backlinks.
-
RE: How important is it to update from a tabular lay-out to a div-layout?
Faster load time and a potentially better ux (specifically on mobile) caused by switching to div's is definitely a plus. The content sounds like it will remain the same, so I would not expect to see much or any organic fluctuations. IMHO, I would do it sooner than later.
-
RE: SEO transfer to new website
Make sure you 301 redirect as many pages as possible (if not all) to the proper landing page (see Moz's redirect tutorial). In most cases this will pass the majority of link juice (over 90%). Once launched, do a site:domain.com and click on each page to make sure it was redirected correctly. Also, monitor your webmaster tools for any alerts.
-
RE: Are directories still an option for linkbuilding
We still get relevant traffic and good leads from some directories. However, we only use reputable directories in certain niches (industrial directories). General rule of thumb, if it's difficult to get in a directory, it's probably a good place to be.
-
RE: When serving a 410 for page gone, should I serve an error page?
Yes, 410's will typically get out of index faster and I recommend serving an error page w/search so that the visitor knows that the site is still around.
-
RE: Is there any tool that can automatically generate backlinks for my website?
There are many tools that do this. However, I would suggest staying away from them. Spend that money on create valuable content for the users. There will be a better long-term roi and the risk of getting a algorithmic or manual action taken against your site is greatly minimized. Good luck!
-
RE: PDFs and SEO
Yes, the content is indexed (specifically if the spider can read the text--ie not embedded text inside an image). However, they are not optimal for a variety of reasons. Alex Fusman had a blog post on this a few years ago where he summarized optimizing pdf's for search. Here is his checklist:
Checklist for PDF Optimization:
- Search-friendly filenames
- Keyword-optimized titles
- Informative, concise descriptions
- Company name in "Author" field
- Use several relevant keywords in "Keyword" field
- Make sure to fill out all available fields - there is an option to view "Additional Metadata" (in Adobe Acrobat)
- Add tags to and accessibility options to your document
- Don't forget about Alt tags for images
- Add links back to relevant pages on the main website
- Write-protect the document
- Offer HTML version of the document
-
RE: COPIED CONTENT IS RANKING Above Orignal
Google doesn't always recognize original source consistently or accurately. However, they are getting better at spotting this. In this case, the scraped content may have been indexed sooner or another factor. First step is to attempt to contact the site owner. Most likely, this will not work. If it doesn't, file a DMCA. Also, implement a self-referral canonical tag and some internal links in the copy. Good luck!
-
RE: What effect does HTTPS have on SEO for a public site?
No, http/https doesn't heavily matter in the serps (https is a bit slower, so the answer is actually a bit). However, it may have to do w/the implementation of the ssl.
Great post on it from Franz: http://moz.com/ugc/seo-for-https-with-s-like-secure
"Do you work on a big company site where everything is within the secure protocol? If so, these are what you have to keep in mind for your onsite optimization:
- If done right, an https site can rank as well as an http site.
- Keep all your certificates valid, all the time, 24/7/365.
- Https sites are more work for the servers, so there is the possibility of a slower download. Keep your site and the certificate on a high performance host. If it takes too long to download a site Googlebot might go bye bye.
- The https-site and the http-site are two completely different sites; it’s a little bit like having a www version of the site and a non-www version. Make sure you have 301 redirects from the http URLs to the https ones.
- The robots.txt is only valid if placed within the same protocol as the site. Place it within the https protocol.
- The same goes for the sitemap.xml, Google verification file, and the other stuff.
Additionally: Keep all your files (images, iframes, etc) within the https protocol. There still is a nasty pop-up in Internet Explorer (6, maybe 7 too) if done otherwise. "
I enjoy digital marketing, family and running.