"I have nooooo idea what to do with all this @screamingfrog information" - @danielkcheung, circa 2018-2020

If you want to learn to do some #technicalseo with one of the best crawling tools out there, here is a 🧵for you.
.@screamingfrog is a crawler that you download and install onto a local or virtual machine. It allows you to crawl almost any website.

I use it as part of my technical seo audit and their customer service is 🔥

However, the initial learning curve is steep.
If you've looked at a finished @screamingfrog crawl and felt overwhelmed - you're NOT alone.

I, and many others, were in your EXACT shoes.

It's ok
You've got this💪🏽

⏩ Here a 9 tips to make your life with Screaming Frog easier, more enjoyable, and F-U-N👇
1. Run in database storage mode

Configuration > System > Storage mode

Doing so will save your crawl data automatically to your computer and allows you to compare between crawls of the same site.
2. You don't need to crawl an entire site

Crawling an entire website can take hours (sometimes days).

🧠You don't *always* need to have full information to make recommendations. Often, a sample is all you need to ID symptoms and make relevant recommendations.
3. Add sitemap address to get sitemap insights

Because many robots.txt files do not reference the sitemap index URL you will have to add it yourself👇
4. Run crawl analysis to see sitemap issues such as:

⚠️ orphaned URLs
⚠️ URLs not in sitemap
⚠️ non-indexable URLs in sitemap

Once your crawl has finished, go to Crawl Analysis > Start then click on the Sitemaps tab and the right hand side panel will display sitemap stats👇
💡FYI, you can run a crawl analysis at any stage of the crawl - you do not have to wait to crawl the entire site.

This is great for crawling large sites and you want to get a sense if there are issues with the its sitemap.

Click Pause > Crawl Analysis > Start👇
💡A website's sitemap is one of the first things I look at.

🧠If I see many non-indexable URLs in the sitemap or many orphaned URLs, this is a strong signal that there are a host of technical issues and warrants a full audit.

Learn more about sitemaps👇
danielkcheung.com.au/xml-sitemap-fo…
5. Crawl a *specific* subfolder of a site (and nothing else)

Eg, I want to crawl all the URLs within /blog/ subfolder of a website.

To do this, go to Configuration > Include then put in https://domain/blog/.*

⚠️Don't forget the ‼️.*‼️
6. On your first crawl, crawl HTML files only

By default, @screamingfrog will crawl all image, CSS, JS and SWF files.

💡Uncheck these for your initial crawl to get a quick sense of the URLs and will save you some time.
7. Run crawler in Javascript mode

Many websites are single-page applications or rely on JS to render ⚡️important⚡️ content. Running in JS mode in @screamingfrog will give you a proxy of how well (or not) Google can crawl and render any website.
To do this:

1⃣ Configuration > Spider > Rendering
2⃣ Change "text only" to "javascript"

When crawling in JS mode:
@screamingfrog takes a lot longer (so be patient)
• go to Rendered Page to see if all important content can be rendered
It *literally* took me 3.5 years to discover the Rendered Page tab (👏@myriamjessier for showing me👏).

This is where you'll find the rendered view when you run a JS crawl in @screamingfrog 👇
💡If you don't see the body content shown in the Rendered Page panel, this is probably a strong signal that Google will have issues rendering text and internal links on the website.

💡If this happens, corroborate with Google Search Console data.
Since @screamingfrog takes⏳to crawl in JS mode, TameTheBots by @davewsmart is ⚡️ah-mazing⚡️ for quick diagnosis on a page-by-page basis.

IE, if important content is not displayed - this indicates a potential issue with rendering, indexing and ranking.

tamethebots.com/tools/fetch-re…
8. Look for 301, 302, 404 response codes

💡Most sites will have URLs that 301/302 redirect or have pages with 404 response codes.

After a crawl go to the Response Codes tab and filter URLs in ascending/descending order. This will show you all the URLs that are non-200👇
💡A really quick way to see this is by running as little extractions as possible.

You can do this by unchecking all "Resource Links" in Crawl Configuration and unchecking everything in Page Details in "Extraction".
9. Find low-content pages

URLs with less than 200 words often lack depth and may contribute to poor indexing of URLs. Luckily @screamingfrog can easily show you these URLs.

💡To find these run a Crawl Analysis, go to Content Tab then look for Low Content Pages.
💡 Not *all* pages with less than 200 words = bad. You'll have to use your own judgement to decide if the ones @screamingfrog has shown you are appropriate or not.

Recommended reading re: thin content pages (via @JonasSickler)👇
terakeet.com/blog/what-is-t…
And that's a wrap! I trust you found this insightful and inspired you to want to work with @screamingfrog even if you don't fully understand it.

For more, give me a follow
@danielkcheung to see how else I can inspire you in your SEO journey👇

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Daniel K Cheung (He/Him)

Daniel K Cheung (He/Him) Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @danielkcheung

May 13
This is a FAQpage rich result as displayed on the SERPs.

In this 🧵I'm going to show you how you can do it yourself.
For those who already know how to get FAQpage rich result using FAQPage markup, skip straight to the inserting🔗into your JSON-LD part - I've got a blog post on it👇
danielkcheung.com.au/links-in-faq-s…
First, what is a FAQpage rich result?

"A Frequently Asked Question (FAQ) page contains a list of questions and answers pertaining to a particular topic. Properly marked up FAQ pages may be eligible to have a rich result on Search" - Google documentation👇
developers.google.com/search/docs/ad…
Read 27 tweets
May 13
Do you want to use Help A Reporter Out (HARO) to land links?
HARO is a free service provided by Cision that enables journalists and PR sources to collaborate.

That is, journalists seek commentary from a source and in return, you *may* get a backlink from the publication they write for.
Anyone can sign up to HARO as a PR source. And it is completely FREE.

There is a paid version but in all honesty, you do NOT need it.

1. Go on, sign up now👇

helpareporter.com/subscriptions/…
Read 19 tweets
May 13
There are *very few things* you have control over in SEO. The depth, quality and expanse of your content is one of these.

Here are 3 examples of how onpage/onsite SEO produced results.

🧵 ..
1/ I researched and published popular wedding venue content for a wedding photographer. We chose venues that would align with his target audience (re: budget, values and visual style).

It took a few months to rank👇
Even throughout COVID and lockdown, the page continued to get organic traffic from searches from potential customers.

More importantly, I *know* my client has booked weddings from this onsite content.
Read 12 tweets
May 12
Hi, I'm Dan.

Here are a few things about me:
• born🇭🇰
• raised🇦🇺
• married to🇰🇷

I share things to do with #SEO and my DMs are open so if you have a question about search or working in search, ask me!

Here are a few of my threads👇
Copy my 🦺 technical SEO checklist to audit your next Wordpress website 👇

How to using a 🌊 rising tide strategy of long tail keywords to rank your website for competitive head terms👇

Read 9 tweets
May 12
What is one thing hiring managers look for when recruiting for a junior SEO position?
✔️ Transferrable skills + experience
✖️ College degree
✔️ Direct experience

But you may not have direct experience - let's solve that right now!

A 🧵 for those of you who are thinking of moving joining the SEO profession 👇
Having your own website teaches you empathy.

Eg, when you realise GoDaddy/Bluehost is a 💩 and need to migrate to another host.

You will realise that it is not so easy.

Can you imagine how confusing and frustrating it is for non-tech business owners?
Read 16 tweets
May 11
How to rank for a head term keyword via long tail keywords, without link-building, for new niche websites, service-based businesses, and any type of website.

🎁Plus a bonus FREE @googledocs content brief template.

🧵👇
First of all, head terms are popular searches.

For example:
▪️ wedding dress
▪️ bitcoin
▪️ credit card
▪️ iPhone
▪️ SEO freelancer
▪️ dentist near me
▪️ jeans

But here's the thing: ranking for head terms are f*cking competitive. Like seriously 💀😅
Eg, using @semrush, "wedding dresses" has a global search volume of 1.2 million PER MONTH and the regional breakdowns are *very* respectable.
Read 26 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(