How answering People Also Ask questions correlates with organic rankings (Study)

Feb 16, 2026 8 min read By Chris Green
Chris Green

Chris Green

Chris Green is the Technical Director at Torque Partnership, with over 15 years of experience in technical SEO and digital strategy. A frequent speaker at major industry conferences like BrightonSEO, he specialises in complex site migrations, auditing – working with for enterprise brands while actively mentoring the next generation of SEOs.

Table of contents

Understanding user needs is critical for search engines, LLMs, and every industry that relies on these platforms to get in front of audiences. A huge part of this is understanding the questions users want answered.

But, discerning the full intent driving a user with just the keyword they searched is a challenge—words have different meanings, people don’t know exactly what they’re looking for, or sometimes do not articulate their needs that well.

It is now the search engines/LLM’s job to simultaneously read minds and judge what content answers those questions and (more recently) preempt the subsequent questions/queries. So, if you know (or can predict) the questions users will ask, you can understand what content is the most valuable.

A lot is being written right now about how SEOs can potentially simulate some of these more sophisticated methods of synthesising questions (i.e., ‘query fan-out’ from Google’s AI Mode). But I think we have an even more accessible proxy than that—People Also Ask (PAA) questions.

If we take a seed keyword, build a list of potential questions related to that search term, and then judge the ranking websites against how well they answer those questions, can we learn more about the value of this approach?

Takeaways

  • Across the top 5 positions, pages that fully answer more PAA questions tend to rank higher.
  • Commercial and transactional queries show the strongest correlations; navigational and informational queries are weaker.
  • Beyond page 1, the correlation drops sharply and can even reverse in lower page 2 positions—likely due to small sample sizes and SERP variability.
  • Answering PAA questions alone isn’t enough to rank well; technical quality, external signals, and user experience still matter.
  • Covering a wide range of related questions can make content semantically richer, which may support ranking potential.

Methodology

Armed with AlsoAsked, SerpAPI, and OpenAI’s API, I think we can explore the correlation between rankings and PAA questions answered. Here’s how:

  1. Start with a list of seed keywords.
  2. Fetch the top 20 organic results.
  3. For each keyword, check PAA questions using the AlsoAsked API.
  4. For each organic result, extract the main readable content.
  5. Send the content and each question to GPT-3.5 Turbo, asking, “Does this page answer this question? Fully, partially, or not at all?”.

This enables us to see:

  • The number of PAA questions answered by results ranking within the same SERP.
  • Whether the number of PAA questions answered correlates with rankings.
  • How all of the above is impacted by search intent.

Note: During this study, I found several areas that I could improve upon in the future. You can find those opportunities and caveats following the conclusion section of this study.

The detailed process

If you want to replicate this study, I’ve covered the major portions of the process below:

  • Keyword selection
  • Working pipeline
  • API specifics

Keyword selection

For this test, I needed a significant number of keywords to smooth over potential biases or low-level data issues. I conducted this test using:

  • 563 keywords
  • 4 search intent stages
  • 3 niches

The keywords were seeded per niche using the DataForSEO’s Keyword Ideas API input, so they were broadly related, essentially random, and grouped around the four standard intent stages (informational, navigational, commercial, and transactional).

Intent

Keywords

Questions

URLs

Intent

Commercial

Keywords

142

Questions

2274

URLs

1470

Intent

Informational

Keywords

139

Questions

2213

URLs

1455

Intent

Navigational

Keywords

137

Questions

2282

URLs

1443

Intent

Transactional

Keywords

145

Questions

2249

URLs

1580

Intent

Grand total

Keywords

563

Questions

7905

URLs

5415

Working pipeline

To conduct my analysis, I followed the steps below:

  1. Preparation and setup
    • Connect to data services by API.
    • Build a list of seed queries (keywords) to research.
  2. For each keyword (seed query)
    • Load AlsoAsked related questions for each keyword.
    • Get Google results — Uses the SerpAPI to fetch the top Google search results for that keyword.
  3. For each result URL
    • Extract main content: Uses a browser (headless Chrome) to load the page and extract the main readable text using a ‘readability’ API.
      • Extract main content: Uses a browser (headless Chrome) to load the page and extract the main readable text using a ‘readability’ API.
      • If content is insufficient (too short or missing), skip the page.
  4. For each question
    • Send the extracted content to GPT-3.5 Turbo and prompt it to evaluate whether the content answers the question “fully”, “partially”, or “not at all”, and provide reasoning for its answer.
    • Record results and store the status and reasoning for each keyword, question, and URL.
    • Save all results for each keyword and save the collected results into a CSV file.
  5. Logging and error management
    • Log any errors (API failures, extraction issues, etc.) with timestamps for troubleshooting later.

API specifics

If you want to replicate this yourself, here (or conduct similar research) here are the specifics around the APIs used for the pipeline.

  • SerpAPI fetches organic URLs from the top 20 results.
  • OpenAI’s GPT-3.5 Turbo is more than adequate for the task at hand, whilst keeping costs manageable.
  • AlsoAsked fetches to a depth of 20 questions to keep them closely associated with the seed term, whilst not being too limiting.

Results

The results below show the number of questions answered by the top 20, 10, and 5 positions across the four intent stages.

For further analysis, I conducted the same test with partially answered questions (using a ‘question score’). The results of that analysis are below as well.

Questions answered by position

Questions answered by intent top 20A line chart showing "Questions answered by intent" across ranking positions 1 through 20. The first half of the chart (positions 1-10) is relatively stable, while the second half (positions 11-20) shows high volatility. Notable spikes occur at the end of the chart, with Navigational (pink) spiking at position 19 and Transactional (dark blue) spiking sharply at position 20.Questions answered by intent top 10A line chart extending the "Questions answered by intent" data to ranking positions 1 through 10. Informational intent (blue) remains the highest overall, with a rebound peak at position 7. Transactional (dark blue) remains the lowest. Commercial (green) and Navigational (pink) show fluctuations in the middle range, with all lines showing a slight general downward trend or volatility as position numbers increase.Questions answered by intent top 5A line chart displaying "Questions answered by intent" for ranking positions 1 through 5. Informational intent (blue line) is consistently the highest, starting at roughly 1.35. Transactional intent (dark blue line) is the lowest, trending downward from 0.8 to 0.5. Commercial (green) and Navigational (pink) hover in the middle, with Commercial seeing a noticeable dip at position 3.

Because this methodology is binary (the content either does or doesn’t answer the question), we’re potentially writing off a lot of half-answered questions, which are valuable if there is no better answer.

Question score by position

Rather than only using a simple yes/no approach, I also assigned points based on whether a question is wholly or partially answered to see how this might play into search rankings:

  • Answered "yes" = 1pt
  • "partial" = 0.5pts
  • "no" = 0pts

This gives less credit to partially answered questions, but may help us see if we’re missing any nuance.

Questions score by intent top 20A line chart showing "Questions score by intent" extending to ranking positions 1 through 20. Informational intent (blue) remains dominant, hovering around 4.0 throughout the chart. The other three intents (Commercial, Navigational, and Transactional) are grouped lower (roughly 2.5 to 3.5) and intertwine frequently, showing varied performance across the lower rankings.Questions score by intent top 10A line chart showing "Questions score by intent" for positions 1 through 10. Informational intent (blue) consistently maintains the highest score, staying above 4.0. Commercial (green) fluctuates between 3.0 and 4.0. Navigational (pink) and Transactional (dark blue) remain the lowest, mostly oscillating between 2.5 and 3.0.Questions score by intent top 5A line chart titled "Questions score by intent" for ranking positions 1 through 5. The Y-axis ranges from 0 to 5. Informational (blue) and Commercial (green) intents score higher, hovering near the 4.0 mark. Navigational (pink) and Transactional (dark blue) score lower, averaging around 3.0. The trends are relatively flat across the top 5 positions.

How do questions answered/question score correlate with ranking position?

Here’s how questions answered correlates with ranking position for the top 20, 10, and 5 search positions.

(Note: As the positions descend but the averages ascend, it means a negative correlation is actually a good thing.)

How questions answered correlated with ranking position

Intent

Top 20

Top 10

Top 5

Intent

Commercial

Top 20

-0.4851

Top 10

-0.5683

Top 5

-0.8941

Intent

Informational

Top 20

-0.1246

Top 10

-0.6226

Top 5

-0.8867

Intent

Navigational

Top 20

0.2472

Top 10

-0.8213

Top 5

-0.7414

Intent

Transactional

Top 20

0.4532

Top 10

-0.5999

Top 5

-0.9573

And, here’s how they correlated when I used the question score method.

How question score correlated with ranking position

Intent

Top 20

Top 10

Top 5

Intent

Commercial

Top 20

-0.2393

Top 10

-0.3806

Top 5

-0.9097

Intent

Informational

Top 20

-0.3027

Top 10

-0.4463

Top 5

-0.4364

Intent

Navigational

Top 20

0.4979

Top 10

-0.5513

Top 5

0.1745

Intent

Transactional

Top 20

0.4482

Top 10

-0.1241

Top 5

-0.5460

One striking pattern is that the correlation scores climb when we look at the top 20 positions. It’s also worth noting that we have far fewer results recorded from the bottom of page 2. The change in average could be impacted by a lower number of results skewing overall averages.

A simple line chart titled "URL per position" plotting a count (Y-axis 0-400) against ranking positions 1 through 20. The line remains stable and high, fluctuating slightly between 300 and 380 for positions 1 through 17. At position 17, the line begins a steep, dramatic crash, dropping to below 100 by position 20.

Key findings

Across the top 5 organic search positions, there is a strong, positive correlation between the number of fully answered questions and ranking.

When I introduced question scoring (which accounts not only for fully answered questions, but also for partial or in-depth snippets) it added useful nuance to the analysis. However, its correlation with rank is weaker than the raw ‘questions answered’ metric. This suggests that, while partial coverage and depth matter, outright completeness appears to be more important.

For navigational queries (i.e., where users seek a specific site or page), the relationships weaken further. There is still a clear trend, where higher questions-answered counts and stronger question scores tend to align with better rankings, but neither metric has as much influence here as it does for other intent query types.

Surprisingly, informational queries don’t display the strongest correlations (despite my expectation that thoroughly informative content would naturally answer more questions). Several factors could be at play, but our sample size makes me reluctant to make any solid conclusions. It is also possible that there are more questions at play, or perhaps that answering fewer, more specific questions is more significant than a breadth of answers.

Finally, beyond the first page—and especially in the lower slots of page 1 and throughout page 2—the link between either metric and ranking diminishes sharply. By the bottom half of page 2, the trend even reverses: pages with fewer answers or lower question scores sometimes rank better.

Part of this anomaly stems from a statistical element, because positions 18–20 appear less frequently in the dataset, each additional answered question (full or partial) disproportionately inflates the average, creating a misleading spike in the ‘questions answered’ metric without reflecting a real ranking advantage.

Answering questions is crucial, but you still need a robust user experience and external quality signals

It seems as though questions answered correlate with ranking position—specifically with regard to the top 5 positions. This may not necessarily be because a number of questions is answered on the page, but because you are covering a lot of the subject with the content that could lead to a positive experience for the user. Clearly covering questions around an area makes that page semantically richer, too.

More time spent on crawling, scraping, and analysis would likely yield more complete results, but the directionality of the results would be the same. EVERY step of this process (keyword selection, niche selection, intent designation, and analysis) would likely have more of an impact on results, so if we wanted to be more concrete, we would need to significantly increase the number of keywords, ensure the pipeline is more robust, and create a more nuanced view of the intent stages.

Put simply, if you take ALL the PAAs for a niche, answer all those questions, you STILL need to have stronger technical standards, well written copy, external quality signals, and evidence of user satisfaction to rank well. So in isolation answering PAA questions is not enough, but it could be a significant factor when all are compared.

Methodology: Room for improvement

Below are some obstacles that I was not able to completely navigate during the first iteration of this study:

  • Readability errors — Some pages may not extract correctly (due to paywalls, heavy JS, or unusual layouts).
  • Low-content filtering — Skips any page with <100 words, which might discard pages that answer succinctly.
  • Bot/scraping protection — Some websites actively block or serve different content to bots or headless browsers.
  • There’s no built-in manual review to check if ChatGPT 3.5-Turbo’s answers are accurate.
  • If a top Google result can’t be scraped or the content is empty, that result was ignored.
  • If the questions can’t be loaded (i.e., because there are none), that keyword was skipped for the gap analysis.
  • A lot of sites (Youtube, Spotify, etc.), do not have content to scrape in the same sense for this kind of analysis. So for this, they are judged to not answer a question, even if they may have.

Some more general caveats about the integrity of the data itself, keyword selection, and potential risks in using LLMs include:

  • Correlation is not causation — Answering questions might not cause high rankings, but could be a side effect of good content.
  • SERP features bias — Featured snippets or People Also Ask boxes themselves might affect which pages get selected and ranked.
  • Sample bias — If the keywords/questions aren’t representative, it may not be appropriate to generalize results.
  • Reliability of “answer status” — As mentioned before, LLM judgment is strong, but not perfect.