Techniques for Evaluating the Completed Investigation - StudyPulse
Boost Your VCE Scores Today with StudyPulse
8000+ Questions AI Tutor Help
Home Subjects Extended Investigation Evaluate completed investigation

Techniques for Evaluating the Completed Investigation

Extended Investigation
StudyPulse

Techniques for Evaluating the Completed Investigation

Extended Investigation
01 May 2026

Techniques for Evaluating the Completed Investigation

Evaluating your own investigation is one of the most intellectually demanding parts of the Extended Investigation — it requires you to step back from your own work and assess it with the same rigour you would apply to any external source. A high-quality self-evaluation is honest, specific and analytically substantive.

Why Self-Evaluation Matters

Self-evaluation serves multiple functions:
1. Academic integrity: Honest acknowledgement of limitations is a mark of intellectual honesty
2. Contribution to knowledge: Telling readers what your findings cannot tell them is as important as what they can
3. Assessment: VCAA assessors specifically look for evidence of critical self-reflection in the evaluation section
4. Research community value: Future researchers need to understand the limitations of your work before building on it

KEY TAKEAWAY: Evaluation is not a modesty exercise. It is a substantive analytical task. Vague statements (“the sample was small”) score poorly. Specific, analytical statements (“the convenience sample of 40 students from one school limits the generalisability of findings to diverse school environments”) score well.

What to Evaluate

1. The Research Question

  • Was the question well-formed and appropriately scoped for the available time and resources?
  • Did the question remain the right one throughout, or did it need revision?
  • In hindsight, is there a better or more precise way the question could have been framed?

2. The Methods

  • Were the chosen methods appropriate for the research question?
  • Were they implemented as planned? If not, what changes occurred and why?
  • What are the key methodological limitations?
  • How do these limitations affect the reliability and validity of the data?

3. The Evidence

  • Is the evidence sufficient to answer the research question?
  • How reliable is the data (would the same results occur if repeated)?
  • How valid is the data (does it measure what you intended to measure)?
  • What sources of error or bias may have affected the data?

4. The Conclusions

  • Are the conclusions proportionate to the evidence — not overclaiming?
  • Do the conclusions directly address the research question?
  • What are the boundaries of generalisation — to whom and in what contexts do the findings apply?
  • What remains unanswered or uncertain?

5. The Research Process

  • Was the project managed effectively?
  • Were there unexpected difficulties? How were they handled?
  • What would you do differently?
  • What ethical issues arose and how were they managed?

EXAM TIP: Evaluation questions often ask: “Identify TWO limitations of the investigation and explain how each affects the conclusions.” This requires (1) naming the limitation, (2) explaining the specific mechanism by which it limits validity/reliability/generalisability, and (3) stating the consequence for the conclusions. Aim for at least three sentences per limitation.

The Strength of Self-Evaluation: Four Moves

Strong evaluation answers use these moves:
1. Identify: Name the limitation or strength specifically
2. Explain: Describe the mechanism — why is this a limitation?
3. Impact: How does this affect the reliability, validity or generalisability of findings?
4. Address: What was done to mitigate it, and/or what would stronger research look like?

Example of a weak evaluation statement:
“The sample size was small.”

Example of a strong evaluation statement:
“The survey sample comprised 35 students from a single metropolitan private school, limiting generalisability. This convenience sample likely overrepresents high-SES students with strong academic support structures, meaning findings about study habits may not apply to students in regional or lower-SES school contexts. A stratified random sample drawn from multiple school types would produce more generalisable findings.”

Evaluating What the Investigation Did Well

Evaluation is not only about weaknesses. Acknowledge genuine strengths:
- Methodological strengths that increase confidence in specific findings
- Areas where the investigation extends or adds to existing literature
- Parts of the process that worked effectively
- Aspects where you managed limitations successfully

APPLICATION: Complete an evaluation framework table before writing the evaluation section:

Aspect Strength Limitation Impact on Conclusions Suggested Improvement
Research question Specific and focused May be too narrow Findings only address X context Broader comparative design
Method Well-suited to question Convenience sample Limited generalisability Random sample
Data Rich qualitative data Small n Cannot quantify prevalence Supplement with survey

COMMON MISTAKE: Writing the evaluation section as a list of apologies rather than an analytical assessment. The tone should be confident and analytical — “this limitation means…” not “unfortunately my research wasn’t perfect because…”

Table of Contents