← Back to Hub

FPL Analysis Methodology

How the club league comparison is built — the metrics, the reasoning, and the limitations · 2025/26 Season

Contents

  1. Introduction
  2. Data Sources & Collection
  3. Metric Definitions
  4. Strategy Archetype Classification
  5. Chip Analysis Methodology
  6. Transfer Discipline Scoring
  7. Squad Value as a Proxy Metric
  8. H1/H2 Split Rationale
  9. Season Trajectory Heatmap
  10. Limitations & Caveats

1. Introduction

The Club League Strategy Comparison report draws conclusions about six (and counting) managers — their strategies, their relative performance, and their prospects for the run-in. This document explains exactly how those conclusions are reached: where the data comes from, what each metric measures, how managers are classified into archetypes, and where the analysis has blind spots.

The intended audience is any manager who wants to go beyond the headline numbers and understand the reasoning. Nothing here requires a statistical background — the goal is transparency, not complexity.

Relationship to the comparison report: This document is the why; the comparison report is the what. Read this if a conclusion in the report surprised you or you want to evaluate whether a methodology decision was right for your situation.

2. Data Sources & Collection

What files are used

Every metric in the comparison report derives from one source: the FPL history API endpoint captured as JSON files per manager.

File patternAPI endpointContains
fpl_*_history.json /api/entry/{id}/history/ GW-by-GW points, OR, squad value, transfers, hits, bench points. One row per gameweek.
fpl_overall_top10.json /api/leagues-classic/314/standings/ Global overall league standings — the top 50 managers worldwide by total points.

Why the Network tab is the only safe method

The FPL website uses Datadome, a bot-detection system that intercepts requests made by injected scripts — browser extensions, MCP tools, DevTools console commands. The danger is not that Datadome blocks these requests. It is that Datadome silently serves wrong data.

Silent corruption, not a clean error: A Datadome-intercepted request returns HTTP 200 with valid JSON structure. The data looks correct. It is not. This was the root cause of the SlowBuild squad being completely wrong in the GW29 reports — the script ran without errors and produced a plausible-looking output from poisoned data. There is no flag in the response that indicates corruption. The only defence is to never use injected scripts.

The correct method is to open the FPL website in a normal browser session, open DevTools → Network tab, navigate to trigger the relevant API calls, and save the raw JSON response. The browser makes the request natively with full session context and real cookies — Datadome cannot distinguish this from a human user.

What the data does not include

The history endpoint is aggregate only. It tells you how many points a manager scored each GW, what their OR was, and how many transfers they made — but not which players they picked, who they captained, or what specific transfers occurred. Player-level detail requires the picks endpoint (`/api/entry/{id}/event/{gw}/picks/`), which is not used in this analysis.


3. Metric Definitions

Every column in the standings table and every bar in the charts maps to one of these fields from the history JSON. The field name shown is the exact key from the API response.

MetricDefinitionWhy it matters
overall_rank (OR) The manager's position in the global standings after each gameweek, ranked by total_points descending. Normalises performance against the full player pool each week. A 70-pt GW during a high-scoring week might drop your OR; the same score during a blank week might raise it. OR captures context that raw points miss.
total_points Cumulative sum of points scored across all played GWs. Hit deductions are already applied by the API — i.e., a 4-pt hit GW shows the net score. The primary competitive metric. All other metrics exist to explain why total_points ended up where it did.
points (GW score) Points scored in a single gameweek, net of any transfer hit deductions applied that GW. Used to identify high/low-water marks and to correlate chip plays with scoring spikes.
value Sum of the current sell prices of all 15 squad players at the end of each GW, in tenths of £m (i.e., 1065 = £106.5m). Proxy for asset quality. Players rise in price as ownership increases; holding winners and selling losers early compounds the value advantage.
Squad Value Growth value at GW31 minus the starting budget of 1000 (£100m), expressed in £m. E.g., value 1086 → growth of +8.6m. Isolates the accumulated asset management advantage. Managers who started with the same £100m ended GW31 with squad values ranging from £104.9m to £108.6m — a 3.7m structural gap.
points_on_bench Total points scored by the four bench players in a given GW, regardless of whether any were used as automatic substitutes. Measures the cost of incorrect lineup ordering. High bench waste does not always indicate poor play — it can mean your bench was unexpectedly strong. Context is required.
Bench Waste (season) Sum of points_on_bench across all 31 GWs. Cumulative measure of lineup mis-ordering. Lower is better, but must be read alongside OR — Ibsen has the highest bench waste (380pts) yet holds OR #1, because his starting XI dominated.
event_transfers_cost Point deduction applied that GW for additional transfers beyond the free transfer allowance. Each extra transfer costs 4 points. The direct cost of reactive transfers. Shown both per-GW and as a season total.
Hit Cost (season) Sum of event_transfers_cost across all 31 GWs. Headline measure of transfer discipline. Zero is achievable — four managers in this club achieved it.
event_transfers Total number of transfers made in a given GW, including free transfers. Combined with hit cost, reveals whether a manager's activity was disciplined (high transfers, zero hits = rolling free transfers) or reactive (high transfers, high hits = impulsive churning).
H1 / H2 Points Total points scored in GW1–19 (H1) and GW20–31 (H2) respectively. See Section 8 for why GW19 is the dividing line. Reveals whether performance is structurally consistent across both halves or concentrated in one phase, often due to chip deployment timing.

4. Strategy Archetype Classification

Each manager in the comparison report is assigned an archetype — a short label that summarises their dominant strategic behaviour across the season. Archetypes are not prescriptive; they are descriptive. They are assigned after the fact by looking at four independent signals and finding the dominant pattern.

The four classification signals

Signal 1 — Transfer discipline (hit threshold)

Season hit costClassificationInterpretation
0 ptsElite disciplineNever took a hit. Every transfer was planned within free transfer budget.
1–12 ptsNear-eliteAt most 3 hits across 31 GWs — occasional tactical exception, not a pattern.
13–24 ptsAverage3–6 hits across the season — reactive transfers were a recurring tendency.
25+ ptsReactive7+ hits — transfer activity driven by short-term performance anxiety rather than long-term planning.

Signal 2 — Transfer rate (churn index)

Avg transfers/GWClassificationInterpretation
< 1.0SurgicalRegularly rolls free transfers — deploys them in batches when genuinely needed.
1.0 – 1.1NormalUses the free transfer most weeks. Balanced activity.
1.11 – 1.19ActiveRegularly uses accumulated transfers plus occasional hits.
> 1.19ChurnConsistent overactivity. More transfers than free transfers available, not justified by OR outcomes.

Signal 3 — Chip timing

PatternClassification
All H1 chips played by GW14H1 Blitz — commits early, builds from a clear position
Chips spread across H1 and H2Structured — staged deployment aligned with the season calendar
Most or all chips concentrated in H2Contrarian / Late — high-risk high-reward approach; depends on timing being right
Chips used reactively without a clear planMixed / Reactive — chips deployed in response to events rather than in anticipation of them

Signal 4 — OR trajectory

PatternArchetype indicator
Top 100 from early GWs, sustained all seasonEarly Detonator or Frontrunner
Gradual, consistent climb through all 31 GWsPatient Accumulator
Flat/mid-table H1, then steep H2 climbLate Bloomer / Contrarian
Volatile — large OR swings week to weekReactive / Hit-Taker
Archetypes are retrospective: No manager sets out to be "The Reactive Churner." The archetype describes the emergent pattern of their decisions over 31 GWs — it is a diagnosis, not a judgement. The same strategic DNA that labels a manager a "Contrarian Late Bloomer" would look like "catastrophic planning" if the H2 chips had been mistimed by two GWs.

5. Chip Analysis Methodology

The 4-layer decision model

All analysis in this project is anchored to a four-layer decision framework, where each layer must be resolved before the next begins:

1
Chip deployment — highest leverage decision When to play Free Hit, Bench Boost, Wildcard, Triple Captain. This decision constrains all downstream choices. A wrong chip play cannot be undone and invalidates layers 2–4.
2
Player triage Score each squad member using: Form 30% · FDR 25% · DGW exposure 20% · Pts/£m 15% · Injury risk 10%. Verdicts: HOLD (≥6.0), MONITOR (4.5–5.9), SELL (<4.5).
3
Replacement selection For each SELL, find the best replacement by price, fixture run, DGW status, and top-10 ownership signals.
4
Sequencing Order trades across GWs to preserve free transfers, avoid unnecessary hits, and align with chip activation timing.

Why GW6 Triple Captain was identified as optimal

The report identifies the GW6 TC as the single most impactful shared decision of the season. This conclusion is arrived at retrospectively, not predictively: 5 of 6 club managers played TC on GW6, and all five saw significant OR improvements that week. This convergence — managers making the same call independently — is strong evidence that GW6 was a genuinely high-ceiling captain week, likely featuring a premium asset (Salah, Haaland, or similar) in a highly favourable fixture.

The estimated impact of 40–60 pts above an average manager assumes the TC captain returned approximately 3× their normal expected score (a realistic ceiling for a premium asset in a great fixture). At 10M managers, even modest captain differentiation compounds into large OR movements.

Why early Bench Boost outperforms late Bench Boost

The optimal BB window is when the squad is deepest and most in-form. Early in the season (GW1–10), managers typically have 15 genuinely viable players before injuries and form divergence thin the squad. Playing BB on GW1 (Trundlers, 115 pts) or GW9–10 (Mirinda, Ibsen) captures the full 15-player depth. Playing BB in GW18+ risks having 3–4 bench players who are injured, blanking, or in poor form — shrinking the effective return.

Remaining chips as a run-in multiplier

Chips are not just about the immediate GW they're played — they are structural advantages for the rest of the season. As of GW31, the chips-remaining picture is:

ManagerChips RemainingRun-in ceiling
Ibsen, Trundlers, MirindaWC2 + TC2Highest — can restructure squad entirely AND play an elite TC on any GW
Crème de la CrèmeFHStrong — FH is most valuable during a blank GW (fields a full 15-player squad unaffected by blanks)
BrunoWC2Moderate — can reset the squad but no TC multiplier remaining
HowittsTC2Limited — one elite captain play but no squad restructure available
DGW33 is the key event: Arsenal, Man City, Aston Villa, and Crystal Palace all play twice in GW33. Managers with WC2 can load up on DGW33 assets beforehand. Managers with TC2 can triple their best DGW33 asset. The combination of both is the highest possible run-in advantage.

6. Transfer Discipline Scoring

Why hits are compound losses, not simple deductions

The common framing of a transfer hit is: "I paid 4 points to get Player B instead of Player A." This understates the true cost. A transfer hit has three components:

1
Immediate cost: −4 pts Applied directly to that GW's score. If Player B scores 8 pts and Player A would have scored 6 pts, the net gain from the hit is 8 − 6 − 4 = −2 pts. The hit player needs to outscore the outgoing player by more than 4 pts just to break even.
2
Opportunity cost: one lost free transfer An unused free transfer rolls to the next GW (max 2 banked). By taking a hit, you burn that future free transfer. The genuine cost is therefore at least 4 pts (immediate) plus the value of whatever that rolled free transfer would have bought — often another 4+ pts of improvement in a later GW.
3
Regression risk: buying high, selling low The most common trigger for a reactive hit is a player haul. The manager buys the haul player immediately after a 20-pt return. Statistically, elite players do not haul in back-to-back weeks at the same rate — the next GW is likely to be a 4–6 pt return. The hit player underperforms expectations while the sold player, freed from the pressure of being in the squad, often returns in the same GW.
True hit cost ≈ 4 (immediate) + 4–8 (opportunity cost of lost FT) + expected regression deficit

Season total framing

Howitts paid 28 pts in hits over the season. The average GW score across the club is approximately 65 pts. That means Howitts surrendered the equivalent of nearly half an average gameweek to transfer penalties — before accounting for opportunity and regression costs. In a season where the gap between OR #1 (Ibsen, 2061 pts) and OR #33,068 (Howitts, 1880 pts) is 181 pts, the hit tax is not the only explanation, but it is a measurable and entirely avoidable contributor.


7. Squad Value as a Proxy Metric

What squad value growth signals

Every manager starts the season with £100m. The ending squad value is a function of which player prices rose (held correctly) and which fell (sold before the drop or avoided entirely). A manager who ended GW31 with £108.6m (Ibsen) has effectively "earned" 8.6m in player appreciation — meaning his squad generates more points-per-pound than a squad worth £104.9m (Howitts), because each of his 15 positions is filled with a more expensive (and generally more in-form) asset.

Squad value advantage ≈ (value_A − value_B) / 10 × avg_pts_per_£m_per_GW × remaining_GWs

The 3.7m gap between Ibsen and Howitts at GW31 equates to roughly 2–3 extra points per GW in expected scoring, compounded across 7 remaining gameweeks. This is a structural advantage that cannot be erased by a single good captain pick.

Limitations of squad value as a metric


8. H1/H2 Split Rationale

Why GW19 is the dividing line

GW19 is not an arbitrary midpoint. It is the last gameweek in which Wildcard 1 (WC1) can be played — the FPL rules restrict WC1 to the first half of the season (GW1–19) and WC2 to the second half (GW20+). This makes GW19 a genuine structural boundary in the season: chip availability, squad composition philosophy, and the fixture calendar all shift at this point.

What each half reveals

The Crème paradox: Crème de la Crème had the lowest H1 OR of the four global top-4 managers (peaked around OR 5,000 at end of H1) but fired every chip in a compressed H2 window and reached OR #3 globally. This shows that a weak H1 can be recovered with perfectly timed H2 chips — but also that the same strategy would have failed entirely if the chip GWs had been misread.

9. Season Trajectory Heatmap

How OR cells are colour-coded

The GW-by-GW heatmap in the comparison report colours each cell based on the manager's overall rank that week. The bands are not evenly distributed — they are weighted towards the elite end because that is where meaningful differentiation occurs between the managers in this club.

OR rangeColourWhat it means
1 – 100 Bright green Elite — top 0.001% of all managers globally. Only Trundlers (GW6–7), Mirinda (GW9), and Ibsen (GW29+) have spent time here.
101 – 1,000 Light green World-class — top 0.01%. Consistent performance at this level signals elite squad construction.
1,001 – 10,000 Teal Excellent — top 0.1%. Most elite players cycle through this band during their best weeks.
10,001 – 100,000 Amber Good — top 1%. A single poor GW or unlucky blank typically pushes a top-10k manager into this band.
100,001 – 1,000,000 Light red Average — top 10%. Where most managers live week-to-week during ordinary GWs.
1,000,001+ Dark red Below average week — bottom half of the player pool. Common in GW1–2 before the field stratifies.

Why OR is used instead of raw points

Raw GW scores are not comparable across gameweeks. A 70-point GW during a blank round (fewer players scoring) might rank higher globally than a 75-point GW during a double gameweek (everyone scoring more). Overall rank normalises the score against the actual field each week, making comparisons meaningful across the season.


10. Limitations & Caveats

What this analysis cannot show

Known data quirks

Wildcard double-counting bug: The FPL API logs both WC1 and WC2 as "wildcard" in chip history, distinguished only by the event number. Any tool that counts "wildcard" entries without checking the GW will report zero wildcards remaining even when WC2 was never played. In this analysis, chip status is always cross-checked against the event number (events 1–19 = WC1, events 20+ = WC2).
The right way to use this analysis: Treat the metrics as lenses, not verdicts. The comparison report identifies patterns and correlations — managers who took zero hits also ranked highest globally. Correlation is not causation. Zero hits is a marker of discipline, but discipline alone does not guarantee OR #1. The report is most useful for identifying structural differences (squad value gaps, hit totals) that persist across 31 GWs, rather than for attributing outcomes to any single decision.

View the Club League Comparison Report →