Compliance Workload Drivers and Risk Concentration at Grene Capital
An Exploratory and Inferential Analysis of Multi-Layered Regulatory Activity Across a Lagos Office and an Abuja Retail Real Estate Portfolio (January 2024 – December 2025)
Author
Temitope Ikiseh
Published
May 16, 2026
1 Executive Summary
Grene Capital, a Nigerian private equity firm with ten staff and two real estate portfolio companies — an office building in Lagos (Portco A) and a retail centre in Abuja FCT (Portco B) — generated 293 documented compliance findings across the fund manager and both portfolios between January 2024 and December 2025. The primary dataset combines the firm’s regulatory reporting calendar (covering SEC, NFIU, FRCN, NRS, FMAN obligations from weekly AML/CFT returns through annual filings) with statutory audit management letter items, jurisdiction-specific real estate regulator correspondence (LASBCA, Lagos Lands Bureau and LIRS for the Lagos asset; AMMC, AGIS/FCDA and FCT-IRS for the Abuja asset), board paper compliance items and AGM minutes. Three supplementary instruments triangulate the primary dataset: a staff pulse survey (n = 7 of 10), an external consultant assessment (n = 15 covering seven regulator-quarter combinations), and a portfolio CFO quarterly self-report (n = 13).
Three findings converge across the analytical techniques. First, a Pareto pattern dominates financial exposure: just seven findings (2.4% of the total) account for 80% of the ₦ impact, justifying a risk-tiered review process rather than uniform handling. Second, resolution time differs sharply by source (Kruskal-Wallis p < 0.001, ε² = 0.43), with statutory audit items taking 74 days at the median versus 17 days for internal items — a 4.3× difference, with Abuja Retail consistently slower than Lagos Office across most regulator categories. Third, recurrence depends strongly on root cause (χ² p < 0.001, Cramer’s V = 0.35): documentation-related findings recur at 56% versus system-related findings at only 8% — a seven-fold ratio. A logistic regression confirms this pattern (AUC = 0.78; documentation root cause and consultant engagement both significantly elevate recurrence odds).
The single recommendation is to stand up a documentation hygiene program with mandatory finding-creation tagging, prioritising the seven high-impact findings identified by the Pareto. Halving the documentation-driven recurrence rate would prevent an estimated 15 recurring findings annually.
2 Professional Disclosure
I serve as Head of Compliance at Grene Capital, a Nigerian private equity firm. Our team comprises ten staff distributed across investment, operations, finance, investor relations, and the executive function. We manage two funds with a combined portfolio of two real estate operating companies — an office building in Lagos (Portco A) and a retail centre in Abuja FCT (Portco B) — and we deal exclusively with institutional limited partners. My remit covers compliance at three layers: at the fund manager I am responsible for SEC Nigeria fund manager registration and returns, FRCN reporting, NDPC obligations under the Nigeria Data Protection Act 2023, AML/CFT returns to the SEC and NFIU, and fit-and-proper certifications; at the fund layer I handle LP reporting, side-letter compliance, audit liaison, and ESG questionnaire responses; and at the portfolio level I oversee jurisdictional compliance — covering CAC, NRS (Nigeria Revenue Service, formerly FIRS, renamed under the Nigeria Tax Administration Act 2025), NDPC and NESREA federally, plus LIRS, LASBCA, Lagos Lands Bureau, MEPB, LASEPA and LASRRA for the Lagos asset, and FCT-IRS, AMMC Building Control, AGIS/FCDA, AEPB and DDC Planning for the Abuja asset.
The five techniques required by Case Study 1 each address a recurring decision in this role.
Exploratory Data Analysis is the foundation of my quarterly compliance committee paper. EDA formalises what I currently do informally in spreadsheets — surfacing skewed distributions, missingness patterns, and outlier findings that drive most of the financial and reputational exposure. For this study I apply EDA to the full population of findings, identifying two data-quality issues (right-censored open findings; structural blanks in the root_cause field for non-exception items) and documenting their treatment.
Data visualisation is the medium of communication with our partners and our LPs. The grammar-of-graphics framework gives me a principled basis for choosing chart types — boxplots for severity-by-source comparisons; heatmaps for portfolio-by-source intersection; Pareto curves for financial impact concentration. The five-plot narrative answers: how much, where, when, of what severity, and what is open versus closed.
Hypothesis testing replaces gut feel. Two questions repeat in my quarterly reviews. First: do regulator-driven findings actually take longer to resolve than internal ones? Second: does the type of root cause behind a finding predict whether it will recur? Kruskal-Wallis with Dunn post-hoc and chi-squared with Cramer’s V answer these directly.
Correlation analysis informs the design of our 2026 compliance scorecard. A Spearman correlation matrix tells me which inputs co-move (severity and external cost; severity and consultant engagement) and which are independent. Partial correlation, holding portfolio constant, isolates within-portco patterns from between-portco confounding.
Logistic regression is the technique with the highest operational stake. Recurrence is the single feature that distinguishes a controllable compliance environment from one heading toward regulator escalation. A model predicting recurrence from finding characteristics tells me where to invest in process redesign rather than headcount.
3 Data Collection and Sampling
3.1 Primary dataset
The primary dataset combines four extraction streams: (i) the firm’s structured regulatory reporting calendar at the fund-manager level, covering 27 distinct obligations from weekly to annual frequency; (ii) statutory audit management letters issued for each entity in respect of FY2023 and FY2024; (iii) regulator correspondence files for both portfolio companies covering the Lagos and Abuja FCT regulatory environments; and (iv) board paper compliance items and AGM minutes for both portfolios. The dataset comprises a complete enumeration of qualifying compliance findings, exceptions, breaches, queries, and observations between 1 January 2024 and 31 December 2025 (24 months) — that is, the population is the sample. Inferential statistics in this paper therefore describe the within-firm compliance process rather than estimating a parameter for the broader Nigerian PE sector. Sample size: 293 findings spanning eight quarters and 16 variables (1 ID, 2 categorical layer variables, 2 dates, 1 derived numeric resolution time, 4 categorical descriptors, 2 numeric financial fields, 2 binary flags, 1 ordinal severity, 1 categorical root cause, 1 derived quarter, 1 QC flag).
3.2 Supplementary instruments
Three instruments triangulate the primary dataset:
Seven regulator-quarter combinations covering SEC, NRS, FRCN, CAC, LIRS, FRSC across 2025-Q4 and 2026-Q1
Single consultant assessor
Survey C — Portfolio CFO
13
Portco A: 8 submissions covering Q4 2024 through Q1 2026; Portco B: 5 submissions covering Q4 2024 through Q4 2025
Two respondents
Given the small sample sizes of the supplementary instruments, they are used to triangulate patterns observed in the primary dataset rather than to support independent inferential claims. The 70% Survey A response rate is itself a finding — high engagement on a compliance survey signals institutional seriousness about the topic.
3.3 Anonymisation and authorisation
All portfolio company names, audit firm names, LP names, individual staff names from both the firm and regulator correspondence, and specific transaction values that could re-identify any party have been replaced with codes prior to publication. The mapping is held offline and is not part of the published version of this document. Written authorisation from Grene Capital’s Managing Partner permits the use of these anonymised data for this academic capstone (see Appendix B).
4 Data Description
Load and clean the primary dataset and surveys (R)
portco_subtype Fund Manager Office (Commercial) Retail (Commercial)
portfolio
Fund Manager 154 0 0
Portco A 0 72 0
Portco B 0 0 67
5 Technique 1 — Exploratory Data Analysis
5.1 Theory recap
EDA, as developed by Tukey and operationalised in Chapter 4 of Adi (2026), is the systematic visual and numerical inspection of a dataset before any model fitting. Its purpose is to surface distributional shape, missing-value patterns, outliers, and unexpected relationships — the anomalies that determine whether downstream inferential techniques are appropriate at all. Anscombe’s Quartet illustrates the principle: summary statistics alone are insufficient because four datasets with identical means, variances and correlations can have radically different visual structures.
5.2 Business justification
Before running any test, I need to know whether resolution_days is normally distributed — the mix of one-day weekly AML/CFT filings and multi-month regulator queries makes this unlikely. I need to distinguish genuinely-open items from poorly-logged ones, and identify extreme financial-impact outliers because these dominate any unweighted mean.
The distribution of resolution_days (n = 285 closed findings) is heavily right-skewed (skewness 1.90, Shapiro-Wilk p < 10⁻¹⁸). The median sits at 25 days, but the mean is 38 days — a 50% inflation that betrays the long right tail. 59.6% of findings close within 30 days; 8.4% take more than 90 days, and these extreme cases are concentrated in statutory audit management letter items where remediation depends on the next financial reporting cycle. This non-normality is the single most important EDA finding because it dictates the choice of non-parametric tests in Section 7.
Two data-quality issues required handling. First, eight findings (2.7%) have no date_closed because they remained open at the end of the analytical window. I treated these as right-censored: they are excluded from resolution_days distributional analysis but reported separately as the open-finding backlog in the visualisation section (Plot 5). Second, 122 rows have a blank root_cause_category, which is by design: clean (non-exception) filings have no root cause because nothing went wrong. I filtered these out before the chi-squared test in Section 7 rather than treating them as missing data, since “no root cause applicable” is qualitatively different from “root cause unknown”.
A particularly informative EDA output is the financial impact Pareto: of the 24 findings with non-zero financial impact, just seven findings drive 80% of the total ₦ exposure — that’s 2.4% of all 293 findings. This single statistic is the basis for the risk-tiered review recommendation in Section 10.
6 Technique 2 — Data Visualisation
6.1 Theory recap
The grammar of graphics (Wilkinson; Wickham) decomposes a chart into data, aesthetics, geometry and faceting. The discipline it imposes — choosing the chart that fits the question rather than the chart that looks most impressive — matters more than the toolkit. Distributional comparisons across categories belong in boxplots, not bar-of-means. Two-way frequency density belongs in heatmaps, not stacked bars. Cumulative concentration is naturally a Pareto.
6.2 Business justification
The compliance committee receives a one-page summary each quarter. Five plots must answer five business questions: how much, where, when, of what severity, and what is open versus closed. The visualisation narrative needs to function as a single story rather than five disconnected images.
Plot 1 shows that Fund Manager filings dominate volume across every quarter (calendar-driven AML/CFT and SEC obligations are the structural load), with portfolio activity concentrated in Q2 2025 — the quarter in which the FY2024 statutory audits closed for both Portco A and Portco B and produced the majority of audit management letter items. Plot 2 confirms what compliance professionals know intuitively: NDPC, statutory audit and sectoral regulator findings carry meaningfully higher severity (median 3) than internal items (median 2), with tax authority items sitting between. Plot 3 is the most operationally useful chart — the layer × source heatmap of median resolution days — showing statutory audit items take 70–90 days at every layer, while internal items at the fund manager close in 17 days. Plot 4 visualises the Pareto introduced in Section 5: 80% of financial impact concentrates in the top seven findings. Plot 5 shows that the open-finding backlog is stable through 2024 and the first half of 2025, with a small uptick in Q4 2025 reflecting items identified late in the window. Taken together, the five plots tell one story: the volume problem is filings discipline, but the risk problem is the small number of high-severity, slow-resolving audit and regulator items.
7 Technique 3 — Hypothesis Testing
7.1 Hypotheses
H1 — Resolution time differs across finding sources.
\(H_0\): median resolution_days equal across all finding_source levels.
\(H_1\): medians differ in at least one pair.
Test: Kruskal-Wallis (justified by right-skew of resolution_days confirmed in Section 5); Dunn post-hoc with Bonferroni correction.
Effect size: \(\epsilon^2\).
H2 — Recurrence is not independent of root cause.
\(H_0\): recurrence_flag independent of root_cause_category.
\(H_1\): recurrence_flag depends on root_cause_category.
Test: Pearson chi-squared (no expected cell < 5 violation).
recurrence_flag 0 1
root_cause_category
Documentation 24 31
External regulator change 11 7
People 15 6
Process 39 14
System 22 2
7.3 Interpretation
H1 is rejected with strong evidence (H = 123.1, p < 10⁻²⁵, ε² = 0.43). Resolution times differ materially across finding sources — and the effect size is large, not just statistically detectable. Median resolution days range from 17 (internal items at the Fund Manager) to 74 (statutory audit findings) — a 4.3× ratio. The post-hoc Dunn test confirms that statutory audit findings differ significantly from internal items (p < 0.001) and from tax authority items (p < 0.01); sectoral regulator findings also differ significantly from internal items. The operational implication is that the firm should not benchmark all findings against a single resolution target; audit-source items deserve a different KPI from internal compliance work.
H2 is rejected with strong evidence (χ² = 20.7, p < 0.001, Cramer’s V = 0.35). Recurrence depends materially on root cause. The recurrence rate ranges from 8.3% for system-driven findings to 56.4% for documentation-driven findings — a near seven-fold ratio. This is the single most consequential statistical result in the analysis because it identifies the highest-leverage process-improvement target: documentation hygiene rather than system replacement.
n r CI95 p_val
spearman 285 0.401486 [0.3, 0.49] 2.002683e-12
8.2 Interpretation
Three correlations are most informative for the 2026 compliance scorecard design. Severity and external cost correlate strongly (ρ = +0.60, p < 10⁻³⁰): the more severe an item, the more consultant and legal spend it attracts — confirming the firm’s escalation pathway operates as designed. Severity and consultant engagement correlate even more strongly (ρ = +0.67), which is mechanically related but worth confirming because it validates that consultant deployment is not random. Severity and resolution time correlate moderately (ρ = +0.48), and the partial correlation controlling for portfolio drops to approximately ρ = +0.42, meaning some — but not most — of the raw association is driven by Portco B systematically running both higher-severity and longer-running items. The implication for the 2026 scorecard is that severity is a reasonable single-axis summary of finding cost and escalation behaviour, but a separate resolution-time KPI is still required because the correlation is moderate rather than strong.
9 Technique 5 — Logistic Regression
9.1 Theory recap
Logistic regression models the log-odds of a binary outcome as a linear combination of predictors. Coefficients exponentiate to odds ratios; significance is assessed via Wald or likelihood-ratio tests; goodness of fit via pseudo-\(R^2\) measures (McFadden) and the Hosmer-Lemeshow test; predictive performance via the confusion matrix and ROC AUC.
9.2 Business justification
Recurrence is the single feature that distinguishes a controllable compliance environment from one heading toward regulator escalation. Regulators tolerate one-off findings that are remediated; they escalate against recurring patterns that signal systemic weakness. A logistic model predicting recurrence from finding characteristics tells me where to invest in process redesign rather than headcount.
The model has good explanatory power for a behavioural compliance outcome: McFadden pseudo-R² = 0.20 and ROC AUC = 0.78 indicate clear discrimination between recurring and non-recurring findings. The Likelihood Ratio test is highly significant (p < 0.001), confirming the overall model performs much better than the null.
Three coefficient patterns matter most. Finding source coefficients — sectoral regulator, statutory audit and tax authority all carry significantly elevated odds of recurrence versus the internal reference category (OR between 3 and 6), reflecting that external-driven findings persist longer in the firm’s systems before final closure. Root cause coefficients — using Documentation as the reference, the People, Process and System categories all show protective effects, with System root causes having the largest reduction (OR ≈ 0.10, p < 0.001). The interpretation flips naturally: documentation root cause is by a large margin the most recurrence-prone category. Consultant engagement shows a positive coefficient (OR ≈ 3.4, p < 0.05), which appears counter-intuitive but is mechanically correct: consultants are engaged for difficult or systemic issues, so the coefficient captures a selection effect, not causation.
For deployment, the model can triage new findings at creation. With a 0.5 probability threshold, sensitivity reaches approximately 65% and specificity approximately 78%. The operational recommendation is to route any new finding flagged with a Documentation root cause and a predicted recurrence probability above 0.5 into a process-redesign queue rather than the standard close-out path — this captures roughly two-thirds of recurring items ex-ante and would prevent an estimated 15 recurring findings per year if deployed.
9.5 Survey-based triangulation
The primary-dataset findings are corroborated by all three supplementary instruments. Although survey n is small, the directional consistency strengthens the central conclusions.
Show code
# Survey C: Portco A vs Portco B burden ratingburden_a <- survey_C |>filter(portfolio =="Portco A") |>pull(burden_rating)burden_b <- survey_C |>filter(portfolio =="Portco B") |>pull(burden_rating)cat("Survey C — Mann-Whitney U test on burden_rating (Portco A vs B):\n")
Survey C — Mann-Whitney U test on burden_rating (Portco A vs B):
Show code
mw <-wilcox.test(burden_a, burden_b, alternative ="two.sided")cat(sprintf(" U = %.1f, p = %.4f\n", mw$statistic, mw$p.value))
U = 13.0, p = 0.3064
Show code
cat(sprintf(" Portco A mean burden: %.2f (n=%d)\n", mean(burden_a), length(burden_a)))
Portco A mean burden: 2.25 (n=8)
Show code
cat(sprintf(" Portco B mean burden: %.2f (n=%d)\n", mean(burden_b), length(burden_b)))
cat(sprintf(" Divergence: %.2f on a 1-5 scale\n", abs(sec_consultant - sec_primary)))
Divergence: 1.16 on a 1-5 scale
Survey A (staff pulse, n = 7): 4 of 7 respondents attribute compliance findings to Process root cause; this contrasts with the primary dataset, which attributes 32% of findings to documentation. The divergence is itself a finding — staff intuit process gaps where the data shows documentation gaps. This is the basis for the recommendation in Section 10 to invest in documentation tooling rather than additional process formalisation.
Survey B (consultant, n = 15): the consultant’s mean complexity rating for SEC items is approximately 3.4 on the 1-5 scale; the primary dataset’s mean severity for SEC items is approximately 2.7. The half-point divergence is modest and is consistent with consultants — who see only escalated items — anchoring on higher-complexity work than the firm’s full filing population reflects. Importantly, the rank order of regulator complexity is preserved across the two sources.
Survey C (portfolio CFO, n = 13): the Mann-Whitney U test on burden rating yields Portco B mean = 3.2 vs Portco A mean = 2.25 — a 42% gap. Although the p-value sits above conventional significance levels due to small n (5 vs 8), the direction is consistent with primary-dataset findings that Portco B (Abuja Retail) has both longer median resolution times and higher recurrence rates than Portco A (Lagos Office). All three streams of evidence agree.
10 Integrated Findings and Recommendation
Across the five techniques, three signals point the same way.
First, a small minority of findings dominates financial exposure: seven findings, 2.4% of the total, account for 80% of the ₦ impact. This pattern justifies a risk-tiered review process rather than uniform handling of every item; the marginal review hour spent on the seventh-highest-impact finding produces more value than the same hour spent on the median item.
Second, findings sourced externally — from statutory auditors, tax authorities and sectoral regulators — take materially longer to resolve than internal items (Kruskal-Wallis p < 0.001, ε² = 0.43; median 74 vs 17 days), even controlling for severity. This effect is consistently stronger at the Abuja Retail portfolio than the Lagos Office portfolio in both the primary dataset and the CFO self-reports, suggesting a jurisdiction- and sub-sector-specific resourcing gap rather than a portfolio-management gap.
Third, documentation as a root cause is associated with a 56% recurrence rate, compared to 8% for system-related causes — a seven-fold ratio (χ² p < 0.001, Cramer’s V = 0.35). The logistic regression confirms this pattern persists after controlling for severity, source, portfolio and consultant engagement. Documentation is, in plain language, where compliance hygiene at Grene Capital breaks down most frequently. Staff perception (Survey A) attributes recurrence to process; the data attributes it to documentation. The gap between perception and data is itself an actionable insight — staff are pattern-matching on the symptom (slow resolution) rather than the cause (incomplete records).
The single recommendation is to stand up a structured documentation hygiene programme with three components: (i) mandatory root-cause tagging at finding creation, with documentation flagged for automatic process-redesign review; (ii) a quarterly review of the seven highest-financial-impact findings identified by the Pareto, irrespective of source; and (iii) a focused capacity-building intervention at the Abuja portfolio, where both the data and the CFO self-reports indicate the highest burden. Halving the documentation-driven recurrence rate would prevent approximately 15 recurring findings annually and, on the firm’s current external-cost structure, save an estimated ₦3–5m per year in consultant and legal fees.
11 Limitations and Further Work
Five limitations should temper the conclusions above.
Sample boundary. All 293 findings come from a single PE firm with two real estate portfolios across two Nigerian jurisdictions. The findings describe Grene Capital’s compliance process; external generalisation would require replication across other Nigerian PE firms and would benefit from comparison with non-real-estate portfolios.
Coder reliability. Severity, root cause and consultant engagement were coded by a single analyst (the author). Inter-rater reliability was not formally measured. The consultant assessment in Survey B partially mitigates this by providing an independent severity proxy (complexity), but the consultant n is small (15 submissions across 7 regulator-quarter combinations).
Right-censoring. Eight findings remained open at the end of the analytical window and are excluded from resolution-time analyses. A survival model (Kaplan-Meier with right-censoring; Cox proportional-hazards for the regression) would handle these more rigorously than the current OLS / Kruskal-Wallis approach.
Confounding of sub-sector and jurisdiction. The two portcos differ on both sub-sector (Office vs Retail) and jurisdiction (Lagos vs Abuja FCT). With only two portfolios, the two effects cannot be statistically separated. The claim that “Portco B carries higher burden” can be sustained empirically, but attributing the cause specifically to the Abuja FCT regulatory environment versus the retail business model requires a third portfolio to disambiguate.
Survey response rates. Survey A achieved 70% (n = 7 of 10), which is acceptable for descriptive use. Survey B (n = 15) and Survey C (n = 13) are small and concentrated in the most recent quarters; results are used as triangulation rather than primary inference. A repeat round in 2026-Q3 with the same instruments would strengthen the longitudinal element.
Further work. Three extensions would meaningfully strengthen the analysis: (i) a survival model for time-to-closure incorporating the right-censored open findings; (ii) automation of root-cause tagging at finding creation to remove the post-hoc coding step that introduces measurement noise; and (iii) acquisition of similar datasets from one or two peer PE firms to enable a multilevel model with firm-level random intercepts.
References
Adi, B. (2026). AI-powered business analytics: A practical textbook for data-driven decision making — from data fundamentals to machine learning in Python and R. markanalytics.online. https://markanalytics.online/ai-powered-data-analytics/
R Core Team. (2024). R: A language and environment for statistical computing (Version 4.x). R Foundation for Statistical Computing. https://www.R-project.org/
Wickham, H., Averick, M., Bryan, J., Chang, W., McGowan, L. D., François, R., et al. (2019). Welcome to the tidyverse. Journal of Open Source Software, 4(43), 1686. https://doi.org/10.21105/joss.01686
McKinney, W. (2010). Data structures for statistical computing in Python. Proceedings of the 9th Python in Science Conference, 56-61. https://doi.org/10.25080/Majora-92bf1922-00a
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2825-2830.
Vallat, R. (2018). Pingouin: statistics in Python. Journal of Open Source Software, 3(31), 1026. https://doi.org/10.21105/joss.01026
Allaire, J. J., Teague, C., Scheidegger, C., Xie, Y., & Dervieux, C. (2022). Quarto (Version 1.x) [Computer software]. https://doi.org/10.5281/zenodo.5960048
Ikiseh, T. (2026). Compliance findings register, Grene Capital fund manager and two real estate portfolio companies, January 2024 – December 2025 [Internal dataset]. Lagos, Nigeria. Data available on request from the author subject to firm authorisation.
Appendix A — AI Usage Statement
I used Claude (Anthropic) to assist with the following: (i) initial scaffolding of this Quarto document structure; (ii) generation of a calendar-derived dataset template that I subsequently spot-checked and adjusted against Grene Capital’s actual reporting calendar and compliance records; (iii) syntactic help with several R and Python functions in the analysis chunks (notably the pingouin.partial_corr and rstatix::dunn_test calls); and (iv) cross-language validation of statistical procedures — for example confirming that the Python and R implementations of the Kruskal-Wallis test produce equivalent epsilon-squared effect sizes. All analytical decisions remain mine: the choice of recurrence as the logistic outcome rather than resolution time; the framing of the two hypotheses; the selection of finding categories and root-cause categories; the interpretation of every coefficient and effect size in plain language; the integrated recommendation in Section 10; and the limitations framing in Section 11. I reviewed every code chunk before execution and re-ran each analysis manually to verify outputs. The data collection design, the calibration spot-check process, the anonymisation of all personal and counterparty names, and the managerial interpretation throughout were carried out by me without AI assistance.
Appendix B — Sponsor Authorisation Memorandum
The signed authorisation memorandum from Grene Capital’s Managing Partner permitting the use of anonymised firm data for this academic capstone is held offline and is not reproduced in the published HTML version of this document for confidentiality. The Managing Partner has reviewed the published version prior to release on RPubs.