Version 0.1 — draft for review. This framework is itself subject to audit.
Purpose
The Authenticity Audit exists because the word authenticity has been so heavily commodified in tourism marketing that it no longer means anything operationally. Every provider claims to be authentic. Every platform claims to support local communities. Every destination claims to value its character.
The audit replaces self-declaration with structural assessment. It asks not what does this operation say about itself, but how is this operation built, where does the money go, who has a voice, and what happens when things go wrong.
It is designed to apply to four kinds of entity:
- Local providers — accommodation, experiences, food and drink, transport, services
- Platforms and intermediaries — booking sites, OTAs, package operators, destination apps, content sites with commercial intent
- Destination organisations — DMOs, tourism boards, municipal tourism strategies
- The author’s own commercial operation —
levifinland.comwill be audited against this framework before it accepts its first paid direct booking
If the framework cannot be applied honestly to the author’s own business, the framework is not honest. That is the test.
What this audit is, and is not
This audit is:
- A structural assessment based on observable, evidenceable facts
- Designed to expose mechanisms, not score marketing language
- Equally applicable to large platforms and small providers
- Revisable — criticism of the framework itself is welcomed and documented
This audit is not:
- A judgement of individuals or their motives
- A declaration that a “failed” operation is morally bad
- A substitute for legal compliance, professional accreditation, or regulatory oversight
- A guarantee — passing this audit does not certify quality, safety, or commercial viability
- Anti-tourism, anti-platform, or anti-growth in principle
The audit’s bias is explicit: pro-local-control, pro-direct-booking, pro-resident-voice, pro-cultural-sovereignty, pro-transparency, pro-long-term place stewardship. Operations that align with those values will score well. Operations whose structure works against those values will score poorly. The framework does not pretend neutrality on this.
Structure
The audit has two parts.
Part One: Hard Criteria. Seven binary, evidence-based questions. An operation must pass all seven to be eligible for any positive verdict. These are gatekeeping criteria — failing any one of them produces a Fail regardless of how the operation scores on the dimensions below.
Part Two: Scored Dimensions. Eight dimensions, each with three to five criteria, each criterion scored on a four-point scale (Strong / Adequate / Weak / Failing). Each dimension is then graded on the basis of its component criteria. The dimension grades, taken together, produce the final verdict.
This two-tier structure exists because some things are categorical (you either disclose your ownership or you don’t) and some things are matters of degree (how locally retained is your money flow). Collapsing both into a single numeric score invites gaming. Keeping them separate forces a defensible judgement.
Part One: Hard Criteria
An operation must pass all seven to be eligible for any verdict above Fail.
HC1. Ownership is not concealed
Test: A reasonable researcher can determine, within 30 minutes of public research, who ultimately owns and controls the operation.
Evidence: Public business registry filings, ownership disclosure on the operation’s own surfaces, named directors or shareholders.
Failure mode: Shell companies, opaque holding structures, or marketing surfaces that imply local ownership while the operation is owned externally.
HC2. No exclusivity clauses preventing direct bookings to providers
Applies to: Platforms, intermediaries, package operators, block-bookers.
Test: The platform does not require providers to route all bookings through the platform, does not penalise providers for accepting direct bookings, does not implement rate parity clauses or equivalent mechanisms that constrain a provider’s pricing on the provider’s own channels.
Evidence: The written agreement governing the platform-provider relationship.
Failure mode: Rate parity clauses, “best price guarantees” enforced against the provider, exclusivity periods, penalty mechanisms for direct competition.
HC3. Commercial model is publicly disclosed
Test: The operation publishes, in plain language, how it makes money — including commission percentages, fee structures, and any commercial relationships that shape what travellers see.
Evidence: A public page describing the commercial model, accessible without signup, written in plain language at the level a 16-year-old can understand.
Failure mode: Commercial model buried in terms and conditions, opaque to travellers, or presented in language designed to obscure rather than disclose.
HC4. Money flow is publicly disclosed
Test: The operation discloses, with reasonable specificity, where money lands first, what proportion is retained locally, what proportion leaves the local economy.
Evidence: A public statement of money flow, including the operation’s own non-commission outflows (payment processing, hosting, software, advertising).
Failure mode: Disclosing only the OTA-versus-direct framing while omitting the operation’s own non-local cost base. This is the criterion that prevents an operation from criticising others for what it does itself.
HC5. Cultural appropriation is not present
Test: The operation does not use Sámi imagery, names, dress, ritual, or symbols in marketing or product without explicit, documented permission from a Sámi-led body or individual with the authority to grant it.
Evidence: Where Sámi cultural elements are present in the operation, documented permission and attribution are publicly available.
Failure mode: Reindeer-and-shaman aesthetic marketing by non-Sámi operators, “Lappish” language that flattens distinct cultures, generic Arctic-mystic positioning that monetises cultural ambience without cultural relationship.
HC6. A written agreement governs commercial relationships
Test: Travellers and providers entering a commercial relationship with the operation do so under written terms that are accessible before they commit.
Evidence: A traveller-facing terms-of-service document. A provider-facing participation agreement. Both publicly available.
Failure mode: Verbal arrangements, “just trust us” partnerships, or agreements existing only after the relationship has begun.
HC7. A functioning right of reply exists
Test: Any individual or entity affected by the operation’s published claims, scores, or content can submit a correction and receive a substantive response within a stated timeframe.
Evidence: A public correction policy with a stated response time, a public log of corrections made.
Failure mode: No correction route, corrections made silently, or a correction policy that exists on paper but has never been used.
Part Two: Scored Dimensions
Each criterion is scored on the following scale:
- Strong. The operation demonstrably and consistently meets the criterion. Evidence is public and verifiable.
- Adequate. The criterion is met, but with caveats, gaps, or inconsistencies. The operation is moving in the right direction but not yet exemplary.
- Weak. The criterion is partially addressed but with material gaps. The operation does not yet meet the standard the dimension requires.
- Failing. The criterion is not met, ignored, or actively worked against.
A dimension is graded based on its component criteria, with the lowest-scoring criterion weighing more than the highest. A dimension with one Failing criterion cannot be graded Strong, regardless of its other scores.
Dimension 1: Ownership & Control
Where do decisions get made, and by whom?
1.1 Beneficial ownership location. Where do the people who ultimately profit from this operation live, pay tax, and raise their families? Higher score: locally resident, tax-domiciled in the municipality. Lower score: non-resident, foreign-domiciled, ultimate parent corporation outside Finland.
1.2 Operational decision-making location. Where are decisions about pricing, hiring, partnerships, and strategy actually made? Higher score: by people physically present in the destination. Lower score: by remote management, head office, or algorithmic system.
1.3 Time horizon and tenure. How long has this operation existed in this place, and how long does its model assume it will continue? Higher score: multi-decade local presence with succession planning. Lower score: short-tenure operations, opportunistic entry, or models predicated on extraction-and-exit.
1.4 Sophisticated failure criterion — local-fronted, externally-controlled. Operations that present a local face while being financially or operationally controlled externally score Failing on this dimension regardless of other criteria. This includes: locally registered subsidiaries of foreign chains, “owner-operator” branding for franchise locations, and storefront localism masking corporate ownership.
Dimension 2: Money Flow & Local Retention
Where does the money actually go?
2.1 First-landing of traveller payment. Where does traveller money first arrive — provider’s account, intermediary’s account, escrow? Higher score: directly to the provider’s local account. Lower score: held by intermediary, especially if held outside Finland.
2.2 Proportion retained locally on a typical transaction. Of every €100 a traveller spends through this operation, how much is paid to people, businesses, and supply chains physically present in the municipality? Higher score: 80%+. Lower score: under 50%.
2.3 Disclosure of non-commission outflows. Does the operation disclose its own non-local cost base — payment processing, hosting, software, advertising spend, professional services — with the same specificity it demands of competitors? Higher score: itemised public disclosure. Lower score: outflows acknowledged only when challenged.
2.4 Sophisticated failure criterion — selective disclosure. Operations that publish “money stays local” claims while omitting their own outflows score Failing on this dimension regardless of other criteria. This is the criterion that distinguishes operations doing the work from operations performing the rhetoric.
Dimension 3: Booking & Distribution Architecture
How do travellers reach providers, and what does the path cost?
3.1 Direct booking accessibility. Can a traveller book the operation’s services directly, without a platform intermediary, at no penalty? Higher score: prominent direct booking, equal or better pricing than intermediated channels. Lower score: direct booking obscured, harder to access, or priced punitively.
3.2 Commission and fee transparency. When an intermediated booking occurs, is the commission or fee structure disclosed to the traveller? Higher score: visible at point of booking. Lower score: invisible, or only inferable from the gap between intermediary and direct prices.
3.3 Provider distribution autonomy. Does the operation respect or constrain providers’ ability to distribute through other channels? Higher score: providers retain full control of their channel mix. Lower score: rate parity, exclusivity, or equivalent constraints.
3.4 Sophisticated failure criterion — pay-to-rank. Does the operation’s visibility, ranking, or recommendation logic correlate with commission tier, advertising spend, or other commercial relationships, in ways that diverge from quality signals? Higher score: ranking is editorially or algorithmically grounded in factors verifiable to travellers. Lower score: higher-paying providers receive better visibility regardless of provider quality. This is the OTA-mechanic at any scale.
Dimension 4: Experience Integrity
Is the experience particular to this place, or replicable anywhere?
4.1 Place-specificity. Is the experience genuinely distinctive to this destination, or is it a templated product applied here? Higher score: experience could not be replicated elsewhere without losing its meaning. Lower score: identical product offered in multiple destinations under different branding.
4.2 Group size and traveller-to-host ratio. Is the experience structured at a human scale — small groups, named hosts, capacity for relationship — or at industrial scale — coach-loads, scripts, processed throughput? Higher score: small-group, named-host. Lower score: mass-handling, anonymous, throughput-optimised.
4.3 Variability and responsiveness. Does the experience respond to weather, group composition, traveller interest, and the rhythms of the place — or does it run to script regardless of circumstance? Higher score: variable, host-led, responsive. Lower score: scripted, time-locked, unchanged across visits.
4.4 Cultural and ecological grounding. Does the experience meaningfully connect travellers to the cultural and ecological reality of the place, or does it offer the aesthetic of place without the substance? Higher score: substantive engagement with local life and land. Lower score: photogenic surface with no underlying engagement.
4.5 Sophisticated failure criterion — performed authenticity. Operations that explicitly market authenticity while running structurally industrialised products score Failing. The combination of authenticity-as-claim and standardisation-as-practice is more harmful than honest standardisation, because it pollutes the term itself.
Dimension 5: Place & Resident Impact
What does this operation do to the community it operates within?
5.1 Housing impact. Does the operation occupy housing stock that would otherwise be available to residents — full-time short-let conversions, staff accommodation displacement, speculative purchase for tourism use? Higher score: no displacement of resident housing. Lower score: material contribution to housing pressure.
5.2 Resident voice in operation. Are residents (not employees, not providers) given structured input into how the operation behaves — content, marketing, presence, partnerships? Higher score: documented resident advisory role with real influence. Lower score: residents are neither consulted nor able to raise concerns.
5.3 Resident objection mechanism. Is there a public, functioning route for a resident to object to specific itineraries, marketing language, partnerships, or behaviours, with a stated response process? Higher score: published mechanism with logged responses. Lower score: no mechanism, or a mechanism that exists nominally but has never produced an outcome.
5.4 Environmental and carrying-capacity awareness. Does the operation publicly recognise the environmental and social carrying capacity of the destination, and does its growth model respect it? Higher score: stated capacity limits, willingness to throttle demand. Lower score: growth-only model, no carrying-capacity language.
5.5 Sophisticated failure criterion — externalised costs. Operations whose business model depends on costs being absorbed by residents — congestion, infrastructure load, inflation of local goods, erosion of social fabric — without contribution to the costs they create score Failing. The test is structural, not rhetorical.
Dimension 6: Cultural Sovereignty
Whose culture is being sold, and on whose terms?
6.1 Sámi cultural relationship. Where Sámi cultural elements appear in marketing, product, or imagery, is there a documented, ongoing relationship with Sámi-led bodies that authorises and shapes that use? Higher score: formal partnership with named Sámi-led organisations, with content authorised by them. Lower score: Sámi imagery used decoratively or generically.
6.2 Finnish-language equivalence. Does the operation present in Finnish as well as in international languages, on substantially equivalent terms? Higher score: full Finnish equivalence on consumer surfaces. Lower score: Finnish content thinner, secondary, or absent.
6.3 Local terminology and knowledge. Does the operation use Finnish and Sámi terms correctly, in context, with explanation rather than as exotic decoration? Higher score: terminology used substantively, with attribution. Lower score: terminology used as marketing flourish without substance.
6.4 Sophisticated failure criterion — extractive cultural use. Operations that use cultural symbols, ritual, or imagery as marketing material while having no operational relationship with the communities those elements come from score Failing. The substantive question is not whether the operation looks culturally connected, but whether the cultures named have any standing to influence the operation.
Dimension 7: Transparency & Disclosure
Does the operation make itself legible to honest scrutiny?
7.1 Commercial model legibility. Beyond the pass/fail of HC3, how clearly does the operation present its commercial model to a non-expert traveller? Higher score: model explained on a dedicated page in plain language, with examples. Lower score: model technically disclosed but practically hidden.
7.2 Editorial-versus-commercial separation. Where the operation produces editorial content alongside commercial offerings, is the separation between the two clearly marked? Higher score: paid placements, partnerships, and commission relationships visibly disclosed at the point of presentation. Lower score: editorial and commercial blurred without disclosure.
7.3 AI and automation disclosure. Where AI, recommendation algorithms, or automated systems shape what travellers see, is this disclosed in language travellers can understand? Higher score: clear, specific disclosure of AI’s role. Lower score: AI use undisclosed or buried in technical language.
7.4 Sophisticated failure criterion — disclosure asymmetry. Operations that demand transparency from competitors while themselves disclosing only what is flattering score Failing. The audit is interested in operations that disclose what they would prefer to hide, not operations that disclose what helps them market.
Dimension 8: Accountability & Right of Reply
What happens when something goes wrong?
8.1 Substantive correction policy. Beyond HC7’s pass/fail, does the correction policy specify response times, escalation routes, and standards for what merits public correction? Higher score: published, specific, time-bound. Lower score: vague commitment without operational specificity.
8.2 Removal-for-cause criteria. Where the operation can remove or downgrade other parties (providers, partners, content), are the criteria for doing so codified and public? Higher score: written criteria, applied consistently, with appeal route. Lower score: admin-discretionary, unwritten, inconsistent.
8.3 External scrutiny. Is there a person, body, or process external to the operation with standing to challenge its decisions, scores, or claims? Higher score: named external scrutineer with public remit. Lower score: all assessment internal.
8.4 Stakeholder feedback loop. Is there a structured process by which affected parties — providers, residents, travellers, communities — can shape the operation’s standards and criteria over time? Higher score: documented feedback loop with logged changes. Lower score: no structured route, or feedback solicited only when convenient.
8.5 Sophisticated failure criterion — single-person bottleneck. Operations whose authenticity, verification, and accountability all depend on a single individual’s judgement, time, or attention score Failing. The criterion exists because solo-administration of trust signals is structurally fragile, and the audit is interested in operations whose standards survive their founder.
Verdicts
The verdict combines the hard-criteria status with the dimension grades.
Pass
- All seven hard criteria met
- All eight dimensions graded Adequate or Strong
- At least five dimensions graded Strong
- No Failing criterion in any dimension
A Pass operation is exemplary. It is structurally aligned with the long-term value of the destination and demonstrably so on public, verifiable evidence.
Pass with Notes
- All seven hard criteria met
- At most one dimension graded Weak
- No dimension graded Failing
- No Failing criterion in any dimension
A Pass with Notes operation is broadly aligned but with specific weaknesses identified. The Notes are public and the operation is expected to address them on a stated timeline.
Conditional Fail
- All seven hard criteria met
- Two or more dimensions graded Weak, OR one dimension graded Failing
- No more than one Failing criterion across all dimensions
A Conditional Fail operation cannot be presented as aligned with the model. The operation may remain operating but should not be referenced as exemplary, recommended without caveats, or used as a positive reference point. The conditions are stated publicly and a re-audit timeline is agreed.
Fail
- One or more hard criteria not met, OR
- Three or more dimensions graded Failing, OR
- Sophisticated failure criterion triggered in two or more dimensions
A Fail operation is structurally misaligned with the model. The audit’s verdict is published with evidence and reasoning. The operation is not recommended, partnered with, or held up as an example.
How scoring is applied
The audit is applied as follows:
- Self-assessment is permitted but not sufficient. Operations may score themselves as a starting point, but no published audit verdict relies on self-assessment alone.
- External scoring is required for a published verdict. A named external scorer, identified in the published audit, assesses the operation independently. Disagreements between self-assessment and external assessment are documented in the published audit.
- Evidence is public. Every criterion’s score must be supported by evidence accessible to a third-party reader. Scores not supported by public evidence are recorded as “Insufficient Evidence” rather than guessed.
- Refusal to provide evidence is itself evidence. An operation that refuses to disclose what is required to score a criterion is recorded as Failing on that criterion, not as Insufficient Evidence. The framework distinguishes between “could not access” and “would not disclose”.
Disputes and revision
Operations have right of reply on every published audit. Disputes are handled as follows:
- The operation may request review of any specific criterion within 30 days of publication.
- The review is conducted by a person other than the original scorer.
- The outcome of the review is published, with reasoning, regardless of whether the score changes.
- If the framework’s criteria themselves are challenged, the challenge is logged in the framework’s revision history, considered for the next revision cycle, and either incorporated or explicitly declined with reasoning.
The framework is revised annually. Each revision is dated, version-numbered, and accompanied by a public changelog.
Limitations
This framework cannot do everything. Specifically:
- It does not measure quality of service, hospitality warmth, or traveller satisfaction. Those are real and important; they are not what this framework is for.
- It does not assess legal compliance, safety, or regulatory standing. Operations passing the audit may still be operating outside the law or unsafely; the audit is not a substitute for proper oversight.
- It is biased toward the specific values of the Levi Tourism Model. Operations operating under different value systems may legitimately score poorly on this framework while being entirely well-suited to other contexts.
- It is necessarily reductive. Eight dimensions cannot capture the full texture of an operation’s relationship with a place. The audit is a starting point for serious conversation, not a substitute for it.
- It is currently designed by one person, drawing on academic literature and primary research, but without yet having been pressure-tested by Sámi-led review, Finnish municipal review, or by application to a body of operations beyond the author’s own. Version 0.1 is a draft, not a settled instrument.
What this framework is not yet
The following are open and are flagged for v0.2:
- Sámi review of Dimension 6 (Cultural Sovereignty). The criteria as drafted are written by a non-Sámi author and require substantive review by Sámi-led bodies before they can be considered authoritative.
- Finnish-language equivalent of the framework. A framework about cultural sovereignty published in English only is itself failing one of its own criteria.
- Calibration against a body of audited operations. The thresholds for Strong / Adequate / Weak / Failing are reasoned but uncalibrated. Application to twenty operations of varying types will reveal where the thresholds are too soft or too hard.
- An external scorer. The framework requires one. The framework’s author has not yet identified, approached, or appointed one. Until this gap is closed, no published audit can be considered fully validated.
Application: the author’s own operation
levifinland.com will be audited against this framework before it accepts its first paid direct booking. The audit will be published in full on future.levifinland.com, including:
- Hard-criteria pass/fail status for all seven HCs
- Dimension grades for all eight dimensions
- Criterion-level scores with evidence
- Notes and conditions where applicable
- The verdict
- A response from the author, including which findings will be addressed and how
- The external scorer’s name and the reasoning behind any disagreement with self-assessment
If the audit produces a verdict of Conditional Fail or Fail, the operation does not launch in the form audited until the failures are resolved or the model changed. This is the published commitment.
If the operation finds itself rewriting the framework to make itself score better, the framework is being corrupted. Independent review of the framework itself, not just of operations against it, is part of how that corruption is prevented.
Author: Colin Harrison Status: v0.1, draft for review. The framework will be revised before v1.0. Contact: colin@levifinland.com Revision history: this is the first version.