# WIA-EDU-022: Implementation Guide

> **弘益人間** (Benefit All Humanity)

## Quick Start

This guide provides practical steps for implementing the WIA E-Sports Education Standard in your institution.

## Prerequisites

### Administrative
- [ ] Leadership buy-in and support
- [ ] Budget allocation ($5,000-$15,000 for startup)
- [ ] Designated program space (computer lab or dedicated room)
- [ ] Faculty advisor or coach identified

### Technical
- [ ] Gaming-capable computers (minimum 5 for team practice)
- [ ] High-speed internet (minimum 100 Mbps download, 20 Mbps upload)
- [ ] Network infrastructure supporting gaming traffic
- [ ] Student information system (SIS) or learning management system (LMS)

### Policy
- [ ] Code of conduct developed
- [ ] Eligibility requirements defined (GPA, attendance, behavior)
- [ ] Parental consent forms prepared
- [ ] Data privacy policies reviewed

## Implementation Phases

### Phase 1: Planning (Months 1-3)

#### Step 1: Form Planning Committee

Assemble stakeholders:
- School administrator (principal, activities director)
- Faculty advisor (can be non-gamer)
- IT staff member
- 2-3 interested students
- 1-2 parents (optional but helpful)

#### Step 2: Define Program Goals

Example goals:
- Engage 30+ students in first year
- Form competitive teams in 2 games
- Join regional esports league
- Integrate esports with STEM curriculum
- Achieve 85% student satisfaction

#### Step 3: Assess Resources

**Facilities Checklist:**
- [ ] Identify practice space with 5+ computer stations
- [ ] Verify adequate electrical outlets and cooling
- [ ] Ensure ergonomic furniture (desks, chairs)
- [ ] Plan for secure equipment storage

**Technology Audit:**
- [ ] Test internet speed and latency
- [ ] Inventory existing computers suitable for gaming
- [ ] Identify network limitations (firewall rules, bandwidth caps)
- [ ] Plan IT support structure

**Budget Development:**

| Category | Low Budget | Medium Budget | High Budget |
|----------|------------|---------------|-------------|
| Computers | $4,000 (refurb) | $7,000 (mid-tier) | $12,000 (high-end) |
| Peripherals | $500 | $1,500 | $3,000 |
| League Fees | $500 | $1,000 | $2,000 |
| Furniture | $500 | $1,500 | $3,000 |
| Software | $200 | $500 | $1,000 |
| **Total** | **$5,700** | **$11,500** | **$21,000** |

#### Step 4: Develop Curriculum

**Core Components:**
1. **Digital Citizenship** (4 weeks)
   - Online safety and privacy
   - Sportsmanship and respectful communication
   - Managing screen time and balance
   - Recognizing and reporting toxic behavior

2. **Game Fundamentals** (6 weeks)
   - Game mechanics and objectives
   - Role identification and specialization
   - Basic strategy and tactics
   - Performance analysis basics

3. **Team Development** (6 weeks)
   - Communication systems
   - Team coordination drills
   - Conflict resolution
   - Leadership and followership

4. **Competitive Preparation** (4 weeks)
   - Tournament formats and rules
   - Pre-match preparation
   - Performance under pressure
   - Post-match analysis and improvement

**Lesson Plan Template:**

```
Lesson: Team Communication Fundamentals
Duration: 60 minutes
Grade Level: 9-12

Objectives:
- Students will identify 3 types of in-game callouts
- Students will demonstrate clear, concise communication
- Students will practice active listening and acknowledgment

Materials:
- Gaming PCs with selected game
- Voice communication software (Discord)
- Communication rubric

Activities:
1. Introduction (10 min): Types of callouts (informational, tactical, emotional)
2. Demo (10 min): Coach demonstrates effective vs. ineffective communication
3. Practice (30 min): Students practice in structured scenarios
4. Debrief (10 min): Reflection on communication quality

Assessment:
- Communication rubric (clarity, conciseness, positivity)
- Peer feedback
- Self-reflection journal
```

### Phase 2: Setup (Months 3-6)

#### Step 5: Acquire Equipment

**Computer Specifications (Minimum):**
- CPU: Intel i5-10400 or AMD Ryzen 5 3600
- GPU: NVIDIA GTX 1660 or AMD RX 5600
- RAM: 16GB DDR4
- Storage: 500GB SSD
- OS: Windows 10/11

**Peripheral Recommendations:**
- Gaming mice: Logitech G203, Razer DeathAdder
- Keyboards: Redragon K552, Corsair K55
- Headsets: HyperX Cloud Stinger, Logitech G432
- Mousepads: Large cloth pads for consistency

#### Step 6: Network Configuration

**IT Checklist:**
- [ ] Whitelist gaming platforms (Steam, Epic, Riot Client, etc.)
- [ ] Open required ports for games
- [ ] Configure QoS to prioritize gaming traffic
- [ ] Set up monitoring for bandwidth usage
- [ ] Create separate VLAN for esports (optional but recommended)

**Common Ports:**
- League of Legends: 5000-5500 UDP
- Valorant: 7000-8000 UDP
- Rocket League: 7000-9000 UDP
- Discord Voice: 50000-65535 UDP

#### Step 7: League Registration

**Major High School Leagues:**

1. **PlayVS**
   - Games: LoL, Rocket League, Smash, Madden, NBA 2K, more
   - Cost: ~$64/player/season
   - Structure: State-based competition
   - Support: Coaching resources, parent communication tools

2. **NASEF (North America Scholastic Esports Federation)**
   - Games: Various free-to-play titles
   - Cost: Free to participate
   - Focus: Educational framework, community building
   - Resources: Curriculum guides, coach training

3. **HSEL (High School Esports League)**
   - Games: 20+ titles
   - Cost: Free tiers available
   - Structure: Regional and national tournaments
   - Features: Scholarship opportunities

**Registration Process:**
1. Create school account on league platform
2. Submit school verification documents
3. Recruit and register students
4. Pay team fees (if applicable)
5. Attend league orientation/coach meeting

### Phase 3: Launch (Months 6-9)

#### Step 8: Student Recruitment

**Recruitment Strategies:**
- School-wide announcements and posters
- Presentations in relevant classes (computer science, game design)
- Open house/demo day with playable stations
- Social media promotion
- Student ambassadors spreading word

**Information Session Agenda:**
1. What is educational esports? (10 min)
2. Program structure and expectations (15 min)
3. Team formation and tryout process (10 min)
4. Q&A (15 min)
5. Signup and consent forms (10 min)

#### Step 9: Team Formation

**Tryout Structure:**
1. **Skills Assessment:**
   - In-game mechanics test
   - Game knowledge quiz
   - 1v1 or skills challenge

2. **Teamwork Evaluation:**
   - Scrimmage observation
   - Communication assessment
   - Attitude and coachability

3. **Interview:**
   - Why do you want to join?
   - Time commitment confirmation
   - Academic standing verification

**Balanced Team Building:**
- Mix skill levels (avoid stacking all best players)
- Consider personality compatibility
- Identify natural leaders for captain roles
- Plan for substitute positions
- Create practice squad for development

#### Step 10: Establish Routines

**Practice Schedule Template:**

| Day | Time | Activity | Duration |
|-----|------|----------|----------|
| Monday | 3:30-5:00 | Individual skill drills | 90 min |
| Tuesday | 3:30-5:30 | Team strategy and scrimmages | 120 min |
| Wednesday | 3:30-4:30 | VOD review and theory | 60 min |
| Thursday | 3:30-5:30 | Full team practice | 120 min |
| Friday | Optional | Open lab/casual play | - |

**Practice Session Structure:**
```
3:30-3:35  Check-in and announcements
3:35-3:50  Warm-up drills (last-hitting, aim training, etc.)
3:50-4:05  Strategy lesson or concept review
4:05-4:50  Structured scrimmages
4:50-5:15  Replay review and feedback
5:15-5:30  Cool-down, goal setting, clean up
```

### Phase 4: Operations (Ongoing)

#### Managing Competition

**Match Day Checklist:**
- [ ] Equipment tested 30 min before match
- [ ] Roster confirmed and submitted
- [ ] Strategy reviewed with team
- [ ] Pre-match team huddle
- [ ] Technical support on standby
- [ ] Streaming setup (if applicable)

**Post-Match Protocol:**
1. GG and handshake (virtual or in-person)
2. Brief immediate feedback (5 min)
3. Cool-down period
4. Detailed VOD review (next practice)
5. Update records and statistics

#### Student Wellness Monitoring

**Weekly Check-ins:**
- Screen time self-reporting
- Academic progress review
- Physical activity logging
- Sleep quality assessment
- Stress/burnout indicators

**Red Flags:**
- Grades declining
- Excessive gaming outside practice
- Withdrawal from friends/family
- Physical complaints (wrist pain, headaches)
- Negative attitude or outbursts

**Intervention Steps:**
1. Private conversation with student
2. Consult parents if concerns persist
3. Adjust practice schedule or role
4. Refer to school counselor if needed
5. Temporary suspension from competition if serious

#### Program Assessment

**Key Metrics:**

| Metric | Target | Measurement |
|--------|--------|-------------|
| Participation Rate | 30+ students year 1 | Enrollment count |
| Retention | 80% stay full season | Track dropouts |
| Academic Eligibility | 95% maintain standards | Grade reports |
| Student Satisfaction | 4.0/5.0 average | End-of-season survey |
| Competitive Success | Win 40% of matches | Record tracking |
| Community Engagement | 3+ events per year | Event attendance |

**End-of-Season Review:**
1. Collect student, parent, and staff feedback
2. Review budget actuals vs. planned
3. Assess achievement of program goals
4. Identify successes and challenges
5. Plan improvements for next season
6. Celebrate achievements (awards ceremony, banquet)

## Best Practices

### Inclusive Environment

- **Gender Inclusion:** Actively recruit underrepresented genders; consider women's/non-binary teams if interest exists
- **Skill Diversity:** Offer multiple tiers (varsity, JV, beginner) so all can participate
- **Non-Player Roles:** Create positions for analysts, content creators, managers, broadcasters
- **Accessibility:** Provide adaptive equipment for students with disabilities
- **Cultural Sensitivity:** Respect diverse backgrounds; translate materials as needed

### Academic Integration

**Cross-Curricular Connections:**

| Subject | Integration Example |
|---------|---------------------|
| Math | Statistics analysis, probability calculations, resource optimization |
| English | Strategic writing, team communication, tournament reports |
| Science | Reaction time experiments, ergonomics studies, nutrition for performance |
| Social Studies | Global gaming culture, esports industry economics, digital citizenship |
| Art | Logo design, stream overlays, promotional graphics |
| Computer Science | Game mechanics analysis, basic scripting, data visualization |

### Sustainability

**Funding Strategies:**
- School activities budget allocation
- Fundraising (car washes, bake sales, crowdfunding)
- Local business sponsorships
- Booster club/PTA support
- Grant applications (education technology, STEM)
- Streaming revenue (if significant following)

**Community Partnerships:**
- Local gaming cafes (space for events)
- Computer retailers (equipment donations)
- Alumni network (mentorship, funding)
- Colleges with esports (recruitment pathway, coaching support)
- Youth organizations (after-school program collaboration)

## Troubleshooting

### Common Challenges

**Challenge: Low student interest**
- Solution: Survey to understand why; adjust games offered; increase visibility; improve recruitment messaging

**Challenge: Parent opposition**
- Solution: Host parent education night; share research on benefits; emphasize academics-first policy; invite parents to observe

**Challenge: Technical difficulties**
- Solution: Have backup equipment; maintain troubleshooting guide; train student tech support; establish ISP relationship for priority support

**Challenge: Toxic behavior**
- Solution: Enforce code of conduct consistently; provide communication training; remove problematic individuals if necessary; model positive behavior

**Challenge: Burnout (students or coach)**
- Solution: Build in off-season; limit practice hours; ensure workload distribution; prioritize wellness over winning

**Challenge: Budget constraints**
- Solution: Start small (one game, one team); use free-to-play games; borrow/repurpose existing equipment; seek community donations

## WIA Standard Compliance

To achieve WIA-EDU-022 certification:

### Bronze Level
- [ ] Implement basic program structure with clear learning objectives
- [ ] Use WIA data models for student and team records
- [ ] Establish code of conduct and safety policies
- [ ] Track participation and basic outcomes

### Silver Level
- [ ] Integrate with school SIS/LMS via WIA APIs
- [ ] Implement comprehensive curriculum with assessments
- [ ] Maintain detailed performance and progress data
- [ ] Conduct regular wellness checks

### Gold Level
- [ ] Demonstrate measurable learning outcomes
- [ ] Implement career pathways program
- [ ] Achieve high student satisfaction (4.0+/5.0)
- [ ] Contribute to WIA community (resources, best practices)

### Platinum Level
- [ ] Excellence in inclusion (diverse participation, accessibility)
- [ ] Innovation in curriculum or program structure
- [ ] Research contribution (publish findings, present at conferences)
- [ ] Exemplary outcomes (academic, competitive, career placement)

## Resources

- **Curriculum**: https://wiastandards.com/edu-022/curriculum
- **Training**: https://wiastandards.com/edu-022/training
- **Community**: https://community.wiastandards.com/esports
- **Support**: support@wiastandards.com

---

© 2025 WIA - World Certification Industry Association
Licensed under MIT License

**弘益人間** · Benefit All Humanity


## Annex E — Implementation Notes for PHASE-4-INTEGRATION

The following implementation notes document field experience from pilot
deployments and are non-normative. They are republished here so that early
adopters can read them in context with the rest of PHASE-4-INTEGRATION.

- **Operational scope** — implementations SHOULD declare their operational
  scope (single-tenant, multi-tenant, federated) in the OpenAPI document so
  that downstream auditors can score the deployment against the correct
  conformance tier in Annex A.
- **Schema evolution** — additive changes (new optional fields, new error
  codes) are non-breaking; renaming or removing fields, even in error
  payloads, MUST trigger a minor version bump.
- **Audit retention** — a 7-year retention window is sufficient to satisfy
  ISO/IEC 17065:2012 audit expectations in most jurisdictions; some
  regulators require longer retention, in which case the deployment policy
  MUST extend the retention window rather than relying on this PHASE's
  defaults.
- **Time synchronization** — sub-second deadlines depend on synchronized
  clocks. NTPv4 with stratum-2 servers is sufficient for most deadlines
  expressed in this PHASE; PTP is recommended for sites that require
  deterministic interlocks.
- **Error budget reporting** — implementations SHOULD publish a monthly
  error-budget summary (latency p95, error rate, violation hours) in the
  format defined by the WIA reporting profile to facilitate cross-vendor
  comparison without exposing tenant-specific data.

These notes are not requirements; they are a reference for field teams
mapping their existing operations onto WIA conformance.

## Annex F — Adoption Roadmap

The adoption roadmap for this PHASE document is non-normative and is intended to set expectations for early implementers about the relative stability of each section.

- **Stable** (sections marked normative with `MUST` / `MUST NOT`) — semantic versioning applies; breaking changes require a major version bump and at minimum 90 days of overlap with the prior major version on all WIA-published reference implementations.
- **Provisional** (sections in this Annex and Annex D) — items are tracked openly and may be promoted to normative status without a major version bump if community feedback supports promotion.
- **Reference** (test vectors, simulator behaviour, the reference TypeScript SDK) — versioned independently of this document so that mistakes in reference material can be corrected without amending the published PHASE document.

Implementers SHOULD subscribe to the WIA Standards GitHub release notifications to track promotions between these tiers. Comments on the roadmap are accepted via the GitHub issues tracker on the WIA-Official organization.

The roadmap is reviewed at every minor version of this PHASE document, and the review outcomes are recorded in the version-history table at the start of the document.

## Annex G — Test Vectors and Conformance Evidence

This annex describes how implementations capture and publish conformance
evidence for PHASE-4-INTEGRATION. The procedure is non-normative; it standardizes the
shape of evidence so that auditors and downstream integrators can compare
implementations without re-running the full test matrix.

- **Test vectors** — every normative requirement in this PHASE has at least
  one positive vector and one negative vector under
  `tests/phase-vectors/phase-4-integration/`. Implementations claiming
  conformance MUST run all vectors in CI and publish the resulting
  pass/fail matrix in their compliance package.
- **Evidence package** — the compliance package is a tarball containing
  the SBOM (CycloneDX 1.5 or SPDX 2.3), the OpenAPI document, the test
  vector matrix, and a signed manifest. Signatures use Sigstore (DSSE
  envelope, Rekor transparency log entry) so that downstream consumers
  can verify provenance without trusting a private CA.
- **Quarterly recheck** — implementations re-publish the evidence package
  every quarter even if no source change occurred, so that consumers can
  detect environmental drift (compiler updates, dependency updates, OS
  updates) without polling vendor changelogs.
- **Cross-vendor crosswalk** — the WIA Standards working group maintains a
  crosswalk that maps each vector to the equivalent assertion in adjacent
  industry programs (where one exists), so an implementer that already
  certifies under one program can show conformance to PHASE-4-INTEGRATION with
  reduced incremental effort.
- **Negative-result reporting** — vendors MUST report negative results
  with the same fidelity as positive ones. A test that is skipped without
  recorded justification is treated by auditors as a failure.

These conventions are intended to make conformance evidence portable and
machine-readable so that adoption of PHASE-4-INTEGRATION does not require bespoke
auditor tooling.

## Annex H — Versioning and Deprecation Policy

This annex codifies the versioning and deprecation policy for PHASE-4-INTEGRATION.
It is non-normative; the rules below describe the policy that the WIA
Standards working group commits to when amending this PHASE document.

- **Semantic versioning** — major / minor / patch components follow
  Semantic Versioning 2.0.0 (https://semver.org/spec/v2.0.0.html).
  Major bump indicates a backwards-incompatible change to a normative
  requirement; minor bump indicates new normative requirements that do
  not break existing implementations; patch bump indicates editorial
  changes only (clarifications, typo fixes, formatting).
- **Deprecation window** — when a normative requirement is removed or
  altered in a backwards-incompatible way, the prior major version is
  maintained in parallel for at least 180 days. During the parallel
  window, both major versions are marked Stable in the WIA Standards
  registry and either may be cited as "WIA-conformant".
- **Sunset notification** — deprecated major versions enter a 12-month
  sunset window during which the WIA registry marks the version as
  Deprecated. The deprecation entry includes a migration note pointing
  to the replacement requirement(s) and an explanation of why the
  change was made.
- **Editorial errata** — patch-level errata are issued without a
  deprecation window because they do not change normative behaviour.
  Errata are tracked in a public errata register and each entry is
  signed by the WIA Standards working group chair.
- **Implementation changelog mapping** — implementations SHOULD publish
  a changelog mapping each PHASE version they support to the specific
  build, container digest, or SDK version that satisfies the version.
  This allows downstream auditors to verify version conformance without
  re-running the entire test matrix on every release.

The policy is reviewed at the same cadence as the PHASE document and
any changes to the policy itself are tracked in the version-history
table at the start of the document.

## Annex I — Interoperability Profiles

This annex describes how implementations declare interoperability profiles
for PHASE-4-INTEGRATION. The profile mechanism is non-normative and exists so that
deployments of varying scope (single tenant, regional cluster, federated
network) can advertise the subset of normative requirements they satisfy
without misrepresenting partial conformance as full conformance.

- **Profile manifest** — every implementation publishes a profile manifest
  in JSON. The manifest enumerates the normative requirement IDs from this
  PHASE that are satisfied (`status: "supported"`), partially satisfied
  (`status: "partial"`, with a reason field), or excluded
  (`status: "excluded"`, with a justification). The manifest is signed
  using the same Sigstore key used for the SBOM in Annex G.
- **Federation profile** — federated deployments publish an aggregated
  manifest summarizing the union and intersection of member-implementation
  profiles. The aggregated manifest is consumed by directory services so
  that callers can route a request to the least common denominator profile
  required for an interaction.
- **Backwards-profile compatibility** — when a deployment migrates from one
  profile to a wider profile, the prior profile manifest remains valid and
  signed for the deprecation window defined in Annex H. This preserves
  audit traceability for auditors evaluating long-term interoperability.
- **Profile registry** — the WIA Standards working group maintains a
  public registry of named profiles. Common deployment shapes (e.g.,
  "Edge-only", "Federated-with-replay") are added to the registry by
  consensus. Registry entries are immutable; new shapes are added under
  new names rather than amending existing entries.
- **Profile versioning** — profile names are versioned with the same
  Semantic Versioning rules described in Annex H. A deployment that
  advertises `WIA-P4-INTEGRATION-Edge-only/2` is asserting conformance with
  the second major version of the named profile, not the second deployment
  of an unversioned profile.

The profile mechanism is intentionally lightweight; it is meant to make
real deployment shapes visible without forcing every deployment to
satisfy every normative requirement.
