This repository contains the reference implementation for the before/after EEG report for Second Attention Workshop participants. It is designed as a visual and UX guide for the development team building the app that processes EEG data and auto-generates participant reports.
- Showcase: This is the intended output for the workshop's before/after comparison report.
- Reference: Use this as a design/UX and data-visualization reference for auto-generating reports from participant EEG data.
https://second-attention-workshop-report.vercel.app/
https://www.loom.com/share/57fbedec20634ef8ab7a66314f5f6f36?sid=669c30cf-2e25-4ae8-b1fe-ff0cb57d661b
- Clone the repo:
git clone https://github.com/Deducer/second-attention-workshop-report.git cd second-attention-workshop-report - Install dependencies and start the app:
cd second-attention-app npm install npm start - Open http://localhost:3000 to view the report.
- All numbers, bars, and graphs are to be generated from participant EEG data.
- The "before" and "after" states correspond to baseline and entrained meditation sessions.
- The report is designed to be auto-generated for each participant after their workshop session.
For each section/element of the report, the following data is required (from EEG analysis and participant input):
- Inputs:
- Aggregate meditative depth metric for each session (e.g., Alpha-Theta power ratio, or a custom score)
- "Before" (baseline) and "After" (entrained) values
- How to compute:
- Typically derived from band power analysis (e.g., mean Alpha+Theta power during meditation)
- Inputs:
- Band power (Alpha, Theta, etc.) for each electrode/region (O1, O2, T3, T4) for both sessions
- (Optional) Coherence/connectivity values between regions
- How to compute:
- Compute average band power per region for each session
- (Optional) Compute pairwise coherence between regions
- Inputs:
- Global or average coherence value for each session (e.g., mean pairwise coherence across all regions)
- How to compute:
- Use standard EEG coherence analysis (e.g., magnitude-squared coherence in Alpha/Theta bands)
- Inputs:
- Mean band power for Alpha, Theta, Beta for each session (whole brain or specific regions)
- How to compute:
- Compute mean power in each band for each session
- Calculate % change (After vs. Before)
- Inputs:
- State Mindfulness Scale (SMS) scores: total and subscales (Mind Awareness, Body Awareness) for each session
- Individual item scores for notable improvements
- How to compute:
- Collect via participant self-report (questionnaire)
- Calculate subscale and total scores for before/after
- Inputs:
- Any standout EEG metric (e.g., exceptional increase in a specific band or region, or unique coherence pattern)
- How to compute:
- Identify participant-specific outliers or strengths in the EEG data
- Inputs:
- Participant's metrics (depth, enhancement, time to deep state) compared to a reference group/distribution
- How to compute:
- Place participant's results in context of group data (e.g., percentile rank)
- Color Mapping:
- Blue → Purple: "After"/"Improvement"/"Higher is better"
- Peach: "Before"/"Baseline"
- Section Structure:
- Meditative Depth Score (gradient card)
- Brain Activity Comparison (spatial EEG, coherence)
- Neural Coherence (progress bar)
- Wave Metrics (Alpha, Theta, Beta)
- Subjective Experience (SMS)
- Unique Neural Strength & Recommendations
- Comparative Results
- Responsiveness:
- Layout is designed for desktop/tablet, but can be adapted for mobile.
- Accessibility:
- Color contrast and font sizes are chosen for readability.
- This repo is a reference implementation: use the structure, styles, and data-visualization logic as a guide for the production app.
- All data is currently static; in production, replace with dynamic values from EEG processing pipeline.
- See Loom walkthrough for intent, dynamic data mapping, and design rationale.
- For questions, contact the report designer or project lead.
second-attention-app/— React app source codepublic/— Static assets (images, logos)README.md— This file
- If you deploy a live demo, update the link above.
- If you record a Loom walkthrough, add the link above.
- For any design/UX clarifications, refer to the Loom or contact the designer.
© Vayu Labs • Neural State Training • www.vayulabs.com