From Project to Product: How cant.study → wehelpyou.study Transforms Student Note‑Taking and Meetings with Open‑Source LLMs


For three semesters I worked as an on‑campus student note‑taker, clocking in at 8 a.m. every Monday, Wednesday, and Friday. While the paycheck helped, the routine exposed a deeper accessibility gap: many courses never secure a note‑taker, leaving NTID classmates to piece together scattered slides, hurried phone photos, or—too often—nothing at all.
Why can’t note‑taking be automated, reliable, and context‑aware?
That question became CANT — the Context‑Aware Note Taker. (http://cant.study/)
From Manual Notes to Multimodal Intelligence
CANT’s earliest prototype hinged on three technical insights:
Context Correction Layer
Raw audio is first transcribed by NVIDIA Canary. A Context Correction Layer then aligns jargon, acronyms, and speaker disfluencies with the instructor’s actual slides and course materials, transforming transcripts into task‑ready text.Iterative Chunk Summarization
Lectures are streamed in ~10 k‑token chunks. Each chunk is summarized, then adjacent summaries are recursively refined to build a cohesive document—reducing LLM hallucinations and keeping sections tight and well‑labeled.24‑Hour Retention Window
To balance study needs with privacy, only the generated structured notes are retained on the server after the first 24 hours. The system performs an auto‑purge unless a student explicitly opts to archive them.
Design Decisions & Why They Won
Challenge | Design Decision | Why It Won |
Transcription accuracy | Context Correction Layer over Canary output | Injects slide headings, technical terms, and abbreviations for near‑human readability. |
Token limits in LLMs | Chunk‑then‑iterate pipeline | Produces tighter summaries and scales to multi‑hour seminars. |
Observability & debugging | Langfuse tracing + log streams (file & Elasticsearch) | One-click replay of any session’s LLM calls; Kibana dashboards spot drift or latency spikes. |
Student trust & compliance | 24‑hour note retention | Guarantees automatic cleanup without manual admin intervention. |
A (Very) Brief Tour of the Deployment Pipeline
Frontend – React 18 + Tailwind. A single “Start Session” button streams 30 s audio chunks and visualizes progress.
API Layer (Node.js + Express) – Handles JWT auth, writes Langfuse traces, handles other business logic.
Sandbox Container
Speech: NVIDIA Canary
Context Parser: OCR for slides/PDFs
LLM: Llama 3 8B‑Instruct‑128k behind an Nginx reverse proxy
Storage – MongoDB for transcripts & structured notes (ephemeral collections respect the 24‑hour policy).
CI/CD – GitHub Actions → Docker Buildx → zero‑downtime blue‑green deploys on Linode’s dedicated GPU cloud.
Launch Day: 30 April 2025 — Capstone Poster & Public Release
We unveiled CANT during the RIT Data Science Capstone poster session on 30 April 2025 and simultaneously opened public access:
Domain | Purpose |
cant.study | The problem space—where students can’t study without equitable notes. |
wehelpyou.study | The solution space—CANT’s live app and onboarding hub. |
First‑week impact
42 new registered users
100 + unique visitors
8 + hours of audio captured
14 structured note sets generated
Improvements and Results
What’s Next?
Federated Learning
Train the Context Correction Layer on‑device so each cohort’s jargon improves future transcripts—without centralizing raw data.Edge Deployment using Service Workers
Packaging the full pipeline for offline field trips and patchy Wi‑Fi environments.Open‑Source Modules
Releasing the chunk‑iterative summarizer, Context Aware ASR Correction so researchers can plug in custom LLMs.
Thank you for following our journey from sleepy lecture halls to an AI‑powered accessibility platform. Have feedback or want to contribute? Visit cant.study—because when you CANT.study, WEHELPYOU.study.
Email: gmail@sanathswaroop.com
Contact: +1 (585)-565-2567
Subscribe to my newsletter
Read articles from Sanath Swaroop Mulky directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
