\pdfminorversion=7 \documentclass[12pt,letterpaper]{article} \usepackage[T1]{fontenc} \usepackage[utf8]{inputenc} \usepackage[american]{babel} \usepackage{csquotes} \usepackage[ style=science, backend=biber, sorting=none, articletitle=true, url=false, doi=false, eprint=false, isbn=false ]{biblatex} \addbibresource{references.bib} \AtEveryBibitem{\clearlist{language}} \usepackage{newtxtext,newtxmath} \usepackage{microtype} \usepackage{amsmath} \usepackage{graphicx} \usepackage{booktabs} \usepackage[margin=0.92in,top=0.82in,bottom=0.95in,headsep=18pt,footskip=24pt]{geometry} \usepackage[font=small,labelfont=bf,labelsep=period]{caption} \usepackage{array} \usepackage{tabularx} \usepackage{setspace} \usepackage{titlesec} \usepackage{enumitem} \usepackage{fancyhdr} \usepackage{needspace} \usepackage[table]{xcolor} \definecolor{ClayLine}{HTML}{EEDBDD} \definecolor{SumiInk}{HTML}{2D2426} \definecolor{DustyTaupe}{HTML}{866A6D} \usepackage[hidelinks]{hyperref} \usepackage[switch]{lineno} \color{SumiInk} \setstretch{1.45} \setlength{\parindent}{0pt} \setlength{\parskip}{0.42em} \setlength{\tabcolsep}{9pt} \renewcommand{\arraystretch}{1.14} \setlist[itemize]{leftmargin=1.3em,itemsep=0.18em,topsep=0.25em} \setlength{\headheight}{28pt} \emergencystretch=2em \raggedbottom \titleformat{\section} {\Large\sffamily\bfseries\color{SumiInk}} {\thesection} {0.6em} {} [\vspace{0.25em}\color{ClayLine}\titlerule] \titleformat{\subsection} {\normalsize\sffamily\bfseries\color{SumiInk}} {\thesubsection} {0.55em} {} \titlespacing*{\section}{0pt}{1.25em}{0.55em} \titlespacing*{\subsection}{0pt}{0.9em}{0.25em} \pagestyle{fancy} \fancyhf{} \lhead{\small\sffamily Curtis Media Review} \rhead{\small\sffamily \thepage} \renewcommand{\headrulewidth}{0.35pt} \renewcommand{\headrule}{\hbox to\headwidth{\color{ClayLine}\leaders\hrule height \headrulewidth\hfill}} \newcommand{\metric}[2]{\textbf{#1} & #2\\} \begin{document} \linenumbers \thispagestyle{empty} \begin{center} \vspace*{-0.6em} {\fontsize{20}{22}\selectfont\bfseries\color{SumiInk} Longitudinal media review for audition-oriented violin practice\par} \vspace{0.95em} {\normalsize\color{SumiInk} Alan N. Pham$^{1}$\par} \vspace{0.38em} {\small\color{DustyTaupe} $^{1}$AO Labs, Worcester, MA, USA\par} \end{center} \vspace{0.35em} \noindent\textbf{High-level music training produces dense evidence: daily practice videos, repeated repertoire, changing technical constraints, and periodic performance goals. Most of that evidence is difficult to use because it is stored as unstructured media rather than as a longitudinal record. We report Curtis Media Review, a live AO Labs system that indexes public practice media, identifies practice candidates, records model-assisted section reviews, and converts the current state into a compact practice record. The checked live state on 7 May 2026 used the YouTube channel \texttt{@nalalan} as the primary source, indexed 207 public videos, marked 102 as practice candidates, marked 82 as long-form candidates, stored 6 media samples, retained 11 reviewed sections, and recorded 51 Curtis-focused findings. No repertoire work is currently named at clear confidence. The first screen therefore reports \textit{Piece being identified}, an identifying state instead of a piece-completion percentage, and one major recording/practice constraint. Possible labels and user-rejected false labels are not allowed to become current repertoire titles, and new piece-title output requires corroboration from a second audio verifier before it can update the piece list. The system uses GPT-5.5 for vision and text review and separate audio models for identification and verification, while preserving explicit limits: Curtis admission cannot be predicted from current samples, weak audio/video evidence is marked unjudged, and the system reports practice signals rather than admissions probability. The contribution is an experimental method for turning accumulated practice media into auditable feedback loops, progress history, and source-bounded coaching suggestions.} \newpage \pagestyle{fancy} \section{Introduction} Elite music preparation depends on repeated, deliberate practice and feedback over long time horizons~\cite{ericsson_role_1993,williamon_musical_2004}. The useful data are not only polished performances. They include ordinary practice sessions, failed takes, camera angle limitations, recurring technical problems, and changes in focus across days. A violinist recording daily practice can therefore produce a large longitudinal dataset, but the raw dataset does not automatically become an operating record. Curtis Media Review addresses this problem as a live software system rather than as a portfolio page~\cite{curtis_app_2026}. The system indexes public practice media, classifies candidate practice logs, records technical observations, and keeps the current practice constraint visible. It is designed for the Curtis Institute goal, but it does not claim to estimate an acceptance probability. The current role of the system is narrower: observe practice evidence, retain history, separate judged from unjudged evidence, and return a small number of practice constraints that can be tested in subsequent recordings. The system is also part of AO Labs' broader progress architecture. In May 2026, AO Progress was introduced as a shared ledger for cross-application state and long-term changes across public apps, papers, CV artifacts, Curtis practice state, Imagineer state, and Relay state~\cite{progress_app_2026}. Curtis supplies one stream in that ledger: timestamped media and review evidence for musical development. \section{System state} The first production state used the live backend at \url{https://curtis.aolabs.io/api/curtis/media-status}. The checked state on 7 May 2026 reported the values in Table~\ref{tab:state}. These are operating facts, not outcome claims. \begin{table}[htbp] \centering \caption{Curtis production state on 7 May 2026.} \label{tab:state} \begin{tabularx}{\textwidth}{>{\bfseries}p{0.31\textwidth}X} \toprule Signal & Current value \\ \midrule \metric{Primary source}{YouTube channel \texttt{@nalalan}} \metric{Inventory}{207 indexed videos} \metric{Practice candidates}{102 videos} \metric{Long-form candidates}{82 videos} \metric{Media samples}{6 captured samples} \metric{Reviewed sections}{11 sections} \metric{Curtis-focused findings}{51 findings} \metric{Latest dated practice log}{5-1-26} \metric{Review model}{GPT-5.5 for vision/text, separate audio models for audio evidence and piece verification} \metric{Current piece}{Piece being identified; no clear repertoire title} \metric{Curtis-level completion}{Withheld until the piece is identified at clear confidence} \metric{Current focus}{Audition-style setup consistency} \metric{Current constraint}{Open strings only, mirror-check contact point} \metric{Boundary}{Curtis admission cannot be predicted from current samples} \bottomrule \end{tabularx} \end{table} \subsection{Media indexing} The source inventory stores platform, title, publication time, duration, view count, candidate reasons, media kind, and blocker state. Dated long-form practice logs are treated as high-value longitudinal records because the title contains a practice date and the duration preserves session-level evidence. Other videos are retained but not forced into the practice lane. \subsection{Review evidence} The review layer stores section-level findings with a dimension, evidence source, judgment, and practice constraint. The current dimensions include tone, intonation, shifts, time, articulation, and audition delivery. Findings can be strong, needing work, or unjudged. The unjudged state is part of the system design: if the sound is not available, the hand is obscured, the camera angle hides a transition, or still frames do not support a timing claim, the system should not convert weak evidence into a confident critique. \section{First readout} The first live readout is experimental but usable. The strongest immediate value is that the system has already converted a large, timestamped practice archive into a queryable corpus with captured media samples and stored section reviews. The current piece state is not a named repertoire work. The home screen reduces the current state to the active piece-identification state, a withheld completion percentage, and one major practice or recording constraint. Generic etude/caprice evidence remains outside the piece-title field, and possible or uncorroborated repertoire labels are suppressed. The current active focus is audition-style setup consistency, with open strings only and mirror-check contact point as the practice condition. These statements are not generic pedagogy; they are extracted from the stored review findings, piece record, and current progress plan. \begin{table}[htbp] \centering \caption{Current evidence classes and limits.} \label{tab:evidence} \begin{tabularx}{\textwidth}{>{\bfseries}p{0.23\textwidth}X X} \toprule Class & Stored signal & Limit \\ \midrule Audio & Tone, intonation, shifts, time, articulation & Some audio reviews fail or cannot inspect the excerpt cleanly \\ Video & Setup, posture, bow lane, visible hand/bow evidence & Still frames can hide motion, sound, and transition accuracy \\ Metadata & Dates, duration, channel, view count, practice-candidate labels & Metadata cannot judge playing quality by itself \\ Progress plan & One focus, practice constraint, short session plan & The plan is a next experiment, not an outcome prediction \\ \bottomrule \end{tabularx} \end{table} \section{Discussion} Curtis Media Review is intentionally conservative. It does not rank the player against applicants, infer admissions probability, or convert every video into a claim. Instead, it keeps a running technical record. This makes the system useful even before it becomes musically sophisticated: it can preserve practice volume, surface recurring constraints, identify missing evidence, and make subsequent recordings easier to compare against earlier ones. The main open problem is repertoire identification and evidence quality. YouTube metadata can identify candidate videos, and the current owner-sync path can provide direct audio/video samples, but unclear camera framing, missing score/title context, and short excerpts still prevent reliable piece naming. The backend therefore keeps generic labels out of the piece-title field, stores them as candidate evidence, and marks weak evidence explicitly. A better system would combine reliable excerpt extraction, audio fingerprinting, score/title context, visual posture analysis, and longitudinal trend scoring. The broader experiment is whether a practice system can make musical development legible over time. A useful Curtis system should show whether tone, intonation, rhythm, articulation, shifts, memory, endurance, and audition delivery are improving across dated recordings. It should also preserve the source evidence behind each claim so that the dashboard does not become motivational text detached from the actual playing. \section{Materials and methods} \subsection*{Backend} The backend is a FastAPI service deployed on Railway. It exposes \url{https://curtis.aolabs.io/api/curtis/media-status} for compact state, \url{https://curtis.aolabs.io/api/curtis/ops-check} for fuller operational state, sample indexes for duplicate-window avoidance, and run endpoints for scans, media probing, analysis, and coaching. Persistent runtime state stores sources, inventory, review findings, media samples, analysis runs, and scan history. \subsection*{Source scan} The default public source is \url{https://www.youtube.com/@nalalan}. The scanner queries public video metadata, classifies videos by title, duration, and candidate reasons, and stores inventory records. Dated practice logs and long-form videos are treated as practice candidates. The scan does not itself inspect performance quality. \subsection*{Model review} The model layer reviews sampled media sections where usable excerpts exist. GPT-5.5 is used for text and visual review, and separate audio models are configured for audio review and piece-title verification. Findings are sanitized so weak evidence terms such as obscured, not audible, or no clear evidence produce an unjudged state rather than an unsupported critique. Piece labels are also sanitized: only clear-confidence composer, work, movement, etude/caprice number, opus, key, or catalog evidence can become a piece title, and the proposed title must be corroborated by an independent verifier before it updates the current piece list. Possible labels are treated as unidentified. The owner-browser sync uses the sample index to prioritize uncaptured public YouTube windows rather than re-uploading already reviewed excerpts, and the piece-identification pass can test multiple active moments inside each captured sample before withholding the title. Piece records retain day-level progress entries so the home screen can show today's percent separately from the longer-running piece record when a piece is actually identified. \subsection*{Progress ledger} AO Progress tracks Curtis as part of a larger multi-application state record. The first progress snapshot included Curtis home, Curtis media state, other AO Labs apps, Imagineer state, Imagineer paper, Relay state, and public project surfaces. Curtis contributes dated practice inventory, current focus, review status, and future trend signals. \section{Limitations} The current system cannot predict acceptance at Curtis. It also cannot replace a musician, teacher, jury, studio lesson, or audition process. It can be wrong when the source media are incomplete, when audio is unavailable, when video frames hide the physical action, or when model review fails. The useful claim is therefore limited: Curtis Media Review turns practice media into a persistent, source-bounded record that can support iteration over time. \printbibliography[title={References}] \end{document}