Saving three weeks of manual scheduling through AI-assisted optimization

Lyriq schedules more than 300 students across multiple teachers, instruments, locations, and family constraints. I designed a hybrid AI-assisted system that generates an initial schedule, scores placement quality, and allows administrators to review, adjust, and finalize the result.

Students couldn’t simply choose a class - they could only submit preferences.

Lyriq operates more like a university than a typical music school. Students don’t enroll directly into lesson slots. Instead they submit preferences: instrument, teacher, location, and availability.

Administrators then build the actual schedule from those preferences while balancing teacher availability, room locations, and cases where siblings need to be scheduled together.

With more than 300 students and dozens of overlapping constraints, assembling that schedule manually could take weeks.

Scheduling across overlapping constraints

Lyriq students could only submit preferences. The system had to turn those preferences into an actual schedule.

Student Preferences
  • Instrument
  • Teacher preference
  • Location preference
  • Availability windows
  • Sibling coordination
Operational Constraints
  • Teacher availability
  • Instruments taught
  • Location capacity
  • Time-slot conflicts
  • Final admin approval
AI-Assisted Scheduling Engine

Matches preferences against real-world constraints

Workable lesson schedule

AI-generated first pass, reviewed and finalized by administrators

Constraint system balancing student preferences, teacher availability, locations, and sibling coordination.

AI-generated schedule showing placement confidence and student conflicts

Automating the schedule was not the hard part. Trusting it was.

An algorithm can generate a schedule quickly, but administrators still need to understand the decisions it makes. A system that simply outputs a final schedule creates a new problem: no visibility into why students were placed where they were.

Instead of fully automating the process, I designed a hybrid workflow. The AI generates the initial schedule, evaluates placement quality, and highlights compromises.

Administrators then review the results, adjust edge cases, and approve placements before schedules are sent to teachers.

AI generates the schedule. Humans finalize it.

The system generates the first pass automatically, but administrators review and adjust placements before schedules are confirmed.

Admin starts scheduling
Run AI Scheduler
AI generates placements
Students appear on calendar
Confidence scoring applied
Green / Yellow / Red placements
Admin reviews placements
Locks or adjusts students
AI re-runs optimization
Improves unresolved conflicts

AI-generated scheduling followed by review, locking, and iterative re-optimization.

If the output isn’t scannable, the system fails.

The AI could place hundreds of students at once. Administrators needed to understand the quality of those placements immediately.

Instead of coloring cards by teacher or instrument, I used confidence-based color coding: green for strong matches, yellow for acceptable placements, and red for compromises.

That made the schedule readable at a glance and directed attention exactly where human judgment was needed.

AI confidence scoring makes scheduling scannable

Instead of hiding scheduling compromises, the system surfaces them visually so administrators can focus attention where it’s needed.

All constraints satisfied
Minor compromise
Admin review required

Good automation preserves the decisions humans want to keep.

Most of the schedule often worked well. The problem was the handful of compromised placements.

I introduced lock controls so administrators could preserve strong placements before re-running optimization on the rest.

That let the system improve problem areas without undoing the progress already made.

Admin locks strong placements
AI re-optimizes remaining students

Locked placements remain unchanged while the AI improves unresolved conflicts.

The result was not blind automation. It was guided optimization.

The AI handled the heavy scheduling logic across hundreds of constraints. Administrators retained control over edge cases, conflicts, and final approval..

The system turned scheduling from a fragile manual task into a faster, iterative collaboration between automation and judgment.

AI-optimized schedule after administrator review and locked placements.

What I would refine further

The next step would be deeper teacher comparison tools and availability overlays, allowing administrators to evaluate alternatives before committing to a compromise placement.

The core workflow already established the right balance. The next iteration would make that decision-making layer even more transparent.

See more of my work