top of page

AI x Audio

Portfolio of AI Training & Collaboration

Over the past two years, I've integrated AI tools into my professional audio workflow—not just as a user, but as a trainer. I've utilized GPTs for audio analysis, coordinated multi-tool AI workflows for complex creative projects, and trained language models to understand nuanced concepts like perceived loudness, genre-specific mixing conventions, and audio-visual synchronization.

This portfolio showcases three examples of how I work with AI to solve real audio challenges. Each demonstrates the kind of detailed, technical communication and iterative problem-solving that's essential for training next-generation audio AI systems.

Tools Used: Runway, CapCut, ChatGPT, Claude.ai, Audio Analyzer GPT, Reason Studios,

Film Scoring Workflow
🎬 Project 1: "Science is Easy" — AI-Coordinated Film Scoring

Challenge: Score a 16-second AI-generated film clip with tempo automation synced to visual beats, requiring frame-by-frame analysis and precise musical hit points.

 

Approach: I coordinated between multiple AI tools to create a complete scoring workflow:

  • Claude.ai for frame-by-frame visual analysis with timestamps

  • ChatGPT for audio dynamics consultation

  • Audio Analyzer GPT for final mix verification

 

What I provided to the AI:

  • Detailed creative brief (whimsical, quirky, science theme)

  • Technical specifications (G major, 125 BPM base, chord progression)

  • Instrumentation plan (acoustic guitar, glockenspiel, brass hits)

  • 19 video screenshots at 0.5-1 second intervals with timecode

What the AI delivered:

  • Segment-by-segment breakdown with pacing, energy, and tone analysis

  • Specific timestamp recommendations for musical accents (e.g., "thumbs-up at 0:02:155 should align with rhythmic accent")

  • Tempo automation suggestions mapped to visual intensity (129 BPM → 146 BPM → 155 BPM at peak)

  • Instrumentation recommendations for each emotional shift

 

Why this matters for AI training: This demonstrates my ability to structure complex creative problems for AI comprehension, provide detailed technical context, and iterate toward production-ready solutions. The AI learned to connect visual timing with musical structure—exactly the kind of multimodal understanding next-gen models need.

IMG_0012
IMG_0014
IMG_0013
IMG_9989
IMG_9990
IMG_9991
IMG_9992
IMG_9993
Science is easy vo cue sheet_edited.jpg
Mixing & Mastering
🔊 Project 2: Mastering Loudness Troubleshooting

Challenge: My mixes sounded clean and balanced but weren't achieving competitive loudness compared to professional references.

 

Approach: I used ChatGPT as a diagnostic partner, working through a complex technical problem over multiple iterations.

What I provided to the AI:

  • Detailed problem description ("My mix is clean but not loud enough")

  • Comparative context ("Other producers' beats sound crisper, exponentially louder")

  • Technical observations ("No clipping, not too much low/mid/high, but they feel weightless")

  • Reference comparison (reference beat vs. my production)

What the AI delivered:

  • Diagnosis of multiple contributing factors (gain staging, perceived vs. actual loudness, master bus processing)

  • Specific technical solutions with values (-6dB to -3dB pre-mastering peaks, -9 to -6 LUFS targets)

  • Explanations of psychoacoustic concepts (saturation for harmonic richness, transient shaping for punch without volume)

  • Differentiation between overthinking and genuine issues (comparison table format)

 

Why this matters for AI training: This shows I can teach AI to diagnose subjective audio issues objectively, provide context for comparative analysis, and refine understanding through iterative questioning. These conversations create training data for AI to understand real-world mixing challenges.

IMG_0016
IMG_0018
IMG_0017
IMG_0020
IMG_9974
IMG_9975
IMG_9976
IMG_9977
🎛️ Project 3: Genre-Specific Production Consulting

Challenge: Determine appropriate hi-hat panning for a soulful R&B production with live instrumentation, balancing modern trends with organic aesthetics.

 

Approach: I used ChatGPT to understand current industry standards, then refined the AI's recommendations by providing specific genre and instrumentation context.

 

What I provided to the AI:

  • Initial question: "What's the industry standard for hi-hat panning right now?"

  • Genre clarification after initial response: "It's an R&B beat but not like r&b pop it's soul for sure I have guitars and strings and a unison octave piano countermelody electric bass acoustic type sounds"

  • Creative context: Modern bounce with nostalgic vibe

What the AI delivered:

  • Current Industry Trends: Breakdown by genre (Hip-Hop/Trap, Pop/R&B, Boom Bap/Old School)

  • Technical Specifications: Slight off-center (10-15%), alternating/stereo bounce, stereo widening techniques, super narrow/mono for minimalist styles

  • Adapted Recommendations: "Hi-Hat Placement Options for Soul/R&B" with specific percentages and reasoning

  • Context-Aware Guidance: How live instrumentation affects panning decisions ("works especially well with acoustic textures so the hats don't feel artificial")

 

Why this matters for AI training: This demonstrates context-aware technical consulting—teaching AI to understand genre nuance where "soulful R&B with live instrumentation" requires different treatment than trap. Shows my ability to refine AI understanding through clarification and guide it from general knowledge to specific creative application.

IMG_0021
IMG_0022
IMG_0023
IMG_0024
IMG_0025
IMG_0026
IMG_0027
IMG_0028

View my complete music and creative portfolio, featuring original compositions, production work, and additional case studies.

© by CH Mitchell

bottom of page