# https://huggingface.co/spaces/asigalov61/Harmonic-Melody-MIDI-Mixer

import os.path

import time as reqtime
import datetime
from pytz import timezone

from itertools import groupby
import copy

import gradio as gr

import random

from midi_to_colab_audio import midi_to_colab_audio
import TMIDIX

import matplotlib.pyplot as plt
    
# =================================================================================================

def pitches_counts(melody_score):

  pitches = [p[4] for p in melody_score]

  pcounts = []

  count = 0
  pp = -1
    
  for p in pitches:
    if p == pp:
      count += 1
      pcounts.append(count)
    else:
      count = 0
      pcounts.append(count)
    pp = p

  return pcounts

# =================================================================================================

def find_similar_song(songs, src_melody):

  src_pcount = pitches_counts(src_melody)

  ratios = []

  for s in songs:
    patch = s[1]

    trg_melody = [e for e in s[3] if e[6] == patch]
    trg_pcount = pitches_counts(trg_melody)

    pcount = 0

    for i, c in enumerate(src_pcount):
      if c == trg_pcount[i]:
        pcount += 1

    ratios.append(pcount / len(src_pcount))

  max_ratio = max(ratios)

  return songs[ratios.index(max_ratio)], max_ratio, ratios.count(max_ratio)

# =================================================================================================

def mix_chord(chord, tones_chord, mel_patch, mel_pitch, next_note_dtime):

  cho = []

  for k, g in groupby(sorted(chord, key=lambda x: x[6]), lambda x: x[6]):

    if k != 128:
      if k == mel_patch:
          
        cg = list(g)
          
        c = copy.deepcopy(cg[0])
          
        if cg[0][2] > next_note_dtime:
            c[2] = next_note_dtime
            
        c[4] = mel_pitch
        c[5] = 105 + (mel_pitch % 12)
          
        cho.append(c)

      else:
        cg = list(g)

        tclen = len(tones_chord)

        if len(cg) > tclen:
          tchord = tones_chord + [random.choice(tones_chord) for _ in range(len(cg)-tclen)]
        
        else:
          tchord = tones_chord

        seen = []

        for i, cc in enumerate(cg):

            if [cc[4], cc[6]] not in seen:
            
                c = copy.deepcopy(cc)
                
                if cc[2] > next_note_dtime:
                    c[2] = next_note_dtime
                  
                c[4] = ((c[4] // 12) * 12) + tchord[i]
                c[5] += c[4] % 12
                
                cho.append(c)

                seen.append([cc[4], cc[6]])

    else:
      cho.extend(list(g))

  return cho

# =================================================================================================

def Mix_Melody(input_midi, 
               input_find_best_match, 
               input_adjust_melody_notes_durations, 
               input_adjust_accompaniment_notes_durations,
               input_output_as_solo_piano,
               input_remove_drums,
               input_output_tempo,
               input_transform,
               input_transpose_to_C4,
               input_transpose_value
             ):
    
    print('=' * 70)
    print('Req start time: {:%Y-%m-%d %H:%M:%S}'.format(datetime.datetime.now(PDT)))
    start_time = reqtime.time()

    print('=' * 70)

    fn = os.path.basename(input_midi.name)
    fn1 = fn.split('.')[0]

    print('-' * 70)
    print('Requested settings:')
    print('-' * 70)
    print('Input file name:', fn)
    print('Find best matches', input_find_best_match)
    print('Adjust melody notes durations:', input_adjust_melody_notes_durations)
    print('Adjust accompaniment notes durations:', input_adjust_accompaniment_notes_durations)
    print('Output as Solo Piano:', input_output_as_solo_piano)
    print('Remove drums:', input_remove_drums)
    print('Output tempo:', input_output_tempo)
    print('Transform:', input_transform)
    print('Transpose to C4:', input_transpose_to_C4)
    print('Transpose value:', input_transpose_value)
    print('-' * 70)

    #===============================================================================
    raw_score = TMIDIX.midi2single_track_ms_score(input_midi.name)
    
    #===============================================================================
    # Enhanced score notes
    
    raw_escore = TMIDIX.advanced_score_processor(raw_score, return_enhanced_score_notes=True)[0]
    
    if len(raw_escore) > 0:
    
        #===============================================================================
        # Augmented enhanced score notes
        
        src_escore = TMIDIX.recalculate_score_timings(TMIDIX.augment_enhanced_score_notes([e for e in raw_escore if e[6] < 80]))

        src_cscore = TMIDIX.chordify_score([1000, src_escore])
        
        src_melody = [c[0] for c in src_cscore][:256]

        if input_transform == 'Flip Melody':
            src_melody = TMIDIX.flip_enhanced_score_notes(src_melody)
            
        elif input_transform == 'Reverse Melody':
            src_melody = TMIDIX.reverse_enhanced_score_notes(src_melody)

        mel_avg_time = TMIDIX.escore_notes_averages(src_melody)[0]
        
        src_melody_pitches = [p[4] for p in src_melody]
        
        src_harm_tones_chords = TMIDIX.harmonize_enhanced_melody_score_notes(src_melody)
        
        #===============================================================================
        
        matched_songs = [a for a in all_songs if a[2] == max(32, len(src_melody))]
        
        random.shuffle(matched_songs)

        max_match_ratio = -1
        max_match_ratios_count = len(matched_songs)

        if input_find_best_match:
            new_song, max_match_ratio, max_match_ratios_count = find_similar_song(matched_songs, src_melody)
        else:
            new_song = random.choice(matched_songs)
        
        print('Selected Monster Mono Melodies MIDI:', new_song[0])
        print('Selected melody match ratio:', max_match_ratio)
        print('Selected melody instrument:', TMIDIX.Number2patch[new_song[1]], '(', new_song[1], ')')
        print('Melody notes count:', new_song[2])
        print('Matched melodies pool count', max_match_ratios_count)
        
        MIDI_Summary = 'Selected Monster Mono Melodies MIDI: ' + str(new_song[0]) + '\n'
        MIDI_Summary += 'Selected melody match ratio: ' + str(max_match_ratio) + '\n'
        MIDI_Summary += 'Selected melody instrument: ' + str(TMIDIX.Number2patch[new_song[1]]) + ' (' + str(new_song[1]) + ')' + '\n'
        MIDI_Summary += 'Melody notes count: ' + str(new_song[2]) + '\n'
        MIDI_Summary += 'Matched melodies pool count: ' + str(max_match_ratios_count)

        fn1 += '_' + str(new_song[0]) + '_' + str(TMIDIX.Number2patch[new_song[1]]) + '_' + str(new_song[1]) + '_' + str(new_song[2])
        
        trg_patch = new_song[1]
        
        trg_song = copy.deepcopy(new_song[3])
        
        mix_avg_time = TMIDIX.escore_notes_averages(trg_song)[0]
        mix_mel_avg_time = TMIDIX.escore_notes_averages([e for e in trg_song if e[6] == trg_patch])[0]
        
        TMIDIX.adjust_score_velocities(trg_song, 95)
        
        cscore = TMIDIX.chordify_score([1000, trg_song])
        
        print('=' * 70)
        print('Done loading source and target MIDIs...!')
        print('=' * 70)
        print('Mixing...')

        mixed_song = []
        
        midx = 0
        next_note_dtime = 255
        
        for i, c in enumerate(cscore):
            cho = copy.deepcopy(c)
            
            patches = sorted(set([e[6] for e in c]))
            
            if trg_patch in patches:

                if input_adjust_melody_notes_durations:
                    if midx < len(src_melody)-1:
                        next_note_dtime = src_melody[midx+1][1] - src_melody[midx][1]
                    else:
                        next_note_dtime = 255
                    
                mixed_song.extend(mix_chord(c, src_harm_tones_chords[midx], trg_patch, src_melody_pitches[midx], next_note_dtime))
                
                midx += 1
            
            else:
                if input_adjust_accompaniment_notes_durations:
                    if i < len(cscore)-1:
                        next_note_dtime = cscore[i+1][0][1] - cscore[i][0][1]
                    else:
                        next_note_dtime = 255
                
                mixed_song.extend(mix_chord(cho, src_harm_tones_chords[midx], trg_patch, src_melody_pitches[midx], next_note_dtime))

            if midx == len(src_melody):
                break      

        print('=' * 70)
        print('Done!')
        print('=' * 70)

        #===============================================================================

        if input_output_as_solo_piano:
            
            csong = TMIDIX.chordify_score([1000, mixed_song])

            mixed_song = []

            for c in csong:
                
                pitches = [e[4] for e in c if e[6] == trg_patch]
                
                for cc in c:
                    
                    ccc = copy.deepcopy(cc)
                    
                    if cc[3] != 9: 
                        if cc[6] == trg_patch:
                            ccc[3] = 3
                            ccc[6] = 0
                            mixed_song.append(ccc)
                            
                        else:
                            if cc[4] not in pitches:
                                ccc[3] = 0
                                ccc[6] = 0
                                mixed_song.append(ccc)
                                pitches.append(cc[4])
            
                    else:
                        mixed_song.append(ccc)    
                    
        if input_remove_drums:
            mixed_song = [e for e in mixed_song if e[3] != 9]

        if input_output_tempo == 'Mix':
    
            time_k = mel_avg_time / mix_avg_time
    
            mixed_song = TMIDIX.adjust_escore_notes_timings(mixed_song, time_k)

        elif input_output_tempo == 'Source Melody':
            
            time_k = mel_avg_time / mix_mel_avg_time
    
            mixed_song = TMIDIX.adjust_escore_notes_timings(mixed_song, time_k)

        if input_transform == 'Flip Mix':
            mixed_song = TMIDIX.flip_enhanced_score_notes(mixed_song)
            
        elif input_transform == 'Reverse Mix':
            mixed_song = TMIDIX.reverse_enhanced_score_notes(mixed_song)        

        if input_transpose_value != 0:
            mixed_song = TMIDIX.transpose_escore_notes(mixed_song, input_transpose_value)

        if input_transpose_to_C4:
            mixed_song = TMIDIX.transpose_escore_notes_to_pitch(mixed_song)
            
        #===============================================================================
        print('Rendering results...')
        
        print('=' * 70)
        print('Sample INTs', mixed_song[:5])
        print('=' * 70)
        
        output_score, patches, overflow_patches = TMIDIX.patch_enhanced_score_notes(mixed_song)

        detailed_stats = TMIDIX.Tegridy_ms_SONG_to_MIDI_Converter(output_score,
                                                                  output_signature = 'Harmonic Melody MIDI Mixer',
                                                                  output_file_name = fn1,
                                                                  track_name='Project Los Angeles',
                                                                  list_of_MIDI_patches=patches,
                                                                  timings_multiplier=16
                                                                  )
        
        new_fn = fn1+'.mid'
                
        
        audio = midi_to_colab_audio(new_fn, 
                            soundfont_path=soundfont,
                            sample_rate=16000,
                            volume_scale=10,
                            output_for_gradio=True
                            )
        
        print('Done!')
        print('=' * 70)
    
        #========================================================
    
        output_midi_title = str(fn1)
        output_midi_summary = str(MIDI_Summary)
        output_midi = str(new_fn)
        output_audio = (16000, audio)

        for o in output_score:
            o[1] *= 16
            o[2] *= 16
        
        output_plot = TMIDIX.plot_ms_SONG(output_score, plot_title=output_midi_title, return_plt=True)
    
        print('Output MIDI file name:', output_midi)
        print('Output MIDI title:', output_midi_title)
        print('Output MIDI summary:', MIDI_Summary)
        print('=' * 70) 
        
    
        #========================================================
        
        print('-' * 70)
        print('Req end time: {:%Y-%m-%d %H:%M:%S}'.format(datetime.datetime.now(PDT)))
        print('-' * 70)
        print('Req execution time:', (reqtime.time() - start_time), 'sec')
    
        return output_midi_title, output_midi_summary, output_midi, output_audio, output_plot

# =================================================================================================

if __name__ == "__main__":
    
    PDT = timezone('US/Pacific')
    
    print('=' * 70)
    print('App start time: {:%Y-%m-%d %H:%M:%S}'.format(datetime.datetime.now(PDT)))
    print('=' * 70)

    soundfont = "SGM-v2.01-YamahaGrand-Guit-Bass-v2.7.sf2"
    
    all_songs = TMIDIX.Tegridy_Any_Pickle_File_Reader('Monster_Mono_Melodies_MIDI_Dataset_65536_32_256')
    print('=' * 70)
    
    app = gr.Blocks()
    with app:
        gr.Markdown("<h1 style='text-align: center; margin-bottom: 1rem'>Harmonic Melody MIDI Mixer</h1>")
        gr.Markdown("<h1 style='text-align: center; margin-bottom: 1rem'>Harmonize and mix any MIDI melody</h1>")
        gr.Markdown(
            "![Visitors](https://api.visitorbadge.io/api/visitors?path=asigalov61.Harmonic-Melody-MIDI-Mixer&style=flat)\n\n"
            "This is a demo for TMIDIX Python module from tegridy-tools and Monster Mono Melodies MIDI Dataset\n\n"
            "Check out [tegridy-tools](https://github.com/asigalov61/tegridy-tools) on GitHub!\n\n"
            "Check out [Monster-MIDI-Dataset](https://github.com/asigalov61/Monster-MIDI-Dataset) on GitHub!\n\n"
        )
        gr.Markdown("## Upload your MIDI or select a sample example MIDI below")
        
        input_midi = gr.File(label="Input MIDI", file_types=[".midi", ".mid", ".kar"])

        gr.Markdown("## Select mixing options")
        
        input_find_best_match = gr.Checkbox(label="Find best match", value=False)
        input_adjust_melody_notes_durations = gr.Checkbox(label="Adjust melody notes durations", value=False)
        input_adjust_accompaniment_notes_durations = gr.Checkbox(label="Adjust accompaniment notes durations", value=False)
        input_output_as_solo_piano = gr.Checkbox(label="Output as Solo Piano", value=False)
        input_remove_drums = gr.Checkbox(label="Remove drums from output", value=False)
        input_output_tempo = gr.Radio(["Mix Melody", "Source Melody", "Mix"], value="Mix Melody", label="Output tempo")
        input_transform = gr.Radio(["As-is", "Flip Melody", "Reverse Melody", "Flip Mix", "Reverse Mix"], value="As-is", label="Transform")        
        input_transpose_value = gr.Slider(-12, 12, value=0, step=1, label="Transpose value")
        input_transpose_to_C4 = gr.Checkbox(label="Transpose to C4", value=False)
         
        run_btn = gr.Button("mix melody", variant="primary")

        gr.Markdown("## Output results")

        output_midi_title = gr.Textbox(label="Output MIDI title")
        output_midi_summary = gr.Textbox(label="Output MIDI summary")
        output_audio = gr.Audio(label="Output MIDI audio", format="mp3", elem_id="midi_audio")
        output_plot = gr.Plot(label="Output MIDI score plot")
        output_midi = gr.File(label="Output MIDI file", file_types=[".mid"])


        run_event = run_btn.click(Mix_Melody, [input_midi, 
                                               input_find_best_match, 
                                               input_adjust_melody_notes_durations, 
                                               input_adjust_accompaniment_notes_durations,
                                               input_output_as_solo_piano,
                                               input_remove_drums,
                                               input_output_tempo,
                                               input_transform,
                                               input_transpose_to_C4,
                                               input_transpose_value
                                              ],
                                            [output_midi_title, output_midi_summary, output_midi, output_audio, output_plot])

        gr.Examples(
            [["Abracadabra-Sample-Melody.mid", True, True, True, False, False, "Mix Melody", "As-is", False, 0], 
             ["Sparks-Fly-Sample-Melody.mid", True, True, True, False, False, "Mix Melody", "As-is", False, 0],
            ],
            [input_midi, 
             input_find_best_match, 
             input_adjust_melody_notes_durations, 
             input_adjust_accompaniment_notes_durations,
             input_output_as_solo_piano,
             input_remove_drums,
             input_output_tempo,
             input_transform,
             input_transpose_to_C4,
             input_transpose_value
            ],
            [output_midi_title, output_midi_summary, output_midi, output_audio, output_plot],
            Mix_Melody,
            cache_examples=True,
        )
        
        app.queue().launch()