diff --git a/examples/audio_captioning.py b/examples/audio_captioning.py index e1c78f1..b355e2a 100644 --- a/examples/audio_captioning.py +++ b/examples/audio_captioning.py @@ -4,8 +4,7 @@ This is an example using CLAPCAP for audio captioning. from msclap import CLAP # Load and initialize CLAP -weights_path = "weights_path" -clap_model = CLAP(weights_path, version = 'clapcap', use_cuda=False) +clap_model = CLAP(version = 'clapcap', use_cuda=False) #Load audio files audio_files = ['audio_file'] diff --git a/examples/zero_shot_classification.py b/examples/zero_shot_classification.py index e845ce8..1f8dc59 100644 --- a/examples/zero_shot_classification.py +++ b/examples/zero_shot_classification.py @@ -17,8 +17,7 @@ prompt = 'this is the sound of ' y = [prompt + x for x in dataset.classes] # Load and initialize CLAP -weights_path = "weights_path" -clap_model = CLAP(weights_path, version = '2023', use_cuda=False) +clap_model = CLAP(version = '2023', use_cuda=False) # Computing text embeddings text_embeddings = clap_model.get_text_embeddings(y) diff --git a/examples/zero_shot_predictions.py b/examples/zero_shot_predictions.py index 45b53e9..4d64a18 100644 --- a/examples/zero_shot_predictions.py +++ b/examples/zero_shot_predictions.py @@ -15,9 +15,8 @@ class_prompts = [prompt + x for x in classes] audio_files = ['audio_file'] # Load and initialize CLAP -weights_path = "weights_path" # Setting use_cuda = True will load the model on a GPU using CUDA -clap_model = CLAP(weights_path, version = '2023', use_cuda=False) +clap_model = CLAP(version = '2023', use_cuda=False) # compute text embeddings from natural text text_embeddings = clap_model.get_text_embeddings(class_prompts)