You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is a segmentation starter template from [Banana.dev](https://www.banana.dev) that allows on-demand serverless GPU inference.
6
+
7
+
You can fork this repository and deploy it on Banana as is, or customize it based on your own needs.
8
+
9
+
10
+
# Running this app
11
+
12
+
## Deploying on Banana.dev
13
+
14
+
1.[Fork this](https://github.com/bananaml/demo-segmentation/fork) repository to your own GitHub account.
15
+
2. Connect your GitHub account on Banana.
16
+
3.[Create a new model](https://app.banana.dev/deploy) on Banana from the forked GitHub repository.
17
+
18
+
## Running after deploying
19
+
20
+
1. Wait for the model to build after creating it.
21
+
2. Make an API request using one of the provided snippets in your Banana dashboard. However, instead of sending a prompt as provided in the snippet, adjust the prompt to fit the needs of the segmentation model:
22
+
23
+
```python
24
+
inputs = {
25
+
"audio": "bucket_link_to_wav_file",
26
+
"option": "voice_activity_detection"
27
+
}
28
+
```
29
+
30
+
The `audio` parameter should be substituted with your S3 (or any other provider where you can store .wav files) bucket link that contains the .wav audio file you want to segment. For the `option` parameter, you have to choose between the following options depending on what segmentation information you want to gain from the audio file:
31
+
32
+
* voice_activity_detection
33
+
* overlapped_speech_detection
34
+
* instantaneous_speaker_counting
35
+
* speaker_change_detection
36
+
37
+
In the example above, we chose `voice_activity_detection` as an option.
38
+
39
+
For more info, check out the [Banana.dev docs](https://docs.banana.dev/banana-docs/).
0 commit comments