krishna-k commited on
Commit
eebe4f8
·
verified ·
1 Parent(s): 41b6d5e

port changed for available one

Browse files
Files changed (1) hide show
  1. app.py +52 -40
app.py CHANGED
@@ -1,40 +1,52 @@
1
- # import os
2
-
3
- # from fastrtc import (ReplyOnPause, Stream, get_stt_model, get_tts_model)
4
- # from openai import OpenAI
5
-
6
- # OpenAI.api_key = '5dd0312c-a759-42e8-a21d-6bbe275e7aa2'
7
-
8
- # sambanova_client = OpenAI(
9
- # api_key="5dd0312c-a759-42e8-a21d-6bbe275e7aa2", base_url="https://api.sambanova.ai/v1"
10
- # )
11
- # stt_model = get_stt_model()
12
- # tts_model = get_tts_model()
13
-
14
- # def echo(audio):
15
- # prompt = stt_model.stt(audio)
16
- # response = sambanova_client.chat.completions.create(
17
- # model="Meta-Llama-3.2-3B-Instruct",
18
- # messages=[{"role": "user", "content": prompt}],
19
- # max_tokens=200,
20
- # )
21
- # prompt = response.choices[0].message.content
22
- # for audio_chunk in tts_model.stream_tts_sync(prompt):
23
- # yield audio_chunk
24
-
25
- # stream = Stream(ReplyOnPause(echo), modality="audio", mode="send-receive")
26
- # stream.ui.launch()
27
-
28
- from fastrtc import Stream, ReplyOnPause
29
- import numpy as np
30
- from pydub import AudioSegment
31
- from pydub.utils import which
32
-
33
- AudioSegment.ffmpeg = which("ffmpeg")
34
-
35
- def echo(audio: tuple[int, np.ndarray]) -> tuple[int, np.ndarray]:
36
- yield audio
37
-
38
- stream = Stream(ReplyOnPause(echo), modality="audio", mode="send-receive")
39
-
40
- stream.ui.launch(share=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # import os
2
+
3
+ # from fastrtc import (ReplyOnPause, Stream, get_stt_model, get_tts_model)
4
+ # from openai import OpenAI
5
+
6
+ # OpenAI.api_key = '5dd0312c-a759-42e8-a21d-6bbe275e7aa2'
7
+
8
+ # sambanova_client = OpenAI(
9
+ # api_key="5dd0312c-a759-42e8-a21d-6bbe275e7aa2", base_url="https://api.sambanova.ai/v1"
10
+ # )
11
+ # stt_model = get_stt_model()
12
+ # tts_model = get_tts_model()
13
+
14
+ # def echo(audio):
15
+ # prompt = stt_model.stt(audio)
16
+ # response = sambanova_client.chat.completions.create(
17
+ # model="Meta-Llama-3.2-3B-Instruct",
18
+ # messages=[{"role": "user", "content": prompt}],
19
+ # max_tokens=200,
20
+ # )
21
+ # prompt = response.choices[0].message.content
22
+ # for audio_chunk in tts_model.stream_tts_sync(prompt):
23
+ # yield audio_chunk
24
+
25
+ # stream = Stream(ReplyOnPause(echo), modality="audio", mode="send-receive")
26
+ # stream.ui.launch()
27
+
28
+ from fastrtc import Stream, ReplyOnPause
29
+ import numpy as np
30
+ from pydub import AudioSegment
31
+ from pydub.utils import which
32
+ import socket
33
+
34
+ AudioSegment.ffmpeg = which("ffmpeg")
35
+
36
+
37
+
38
+ def find_free_port():
39
+ s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
40
+ s.bind(('', 0)) # Bind to an available port (0 means any available port)
41
+ port = s.getsockname()[1] # Get the port number
42
+ s.close() # Close the socket
43
+ return port
44
+
45
+ available_port = find_free_port()
46
+
47
+ def echo(audio: tuple[int, np.ndarray]) -> tuple[int, np.ndarray]:
48
+ yield audio
49
+
50
+ stream = Stream(ReplyOnPause(echo), modality="audio", mode="send-receive")
51
+
52
+ stream.ui.launch(share=True,server_port=available_port)