| | import torch |
| | from transformers import pipeline |
| | from cosyvoice.cli.cosyvoice import CosyVoice |
| |
|
| | class EndpointHandler: |
| | def __init__(self, model): |
| | self.model = CosyVoice('path/to/model') |
| | self.device = 'cuda' if torch.cuda.is_available() else 'cpu' |
| | self.model.to(self.device) |
| |
|
| | def __call__(self, data): |
| | inputs = data.get('inputs', 'Default text') |
| | parameters = data.get('parameters', {}) |
| | language = parameters.get('language', 'en') |
| | duration = parameters.get('duration', 30) |
| |
|
| | |
| | audio = self.model.generate(inputs, language=language, singing_mode=True) |
| |
|
| | |
| | import io |
| | import base64 |
| | buffer = io.BytesIO() |
| | torch.save(audio, buffer) |
| | audio_b64 = base64.b64encode(buffer.getvalue()).decode() |
| |
|
| | return {'audio': audio_b64} |
| |
|
| | |
| | def query_endpoint(payload): |
| | handler = EndpointHandler(None) |
| | return handler(payload) |