| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | #include <torch/extension.h> |
| |
|
| | #include "gmm.h" |
| |
|
| | py::tuple init() { |
| | torch::Tensor gmm_tensor = |
| | torch::zeros({GMM_COUNT, GMM_COMPONENT_COUNT}, torch::dtype(torch::kFloat32).device(torch::kCUDA)); |
| | torch::Tensor scratch_tensor = torch::empty({1}, torch::dtype(torch::kFloat32).device(torch::kCUDA)); |
| | return py::make_tuple(gmm_tensor, scratch_tensor); |
| | } |
| |
|
| | void learn( |
| | torch::Tensor gmm_tensor, |
| | torch::Tensor scratch_tensor, |
| | torch::Tensor input_tensor, |
| | torch::Tensor label_tensor) { |
| | c10::DeviceType device_type = input_tensor.device().type(); |
| |
|
| | unsigned int batch_count = input_tensor.size(0); |
| | unsigned int element_count = input_tensor.stride(1); |
| |
|
| | unsigned int scratch_size = |
| | batch_count * (element_count + GMM_COMPONENT_COUNT * GMM_COUNT * (element_count / (32 * 32))); |
| |
|
| | if (scratch_tensor.size(0) < scratch_size) { |
| | scratch_tensor.resize_({scratch_size}); |
| | } |
| |
|
| | float* gmm = gmm_tensor.data_ptr<float>(); |
| | float* scratch = scratch_tensor.data_ptr<float>(); |
| | float* input = input_tensor.data_ptr<float>(); |
| | int* labels = label_tensor.data_ptr<int>(); |
| |
|
| | if (device_type == torch::kCUDA) { |
| | learn_cuda(input, labels, gmm, scratch, batch_count, element_count); |
| | } else { |
| | learn_cpu(input, labels, gmm, scratch, batch_count, element_count); |
| | } |
| | } |
| |
|
| | torch::Tensor apply(torch::Tensor gmm_tensor, torch::Tensor input_tensor) { |
| | c10::DeviceType device_type = input_tensor.device().type(); |
| |
|
| | unsigned int dim = input_tensor.dim(); |
| | unsigned int batch_count = input_tensor.size(0); |
| | unsigned int element_count = input_tensor.stride(1); |
| |
|
| | auto output_size = input_tensor.sizes().vec(); |
| | output_size[1] = MIXTURE_COUNT; |
| | torch::Tensor output_tensor = |
| | torch::empty(c10::IntArrayRef(output_size), torch::dtype(torch::kFloat32).device(device_type)); |
| |
|
| | const float* gmm = gmm_tensor.data_ptr<float>(); |
| | const float* input = input_tensor.data_ptr<float>(); |
| | float* output = output_tensor.data_ptr<float>(); |
| |
|
| | if (device_type == torch::kCUDA) { |
| | apply_cuda(gmm, input, output, batch_count, element_count); |
| | } else { |
| | apply_cpu(gmm, input, output, batch_count, element_count); |
| | } |
| |
|
| | return output_tensor; |
| | } |
| |
|
| | PYBIND11_MODULE(TORCH_EXTENSION_NAME, m) { |
| | m.def("init", torch::wrap_pybind_function(init)); |
| | m.def("learn", torch::wrap_pybind_function(learn)); |
| | m.def("apply", torch::wrap_pybind_function(apply)); |
| | } |
| |
|