/* * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version 2 * of the License, or (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. */ /** \file * \ingroup bpygpu * * - Use ``bpygpu_`` for local API. * - Use ``BPyGPU`` for public API. */ #include #include "GPU_vertex_buffer.h" #include "BLI_math.h" #include "MEM_guardedalloc.h" #include "../generic/py_capi_utils.h" #include "../generic/python_utildefines.h" #include "gpu_py_vertex_buffer.h" /* own include */ #include "gpu_py_vertex_format.h" /* -------------------------------------------------------------------- */ /** \name Utility Functions * \{ */ #define PY_AS_NATIVE_SWITCH(attr) \ switch (attr->comp_type) { \ case GPU_COMP_I8: { \ PY_AS_NATIVE(int8_t, PyC_Long_AsI8); \ break; \ } \ case GPU_COMP_U8: { \ PY_AS_NATIVE(uint8_t, PyC_Long_AsU8); \ break; \ } \ case GPU_COMP_I16: { \ PY_AS_NATIVE(int16_t, PyC_Long_AsI16); \ break; \ } \ case GPU_COMP_U16: { \ PY_AS_NATIVE(uint16_t, PyC_Long_AsU16); \ break; \ } \ case GPU_COMP_I32: { \ PY_AS_NATIVE(int32_t, PyC_Long_AsI32); \ break; \ } \ case GPU_COMP_U32: { \ PY_AS_NATIVE(uint32_t, PyC_Long_AsU32); \ break; \ } \ case GPU_COMP_F32: { \ PY_AS_NATIVE(float, PyFloat_AsDouble); \ break; \ } \ default: \ BLI_assert(0); \ } \ ((void)0) /* No error checking, callers must run PyErr_Occurred */ static void fill_format_elem(void *data_dst_void, PyObject *py_src, const GPUVertAttr *attr) { #define PY_AS_NATIVE(ty_dst, py_as_native) \ { \ ty_dst *data_dst = data_dst_void; \ *data_dst = py_as_native(py_src); \ } \ ((void)0) PY_AS_NATIVE_SWITCH(attr); #undef PY_AS_NATIVE } /* No error checking, callers must run PyErr_Occurred */ static void fill_format_sequence(void *data_dst_void, PyObject *py_seq_fast, const GPUVertAttr *attr) { const uint len = attr->comp_len; PyObject **value_fast_items = PySequence_Fast_ITEMS(py_seq_fast); /** * Args are constants, so range checks will be optimized out if they're nop's. */ #define PY_AS_NATIVE(ty_dst, py_as_native) \ ty_dst *data_dst = data_dst_void; \ for (uint i = 0; i < len; i++) { \ data_dst[i] = py_as_native(value_fast_items[i]); \ } \ ((void)0) PY_AS_NATIVE_SWITCH(attr); #undef PY_AS_NATIVE } #undef PY_AS_NATIVE_SWITCH #undef WARN_TYPE_LIMIT_PUSH #undef WARN_TYPE_LIMIT_POP static bool bpygpu_vertbuf_fill_impl(GPUVertBuf *vbo, uint data_id, PyObject *seq, const char *error_prefix) { const char *exc_str_size_mismatch = "Expected a %s of size %d, got %u"; bool ok = true; const GPUVertAttr *attr = &vbo->format.attrs[data_id]; if (PyObject_CheckBuffer(seq)) { Py_buffer pybuffer; if (PyObject_GetBuffer(seq, &pybuffer, PyBUF_STRIDES | PyBUF_ND) == -1) { /* PyObject_GetBuffer raise a PyExc_BufferError */ return false; } uint comp_len = pybuffer.ndim == 1 ? 1 : (uint)pybuffer.shape[1]; if (pybuffer.shape[0] != vbo->vertex_len) { PyErr_Format( PyExc_ValueError, exc_str_size_mismatch, "sequence", vbo->vertex_len, pybuffer.shape[0]); ok = false; } else if (comp_len != attr->comp_len) { PyErr_Format(PyExc_ValueError, exc_str_size_mismatch, "component", attr->comp_len, comp_len); ok = false; } else { GPU_vertbuf_attr_fill_stride(vbo, data_id, pybuffer.strides[0], pybuffer.buf); } PyBuffer_Release(&pybuffer); } else { GPUVertBufRaw data_step; GPU_vertbuf_attr_get_raw_data(vbo, data_id, &data_step); PyObject *seq_fast = PySequence_Fast(seq, "Vertex buffer fill"); if (seq_fast == NULL) { return false; } const uint seq_len = PySequence_Fast_GET_SIZE(seq_fast); if (seq_len != vbo->vertex_len) { PyErr_Format(PyExc_ValueError, exc_str_size_mismatch, "sequence", vbo->vertex_len, seq_len); } PyObject **seq_items = PySequence_Fast_ITEMS(seq_fast); if (attr->comp_len == 1) { for (uint i = 0; i < seq_len; i++) { uchar *data = (uchar *)GPU_vertbuf_raw_step(&data_step); PyObject *item = seq_items[i]; fill_format_elem(data, item, attr); } } else { for (uint i = 0; i < seq_len; i++) { uchar *data = (uchar *)GPU_vertbuf_raw_step(&data_step); PyObject *seq_fast_item = PySequence_Fast(seq_items[i], error_prefix); if (seq_fast_item == NULL) { ok = false; goto finally; } if (PySequence_Fast_GET_SIZE(seq_fast_item) != attr->comp_len) { PyErr_Format(PyExc_ValueError, exc_str_size_mismatch, "sequence", attr->comp_len, PySequence_Fast_GET_SIZE(seq_fast_item)); ok = false; Py_DECREF(seq_fast_item); goto finally; } /* May trigger error, check below */ fill_format_sequence(data, seq_fast_item, attr); Py_DECREF(seq_fast_item); } } if (PyErr_Occurred()) { ok = false; } finally: Py_DECREF(seq_fast); } return ok; } static int bpygpu_attr_fill(GPUVertBuf *buf, int id, PyObject *py_seq_data, const char *error_prefix) { if (id < 0 || id >= buf->format.attr_len) { PyErr_Format(PyExc_ValueError, "Format id %d out of range", id); return 0; } if (buf->data == NULL) { PyErr_SetString(PyExc_ValueError, "Can't fill, static buffer already in use"); return 0; } if (!bpygpu_vertbuf_fill_impl(buf, (uint)id, py_seq_data, error_prefix)) { return 0; } return 1; } /** \} */ /* -------------------------------------------------------------------- */ /** \name VertBuf Type * \{ */ static PyObject *bpygpu_VertBuf_new(PyTypeObject *UNUSED(type), PyObject *args, PyObject *kwds) { struct { PyObject *py_fmt; uint len; } params; static const char *_keywords[] = {"format", "len", NULL}; static _PyArg_Parser _parser = {"O!I:GPUVertBuf.__new__", _keywords, 0}; if (!_PyArg_ParseTupleAndKeywordsFast( args, kwds, &_parser, &BPyGPUVertFormat_Type, ¶ms.py_fmt, ¶ms.len)) { return NULL; } const GPUVertFormat *fmt = &((BPyGPUVertFormat *)params.py_fmt)->fmt; GPUVertBuf *vbo = GPU_vertbuf_create_with_format(fmt); GPU_vertbuf_data_alloc(vbo, params.len); return BPyGPUVertBuf_CreatePyObject(vbo); } PyDoc_STRVAR(bpygpu_VertBuf_attr_fill_doc, ".. method:: attr_fill(id, data)\n" "\n" " Insert data into the buffer for a single attribute.\n" "\n" " :param id: Either the name or the id of the attribute.\n" " :type id: int or str\n" " :param data: Sequence of data that should be stored in the buffer\n" " :type data: sequence of values or tuples\n"); static PyObject *bpygpu_VertBuf_attr_fill(BPyGPUVertBuf *self, PyObject *args, PyObject *kwds) { PyObject *data; PyObject *identifier; static const char *_keywords[] = {"id", "data", NULL}; static _PyArg_Parser _parser = {"OO:attr_fill", _keywords, 0}; if (!_PyArg_ParseTupleAndKeywordsFast(args, kwds, &_parser, &identifier, &data)) { return NULL; } int id; if (PyLong_Check(identifier)) { id = PyLong_AsLong(identifier); } else if (PyUnicode_Check(identifier)) { const char *name = PyUnicode_AsUTF8(identifier); id = GPU_vertformat_attr_id_get(&self->buf->format, name); if (id == -1) { PyErr_SetString(PyExc_ValueError, "Unknown attribute name"); return NULL; } } else { PyErr_SetString(PyExc_TypeError, "expected int or str type as identifier"); return NULL; } if (!bpygpu_attr_fill(self->buf, id, data, "GPUVertBuf.attr_fill")) { return NULL; } Py_RETURN_NONE; } static struct PyMethodDef bpygpu_VertBuf_methods[] = { {"attr_fill", (PyCFunction)bpygpu_VertBuf_attr_fill, METH_VARARGS | METH_KEYWORDS, bpygpu_VertBuf_attr_fill_doc}, {NULL, NULL, 0, NULL}, }; static void bpygpu_VertBuf_dealloc(BPyGPUVertBuf *self) { GPU_vertbuf_discard(self->buf); Py_TYPE(self)->tp_free(self); } PyDoc_STRVAR(py_gpu_vertex_buffer_doc, ".. class:: GPUVertBuf(len, format)\n" "\n" " Contains a VBO.\n" "\n" " :param len: Amount of vertices that will fit into this buffer.\n" " :type type: `int`\n" " :param format: Vertex format.\n" " :type buf: :class:`gpu.types.GPUVertFormat`\n"); PyTypeObject BPyGPUVertBuf_Type = { PyVarObject_HEAD_INIT(NULL, 0).tp_name = "GPUVertBuf", .tp_basicsize = sizeof(BPyGPUVertBuf), .tp_dealloc = (destructor)bpygpu_VertBuf_dealloc, .tp_flags = Py_TPFLAGS_DEFAULT, .tp_doc = py_gpu_vertex_buffer_doc, .tp_methods = bpygpu_VertBuf_methods, .tp_new = bpygpu_VertBuf_new, }; /** \} */ /* -------------------------------------------------------------------- */ /** \name Public API * \{ */ PyObject *BPyGPUVertBuf_CreatePyObject(GPUVertBuf *buf) { BPyGPUVertBuf *self; self = PyObject_New(BPyGPUVertBuf, &BPyGPUVertBuf_Type); self->buf = buf; return (PyObject *)self; } /** \} */