.. _sdkruntime-api-reference: SdkRuntime API Reference ======================== This section presents the ``SdkRuntime`` Python host API reference and associated utilities to develop kernels for the Cerebras Wafer Scale Engine. SdkRuntime ---------- .. py:module:: cerebras.sdk.runtime.sdkruntimepybind Python API for ``SdkRuntime`` functions. .. py:class:: SdkRuntime(bindir: Union[pathlib.Path, str], **kwargs) :module: cerebras.sdk.runtime.sdkruntimepybind Bases: :class:`object` Manages the execution of SDK programs on the Cerebras Wafer Scale Engine (WSE) or simfabric. The constructor analyzes the WSE ELFs in the ``bindir`` and prepares the WSE or simfabric for a run. Requires CM IP address and port for WSE runs. :param bindir: Path to ELF files which is compiled by ``cslc``. The runtime collects the I/O and fabric parameters automatically, including height, width, number of channels, width of buffers,... etc. :type bindir: ``Union[pathlib.Path, str]`` :Keyword Arguments: * **cmaddr** (``str``) -- ``'IP_ADDRESS:PORT'`` string of CM. Omit this ``kwarg`` to run on simfabric. * **suppress_simfab_trace** (``bool``) -- If ``True``, suppresses generation of ``simfab_traces`` when running. Default value is ``False``, i.e., ``simfab_traces`` are produced. * **simfab_numthreads** (``int``) -- Number of threads to use if running on simfabric. Maximum value is ``64``. Default value is ``5``, i.e., the simulator uses 5 threads. * **msg_level** (``str``) -- Message logging output level. Available output levels are ``DEBUG``, ``INFO``, ``WARNING``, and ``ERROR``. Default value is ``WARNING``. **Example**: In the following example, an ``SdkRuntime`` runner object is instantiated. If ``args.cmaddr`` is non-empty, then the kernel code will run on the WSE pointed to by that address; otherwise, the kernel code will run on simfabric. The compiled kernel code in the directory ``args.name`` has exported symbols ``A`` and ``B`` pointing to arrays on the device. After loading the code and starting the run with ``load()`` and ``run()``, data on the host stored in ``data`` is copied to ``A`` on the device, and then ``B`` on the device is copied back into ``data`` on the host. .. code-block:: python runner = SdkRuntime(args.name, cmaddr=args.cmaddr) symbol_A = runner.get_id("A") symbol_B = runner.get_id("B") runner.load() runner.run() runner.memcpy_h2d(symbol_A, data, px, py, w, h, l, streaming=False, data_type=memcpy_dtype, order=memcpy_order, nonblock=False) runner.memcpy_d2h(data, symbol_B, px, py, w, h, l, streaming=False, data_type=memcpy_dtype, order=memcpy_order, nonblock=False) .. py:method:: coord_logical_to_physical(logical_coords: (int, int)) :module: cerebras.sdk.runtime.sdkruntimepybind Convert a logical coordinate to a physical coordinate. For a program with fabric offsets (``offset_x``, ``offset_y``), and program rectangle coordinate (``x``, ``y``), this function returns (``offset_x + x``, ``offset_y + y``). :param logical_coords: Tuple containing logical coordinates. :type symbol: ``(int, int)`` :returns: **physical_coords** (``(int, int)``) -- Tuple containing physical coordinates. .. py:method:: dump_core(corefile: str) :module: cerebras.sdk.runtime.sdkruntimepybind Dump the core of a simulator run, to be used for debugging with ``csdb``. Note that the specified name of the corefile MUST be "corefile.cs1" to use with ``csdb``, and this method can only be called after calling ``stop()``. :param corefile: Name of corefile. Must be "corefile.cs1" to use with ``csdb``. :type symbol: ``str`` .. py:method:: get_id(symbol: str) :module: cerebras.sdk.runtime.sdkruntimepybind Retrieve the integer representation of an exported symbol which is exported in the kernel. Possible symbols include a data tensor or a host-callable function. :param symbol: The exported name of the symbol. :type symbol: ``str`` .. py:method:: is_task_done(task_handle: Task) -> bool :module: cerebras.sdk.runtime.sdkruntimepybind Query if task ``task_handle`` is complete :param task_handle: Handle to a task previously launched by ``SdkRuntime``. :type task_handle: ``Task`` :returns: **task_done** (``bool``) -- ``True`` if task is done, and ``False`` otherwise. .. py:method:: launch(symbol: str, *args, **kwargs) -> Task :module: cerebras.sdk.runtime.sdkruntimepybind Trigger a host-callable function defined in the kernel, with type checking for arguments. :param symbol: The exported name of the symbol corresponding to a host-callable function. :type symbol: ``str`` :Positional Arguments: Matches the arguments of the host-callable function. ``launch`` will perform type checking on the arguments. :Keyword Arguments: **nonblock** (``bool``) -- Nonblocking if ``True``, blocking otherwise. :returns: **task_handle** (``Task``) -- Handle to the task launched by ``launch``. **Example**: Consider a kernel which defines a host-callable function ``fn_foo`` by: .. code-block:: csl comptime { @export_symbol(fn_foo); } The host calls ``fn_foo`` by ``runner.launch("fn_foo", nonblock=False)``. .. py:method:: load() :module: cerebras.sdk.runtime.sdkruntimepybind Load the binaries to simfabric or WSE. It may takes 80+ seconds to load the binaries onto the WSE. .. py:method:: memcpy_d2h(dest: numpy.ndarray, src: int, px: int, py: int, w: int, h: int, elem_per_pe: int, **kwargs) -> Task :module: cerebras.sdk.runtime.sdkruntimepybind Receive a host tensor to the device via either copy mode or streaming mode. The data is distributed into the region of interest (ROI) which is a bounding box starting at coordinate ``(px, py)`` with width ``w`` and height ``h``. :param dest: A 3-D host tensor ``A[h][w][l]``, wrapped in a 1-D array according to keyword argument ``order``. :type dest: ``numpy.ndarray`` :param src: A user-defined color if keyword argument ``streaming=True``, symbol of a device tensor otherwise. :type src: ``int`` :param px: ``x``-coordinate of start point of the ROI. :type px: ``int`` :param py: ``y``-coordinate of start point of the ROI. :type py: ``int`` :param w: Width of the ROI. :type w: ``int`` :param h: Height of the ROI. :type h: ``int`` :param elem_per_pe: Number of elements per PE. The data type of an element is 16-bit and 32-bit only. If the tensor has ``k`` elements per PE, ``elt_per_pe`` is ``k`` even if the data type is 16-bit. If the data type is 16-bit, the user has to extend the tensor to a 32-bit one, with zero filled in the higher 16 bits. :type elem_per_pe: ``int`` :Keyword Arguments: * **streaming** (``bool``) -- Streaming mode if ``True``, copy mode otherwise. * **data_type** (``MemcpyDataType``) -- 32-bit if ``MemcpyDataType.MEMCPY_32BIT`` or 16-bit if ``MemcpyDataType.MEMCPY_16BIT``. Note that this argument has no effect if ``streaming`` is ``True``, and the user must handle the data appropriately in the receiving wavelet-triggered task. Additionally, the underlying type of the tensor ``dest`` must be 32-bit. The tensor must be extended to a 32-bit one with zero filled in the higher 16 bits. * **order** (``MemcpyOrder``) -- Row-major if ``MemcpyOrder.ROW_MAJOR`` or column-major if ``MemcpyOrder.COL_MAJOR``. * **nonblock** (``bool``) -- Nonblocking if ``True``, blocking otherwise. :returns: **task_handle** (``Task``) -- Handle to the task launched by ``memcpy_d2h``. .. py:method:: memcpy_h2d(dest: int, src: numpy.ndarray, px: int, py: int, w: int, h: int, elem_per_pe: int, **kwargs) -> Task :module: cerebras.sdk.runtime.sdkruntimepybind Send a host tensor to the device via either copy mode or streaming mode. The data is distributed into the region of interest (ROI) which is a bounding box starting at coordinate ``(px, py)`` with width ``w`` and height ``h``. :param dest: A user-defined color if keyword argument ``streaming=True``, symbol of a device tensor otherwise. :type dest: ``int`` :param src: A 3-D host tensor ``A[h][w][l]``, wrapped in a 1-D array according to parameter ``order``. :type src: ``numpy.ndarray`` :param px: ``x``-coordinate of start point of the ROI. :type px: ``int`` :param py: ``y``-coordinate of start point of the ROI. :type py: ``int`` :param w: Width of the ROI. :type w: ``int`` :param h: Height of the ROI. :type h: ``int`` :param elem_per_pe: Number of elements per PE. The data type of an element is 16-bit and 32-bit only. If the tensor has ``k`` elements per PE, ``elt_per_pe`` is ``k`` even if the data type is 16-bit. If the data type is 16-bit, the user has to extend the tensor to a 32-bit one, with zero filled in the higher 16 bits. :type elem_per_pe: ``int`` :Keyword Arguments: * **streaming** (``bool``) -- Streaming mode if ``True``, copy mode otherwise. * **data_type** (``MemcpyDataType``) -- 32-bit if ``MemcpyDataType.MEMCPY_32BIT`` or 16-bit if ``MemcpyDataType.MEMCPY_16BIT``. Note that this argument has no effect if ``streaming`` is ``True``, and the user must handle the data appropriately in the receiving wavelet-triggered task. Additionally, the underlying type of the tensor ``src`` must be 32-bit. The tensor must be extended to a 32-bit one with zero filled in the higher 16 bits. * **order** (``MemcpyOrder``) -- Row-major if ``MemcpyOrder.ROW_MAJOR`` or column-major if ``MemcpyOrder.COL_MAJOR``. * **nonblock** (``bool``) -- Nonblocking if ``True``, blocking otherwise. :returns: * **task_handle** (``Task``) -- Handle to the task launched by ``memcpy_h2d``. .. py:method:: run() :module: cerebras.sdk.runtime.sdkruntimepybind Start the simfabric or WSE run and wait for commands from the host runtime. .. py:method:: stop() :module: cerebras.sdk.runtime.sdkruntimepybind Wait for all pending commands (data transfers and kernel function calls) to complete and then stop simfabric or WSE. After this call is complete, no new commands will be accepted for this ``SdkRuntime`` object. ``stop`` must be called to end a program. Otherwise, the runtime will emit an error. .. py:method:: task_wait(task_handle: Task) :module: cerebras.sdk.runtime.sdkruntimepybind Wait for the task ``task_handle`` to complete. :param task_handle: Handle to a task previously launched by ``SdkRuntime``. :type task_handle: ``Task`` .. py:class:: MemcpyDataType :module: cerebras.sdk.runtime.sdkruntimepybind Bases: :class:`Enum` Specifies the data size for transfers using ``memcpy_d2h`` and ``memcpy_h2d`` copy mode. :Values: * **MEMCPY_16BIT** * **MEMCPY_32BIT** .. py:class:: MemcpyOrder :module: cerebras.sdk.runtime.sdkruntimepybind Bases: :class:`Enum` Specifies mapping of data for transfers using ``memcpy_d2h`` and ``memcpy_h2d``. :Values: * **ROW_MAJOR** * **COL_MAJOR** .. py:class:: Task :module: cerebras.sdk.runtime.sdkruntimepybind Handle to a task launched by ``SdkRuntime``. .. _sdkruntime-sdk-utils: sdk_utils --------- Utility functions for common operations with ``SdkRuntime``. .. py:module:: cerebras.sdk.sdk_utils .. py:function:: calculate_cycles(timestamp_buf: numpy.ndarray) -> numpy.int64: :module: cerebras.sdk.sdk_utils Converts values in ``timestamp_buf`` returned from device into a human-readable elapsed cycle count. :param timestamp_buf: array returned from device containing elapsed timestamp data :type timestamp_buf: ``numpy.ndarray`` :returns: **elapsed_cycles** (``numpy.int64``) -- Elapsed cycle count. **Example**: Consider the following CSL snippet which records timestamps and produces a single array to copy back to the host, to generate an elapsed cycle count: .. code-block:: csl // import time module and create timestamp buffers const timestamp = @import_module("