text
stringlengths
1
2.05k
use regex::Regex; use std::fs; use std::path::Path;
fn main() { let trait_path = "src/operators/tensor/core.cairo"; let doc_path = "docs/framework/operators/tensor"; let label = "tensor"; let trait_name = "TensorTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/nn/core.cairo"; let doc_path = "docs/framework/operators/neural-network"; let label = "nn"; let trait_name = "NNTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/sequence/core.cairo"; let doc_path = "docs/framework/operators/sequence"; let label = "sequence"; let trait_name = "SequenceTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/numbers/fixed_point/core.cairo"; let doc_path = "docs/framework/numbers/fixed-point"; let label = "fp"; let trait_name = "FixedTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/numbers/complex_number/complex_trait.cairo"; let doc_path = "docs/framework/numbers/complex-number"; let label = "complex"; let trait_name: &str = "ComplexTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/tree_ensemble/tree_ensemble_classifier.cairo"; let doc_path = "docs/framework/operators/machine-learning/tree-ensemble-classifier"; let label = "tree_ensemble_classifier"; let trait_name: &str = "TreeEnsembleClassifierTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/tree_ensemble/tree_ensemble_regressor.cairo"; let doc_path = "docs/framework/operators/machine-learning/tree-ensemble-regressor"; let label = "tree_ensem
ble_regressor"; let trait_name: &str = "TreeEnsembleRegressorTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/linear/linear_regressor.cairo"; let doc_path = "docs/framework/operators/machine-learning/linear-regressor"; let label = "linear_regressor"; let trait_name: &str = "LinearRegressorTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/linear/linear_classifier.cairo"; let doc_path = "docs/framework/operators/machine-learning/linear-classifier"; let label = "linear_classifier"; let trait_name: &str = "LinearClassifierTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/svm/svm_regressor.cairo"; let doc_path = "docs/framework/operators/machine-learning/svm-regressor"; let label = "svm_regressor"; let trait_name: &str = "SVMRegressorTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/svm/svm_classifier.cairo"; let doc_path = "docs/framework/operators/machine-learning/svm-classifier"; let label = "svm_classifier"; let trait_name: &str = "SVMClassifierTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); let trait_path = "src/operators/ml/normalizer/normalizer.cairo"; let doc_path = "docs/framework/operators/machine-learning/normalizer"; let label = "normalizer"; let trait_name: &str = "NormalizerTrait"; doc_trait(trait_path, doc_path, label); doc_functions(trait_path, doc_path, trait_name, label); }
fn doc_trait(trait_path: &str, doc_path: &str, label: &str) { let path_str = format!("../{}", trait_path); let path = Path::new(&path_str); let contents = fs::read_to_string(&path).expect("Could not read the file"); let re = Regex::new(r let mut table = String::from("| function | description |\n| --- | --- |\n"); for cap in re.captures_iter(&contents) { if &cap[1] == "Trait" { continue; } let func_name = format!( "[`{}.{}`]({}.{}.md)", label, &cap[1], label, &cap[1].replace('_', r"\_") ); let func_desc = &cap[2]; table += &format!("| {} | {} |\n", func_name, func_desc); } let readme_path_str = format!("../{}/README.md", doc_path); let readme_path = Path::new(&readme_path_str); let readme = fs::read_to_string(&readme_path).expect("Could not read the file"); let re_table = Regex::new(r"(?ms)\n\n\| fun.*?(\n[^|]|\z)").unwrap(); let new_readme = re_table.replace(&readme, &("\n\n".to_owned() + &table + "\n")); fs::write(&readme_path, &*new_readme).expect("Could not write the file"); }
fn doc_functions(trait_path: &str, doc_path: &str, trait_name: &str, label: &str) { let filepath_str = format!("../{}", trait_path); let filepath = Path::new(&filepath_str); let contents = fs::read_to_string(filepath).expect("Something went wrong reading the file"); let trait_re = Regex::new(&format!( r"(?s)trait\s+{}\s*(<[\w\s,]*>)?\s*\{{.*?\n\s*\}}", trait_name )) .unwrap(); let trait_match = trait_re.captures(&contents).unwrap(); let trait_block = trait_match.get(0).unwrap().as_str(); let func_re = Regex::new(r"(?s)( for func_match in func_re.captures_iter(trait_block) { let func_name = func_match.get(2).unwrap().as_str(); let doc_comment = func_match.get(1).unwrap().as_str(); let markdown_filename = format!("../{}/{}.{}.md", doc_path, label, func_name); let transformed_comment = doc_comment .lines() .map(|line| { line.trim_start().strip_prefix(" line.trim_start() .strip_prefix(" .unwrap_or(line.trim_start()), ) }) .collect::<Vec<_>>() .join("\n"); fs::write(markdown_filename, transformed_comment).expect("Unable to write file"); } }
import os from pathlib
import Path BASE_PATH = "./tests/nodes" class ModFile: def __init__(self): """ Initialize a ModFile object. This method creates a new file with a .cairo extension in the BASE_PATH directory. If the directory doesn't exist, it's created. The contents of the file are then read into the buffer attribute. """ self.path = Path(f"{BASE_PATH}.cairo") self.path.parent.mkdir(parents=True, exist_ok=True) with self.path.open("r") as f: self.buffer = f.readlines() def update(self, name: str): """ Update the .cairo file with a new module statement. Args: name (str): The name of the module to be added. This method checks if a module statement for the given name already exists in the buffer. If it doesn't, the new module statement is appended to the file. """ statement = f"mod {name};" if any([line.startswith(statement) for line in self.buffer]): return with self.path.open("a") as f: f.write(f"{statement}\n") class File: def __init__(self, path: str): """ Initialize a File object. Args: path (str): The file path where the File object will operate. This method creates a new file at the specified path. If the file already exists, its contents are read into the buffer attribute. """ self.path = Path(path) self.path.parent.mkdir(parents=True, exist_ok=True) self.buffer = [] if os.path.isfile(path): with self.path.open("r") as f: self.buffer = f.readlines() def dump(self): """ Write the contents of the buffer to the file. This method writes each line in the buffer to the file, ensuring each line is properly terminated with a newline character. """ with self.path.open("w") as f: f.writelines([f"{line}\n" for line in self.buffer])
class CairoTest(File): def __init__(self, file: str): super().__init__(os.path.join(BASE_PATH, file)) @classmethod def base_template( cls, name: str, arg_cnt: int, refs: list[str], func_sig: str, out_cnt: int = 1 ) -> list[str]: """ Create a template for a Cairo test function which expects a tensor output. Args: name (str): Name of the test function. arg_cnt (int): Number of arguments for the function. refs (list[str]): List of references (modules) to be used in the function. func_sig (str): The function signature. out_cnt (int): Number of outputs for the function. Defaults to 1. Returns: list[str]: A list of strings that together form the template of a Cairo test function. This method generates a list of strings that form the template of a Cairo test function, including module imports, function definition, and assertions. """ template = [ *[f"mod input_{i};" for i in range(arg_cnt)], *[f"mod output_{i};" for i in range(out_cnt)], "", "", *[f"use {ref};" for ref in refs], "", " " f"fn test_{name}()" + " {", *[f" let input_{i} = input_{i}::input_{i}();" for i in range(arg_cnt)], *[f" let z_{i} = output_{i}::output_{i}();" for i in range(out_cnt)], "" ] if out_cnt > 1: template.append(f" let ({', '.join(f'y_{i}' for i in range(out_cnt))}) = {func_sig};") else: template.append(f" let y_0 = {func_sig};") template.extend([ "", *[f" assert_eq(y_{i}, z_{i});" for i in range(out_cnt)], "}" ]) return template @classmethod def sequence_template(cls, name: str, arg_cnt: int, refs: list[str], func_sig: str) -> list[str]: """ Create a template for a
Cairo test function which expects a tensor sequence. Args: name (str): Name of the test function. arg_cnt (int): Number of arguments for the function. refs (list[str]): List of references (modules) to be used in the function. func_sig (str): The function signature. Returns: list[str]: A list of strings that together form the template of a Cairo test function. This method generates a list of strings that form the template of a Cairo test function, including module imports, function definition, and assertions. """ return [ *[f"mod input_{i};" for i in range(arg_cnt)], *[ "mod output_0;"], *[ ""], *[ ""], *[f"use {ref};" for ref in refs], *[ ""], *[ " *[ " *[f"fn test_{name}()"+" {"], *[f" let input_{i} = input_{i}::input_{i}();" for i in range(arg_cnt)], *[ " let z = output_0::output_0();"], *[ ""], *[f" let y = {func_sig};"], *[ ""], *[ " assert_seq_eq(y, z);"], *[ "}"], ]
class CairoData(File): def __init__(self, file: str): super().__init__(os.path.join(BASE_PATH, file)) @classmethod def base_template( cls, func: str, dtype: str, refs: list[str], data: list[str], shape: tuple ) -> list[str]: """ Create a base template for data representation in Cairo. Args: func (str): The function name. dtype (str): The data type of the tensor. refs (list[str]): A list of module references. data (list[str]): The data to be included in the tensor. shape (tuple): The shape of the tensor. Returns: list[str]: A list of strings that together form the template of a data function in Cairo. This method generates a list of strings representing a function in Cairo for data handling, defining the shape and contents of a tensor. """ template = [ *[f"use {ref};" for ref in refs], *[""], *[f"fn {func}() -> Tensor<{dtype}>" + " {"], *[" let mut shape = ArrayTrait::<usize>::new();"], *[f" shape.append({s});" for s in shape], *[""], *[" let mut data = ArrayTrait::new();"], *[f" data.append({d});" for d in data], *[" TensorTrait::new(shape.span(), data.span())"], *["}"], ] return template @classmethod def sequence_template( cls, func: str, dtype: str, refs: list[str], data: list[list[str]], shape: list[tuple], ) -> list[str]: """ Create a template for handling tensor sequences in Cairo. Args: func (str): The function name. dtype (str): The data type of the tensor sequence. refs (list[str]): A list of module references. data (list[list[str]]): The data to be included in each tensor. shape (list[tuple]): The shapes of each tensor in the sequence.
Returns: list[str]: A list of strings that together form the template of a sequence tensor function in Cairo. This method generates a list of strings representing a function in Cairo for handling a sequence of tensors, each with its own data and shape. """ def expand_sequence_init(s: list[tuple], d: list[list[str]]) -> list[str]: snippet = [] for i in range(len(s)): snippet += [ *[" let mut shape = ArrayTrait::<usize>::new();"], *[f" shape.append({s});" for s in s[i]], *[""], *[" let mut data = ArrayTrait::new();"], *[f" data.append({d});" for d in d[i]], *[""], *[ " sequence.append(TensorTrait::new(shape.span(), data.span()));" ], *[""], ] return snippet template = [ *[f"use {ref};" for ref in refs], *[""], *[f"fn {func}() -> Array<Tensor<{dtype}>>" + " {"], *[" let mut sequence = ArrayTrait::new();"], *[""], *expand_sequence_init(shape, data), *[" sequence"], *["}"], ] return template
from enum
import Enum
import os from typing
import List from .file_manager
import CairoTest, CairoData, ModFile
import numpy as np
class FixedImpl(Enum): FP8x23 = 'FP8x23' FP16x16 = 'FP16x16' FP32x32 = 'FP32x32' def to_fp(x: np.ndarray, fp_impl: FixedImpl): match fp_impl: case FixedImpl.FP8x23: return (x * 2**23).astype(np.int64) case FixedImpl.FP16x16: return (x * 2**16).astype(np.int64) case FixedImpl.FP32x32: return (x * 2**32).astype(np.int64)
class Dtype(Enum): FP8x23 = 'FP8x23' FP16x16 = 'FP16x16' FP32x32 = 'FP32x32' I8 = 'i8' I32 = 'i32' U32 = 'u32' BOOL = 'bool' COMPLEX64 = 'complex64' class Tensor: def __init__(self, dtype: Dtype, shape: tuple, data: np.ndarray): self.dtype = dtype self.shape = shape self.data = data Sequence = List[Tensor]
class Trait(Enum): TENSOR = 'TENSOR' NN = 'NN' SEQUENCE = 'SEQUENCE' def make_test(inputs: list[Tensor | Sequence], output: Tensor | Sequence, func_sig: str, name: str, trait: Trait = Trait.TENSOR): """ Generate and write Cairo tests based on the provided inputs and output. Args: inputs (list[Tensor | list[Tensor]]): A list of input tensors or tensor sequences. output (Tensor | list[Tensor]): The expected output tensor or tensor sequences. func_sig (str): The signature of the function to be tested. name (str): The name of the test. trait (Trait, optional): The trait of the tensors. Defaults to Trait.TENSOR. """ ModFile().update(name) for i, input in enumerate(inputs): input_data = CairoData(os.path.join(name, f"input_{i}.cairo")) match input: case list(): input_data.buffer = CairoData.sequence_template( func=f"input_{i}", dtype=input[0].dtype.value, refs=get_data_refs(input[0].dtype), data=get_data_statement_for_sequences( input, input[0].dtype), shape=[x.shape for x in input], ) case Tensor(): input_data.buffer = CairoData.base_template( func=f"input_{i}", dtype=input.dtype.value, refs=get_data_refs(input.dtype), data=get_data_statement(input.data, input.dtype), shape=input.shape, ) input_data.dump() match output: case list(): output_data = CairoData(os.path.join(name, "output_0.cairo")) output_data.buffer = CairoData.sequence_template( func="output_0", dtype=output[0].dtype.value, refs=get_data_refs(output[0].dtype), data=get_data_statement_for_sequences(output, output[0].dtype), shape=[
x.shape for x in output], ) output_data.dump() case tuple(): for i, out in enumerate(output): output_data = CairoData( os.path.join(name, f"output_{i}.cairo")) output_data.buffer = CairoData.base_template( func=f"output_{i}", dtype=out.dtype.value, refs=get_data_refs(out.dtype), data=get_data_statement(out.data, out.dtype), shape=out.shape, ) output_data.dump() case Tensor(): output_data = CairoData(os.path.join(name, "output_0.cairo")) output_data.buffer = CairoData.base_template( func="output_0", dtype=output.dtype.value, refs=get_data_refs(output.dtype), data=get_data_statement(output.data, output.dtype), shape=output.shape, ) output_data.dump() test_file = CairoTest(f"{name}.cairo") match output: case list(): test_file.buffer = CairoTest.sequence_template( name=name, arg_cnt=len(inputs), refs=get_all_test_refs(find_all_types([*inputs, *output]), trait), func_sig=func_sig, ) case Tensor(): test_file.buffer = CairoTest.base_template( name=name, arg_cnt=len(inputs), refs=get_all_test_refs(find_all_types([*inputs, output]), trait), func_sig=func_sig, ) case tuple(): test_file.buffer = CairoTest.base_template( name=name, arg_cnt=len(inputs), out_cnt=len(output), refs=get_all_test_refs(find_all_types([*inputs, output]), trait), func_sig=func_sig, ) test_file.dump() def get_data_refs(dtype: Dtype) -> list[str]: refs = [ *trait_to_r
ef[Trait.TENSOR], *dtype_to_tensor[dtype], *dtype_to_numbers[dtype], ] return refs def get_data_statement(data: np.ndarray, dtype: Dtype) -> list[str]: match dtype: case Dtype.U32: return [f"{int(x)}" for x in data.flatten()] case Dtype.I32: return [f"{int(x)}" for x in data.flatten()] case Dtype.I8: return [f"{int(x)}" for x in data.flatten()] case Dtype.FP8x23: return ["FP8x23 { "+f"mag: {abs(int(x))}, sign: {str(x < 0).lower()} "+"}" for x in data.flatten()] case Dtype.FP16x16: return ["FP16x16 { "+f"mag: {abs(int(x))}, sign: {str(x < 0).lower()} "+"}" for x in data.flatten()] case Dtype.FP32x32: return ["FP32x32 { "+f"mag: {abs(int(x))}, sign: {str(x < 0).lower()} "+"}" for x in data.flatten()] case Dtype.BOOL: return [str(x).lower() for x in data.flatten()] case Dtype.COMPLEX64: return ["complex64 { "+"real: FP64x64 { "+f"mag: {abs(int(np.real(x)))}, sign: {str(np.real(x) < 0).lower()} "+"} , img: FP64x64 { "+f"mag: {abs(int(np.imag(x)))}, sign: {str(np.imag(x) < 0).lower()} "+"} }" for x in data.flatten()] def get_data_statement_for_sequences(data: Sequence, dtype: Dtype) -> list[list[str]]: return [get_data_statement(x.data, dtype) for x in data] def get_all_test_refs(dtypes: list[Dtype], trait: Trait) -> list[str]: refs = [] for dtype in dtypes: refs += get_test_refs(dtype, trait) return list(set(refs)) def get_test_refs(dtype: Dtype, trait: Trait) -> list[str]: if trait == Trait.NN and dtype == Dtype.BOOL: raise Exception("NN trait does not support bool dtype") if trait == Trait.NN: dtype_ref = dtype_to_nn[dtype] elif trait == Trait.SEQUENCE: dtype_ref = dtype_to_sequence[dtype] else: dtype_ref = dtype_to_tensor[dtype] refs = [ *trait_to_ref[trait], *dtype_ref, *dtype_to_partial_eq[dtype],
"orion::utils::{assert_eq, assert_seq_eq}", ] return refs def find_all_types(tensors: list[Tensor | Sequence]) -> list[Dtype]: dtypes = [] for tensor in tensors: if isinstance(tensor, list) or isinstance(tensor, tuple): dtypes += [x.dtype for x in tensor] else: dtypes.append(tensor.dtype) return list(set(dtypes)) trait_to_ref = { Trait.TENSOR: [ "core::array::{ArrayTrait, SpanTrait}", "orion::operators::tensor::{TensorTrait, Tensor}", ], Trait.NN: [ "orion::numbers::FixedTrait", "orion::operators::nn::NNTrait", ], Trait.SEQUENCE: [ "core::array::{ArrayTrait, SpanTrait}", "orion::operators::sequence::SequenceTrait", ], } dtype_to_tensor = { Dtype.U32: ["orion::operators::tensor::{U32Tensor, U32TensorAdd}",], Dtype.I32: ["orion::operators::tensor::{I32Tensor, I32TensorAdd}",], Dtype.I8: ["orion::operators::tensor::{I8Tensor, I8TensorAdd}",], Dtype.FP8x23: ["orion::operators::tensor::{FP8x23Tensor, FP8x23TensorAdd}",], Dtype.FP16x16: ["orion::operators::tensor::{FP16x16Tensor, FP16x16TensorAdd}",], Dtype.BOOL: ["orion::operators::tensor::BoolTensor",], Dtype.COMPLEX64: ["orion::operators::tensor::Complex64Tensor",], Dtype.FP32x32: ["orion::operators::tensor::FP32x32Tensor",], } dtype_to_nn = { Dtype.U32: ["orion::operators::nn::U32NN",], Dtype.I32: ["orion::operators::nn::I32NN",], Dtype.I8: ["orion::operators::nn::I8NN",], Dtype.FP8x23: ["orion::operators::nn::FP8x23NN",], Dtype.FP16x16: ["orion::operators::nn::FP16x16NN",], } dtype_to_sequence = { Dtype.U32: ["orion::operators::sequence::U32Sequence",], Dtype.I32: ["orion::operators::sequence::I32Sequence",], Dtype.I8: ["orion::operators::sequence::I8Sequence",], Dtype.FP8x23: ["orion::operators::sequence::FP8x23Sequence",], Dtype.FP16x16: ["orion::operators::sequence::FP16x16Sequence",], } dtype_to_partial_eq = { Dtype.U32: ["orion::oper
ators::tensor::U32TensorPartialEq",], Dtype.I32: ["orion::operators::tensor::I32TensorPartialEq",], Dtype.I8: ["orion::operators::tensor::I8TensorPartialEq",], Dtype.FP8x23: ["orion::operators::tensor::FP8x23TensorPartialEq",], Dtype.FP16x16: ["orion::operators::tensor::FP16x16TensorPartialEq",], Dtype.FP32x32: ["orion::operators::tensor::FP32x32TensorPartialEq",], Dtype.BOOL: ["orion::operators::tensor::BoolTensorPartialEq",], Dtype.COMPLEX64: ["orion::operators::tensor::Complex64TensorPartialEq",], } dtype_to_numbers = { Dtype.U32: ["orion::numbers::NumberTrait"], Dtype.I32: ["orion::numbers::NumberTrait"], Dtype.I8: ["orion::numbers::NumberTrait"], Dtype.FP8x23: ["orion::numbers::{FixedTrait, FP8x23}",], Dtype.FP16x16: ["orion::numbers::{FixedTrait, FP16x16}",], Dtype.FP32x32: ["orion::numbers::{FixedTrait, FP32x32}",], Dtype.BOOL: [], Dtype.COMPLEX64: ["orion::numbers::{NumberTrait, complex64}",], }
import argparse import importlib import os import sys class RunAll: @classmethod def run_all(cls): for method_name in dir(cls): if method_name.startswith('__') or method_name == 'run_all': continue method = getattr(cls, method_name) if callable(method): method() # Add the path to the 'orion' directory to the Python path sys.path.append(os.path.join(os.path.dirname(__file__), '..', '..')) def main(): parser = argparse.ArgumentParser(description="Generate nodes.") parser.add_argument('node_class', help="The class of node to run.") args = parser.parse_args() class_name = args.node_class.capitalize() # Verify that the specified Python file exists filename = os.path.join('nodegen/node', args.node_class + '.py') if not os.path.exists(filename): print(f"Error: {filename} does not exist.") return # Import the module dynamically module = importlib.import_module('nodegen.node.' + args.node_class) # Get the class from the module node_class = getattr(module, class_name) # Instantiate the class and call the run_all method node_instance = node_class() node_instance.run_all() if __name__ == "__main__": main()
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Abs(RunAll): @staticmethod def abs_i32(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int32) y = abs(x) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "abs_i32" make_test([x], y, "input_0.abs()", name) @staticmethod def abs_i8(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int8) y = abs(x) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "abs_i8" make_test([x], y, "input_0.abs()", name) @staticmethod def abs_fp8x23(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int64), FixedImpl.FP8x23) y = abs(x) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.FP8x23, y.shape, y.flatten()) name = "abs_fp8x23" make_test([x], y, "input_0.abs()", name) @staticmethod def abs_fp16x16(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int64), FixedImpl.FP16x16) y = abs(x) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.FP16x16, y.shape, y.flatten()) name = "abs_fp16x16" make_test([x], y, "input_0.abs()", name)
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Acos(RunAll): @staticmethod def acos_fp8x23(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.arccos(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp(x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "acos_fp8x23" make_test([x], y, "input_0.acos()", name) @staticmethod def acos_fp16x16(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.arccos(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "acos_fp16x16" make_test([x], y, "input_0.acos()", name)
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Acosh(RunAll): @staticmethod def acosh_fp8x23(): x = np.random.uniform(1, 5, (2, 2)).astype(np.float64) y = np.arccosh(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) name = "acosh_fp8x23" make_test([x], y, "input_0.acosh()", name) @staticmethod def acosh_fp16x16(): x = np.random.uniform(1, 5, (2, 2)).astype(np.float64) y = np.arccosh(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "acosh_fp16x16" make_test([x], y, "input_0.acosh()", name)
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl
class Add(RunAll): @staticmethod def add_u32(): def default(): x = np.random.randint(0, 3, (3, 3, 3)).astype(np.uint32) y = np.random.randint(0, 3, (3, 3, 3)).astype(np.uint32) z = x + y x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.U32, z.shape, z.flatten()) name = "add_u32" make_test([x, y], z, "input_0 + input_1", name) def broadcast(): x = np.random.randint(0, 3, (3, 3, 3)).astype(np.uint32) y = np.random.randint(0, 3, (1, 3, 1)).astype(np.uint32) z = x + y x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.U32, z.shape, z.flatten()) name = "add_u32_broadcast" make_test([x, y], z, "input_0 + input_1", name) default() broadcast() @staticmethod def add_i32(): def default(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.int32) y = np.random.randint(-3, 3, (3, 3, 3)).astype(np.int32) z = x + y x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) z = Tensor(Dtype.I32, z.shape, z.flatten()) name = "add_i32" make_test([x, y], z, "input_0 + input_1", name) def broadcast(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.int32) y = np.random.randint(-3, 3, (1, 3, 1)).astype(np.int32) z = x + y x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) z = Tensor(Dtype.I32, z.shape, z.flatten()) name = "add_i32_broadcast" make_test([x, y], z, "input_0 + input_1", name) default() broadcast() @staticmethod def add_i8(): def default(): x = np.ran
dom.randint(-3, 3, (3, 3, 3)).astype(np.int8) y = np.random.randint(-3, 3, (3, 3, 3)).astype(np.int8) z = x + y x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) z = Tensor(Dtype.I8, z.shape, z.flatten()) name = "add_i8" make_test([x, y], z, "input_0 + input_1", name) def broadcast(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.int8) y = np.random.randint(-3, 3, (1, 3, 1)).astype(np.int8) z = x + y x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) z = Tensor(Dtype.I8, z.shape, z.flatten()) name = "add_i8_broadcast" make_test([x, y], z, "input_0 + input_1", name) default() broadcast() @staticmethod def add_fp8x23(): def default(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) y = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) z = x + y x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) z = Tensor(Dtype.FP8x23, z.shape, to_fp( z.flatten(), FixedImpl.FP8x23)) name = "add_fp8x23" make_test([x, y], z, "input_0 + input_1", name) def broadcast(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) y = np.random.randint(-3, 3, (1, 3, 1)).astype(np.float64) z = x + y x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) z = Tensor(Dtype.FP8x23, z.shape, to_fp( z.flatten(), FixedImpl.FP8x23)) name = "add_fp8x23_broadcast"
make_test([x, y], z, "input_0 + input_1", name) default() broadcast() @staticmethod def add_fp16x16(): def default(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) y = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) z = x + y x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) z = Tensor(Dtype.FP16x16, z.shape, to_fp( z.flatten(), FixedImpl.FP16x16)) name = "add_fp16x16" make_test([x, y], z, "input_0 + input_1", name) def broadcast(): x = np.random.randint(-3, 3, (3, 3, 3)).astype(np.float64) y = np.random.randint(-3, 3, (1, 3, 1)).astype(np.float64) z = x + y x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) z = Tensor(Dtype.FP16x16, z.shape, to_fp( z.flatten(), FixedImpl.FP16x16)) name = "add_fp16x16_broadcast" make_test([x, y], z, "input_0 + input_1", name) default() broadcast()
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class And(RunAll): @staticmethod def and_bool(): def default(): x = (np.random.randn(3, 4) > 0).astype(bool) y = (np.random.randn(3, 4) > 0).astype(bool) z = np.logical_and(x, y) x = Tensor(Dtype.BOOL, x.shape, x.flatten()) y = Tensor(Dtype.BOOL, y.shape, y.flatten()) z = Tensor(Dtype.BOOL, z.shape, z.flatten()) name = "and_bool" make_test([x, y], z, "BoolTensor::and(@input_0, @input_1)", name) def broadcast(): x = (np.random.randn(3, 4, 5) > 0).astype(bool) y = (np.random.randn(3, 4, 5) > 0).astype(bool) z = np.logical_and(x, y) x = Tensor(Dtype.BOOL, x.shape, x.flatten()) y = Tensor(Dtype.BOOL, y.shape, y.flatten()) z = Tensor(Dtype.BOOL, z.shape, z.flatten()) name = "and_bool_broadcast" make_test([x, y], z, "BoolTensor::and(@input_0, @input_1)", name) default() broadcast()
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl def argmax_use_numpy(data: np.ndarray, axis: int = 0, keepdims: int = 1) -> np.ndarray: result = np.argmax(data, axis=axis) if keepdims == 1: result = np.expand_dims(result, axis) return result.astype(np.int64) def argmax_use_numpy_select_last_index( data: np.ndarray, axis: int = 0, keepdims: int = True ) -> np.ndarray: data = np.flip(data, axis) result = np.argmax(data, axis=axis) result = data.shape[axis] - result - 1 if keepdims: result = np.expand_dims(result, axis) return result.astype(np.int64)
class Argmax(RunAll): @staticmethod def no_keepdims(): data = np.array([[2, 1], [3, 10]], dtype=np.float32) axis = 1 keepdims = 0 result = argmax_use_numpy(data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_no_keepdims" make_test( [x], y, "input_0.argmax(1, Option::Some(false), Option::None(()))", name) @staticmethod def keepdims(): data = np.array([[2, 1], [3, 10]], dtype=np.float32) axis = 1 keepdims = 1 result = argmax_use_numpy(data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_keepdims" make_test( [x], y, "input_0.argmax(1, Option::Some(true), Option::None(()))", name) @staticmethod def default_axes_keepdims(): data = np.array([[2, 1], [3, 10]], dtype=np.float32) keepdims = 1 result = argmax_use_numpy(data, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_default_axes_keepdims" make_test( [x], y, "input_0.argmax(0, Option::Some(true), Option::None(()))", name) @staticmethod def negative_axis_keepdims(): data = np.array([[2, 1], [3, 10]], dtype=np.float32) axis = -1 keepdims = 1 result = argmax_use_numpy(data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_negative_axis_keepdims" make_test( [x], y, "input_0.argmax(-1, Option::Some(true), Option::None(()))", name) @staticmethod def no_keepdims_select_last_index(): data = np.array([[2, 2], [3, 10]], dtyp
e=np.float32) axis = 1 keepdims = 0 result = argmax_use_numpy_select_last_index( data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_no_keepdims_select_last_index" make_test( [x], y, "input_0.argmax(1, Option::Some(false), Option::Some(true))", name) @staticmethod def keepdims_select_last_index(): data = np.array([[2, 2], [3, 10]], dtype=np.float32) axis = 1 keepdims = 1 result = argmax_use_numpy_select_last_index( data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_keepdims_select_last_index" make_test( [x], y, "input_0.argmax(1, Option::Some(true), Option::Some(true))", name) @staticmethod def default_axes_keepdims_select_last_index(): data = np.array([[2, 2], [3, 10]], dtype=np.float32) keepdims = 1 result = argmax_use_numpy_select_last_index(data, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_default_axes_keepdims_select_last_index" make_test( [x], y, "input_0.argmax(0, Option::Some(true), Option::Some(true))", name) @staticmethod def negative_axis_keepdims_select_last_index(): data = np.array([[2, 2], [3, 10]], dtype=np.float32) axis = -1 keepdims = 1 result = argmax_use_numpy_select_last_index(data, axis=axis, keepdims=keepdims) x = Tensor(Dtype.FP16x16, data.shape, data.flatten()) y = Tensor(Dtype.I32, result.shape, result.flatten()) name = "argmax_negative_axis_keepdims_select_last_index" make_test( [x], y, "input_0.argmax(-1, Option::Some(true), Option::Some(t
rue))", name)
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl def argmin_use_numpy(data: np.ndarray, axis: int = 0, keepdims: int = 1, dtype=np.int64) -> np.ndarray: result = np.argmin(data, axis=axis) if keepdims == 1: result = np.expand_dims(result, axis) return result.astype(dtype) def argmin_use_numpy_select_last_index( data: np.ndarray, axis: int = 0, keepdims: int = True, dtype=np.int64 ) -> np.ndarray: data = np.flip(data, axis) result = np.argmin(data, axis=axis) result = data.shape[axis] - result - 1 if keepdims: result = np.expand_dims(result, axis) return result.astype(dtype)
class Argmin(RunAll): @staticmethod def argmin_u32(): def argmin_1D(): def default_params(): x = np.random.randint(0, 255, (3)).astype(np.uint32) y = argmin_use_numpy(x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_1D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(0, 255, (3)).astype(np.uint32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_1D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(0, 255, (3)).astype(np.uint32) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_1D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_1D() def argmin_2D(): def default_params(): x = np.random.randint(0, 255, (2, 2)).astype(np.uint32) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_2D_default" make_test(
[x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(0, 255, (2, 2)).astype(np.uint32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_2D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(0, 255, (2, 2)).astype(np.uint32) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_2D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_2D() def argmin_3D(): def default_params(): x = np.random.randint(0, 255, (2, 2, 2)).astype(np.uint32) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_3D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(0, 255, (2, 2, 2)).astype(np.uint32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_3D_keepdims_false" mak
e_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(0, 255, (2, 2, 2)).astype(np.uint32) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_u32_3D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_3D() @staticmethod def argmin_i32(): def argmin_1D(): def default_params(): x = np.random.randint(-127, 127, (3)).astype(np.int32) y = argmin_use_numpy(x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_1D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (3)).astype(np.int32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_1D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(0, 255, (3)).astype(np.int32) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape
, y.flatten()) name = "argmin_i32_1D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_1D() def argmin_2D(): def default_params(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int32) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_2D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_2D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int32) y = argmin_use_numpy_select_last_index( x, dtype=np.int32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_2D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_2D() def argmin_3D(): def default_params(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int32) y = argmin_use_numpy(x, dtype=np.ui
nt32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_3D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int32) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_3D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int32) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i32_3D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_3D() @staticmethod def argmin_i8(): def argmin_1D(): def default_params(): x = np.random.randint(-127, 127, (3)).astype(np.int8) y = argmin_use_numpy(x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_1D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (3)).astype(np.int8) y = argmin_
use_numpy( x, keepdims=0, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_1D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(0, 255, (3)).astype(np.int8) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_1D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_1D() def argmin_2D(): def default_params(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int8) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_2D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int8) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_2D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(-127, 127, (2, 2)).astype(np.int8)
y = argmin_use_numpy_select_last_index( x, dtype=np.int8) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_2D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_2D() def argmin_3D(): def default_params(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int8) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_3D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int8) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_3D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = np.random.randint(-127, 127, (2, 2, 2)).astype(np.int8) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_i8_3D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmi
n_3D() @staticmethod def argmin_fp16x16(): def argmin_1D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (3) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy(x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_1D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = to_fp(np.random.randint(-127, 127, (3) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_1D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(0, 255, (3)).astype( np.int8), FixedImpl.FP16x16) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_1D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_1D() def argmin_2D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), Fixe
dImpl.FP16x16) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_2D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_2D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy_select_last_index( x, dtype=np.int8) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_2D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_2D() def argmin_3D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_3D_de
fault" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_3D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP16x16) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp16x16_3D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_3D() @staticmethod def argmin_fp8x23(): def argmin_1D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (3) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy(x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_1D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false():
x = to_fp(np.random.randint(-127, 127, (3) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_1D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(0, 255, (3)).astype( np.int8), FixedImpl.FP8x23) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32).reshape((1)) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_1D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_1D() def argmin_2D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_2D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy( x, k
eepdims=0, dtype=np.uint32) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_2D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(-127, 127, (2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy_select_last_index( x, dtype=np.int8) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_2D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_2D() def argmin_3D(): def default_params(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy(x, dtype=np.uint32) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_3D_default" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::None(()))", name) def keepdims_false(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy( x, keepdims=0, dtype=np.uint32) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y
.flatten()) name = "argmin_fp8x23_3D_keepdims_false" make_test( [x], y, "input_0.argmin(0, Option::Some(false), Option::None(()))", name) def last_index(): x = to_fp(np.random.randint(-127, 127, (2, 2, 2) ).astype(np.int8), FixedImpl.FP8x23) y = argmin_use_numpy_select_last_index( x, dtype=np.uint32) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "argmin_fp8x23_3D_last_index" make_test( [x], y, "input_0.argmin(0, Option::None(()), Option::Some(true))", name) default_params() keepdims_false() last_index() argmin_3D()
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl
class Array_feature_extractor(RunAll): @staticmethod def array_feature_extractor_3D(): def array_feature_extractor_i32(): x = np.random.randint(-3, 3, (2, 3, 4)).astype(np.int32) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.I32, z.shape, z.flatten()) name = "array_feature_extractor_3D_i32" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp8x23(): x = np.random.randint(-3, 3, (2, 3, 4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP8x23, z.shape, to_fp( z.flatten(), FixedImpl.FP8x23)) name = "array_feature_extractor_3D_fp8x23" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp16x16(): x = np.random.randint(-3, 3, (2, 3, 4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP16x16, z.shape, to_fp( z.flatten(), FixedImpl.FP16x16)) name = "array_feature_extractor_3D_fp16x16" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) array_feature_extractor_i32() array_feature_extractor_fp8x23() array_feature_extractor_fp16x16() @staticmethod def array_feature_extractor_2D():
def array_feature_extractor_i32(): x = np.random.randint(-3, 3, (3, 4)).astype(np.int32) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.I32, z.shape, z.flatten()) name = "array_feature_extractor_2D_i32" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp8x23(): x = np.random.randint(-3, 3, (3, 4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP8x23, z.shape, to_fp( z.flatten(), FixedImpl.FP8x23)) name = "array_feature_extractor_2D_fp8x23" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp16x16(): x = np.random.randint(-3, 3, (3, 4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP16x16, z.shape, to_fp( z.flatten(), FixedImpl.FP16x16)) name = "array_feature_extractor_2D_fp16x16" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) array_feature_extractor_i32() array_feature_extractor_fp8x23() array_feature_extractor_fp16x16() @staticmethod def array_feature_extractor_1D(): def array_feature_extractor_i32(): x = np.random.randint(-3, 3, (4)).a
stype(np.int32) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.I32, z.shape, z.flatten()) name = "array_feature_extractor_1D_i32" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp8x23(): x = np.random.randint(-3, 3, (4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP8x23, z.shape, to_fp( z.flatten(), FixedImpl.FP8x23)) name = "array_feature_extractor_1D_fp8x23" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) def array_feature_extractor_fp16x16(): x = np.random.randint(-3, 3, (4)).astype(np.float64) y = np.array([1, 3]).astype(np.uint32) z = (x[..., y]) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.U32, y.shape, y.flatten()) z = Tensor(Dtype.FP16x16, z.shape, to_fp( z.flatten(), FixedImpl.FP16x16)) name = "array_feature_extractor_1D_fp16x16" make_test([x, y], z, "TensorTrait::array_feature_extractor(@input_0, input_1)", name) array_feature_extractor_i32() array_feature_extractor_fp8x23() array_feature_extractor_fp16x16()
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Asin(RunAll): @staticmethod def asin_fp8x23(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.arcsin(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp(x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "asin_fp8x23" make_test([x], y, "input_0.asin()", name) @staticmethod def asin_fp16x16(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.arcsin(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "asin_fp16x16" make_test([x], y, "input_0.asin()", name)
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Asinh(RunAll): @staticmethod def asinh_fp8x23(): x = np.random.uniform(1, 5, (2, 2)).astype(np.float64) y = np.arcsinh(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) name = "asinh_fp8x23" make_test([x], y, "input_0.asinh()", name) @staticmethod def asinh_fp16x16(): x = np.random.uniform(1, 5, (2, 2)).astype(np.float64) y = np.arcsinh(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "asinh_fp16x16" make_test([x], y, "input_0.asinh()", name)
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Atan(RunAll): @staticmethod def atan_fp8x23(): x = np.random.uniform(-10, 127, (2, 2)).astype(np.float64) y = np.arctan(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) name = "atan_fp8x23" make_test([x], y, "input_0.atan()", name) @staticmethod def atan_fp16x16(): x = np.random.uniform(-10, 127, (2, 2)).astype(np.float64) y = np.arctan(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "atan_fp16x16" make_test([x], y, "input_0.atan()", name)
import numpy as np from nodegen.node import RunAll from ..helpers import make_node, make_test, to_fp, Tensor, Dtype, FixedImpl class Binarizer(RunAll): @staticmethod def binarizer_fp8x23(): x = np.random.uniform(-3, 3, (3, 3, 3)).astype(np.float64) threshold = np.float64(1) y = (x > threshold).astype(np.float64) x = Tensor(Dtype.FP8x23, x.shape, to_fp( x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp( y.flatten(), FixedImpl.FP8x23)) name = "binarizer_fp8x23" make_node([x], [y], name) make_test([x], y, "TensorTrait::binarizer(@input_0, Option::Some(FixedTrait::new(8388608, false));", name) @staticmethod def binarizer_fp16x16(): x = np.random.uniform(-3, 3, (3, 3, 3)).astype(np.float64) threshold = np.float64(1) y = (x > threshold).astype(np.float64) x = Tensor(Dtype.FP16x16, x.shape, to_fp( x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "binarizer_fp16x16" make_node([x], [y], name) make_test([x], y, "TensorTrait::binarizer(@input_0, Option::Some(FixedTrait::new(65536, false));", name)
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl, Trait, get_data_statement def blackman_window(size, output_datatype=None, periodic=None) -> np.ndarray: if periodic == 1: N_1 = size else: N_1 = size - 1 ni = np.arange(size, dtype=output_datatype) alpha = 0.42 beta = 0.08 y = np.cos((ni * (np.float64(np.pi).astype(output_datatype) * 2)) / N_1).astype(output_datatype) * (-0.5) y += np.cos((ni * (np.float64(np.pi).astype(output_datatype) * 4)) / N_1) * beta y += alpha return y.astype(output_datatype)
class Blackman_window(RunAll): @staticmethod def fp8x23(): args = [3] args_str = get_data_statement(to_fp(np.array(args).flatten(), FixedImpl.FP8x23), Dtype.FP8x23) y = blackman_window(*args, np.float64) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "blackman_window_fp8x23" make_test( [], y, f"TensorTrait::blackman_window({','.join(args_str)}, Option::Some(0))", name ) @staticmethod def fp16x16(): print(get_data_statement(to_fp(np.array([np.pi]).flatten(), FixedImpl.FP16x16), Dtype.FP16x16)) args = [3] args_str = get_data_statement(to_fp(np.array(args).flatten(), FixedImpl.FP16x16), Dtype.FP16x16) y = blackman_window(*args, np.float16, 1) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "blackman_window_fp16x16" make_test( [], y, f"TensorTrait::blackman_window({','.join(args_str)}, Option::Some(1))", name )
import numpy as np from nodegen.node import RunAll from ..helpers import make_test, to_fp, Tensor, Dtype, FixedImpl class Ceil(RunAll): @staticmethod def ceil_fp8x23(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.ceil(x) x = Tensor(Dtype.FP8x23, x.shape, to_fp(x.flatten(), FixedImpl.FP8x23)) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "ceil_fp8x23" make_test([x], y, "input_0.ceil()", name) @staticmethod def ceil_fp16x16(): x = np.random.uniform(-1, 1, (2, 2)).astype(np.float64) y = np.ceil(x) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "ceil_fp16x16" make_test([x], y, "input_0.ceil()", name)
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl
class Clip(RunAll): @staticmethod def clip_u32(): def clip_2D(): x = np.random.randint(0, 255, (2, 4)).astype(np.uint32) y = np.clip(x, np.uint32(10), np.uint32(20)) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "clip_u32_2d" make_test( [x], y, "input_0.clip(Option::Some(10_u32), Option::Some(20_u32))", name) def clip_3D(): x = np.random.randint(0, 255, (20, 10, 5)).astype(np.uint32) y = np.clip(x, np.uint32(10), np.uint32(20)) x = Tensor(Dtype.U32, x.shape, x.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "clip_u32_3d" make_test( [x], y, "input_0.clip(Option::Some(10_u32), Option::Some(20_u32))", name) clip_2D() clip_3D() @staticmethod def clip_i32(): def clip_2D(): x = np.random.randint(-127, 127, (2, 4)).astype(np.int32) y = np.clip(x, np.int32(-10), np.int32(20)) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "clip_i32_2d" make_test( [x], y, "input_0.clip(Option::Some(-10_i32), Option::Some(20_i32))", name) def clip_3D(): x = np.random.randint(-127, 127, (20, 10, 5)).astype(np.int32) y = np.clip(x, np.int32(-10), np.int32(20)) x = Tensor(Dtype.I32, x.shape, x.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "clip_i32_3d" make_test( [x], y, "input_0.clip(Option::Some(-10_i32), Option::Some(20_i32))", name) clip_2D() clip_3D() @staticmethod def clip_i8(): def clip_2D(): x = np.random.randint(-127, 127, (2, 4)).astype(np.int8) y = np.clip(x, np.int8(-10), np.int8(20)) x = Tensor(Dtype.I8, x.shape, x
.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "clip_i8_2d" make_test( [x], y, "input_0.clip(Option::Some(-10_i8), Option::Some(20_i8))", name) def clip_3D(): x = np.random.randint(-127, 127, (20, 10, 5)).astype(np.int8) y = np.clip(x, np.int8(-10), np.int8(20)) x = Tensor(Dtype.I8, x.shape, x.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "clip_i8_3d" make_test( [x], y, "input_0.clip(Option::Some(-10_i8), Option::Some(20_i8))", name) clip_2D() clip_3D() @staticmethod def clip_fp8x23(): def clip_2D(): x = to_fp(np.random.randint(-127, 127, (2, 4) ).astype(np.int64), FixedImpl.FP8x23) y = np.clip(x, to_fp(np.int64(-10), FixedImpl.FP8x23), to_fp(np.int64(20), FixedImpl.FP8x23)) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.FP8x23, y.shape, y.flatten()) name = "clip_fp8x23_2d" make_test( [x], y, "input_0.clip(Option::Some(FP8x23 { mag: 83886080, sign: true }), Option::Some(FP8x23 { mag: 167772160, sign: false }))", name) def clip_3D(): x = to_fp(np.random.randint(-127, 127, (20, 10, 5) ).astype(np.int64), FixedImpl.FP8x23) y = np.clip(x, to_fp(np.int64(-10), FixedImpl.FP8x23), to_fp(np.int64(20), FixedImpl.FP8x23)) x = Tensor(Dtype.FP8x23, x.shape, x.flatten()) y = Tensor(Dtype.FP8x23, y.shape, y.flatten()) name = "clip_fp8x23_3d" make_test( [x], y, "input_0.clip(Option::Some(FP8x23 { mag: 83886080, sign: true }), Option::Some(FP8x23 { mag: 167772160, sign: false }))", name) clip_2D() clip_3D() @staticmethod def clip_fp16x16(): def clip_2D(): x = to_fp(np
.random.randint(-127, 127, (2, 4) ).astype(np.int64), FixedImpl.FP16x16) y = np.clip(x, to_fp(np.int64(-10), FixedImpl.FP16x16), to_fp(np.int64(20), FixedImpl.FP16x16)) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.FP16x16, y.shape, y.flatten()) name = "clip_fp16x16_2d" make_test( [x], y, "input_0.clip(Option::Some(FP16x16 { mag: 655360, sign: true }), Option::Some(FP16x16 { mag: 1310720, sign: false }))", name) def clip_3D(): x = to_fp(np.random.randint(-127, 127, (20, 10, 5) ).astype(np.int64), FixedImpl.FP16x16) y = np.clip(x, to_fp(np.int64(-10), FixedImpl.FP16x16), to_fp(np.int64(20), FixedImpl.FP16x16)) x = Tensor(Dtype.FP16x16, x.shape, x.flatten()) y = Tensor(Dtype.FP16x16, y.shape, y.flatten()) name = "clip_fp16x16_3d" make_test( [x], y, "input_0.clip(Option::Some(FP16x16 { mag: 655360, sign: true }), Option::Some(FP16x16 { mag: 1310720, sign: false }))", name) clip_2D() clip_3D()
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl, Trait def col2im(data, image_shape, block_shape, dilations=None, pads=None, strides=None): if dilations is None: dilations = [1 for s in image_shape] if pads is None: pads = [0 for s in image_shape] * 2 if strides is None: strides = [1 for s in image_shape] bl = np.prod(block_shape) C = data.shape[1] data = data.reshape(data.shape[:1] + (C,) + (bl,) + data.shape[2:]) ks = tuple(block_shape) res = None for n in range(data.shape[0]): for c in range(data.shape[1]): out = col2im_naive_implementation( data[n, c, ...], image_shape, ks, dilations, pads, strides ) if res is None: new_shape = data.shape[:2] + out.shape res = np.empty(new_shape, dtype=data.dtype) res[n, c, ...] = out return (res,) def _get_indices(i, shape): res = np.empty((len(shape),), dtype=np.int64) k = len(shape) - 1 while k > 0: m = i % shape[k] res[k] = m i -= m i /= shape[k] k -= 1 res[0] = i return res def _col2im_shape_check(X, output_shape, kernel_shape, dilations, pads, strides): n_input_plane = X.shape[0] kernel_size = np.prod(kernel_shape) if n_input_plane % kernel_size != 0: raise ValueError( f"Expected size of input's dimension 1 to be divisible by the " f"product of kernel_size={kernel_size}, " f"but got input.size(1)={n_input_plane} " f"and kernel_shape={kernel_shape}, X.shape={X.shape}, output_shape={output_shape}." ) input_length = X.shape[1] n_dims = len(output_shape) n_blocks = [] for i in range(n_dims): n_block = ( output_shape[i] + pads[i, :].sum() - dilations[i] * (kernel_shape[i] - 1) - 1 ) n_blocks.append(n_block) block_size = np.prod(n_blocks) if input_length
!= block_size: raise ValueError( f"Given n_input_plane={n_input_plane}, X.shape={X.shape}, " f"output_shape={output_shape}, kernel_shape={kernel_shape}, " f"dilations={dilations}, pads={pads}, strides={strides}, " f"expected size of input's dimension 2 to match the calculated number of " f"sliding blocks {n_blocks} = {block_size}, " f"but got input.size(2)={input_length}.", ) def col2im_naive_implementation(data, image_shape, kernel_shape, dilations, pads, strides): n_dims = len(pads) new_pads = np.array([(pads[i], pads[i + n_dims]) for i in range(n_dims)]) _col2im_shape_check(data, image_shape, kernel_shape, dilations, new_pads, strides) data_col = data data_im = np.zeros(image_shape, dtype=data.dtype) dim_col = [] for i in range(n_dims): col = ( image_shape[i] + new_pads[i, :].sum() - (dilations[i] * (kernel_shape[i] - 1) + 1) ) dim_col.append(col) kernel_size = np.prod(kernel_shape) col_size = np.prod(dim_col) for c_col in range(kernel_size): offset = _get_indices(c_col, kernel_shape) for col in range(col_size): ind_col = _get_indices(col, dim_col) ind_im = [] for i in range(n_dims): ind = ( ind_col[i] * strides[i] - new_pads[i, 0] + offset[i] * dilations[i] ) ind_im.append(ind) if not _is_out(ind_im, data_im.shape): data_im[tuple(ind_im)] += data_col[c_col, col] return data_im def _is_out(ind, shape): for i, s in zip(ind, shape): if i < 0: return True if i >= s: return True return False
class Col2im(RunAll): @staticmethod def export_col2im() -> None: x = np.array( [ [ [1.0, 6.0, 11.0, 16.0, 21.0], [2.0, 7.0, 12.0, 17.0, 22.0], [3.0, 8.0, 13.0, 18.0, 23.0], [4.0, 9.0, 14.0, 19.0, 24.0], [5.0, 0.0, 15.0, 20.0, 25.0], ] ] ).astype(np.float32) image_shape = np.array([5, 5]).astype(np.int64) block_shape = np.array([1, 5]).astype(np.int64) y = col2im(x,image_shape,block_shape) y = np.array(y[0]) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "col2im" func_sig = "NNTrait::col2im(" func_sig += "@input_0," func_sig += "array![5, 5].span()," func_sig += "array![1, 5].span()," func_sig += "Option::None," func_sig += "Option::None," func_sig += "Option::None)" make_test( [x], y, func_sig, name, Trait.NN) @staticmethod def export_col2im_strides() -> None: x = np.array( [ [ [0.0, 0.0, 0.0, 0.0], [1.0, 1.0, 1.0, 1.0], [1.0, 1.0, 1.0, 1.0], [1.0, 1.0, 1.0, 1.0], [0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0], [1.0, 1.0, 1.0, 1.0], [0.0, 0.0, 0.0, 0.0], ] ] ).astype(np.float32) image_shape = np.array([5, 5]).astype(np.int64) block_shape = np.array([3, 3]).astype(np.int64) y = col2im(x,image_shape,block_shape,strides=[2, 2]) y = np.array(y[0]) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(
y.flatten(), FixedImpl.FP16x16)) name = "col2im_strides" func_sig = "NNTrait::col2im(" func_sig += "@input_0," func_sig += "array![5, 5].span()," func_sig += "array![3, 3].span()," func_sig += "Option::None," func_sig += "Option::None," func_sig += "Option::Some(array![2, 2].span()))" make_test( [x], y, func_sig, name, Trait.NN) @staticmethod def export_col2im_pads() -> None: x = np.array( [ [ [ 1.0, 6.0, 11.0, 16.0, 21.0, 26, 31, 36, 41, 46, 51, 56, 61, 66, 71, ], [ 2.0, 7.0, 12.0, 17.0, 22.0, 27, 32, 37, 42, 47, 52, 57, 62, 67, 72, ], [ 3.0, 8.0, 13.0, 18.0, 23.0, 28, 33, 38, 43, 48, 53, 58, 63, 68, 73, ], [ 4.0, 9.0, 14.0, 19.0, 24.0, 29, 34, 39, 44, 49, 54, 59, 64, 69, 74, ], [ 5.0, 10.0, 15.0, 20.0, 25.0, 30, 35, 40, 45, 50, 55, 60, 65, 70, 75, ], ] ] ).astype(np.float32) image_shape = np.array([5, 5]).astype(np.int64) block_shape = np.array([1, 5]).astype(np.int64) y = col2im(x,image_shape,block_shape,pads=[0, 1, 0, 1]) y = np.array(y[0]) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "col2im_pads" func_sig = "NNTrait::col2im(" func_sig += "@input_0," func_sig += "array![5, 5].span()," func_sig += "array![1, 5].span()," func_sig += "Option::None," func_sig += "Option::Some(array![0, 1, 0, 1].span())," func_sig += "Option::None)" make_test( [x], y, func_sig, name, Trait.NN) @sta
ticmethod def export_col2im_dilations() -> None: x = np.array( [ [ [1.0, 5.0, 9.0, 13.0, 17], [2.0, 6.0, 10.0, 14.0, 18], [3.0, 7.0, 11.0, 15.0, 19], [4.0, 8.0, 12.0, 16.0, 20], ] ] ).astype(np.float32) image_shape = np.array([6, 6]).astype(np.int64) block_shape = np.array([2, 2]).astype(np.int64) y = col2im(x,image_shape,block_shape, dilations=[1, 5]) y = np.array(y[0]) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "col2im_dilations" func_sig = "NNTrait::col2im(" func_sig += "@input_0," func_sig += "array![6, 6].span()," func_sig += "array![2, 2].span()," func_sig += "Option::Some(array![1, 5].span())," func_sig += "Option::None," func_sig += "Option::None)" make_test( [x], y, func_sig, name, Trait.NN) @staticmethod def export_col2im_5D() -> None: x = np.array( [ [ [1, 6, 11, 16, 21, 26, 31, 36, 41, 46, 51, 56], [2, 7, 12, 17, 22, 27, 32, 37, 42, 47, 52, 57], [3, 8, 13, 18, 23, 28, 33, 38, 43, 48, 53, 58], [4, 9, 14, 19, 24, 29, 34, 39, 44, 49, 54, 59], [5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 55, 60], [61, 66, 71, 76, 81, 86, 91, 96, 101, 106, 111, 116], [62, 67, 72, 77, 82, 87, 92, 97, 102, 107, 112, 117], [63, 68, 73, 78, 83, 88, 93, 98, 103, 108, 113, 118], [64, 69, 74, 79, 84, 89, 94, 99, 104, 109, 114, 119], [65, 70, 75, 80, 85, 90, 95, 100, 105, 110, 115, 120], ] ] ).astype(np.float32) image_shape = np.array([3, 4, 5]
).astype(np.int64) block_shape = np.array([1, 1, 5]).astype(np.int64) y = col2im(x,image_shape,block_shape) y = np.array(y[0]) x = Tensor(Dtype.FP16x16, x.shape, to_fp(x.flatten(), FixedImpl.FP16x16)) y = Tensor(Dtype.FP16x16, y.shape, to_fp(y.flatten(), FixedImpl.FP16x16)) name = "col2im_5D" func_sig = "NNTrait::col2im(" func_sig += "@input_0," func_sig += "array![3, 4, 5].span()," func_sig += "array![1, 1, 5].span()," func_sig += "Option::None," func_sig += "Option::None," func_sig += "Option::None)" make_test( [x], y, func_sig, name, Trait.NN)
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl, Trait
class Compress(RunAll): @staticmethod def compress_fp16x16(): def compress_3D(): def default(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int64) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=0) x1 = Tensor(Dtype.FP16x16, x1.shape, to_fp(x1.flatten(), FixedImpl.FP16x16)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "compress_fp16x16_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(0))", name= name) def axis1(): x1 = np.arange(0,180).reshape(3,4,3,5).astype(np.int64) x2 = np.array([1, 1, 1, 0]).astype(np.int64) y = x1.compress(x2, axis=1) x1 = Tensor(Dtype.FP16x16, x1.shape, to_fp(x1.flatten(), FixedImpl.FP16x16)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "compress_fp16x16_3d_axis1" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(1))", name= name) def axis2(): x1 = np.arange(0,48).reshape(4,3,4).astype(np.int64) x2 = np.array([1, 0, 1, 1]).astype(np.int64) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.FP16x16, x1.shape, to_fp(x1.flatten(), FixedImpl.FP16x16)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "compress
_fp16x16_3d_axis2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) def axis3(): x1 = np.arange(0,96).reshape(4,3,4, 2).astype(np.int64) x2 = np.array([1, 0]).astype(np.int64) y = x1.compress(x2, axis=3) x1 = Tensor(Dtype.FP16x16, x1.shape, to_fp(x1.flatten(), FixedImpl.FP16x16)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "compress_fp16x16_3d_axis3" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(3))", name= name) def noaxis(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int64) x2 = np.array([1, 0, 1, 0, 1, 1, 1, 1, 1]).astype(np.int64) y = x1.compress(x2) x1 = Tensor(Dtype.FP16x16, x1.shape, to_fp(x1.flatten(), FixedImpl.FP16x16)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP16x16, y.shape, to_fp( y.flatten(), FixedImpl.FP16x16)) name = "compress_fp16x16_3d_noaxis" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::None(()))", name= name) default() axis1() axis2() axis3() noaxis() compress_3D() @staticmethod def compress_fp8x23(): def compress_3D(): def default(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int64) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1
.compress(x2, axis=0) x1 = Tensor(Dtype.FP8x23, x1.shape, to_fp(x1.flatten(), FixedImpl.FP8x23)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "compress_fp8x23_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(0))", name= name) def axis1(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int64) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=1) x1 = Tensor(Dtype.FP8x23, x1.shape, to_fp(x1.flatten(), FixedImpl.FP8x23)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "compress_fp8x23_3d_axis1" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(1))", name= name) def axis2(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int64) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.FP8x23, x1.shape, to_fp(x1.flatten(), FixedImpl.FP8x23)) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.FP8x23, y.shape, to_fp(y.flatten(), FixedImpl.FP8x23)) name = "compress_fp8x23_3d_axis2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) default() axis1() axis2() compress_3D() @staticmethod def compress_i8():
def compress_3D(): def default(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int8) x2 = np.array([0, 1, 1]).astype(np.uint8) y = x1.compress(x2, axis=0) x1 = Tensor(Dtype.I8, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "compress_i8_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(0))", name= name) def axis1(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int8) x2 = np.array([0, 1, 1]).astype(np.uint8) y = x1.compress(x2, axis=1) x1 = Tensor(Dtype.I8, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "compress_i8_3d_axis1" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(1))", name= name) def axis2(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int8) x2 = np.array([0, 1, 1]).astype(np.uint8) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.I8, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I8, y.shape, y.flatten()) name = "compress_i8_3d_axis2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) default() axis1() axis2() compress_3D() @staticmethod def compress_i32()
: def compress_3D(): def default(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int32) x2 = np.array([0, 1, 1]).astype(np.int32) y = x1.compress(x2, axis=0) x1 = Tensor(Dtype.I32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "compress_i32_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(0))", name= name) def axis1(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int32) x2 = np.array([0, 1, 1]).astype(np.int32) y = x1.compress(x2, axis=1) x1 = Tensor(Dtype.I32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "compress_i32_3d_axis1" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(1))", name= name) def axis2(): x1 = np.arange(0,27).reshape(3,3,3).astype(np.int32) x2 = np.array([0, 1, 1]).astype(np.int32) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.I32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.I32, y.shape, y.flatten()) name = "compress_i32_3d_axis2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) default() axis1() axis2() compress_3D() @staticmet
hod def compress_u32(): def compress_3D(): def default(): x1 = np.arange(0,48).reshape(4,4,3).astype(np.uint32) x2 = np.array([1, 1]).astype(np.uint32) y = x1.compress(x2, axis=0) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "compress_u32_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(0))", name= name) def axis1(): x1 = np.arange(0,36).reshape(3,4,3).astype(np.uint32) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=1) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "compress_u32_3d_axis1" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(1))", name= name) def axis2(): x1 = np.arange(0,48).reshape(3,4,4).astype(np.uint32) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "compress_u32_3d_axis2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) def axis2_2(): x1 = np.arange(0,60).reshap
e(3,4,5).astype(np.uint32) x2 = np.array([0, 1, 1]).astype(np.uint32) y = x1.compress(x2, axis=2) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "compress_u32_3d_axis2_2" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(2))", name= name) def axis3(): x1 = np.arange(0,270).reshape(3,3,5,6).astype(np.uint32) x2 = np.array([0, 1, 1,1,0,1]).astype(np.uint32) y = x1.compress(x2, axis=3) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "compress_u32_3d_axis3" make_test( inputs = [x1, x2], output = y, func_sig = "input_0.compress(condition:input_1, axis:Option::Some(3))", name= name) default() axis1() axis2() axis2_2() axis3() compress_3D()
import numpy as np from nodegen.node
import RunAll from ..helpers
import make_test, to_fp, Tensor, Dtype, FixedImpl, Trait
class Concat(RunAll): @staticmethod def concat_u32(): def concat_1D(): x1 = np.arange(0,3).astype(np.uint32) x2 = np.arange(3,6).astype(np.uint32) y = np.concatenate((x1, x2)) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "concat_u32_1d" make_test( inputs = [x1, x2], output = y, func_sig = "TensorTrait::concat(array![input_0, input_1].span(), 0)", name= name, trait= Trait.TENSOR) def concat_2D(): x1 = np.arange(0,4).astype(np.uint32).reshape(2,2) x2 = np.arange(4,8).astype(np.uint32).reshape(2,2) y = np.concatenate((x1, x2), axis=0) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "concat_u32_2d" make_test( inputs = [x1, x2], output = y, func_sig = "TensorTrait::concat(array![input_0, input_1].span(), 0)", name= name, trait= Trait.TENSOR) def concat_3D(): def default(): x1 = np.arange(0,27).astype(np.uint32).reshape(3,3,3) x2 = np.arange(27,54).astype(np.uint32).reshape(3,3,3) y = np.concatenate((x1, x2), axis=0) x1 = Tensor(Dtype.U32, x1.shape, x1.flatten()) x2 = Tensor(Dtype.U32, x2.shape, x2.flatten()) y = Tensor(Dtype.U32, y.shape, y.flatten()) name = "concat_u32_3d_default" make_test( inputs = [x1, x2], output = y, func_sig = "TensorTrait::concat(array![input_0, input_1].span(), 0)", name= name, trait= Trait.TENSOR) def axis_1(): x1 = np.arange(0,27).astype(np.uint32).resha

This dataset is a truncated version of this one but where the format is compatible with MLX-lora, using {"text": "This is an example for the model."}, and where each entry has been truncated, following some code logic (i.e., following classes, functions etc) to ensure each entry is smaller than 2048 tokens.

Downloads last month
62