forked from M-Labs/nac3
Compare commits
2 Commits
master
...
numpy-anya
Author | SHA1 | Date | |
---|---|---|---|
b1e97aa2b0 | |||
42c5f906fb |
@ -1,32 +0,0 @@
|
|||||||
BasedOnStyle: LLVM
|
|
||||||
|
|
||||||
Language: Cpp
|
|
||||||
Standard: Cpp11
|
|
||||||
|
|
||||||
AccessModifierOffset: -1
|
|
||||||
AlignEscapedNewlines: Left
|
|
||||||
AlwaysBreakAfterReturnType: None
|
|
||||||
AlwaysBreakTemplateDeclarations: Yes
|
|
||||||
AllowAllParametersOfDeclarationOnNextLine: false
|
|
||||||
AllowShortFunctionsOnASingleLine: Inline
|
|
||||||
BinPackParameters: false
|
|
||||||
BreakBeforeBinaryOperators: NonAssignment
|
|
||||||
BreakBeforeTernaryOperators: true
|
|
||||||
BreakConstructorInitializers: AfterColon
|
|
||||||
BreakInheritanceList: AfterColon
|
|
||||||
ColumnLimit: 120
|
|
||||||
ConstructorInitializerAllOnOneLineOrOnePerLine: true
|
|
||||||
ContinuationIndentWidth: 4
|
|
||||||
DerivePointerAlignment: false
|
|
||||||
IndentCaseLabels: true
|
|
||||||
IndentPPDirectives: None
|
|
||||||
IndentWidth: 4
|
|
||||||
MaxEmptyLinesToKeep: 1
|
|
||||||
PointerAlignment: Left
|
|
||||||
ReflowComments: true
|
|
||||||
SortIncludes: false
|
|
||||||
SortUsingDeclarations: true
|
|
||||||
SpaceAfterTemplateKeyword: false
|
|
||||||
SpacesBeforeTrailingComments: 2
|
|
||||||
TabWidth: 4
|
|
||||||
UseTab: Never
|
|
1
.gitignore
vendored
1
.gitignore
vendored
@ -1,4 +1,3 @@
|
|||||||
__pycache__
|
__pycache__
|
||||||
/target
|
/target
|
||||||
/nac3standalone/demo/linalg/target
|
|
||||||
nix/windows/msys2
|
nix/windows/msys2
|
||||||
|
@ -8,17 +8,17 @@ repos:
|
|||||||
hooks:
|
hooks:
|
||||||
- id: nac3-cargo-fmt
|
- id: nac3-cargo-fmt
|
||||||
name: nac3 cargo format
|
name: nac3 cargo format
|
||||||
entry: nix
|
entry: cargo
|
||||||
language: system
|
language: system
|
||||||
types: [file, rust]
|
types: [file, rust]
|
||||||
pass_filenames: false
|
pass_filenames: false
|
||||||
description: Runs cargo fmt on the codebase.
|
description: Runs cargo fmt on the codebase.
|
||||||
args: [develop, -c, cargo, fmt, --all]
|
args: [fmt]
|
||||||
- id: nac3-cargo-clippy
|
- id: nac3-cargo-clippy
|
||||||
name: nac3 cargo clippy
|
name: nac3 cargo clippy
|
||||||
entry: nix
|
entry: cargo
|
||||||
language: system
|
language: system
|
||||||
types: [file, rust]
|
types: [file, rust]
|
||||||
pass_filenames: false
|
pass_filenames: false
|
||||||
description: Runs cargo clippy on the codebase.
|
description: Runs cargo clippy on the codebase.
|
||||||
args: [develop, -c, cargo, clippy, --tests]
|
args: [clippy, --tests]
|
||||||
|
505
Cargo.lock
generated
505
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
6
flake.lock
generated
6
flake.lock
generated
@ -2,11 +2,11 @@
|
|||||||
"nodes": {
|
"nodes": {
|
||||||
"nixpkgs": {
|
"nixpkgs": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1727348695,
|
"lastModified": 1718530797,
|
||||||
"narHash": "sha256-J+PeFKSDV+pHL7ukkfpVzCOO7mBSrrpJ3svwBFABbhI=",
|
"narHash": "sha256-pup6cYwtgvzDpvpSCFh1TEUjw2zkNpk8iolbKnyFmmU=",
|
||||||
"owner": "NixOS",
|
"owner": "NixOS",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "1925c603f17fc89f4c8f6bf6f631a802ad85d784",
|
"rev": "b60ebf54c15553b393d144357375ea956f89e9a9",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
|
30
flake.nix
30
flake.nix
@ -6,7 +6,6 @@
|
|||||||
outputs = { self, nixpkgs }:
|
outputs = { self, nixpkgs }:
|
||||||
let
|
let
|
||||||
pkgs = import nixpkgs { system = "x86_64-linux"; };
|
pkgs = import nixpkgs { system = "x86_64-linux"; };
|
||||||
pkgs32 = import nixpkgs { system = "i686-linux"; };
|
|
||||||
in rec {
|
in rec {
|
||||||
packages.x86_64-linux = rec {
|
packages.x86_64-linux = rec {
|
||||||
llvm-nac3 = pkgs.callPackage ./nix/llvm {};
|
llvm-nac3 = pkgs.callPackage ./nix/llvm {};
|
||||||
@ -16,22 +15,6 @@
|
|||||||
ln -s ${pkgs.llvmPackages_14.clang-unwrapped}/bin/clang $out/bin/clang-irrt
|
ln -s ${pkgs.llvmPackages_14.clang-unwrapped}/bin/clang $out/bin/clang-irrt
|
||||||
ln -s ${pkgs.llvmPackages_14.llvm.out}/bin/llvm-as $out/bin/llvm-as-irrt
|
ln -s ${pkgs.llvmPackages_14.llvm.out}/bin/llvm-as $out/bin/llvm-as-irrt
|
||||||
'';
|
'';
|
||||||
demo-linalg-stub = pkgs.rustPlatform.buildRustPackage {
|
|
||||||
name = "demo-linalg-stub";
|
|
||||||
src = ./nac3standalone/demo/linalg;
|
|
||||||
cargoLock = {
|
|
||||||
lockFile = ./nac3standalone/demo/linalg/Cargo.lock;
|
|
||||||
};
|
|
||||||
doCheck = false;
|
|
||||||
};
|
|
||||||
demo-linalg-stub32 = pkgs32.rustPlatform.buildRustPackage {
|
|
||||||
name = "demo-linalg-stub32";
|
|
||||||
src = ./nac3standalone/demo/linalg;
|
|
||||||
cargoLock = {
|
|
||||||
lockFile = ./nac3standalone/demo/linalg/Cargo.lock;
|
|
||||||
};
|
|
||||||
doCheck = false;
|
|
||||||
};
|
|
||||||
nac3artiq = pkgs.python3Packages.toPythonModule (
|
nac3artiq = pkgs.python3Packages.toPythonModule (
|
||||||
pkgs.rustPlatform.buildRustPackage rec {
|
pkgs.rustPlatform.buildRustPackage rec {
|
||||||
name = "nac3artiq";
|
name = "nac3artiq";
|
||||||
@ -41,7 +24,7 @@
|
|||||||
lockFile = ./Cargo.lock;
|
lockFile = ./Cargo.lock;
|
||||||
};
|
};
|
||||||
passthru.cargoLock = cargoLock;
|
passthru.cargoLock = cargoLock;
|
||||||
nativeBuildInputs = [ pkgs.python3 (pkgs.wrapClangMulti pkgs.llvmPackages_14.clang) llvm-tools-irrt pkgs.llvmPackages_14.llvm.out llvm-nac3 ];
|
nativeBuildInputs = [ pkgs.python3 pkgs.llvmPackages_14.clang llvm-tools-irrt pkgs.llvmPackages_14.llvm.out llvm-nac3 ];
|
||||||
buildInputs = [ pkgs.python3 llvm-nac3 ];
|
buildInputs = [ pkgs.python3 llvm-nac3 ];
|
||||||
checkInputs = [ (pkgs.python3.withPackages(ps: [ ps.numpy ps.scipy ])) ];
|
checkInputs = [ (pkgs.python3.withPackages(ps: [ ps.numpy ps.scipy ])) ];
|
||||||
checkPhase =
|
checkPhase =
|
||||||
@ -49,9 +32,7 @@
|
|||||||
echo "Checking nac3standalone demos..."
|
echo "Checking nac3standalone demos..."
|
||||||
pushd nac3standalone/demo
|
pushd nac3standalone/demo
|
||||||
patchShebangs .
|
patchShebangs .
|
||||||
export DEMO_LINALG_STUB=${demo-linalg-stub}/lib/liblinalg.a
|
./check_demos.sh
|
||||||
export DEMO_LINALG_STUB32=${demo-linalg-stub32}/lib/liblinalg.a
|
|
||||||
./check_demos.sh -i686
|
|
||||||
popd
|
popd
|
||||||
echo "Running Cargo tests..."
|
echo "Running Cargo tests..."
|
||||||
cargoCheckHook
|
cargoCheckHook
|
||||||
@ -168,7 +149,7 @@
|
|||||||
buildInputs = with pkgs; [
|
buildInputs = with pkgs; [
|
||||||
# build dependencies
|
# build dependencies
|
||||||
packages.x86_64-linux.llvm-nac3
|
packages.x86_64-linux.llvm-nac3
|
||||||
(pkgs.wrapClangMulti llvmPackages_14.clang) llvmPackages_14.llvm.out # for running nac3standalone demos
|
llvmPackages_14.clang llvmPackages_14.llvm.out # for running nac3standalone demos
|
||||||
packages.x86_64-linux.llvm-tools-irrt
|
packages.x86_64-linux.llvm-tools-irrt
|
||||||
cargo
|
cargo
|
||||||
rustc
|
rustc
|
||||||
@ -181,11 +162,6 @@
|
|||||||
pre-commit
|
pre-commit
|
||||||
rustfmt
|
rustfmt
|
||||||
];
|
];
|
||||||
shellHook =
|
|
||||||
''
|
|
||||||
export DEMO_LINALG_STUB=${packages.x86_64-linux.demo-linalg-stub}/lib/liblinalg.a
|
|
||||||
export DEMO_LINALG_STUB32=${packages.x86_64-linux.demo-linalg-stub32}/lib/liblinalg.a
|
|
||||||
'';
|
|
||||||
};
|
};
|
||||||
devShells.x86_64-linux.msys2 = pkgs.mkShell {
|
devShells.x86_64-linux.msys2 = pkgs.mkShell {
|
||||||
name = "nac3-dev-shell-msys2";
|
name = "nac3-dev-shell-msys2";
|
||||||
|
@ -12,10 +12,15 @@ crate-type = ["cdylib"]
|
|||||||
itertools = "0.13"
|
itertools = "0.13"
|
||||||
pyo3 = { version = "0.21", features = ["extension-module", "gil-refs"] }
|
pyo3 = { version = "0.21", features = ["extension-module", "gil-refs"] }
|
||||||
parking_lot = "0.12"
|
parking_lot = "0.12"
|
||||||
tempfile = "3.13"
|
tempfile = "3.10"
|
||||||
|
nac3parser = { path = "../nac3parser" }
|
||||||
nac3core = { path = "../nac3core" }
|
nac3core = { path = "../nac3core" }
|
||||||
nac3ld = { path = "../nac3ld" }
|
nac3ld = { path = "../nac3ld" }
|
||||||
|
|
||||||
|
[dependencies.inkwell]
|
||||||
|
version = "0.4"
|
||||||
|
default-features = false
|
||||||
|
features = ["llvm14-0", "target-x86", "target-arm", "target-riscv", "no-libffi-linking"]
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
init-llvm-profile = []
|
init-llvm-profile = []
|
||||||
no-escape-analysis = ["nac3core/no-escape-analysis"]
|
|
||||||
|
@ -1,24 +0,0 @@
|
|||||||
from min_artiq import *
|
|
||||||
from numpy import int32
|
|
||||||
|
|
||||||
|
|
||||||
@nac3
|
|
||||||
class EmptyList:
|
|
||||||
core: KernelInvariant[Core]
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
self.core = Core()
|
|
||||||
|
|
||||||
@rpc
|
|
||||||
def get_empty(self) -> list[int32]:
|
|
||||||
return []
|
|
||||||
|
|
||||||
@kernel
|
|
||||||
def run(self):
|
|
||||||
a: list[int32] = self.get_empty()
|
|
||||||
if a != []:
|
|
||||||
raise ValueError
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
EmptyList().run()
|
|
@ -112,15 +112,10 @@ def extern(function):
|
|||||||
register_function(function)
|
register_function(function)
|
||||||
return function
|
return function
|
||||||
|
|
||||||
|
def rpc(function):
|
||||||
def rpc(arg=None, flags={}):
|
"""Decorates a function declaration defined by the core device runtime."""
|
||||||
"""Decorates a function or method to be executed on the host interpreter."""
|
register_function(function)
|
||||||
if arg is None:
|
return function
|
||||||
def inner_decorator(function):
|
|
||||||
return rpc(function, flags)
|
|
||||||
return inner_decorator
|
|
||||||
register_function(arg)
|
|
||||||
return arg
|
|
||||||
|
|
||||||
def kernel(function_or_method):
|
def kernel(function_or_method):
|
||||||
"""Decorates a function or method to be executed on the core device."""
|
"""Decorates a function or method to be executed on the core device."""
|
||||||
|
@ -1,26 +0,0 @@
|
|||||||
from min_artiq import *
|
|
||||||
from numpy import ndarray, zeros as np_zeros
|
|
||||||
|
|
||||||
|
|
||||||
@nac3
|
|
||||||
class StrFail:
|
|
||||||
core: KernelInvariant[Core]
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
self.core = Core()
|
|
||||||
|
|
||||||
@kernel
|
|
||||||
def hello(self, arg: str):
|
|
||||||
pass
|
|
||||||
|
|
||||||
@kernel
|
|
||||||
def consume_ndarray(self, arg: ndarray[str, 1]):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def run(self):
|
|
||||||
self.hello("world")
|
|
||||||
self.consume_ndarray(np_zeros([10], dtype=str))
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
StrFail().run()
|
|
@ -1,40 +0,0 @@
|
|||||||
from min_artiq import *
|
|
||||||
from numpy import int32
|
|
||||||
|
|
||||||
|
|
||||||
@nac3
|
|
||||||
class Demo:
|
|
||||||
attr1: KernelInvariant[int32] = 2
|
|
||||||
attr2: int32 = 4
|
|
||||||
attr3: Kernel[int32]
|
|
||||||
|
|
||||||
@kernel
|
|
||||||
def __init__(self):
|
|
||||||
self.attr3 = 8
|
|
||||||
|
|
||||||
|
|
||||||
@nac3
|
|
||||||
class NAC3Devices:
|
|
||||||
core: KernelInvariant[Core]
|
|
||||||
attr4: KernelInvariant[int32] = 16
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
self.core = Core()
|
|
||||||
|
|
||||||
@kernel
|
|
||||||
def run(self):
|
|
||||||
Demo.attr1 # Supported
|
|
||||||
# Demo.attr2 # Field not accessible on Kernel
|
|
||||||
# Demo.attr3 # Only attributes can be accessed in this way
|
|
||||||
# Demo.attr1 = 2 # Attributes are immutable
|
|
||||||
|
|
||||||
self.attr4 # Attributes can be accessed within class
|
|
||||||
|
|
||||||
obj = Demo()
|
|
||||||
obj.attr1 # Attributes can be accessed by class objects
|
|
||||||
|
|
||||||
NAC3Devices.attr4 # Attributes accessible for classes without __init__
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
NAC3Devices().run()
|
|
File diff suppressed because it is too large
Load Diff
@ -16,67 +16,63 @@
|
|||||||
clippy::wildcard_imports
|
clippy::wildcard_imports
|
||||||
)]
|
)]
|
||||||
|
|
||||||
use std::{
|
use std::collections::{HashMap, HashSet};
|
||||||
collections::{HashMap, HashSet},
|
use std::fs;
|
||||||
fs,
|
use std::io::Write;
|
||||||
io::Write,
|
use std::process::Command;
|
||||||
process::Command,
|
use std::rc::Rc;
|
||||||
rc::Rc,
|
use std::sync::Arc;
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
use itertools::Itertools;
|
use inkwell::{
|
||||||
use parking_lot::{Mutex, RwLock};
|
|
||||||
use pyo3::{
|
|
||||||
create_exception, exceptions,
|
|
||||||
prelude::*,
|
|
||||||
types::{PyBytes, PyDict, PySet},
|
|
||||||
};
|
|
||||||
use tempfile::{self, TempDir};
|
|
||||||
|
|
||||||
use nac3core::{
|
|
||||||
codegen::{
|
|
||||||
concrete_type::ConcreteTypeStore, gen_func_impl, irrt::load_irrt, CodeGenLLVMOptions,
|
|
||||||
CodeGenTargetMachineOptions, CodeGenTask, WithCall, WorkerRegistry,
|
|
||||||
},
|
|
||||||
inkwell::{
|
|
||||||
context::Context,
|
|
||||||
memory_buffer::MemoryBuffer,
|
memory_buffer::MemoryBuffer,
|
||||||
module::{Linkage, Module},
|
module::{Linkage, Module},
|
||||||
passes::PassBuilderOptions,
|
passes::PassBuilderOptions,
|
||||||
support::is_multithreaded,
|
support::is_multithreaded,
|
||||||
targets::*,
|
targets::*,
|
||||||
OptimizationLevel,
|
OptimizationLevel,
|
||||||
},
|
};
|
||||||
nac3parser::{
|
use itertools::Itertools;
|
||||||
ast::{Constant, ExprKind, Located, Stmt, StmtKind, StrRef},
|
use nac3core::codegen::{gen_func_impl, CodeGenLLVMOptions, CodeGenTargetMachineOptions};
|
||||||
|
use nac3core::toplevel::builtins::get_exn_constructor;
|
||||||
|
use nac3core::typecheck::typedef::{TypeEnum, Unifier, VarMap};
|
||||||
|
use nac3parser::{
|
||||||
|
ast::{ExprKind, Stmt, StmtKind, StrRef},
|
||||||
parser::parse_program,
|
parser::parse_program,
|
||||||
},
|
};
|
||||||
|
use pyo3::create_exception;
|
||||||
|
use pyo3::prelude::*;
|
||||||
|
use pyo3::{exceptions, types::PyBytes, types::PyDict, types::PySet};
|
||||||
|
|
||||||
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
|
||||||
|
use nac3core::{
|
||||||
|
codegen::irrt::load_irrt,
|
||||||
|
codegen::{concrete_type::ConcreteTypeStore, CodeGenTask, WithCall, WorkerRegistry},
|
||||||
symbol_resolver::SymbolResolver,
|
symbol_resolver::SymbolResolver,
|
||||||
toplevel::{
|
toplevel::{
|
||||||
builtins::get_exn_constructor,
|
composer::{ComposerConfig, TopLevelComposer},
|
||||||
composer::{BuiltinFuncCreator, BuiltinFuncSpec, ComposerConfig, TopLevelComposer},
|
|
||||||
DefinitionId, GenCall, TopLevelDef,
|
DefinitionId, GenCall, TopLevelDef,
|
||||||
},
|
},
|
||||||
typecheck::{
|
typecheck::typedef::{FunSignature, FuncArg},
|
||||||
type_inferencer::PrimitiveStore,
|
typecheck::{type_inferencer::PrimitiveStore, typedef::Type},
|
||||||
typedef::{into_var_map, FunSignature, FuncArg, Type, TypeEnum, Unifier, VarMap},
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use nac3ld::Linker;
|
use nac3ld::Linker;
|
||||||
|
|
||||||
|
use tempfile::{self, TempDir};
|
||||||
|
|
||||||
|
use crate::codegen::attributes_writeback;
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{
|
codegen::{rpc_codegen_callback, ArtiqCodeGenerator},
|
||||||
attributes_writeback, gen_core_log, gen_rtio_log, rpc_codegen_callback, ArtiqCodeGenerator,
|
|
||||||
},
|
|
||||||
symbol_resolver::{DeferredEvaluationStore, InnerResolver, PythonHelper, Resolver},
|
symbol_resolver::{DeferredEvaluationStore, InnerResolver, PythonHelper, Resolver},
|
||||||
timeline::TimeFns,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
mod codegen;
|
mod codegen;
|
||||||
mod symbol_resolver;
|
mod symbol_resolver;
|
||||||
mod timeline;
|
mod timeline;
|
||||||
|
|
||||||
|
use timeline::TimeFns;
|
||||||
|
|
||||||
#[derive(PartialEq, Clone, Copy)]
|
#[derive(PartialEq, Clone, Copy)]
|
||||||
enum Isa {
|
enum Isa {
|
||||||
Host,
|
Host,
|
||||||
@ -130,7 +126,7 @@ struct Nac3 {
|
|||||||
isa: Isa,
|
isa: Isa,
|
||||||
time_fns: &'static (dyn TimeFns + Sync),
|
time_fns: &'static (dyn TimeFns + Sync),
|
||||||
primitive: PrimitiveStore,
|
primitive: PrimitiveStore,
|
||||||
builtins: Vec<BuiltinFuncSpec>,
|
builtins: Vec<(StrRef, FunSignature, Arc<GenCall>)>,
|
||||||
pyid_to_def: Arc<RwLock<HashMap<u64, DefinitionId>>>,
|
pyid_to_def: Arc<RwLock<HashMap<u64, DefinitionId>>>,
|
||||||
primitive_ids: PrimitivePythonId,
|
primitive_ids: PrimitivePythonId,
|
||||||
working_directory: TempDir,
|
working_directory: TempDir,
|
||||||
@ -197,8 +193,10 @@ impl Nac3 {
|
|||||||
body.retain(|stmt| {
|
body.retain(|stmt| {
|
||||||
if let StmtKind::FunctionDef { ref decorator_list, .. } = stmt.node {
|
if let StmtKind::FunctionDef { ref decorator_list, .. } = stmt.node {
|
||||||
decorator_list.iter().any(|decorator| {
|
decorator_list.iter().any(|decorator| {
|
||||||
if let Some(id) = decorator_id_string(decorator) {
|
if let ExprKind::Name { id, .. } = decorator.node {
|
||||||
id == "kernel" || id == "portable" || id == "rpc"
|
id.to_string() == "kernel"
|
||||||
|
|| id.to_string() == "portable"
|
||||||
|
|| id.to_string() == "rpc"
|
||||||
} else {
|
} else {
|
||||||
false
|
false
|
||||||
}
|
}
|
||||||
@ -211,8 +209,9 @@ impl Nac3 {
|
|||||||
}
|
}
|
||||||
StmtKind::FunctionDef { ref decorator_list, .. } => {
|
StmtKind::FunctionDef { ref decorator_list, .. } => {
|
||||||
decorator_list.iter().any(|decorator| {
|
decorator_list.iter().any(|decorator| {
|
||||||
if let Some(id) = decorator_id_string(decorator) {
|
if let ExprKind::Name { id, .. } = decorator.node {
|
||||||
id == "extern" || id == "kernel" || id == "portable" || id == "rpc"
|
let id = id.to_string();
|
||||||
|
id == "extern" || id == "portable" || id == "kernel" || id == "rpc"
|
||||||
} else {
|
} else {
|
||||||
false
|
false
|
||||||
}
|
}
|
||||||
@ -265,7 +264,7 @@ impl Nac3 {
|
|||||||
arg_names.len(),
|
arg_names.len(),
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
for (i, FuncArg { ty, default_value, name, .. }) in args.iter().enumerate() {
|
for (i, FuncArg { ty, default_value, name }) in args.iter().enumerate() {
|
||||||
let in_name = match arg_names.get(i) {
|
let in_name = match arg_names.get(i) {
|
||||||
Some(n) => n,
|
Some(n) => n,
|
||||||
None if default_value.is_none() => {
|
None if default_value.is_none() => {
|
||||||
@ -301,64 +300,6 @@ impl Nac3 {
|
|||||||
None
|
None
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns a [`Vec`] of builtins that needs to be initialized during method compilation time.
|
|
||||||
fn get_lateinit_builtins() -> Vec<Box<BuiltinFuncCreator>> {
|
|
||||||
vec![
|
|
||||||
Box::new(|primitives, unifier| {
|
|
||||||
let arg_ty = unifier.get_fresh_var(Some("T".into()), None);
|
|
||||||
|
|
||||||
(
|
|
||||||
"core_log".into(),
|
|
||||||
FunSignature {
|
|
||||||
args: vec![FuncArg {
|
|
||||||
name: "arg".into(),
|
|
||||||
ty: arg_ty.ty,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
}],
|
|
||||||
ret: primitives.none,
|
|
||||||
vars: into_var_map([arg_ty]),
|
|
||||||
},
|
|
||||||
Arc::new(GenCall::new(Box::new(move |ctx, obj, fun, args, generator| {
|
|
||||||
gen_core_log(ctx, &obj, fun, &args, generator)?;
|
|
||||||
|
|
||||||
Ok(None)
|
|
||||||
}))),
|
|
||||||
)
|
|
||||||
}),
|
|
||||||
Box::new(|primitives, unifier| {
|
|
||||||
let arg_ty = unifier.get_fresh_var(Some("T".into()), None);
|
|
||||||
|
|
||||||
(
|
|
||||||
"rtio_log".into(),
|
|
||||||
FunSignature {
|
|
||||||
args: vec![
|
|
||||||
FuncArg {
|
|
||||||
name: "channel".into(),
|
|
||||||
ty: primitives.str,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
},
|
|
||||||
FuncArg {
|
|
||||||
name: "arg".into(),
|
|
||||||
ty: arg_ty.ty,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
ret: primitives.none,
|
|
||||||
vars: into_var_map([arg_ty]),
|
|
||||||
},
|
|
||||||
Arc::new(GenCall::new(Box::new(move |ctx, obj, fun, args, generator| {
|
|
||||||
gen_rtio_log(ctx, &obj, fun, &args, generator)?;
|
|
||||||
|
|
||||||
Ok(None)
|
|
||||||
}))),
|
|
||||||
)
|
|
||||||
}),
|
|
||||||
]
|
|
||||||
}
|
|
||||||
|
|
||||||
fn compile_method<T>(
|
fn compile_method<T>(
|
||||||
&self,
|
&self,
|
||||||
obj: &PyAny,
|
obj: &PyAny,
|
||||||
@ -371,7 +312,6 @@ impl Nac3 {
|
|||||||
let size_t = self.isa.get_size_type();
|
let size_t = self.isa.get_size_type();
|
||||||
let (mut composer, mut builtins_def, mut builtins_ty) = TopLevelComposer::new(
|
let (mut composer, mut builtins_def, mut builtins_ty) = TopLevelComposer::new(
|
||||||
self.builtins.clone(),
|
self.builtins.clone(),
|
||||||
Self::get_lateinit_builtins(),
|
|
||||||
ComposerConfig { kernel_ann: Some("Kernel"), kernel_invariant_ann: "KernelInvariant" },
|
ComposerConfig { kernel_ann: Some("Kernel"), kernel_invariant_ann: "KernelInvariant" },
|
||||||
size_t,
|
size_t,
|
||||||
);
|
);
|
||||||
@ -448,6 +388,7 @@ impl Nac3 {
|
|||||||
pyid_to_type: pyid_to_type.clone(),
|
pyid_to_type: pyid_to_type.clone(),
|
||||||
primitive_ids: self.primitive_ids.clone(),
|
primitive_ids: self.primitive_ids.clone(),
|
||||||
global_value_ids: global_value_ids.clone(),
|
global_value_ids: global_value_ids.clone(),
|
||||||
|
class_names: Mutex::default(),
|
||||||
name_to_pyid: name_to_pyid.clone(),
|
name_to_pyid: name_to_pyid.clone(),
|
||||||
module: module.clone(),
|
module: module.clone(),
|
||||||
id_to_pyval: RwLock::default(),
|
id_to_pyval: RwLock::default(),
|
||||||
@ -478,25 +419,9 @@ impl Nac3 {
|
|||||||
|
|
||||||
match &stmt.node {
|
match &stmt.node {
|
||||||
StmtKind::FunctionDef { decorator_list, .. } => {
|
StmtKind::FunctionDef { decorator_list, .. } => {
|
||||||
if decorator_list
|
if decorator_list.iter().any(|decorator| matches!(decorator.node, ExprKind::Name { id, .. } if id == "rpc".into())) {
|
||||||
.iter()
|
store_fun.call1(py, (def_id.0.into_py(py), module.getattr(py, name.to_string().as_str()).unwrap())).unwrap();
|
||||||
.any(|decorator| decorator_id_string(decorator) == Some("rpc".to_string()))
|
rpc_ids.push((None, def_id));
|
||||||
{
|
|
||||||
store_fun
|
|
||||||
.call1(
|
|
||||||
py,
|
|
||||||
(
|
|
||||||
def_id.0.into_py(py),
|
|
||||||
module.getattr(py, name.to_string().as_str()).unwrap(),
|
|
||||||
),
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
let is_async = decorator_list.iter().any(|decorator| {
|
|
||||||
decorator_get_flags(decorator)
|
|
||||||
.iter()
|
|
||||||
.any(|constant| *constant == Constant::Str("async".into()))
|
|
||||||
});
|
|
||||||
rpc_ids.push((None, def_id, is_async));
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
StmtKind::ClassDef { name, body, .. } => {
|
StmtKind::ClassDef { name, body, .. } => {
|
||||||
@ -504,26 +429,19 @@ impl Nac3 {
|
|||||||
let class_obj = module.getattr(py, class_name.as_str()).unwrap();
|
let class_obj = module.getattr(py, class_name.as_str()).unwrap();
|
||||||
for stmt in body {
|
for stmt in body {
|
||||||
if let StmtKind::FunctionDef { name, decorator_list, .. } = &stmt.node {
|
if let StmtKind::FunctionDef { name, decorator_list, .. } = &stmt.node {
|
||||||
if decorator_list.iter().any(|decorator| {
|
if decorator_list.iter().any(|decorator| matches!(decorator.node, ExprKind::Name { id, .. } if id == "rpc".into())) {
|
||||||
decorator_id_string(decorator) == Some("rpc".to_string())
|
|
||||||
}) {
|
|
||||||
let is_async = decorator_list.iter().any(|decorator| {
|
|
||||||
decorator_get_flags(decorator)
|
|
||||||
.iter()
|
|
||||||
.any(|constant| *constant == Constant::Str("async".into()))
|
|
||||||
});
|
|
||||||
if name == &"__init__".into() {
|
if name == &"__init__".into() {
|
||||||
return Err(CompileError::new_err(format!(
|
return Err(CompileError::new_err(format!(
|
||||||
"compilation failed\n----------\nThe constructor of class {} should not be decorated with rpc decorator (at {})",
|
"compilation failed\n----------\nThe constructor of class {} should not be decorated with rpc decorator (at {})",
|
||||||
class_name, stmt.location
|
class_name, stmt.location
|
||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
rpc_ids.push((Some((class_obj.clone(), *name)), def_id, is_async));
|
rpc_ids.push((Some((class_obj.clone(), *name)), def_id));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
_ => (),
|
_ => ()
|
||||||
}
|
}
|
||||||
|
|
||||||
let id = *name_to_pyid.get(&name).unwrap();
|
let id = *name_to_pyid.get(&name).unwrap();
|
||||||
@ -562,6 +480,7 @@ impl Nac3 {
|
|||||||
pyid_to_type: pyid_to_type.clone(),
|
pyid_to_type: pyid_to_type.clone(),
|
||||||
primitive_ids: self.primitive_ids.clone(),
|
primitive_ids: self.primitive_ids.clone(),
|
||||||
global_value_ids: global_value_ids.clone(),
|
global_value_ids: global_value_ids.clone(),
|
||||||
|
class_names: Mutex::default(),
|
||||||
id_to_pyval: RwLock::default(),
|
id_to_pyval: RwLock::default(),
|
||||||
id_to_primitive: RwLock::default(),
|
id_to_primitive: RwLock::default(),
|
||||||
field_to_val: RwLock::default(),
|
field_to_val: RwLock::default(),
|
||||||
@ -578,10 +497,6 @@ impl Nac3 {
|
|||||||
.register_top_level(synthesized.pop().unwrap(), Some(resolver.clone()), "", false)
|
.register_top_level(synthesized.pop().unwrap(), Some(resolver.clone()), "", false)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
// Process IRRT
|
|
||||||
let context = Context::create();
|
|
||||||
let irrt = load_irrt(&context, resolver.as_ref());
|
|
||||||
|
|
||||||
let fun_signature =
|
let fun_signature =
|
||||||
FunSignature { args: vec![], ret: self.primitive.none, vars: VarMap::new() };
|
FunSignature { args: vec![], ret: self.primitive.none, vars: VarMap::new() };
|
||||||
let mut store = ConcreteTypeStore::new();
|
let mut store = ConcreteTypeStore::new();
|
||||||
@ -619,12 +534,13 @@ impl Nac3 {
|
|||||||
let top_level = Arc::new(composer.make_top_level_context());
|
let top_level = Arc::new(composer.make_top_level_context());
|
||||||
|
|
||||||
{
|
{
|
||||||
|
let rpc_codegen = rpc_codegen_callback();
|
||||||
let defs = top_level.definitions.read();
|
let defs = top_level.definitions.read();
|
||||||
for (class_data, id, is_async) in &rpc_ids {
|
for (class_data, id) in &rpc_ids {
|
||||||
let mut def = defs[id.0].write();
|
let mut def = defs[id.0].write();
|
||||||
match &mut *def {
|
match &mut *def {
|
||||||
TopLevelDef::Function { codegen_callback, .. } => {
|
TopLevelDef::Function { codegen_callback, .. } => {
|
||||||
*codegen_callback = Some(rpc_codegen_callback(*is_async));
|
*codegen_callback = Some(rpc_codegen.clone());
|
||||||
}
|
}
|
||||||
TopLevelDef::Class { methods, .. } => {
|
TopLevelDef::Class { methods, .. } => {
|
||||||
let (class_def, method_name) = class_data.as_ref().unwrap();
|
let (class_def, method_name) = class_data.as_ref().unwrap();
|
||||||
@ -635,7 +551,7 @@ impl Nac3 {
|
|||||||
if let TopLevelDef::Function { codegen_callback, .. } =
|
if let TopLevelDef::Function { codegen_callback, .. } =
|
||||||
&mut *defs[id.0].write()
|
&mut *defs[id.0].write()
|
||||||
{
|
{
|
||||||
*codegen_callback = Some(rpc_codegen_callback(*is_async));
|
*codegen_callback = Some(rpc_codegen.clone());
|
||||||
store_fun
|
store_fun
|
||||||
.call1(
|
.call1(
|
||||||
py,
|
py,
|
||||||
@ -650,11 +566,6 @@ impl Nac3 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
TopLevelDef::Variable { .. } => {
|
|
||||||
return Err(CompileError::new_err(String::from(
|
|
||||||
"Unsupported @rpc annotation on global variable",
|
|
||||||
)))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -714,9 +625,7 @@ impl Nac3 {
|
|||||||
let buffer = buffer.as_slice().into();
|
let buffer = buffer.as_slice().into();
|
||||||
membuffer.lock().push(buffer);
|
membuffer.lock().push(buffer);
|
||||||
})));
|
})));
|
||||||
let size_t = context
|
let size_t = if self.isa == Isa::Host { 64 } else { 32 };
|
||||||
.ptr_sized_int_type(&self.get_llvm_target_machine().get_target_data(), None)
|
|
||||||
.get_bit_width();
|
|
||||||
let num_threads = if is_multithreaded() { 4 } else { 1 };
|
let num_threads = if is_multithreaded() { 4 } else { 1 };
|
||||||
let thread_names: Vec<String> = (0..num_threads).map(|_| "main".to_string()).collect();
|
let thread_names: Vec<String> = (0..num_threads).map(|_| "main".to_string()).collect();
|
||||||
let threads: Vec<_> = thread_names
|
let threads: Vec<_> = thread_names
|
||||||
@ -733,11 +642,8 @@ impl Nac3 {
|
|||||||
|
|
||||||
let mut generator =
|
let mut generator =
|
||||||
ArtiqCodeGenerator::new("attributes_writeback".to_string(), size_t, self.time_fns);
|
ArtiqCodeGenerator::new("attributes_writeback".to_string(), size_t, self.time_fns);
|
||||||
let context = Context::create();
|
let context = inkwell::context::Context::create();
|
||||||
let module = context.create_module("attributes_writeback");
|
let module = context.create_module("attributes_writeback");
|
||||||
let target_machine = self.llvm_options.create_target_machine().unwrap();
|
|
||||||
module.set_data_layout(&target_machine.get_target_data().get_data_layout());
|
|
||||||
module.set_triple(&target_machine.get_triple());
|
|
||||||
let builder = context.create_builder();
|
let builder = context.create_builder();
|
||||||
let (_, module, _) = gen_func_impl(
|
let (_, module, _) = gen_func_impl(
|
||||||
&context,
|
&context,
|
||||||
@ -756,7 +662,7 @@ impl Nac3 {
|
|||||||
membuffer.lock().push(buffer);
|
membuffer.lock().push(buffer);
|
||||||
});
|
});
|
||||||
|
|
||||||
// Link all modules into `main`.
|
let context = inkwell::context::Context::create();
|
||||||
let buffers = membuffers.lock();
|
let buffers = membuffers.lock();
|
||||||
let main = context
|
let main = context
|
||||||
.create_module_from_ir(MemoryBuffer::create_from_memory_range(&buffers[0], "main"))
|
.create_module_from_ir(MemoryBuffer::create_from_memory_range(&buffers[0], "main"))
|
||||||
@ -785,7 +691,8 @@ impl Nac3 {
|
|||||||
)
|
)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
main.link_in_module(irrt).map_err(|err| CompileError::new_err(err.to_string()))?;
|
main.link_in_module(load_irrt(&context))
|
||||||
|
.map_err(|err| CompileError::new_err(err.to_string()))?;
|
||||||
|
|
||||||
let mut function_iter = main.get_first_function();
|
let mut function_iter = main.get_first_function();
|
||||||
while let Some(func) = function_iter {
|
while let Some(func) = function_iter {
|
||||||
@ -871,41 +778,6 @@ impl Nac3 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Retrieves the Name.id from a decorator, supports decorators with arguments.
|
|
||||||
fn decorator_id_string(decorator: &Located<ExprKind>) -> Option<String> {
|
|
||||||
if let ExprKind::Name { id, .. } = decorator.node {
|
|
||||||
// Bare decorator
|
|
||||||
return Some(id.to_string());
|
|
||||||
} else if let ExprKind::Call { func, .. } = &decorator.node {
|
|
||||||
// Decorators that are calls (e.g. "@rpc()") have Call for the node,
|
|
||||||
// need to extract the id from within.
|
|
||||||
if let ExprKind::Name { id, .. } = func.node {
|
|
||||||
return Some(id.to_string());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
None
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Retrieves flags from a decorator, if any.
|
|
||||||
fn decorator_get_flags(decorator: &Located<ExprKind>) -> Vec<Constant> {
|
|
||||||
let mut flags = vec![];
|
|
||||||
if let ExprKind::Call { keywords, .. } = &decorator.node {
|
|
||||||
for keyword in keywords {
|
|
||||||
if keyword.node.arg != Some("flags".into()) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
if let ExprKind::Set { elts } = &keyword.node.value.node {
|
|
||||||
for elt in elts {
|
|
||||||
if let ExprKind::Constant { value, .. } = &elt.node {
|
|
||||||
flags.push(value.clone());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
flags
|
|
||||||
}
|
|
||||||
|
|
||||||
fn link_with_lld(elf_filename: String, obj_filename: String) -> PyResult<()> {
|
fn link_with_lld(elf_filename: String, obj_filename: String) -> PyResult<()> {
|
||||||
let linker_args = vec![
|
let linker_args = vec![
|
||||||
"-shared".to_string(),
|
"-shared".to_string(),
|
||||||
@ -975,7 +847,7 @@ impl Nac3 {
|
|||||||
Isa::RiscV32IMA => &timeline::NOW_PINNING_TIME_FNS,
|
Isa::RiscV32IMA => &timeline::NOW_PINNING_TIME_FNS,
|
||||||
Isa::CortexA9 | Isa::Host => &timeline::EXTERN_TIME_FNS,
|
Isa::CortexA9 | Isa::Host => &timeline::EXTERN_TIME_FNS,
|
||||||
};
|
};
|
||||||
let (primitive, _) = TopLevelComposer::make_primitives(isa.get_size_type());
|
let primitive: PrimitiveStore = TopLevelComposer::make_primitives(isa.get_size_type()).0;
|
||||||
let builtins = vec![
|
let builtins = vec![
|
||||||
(
|
(
|
||||||
"now_mu".into(),
|
"now_mu".into(),
|
||||||
@ -991,7 +863,6 @@ impl Nac3 {
|
|||||||
name: "t".into(),
|
name: "t".into(),
|
||||||
ty: primitive.int64,
|
ty: primitive.int64,
|
||||||
default_value: None,
|
default_value: None,
|
||||||
is_vararg: false,
|
|
||||||
}],
|
}],
|
||||||
ret: primitive.none,
|
ret: primitive.none,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
@ -1011,7 +882,6 @@ impl Nac3 {
|
|||||||
name: "dt".into(),
|
name: "dt".into(),
|
||||||
ty: primitive.int64,
|
ty: primitive.int64,
|
||||||
default_value: None,
|
default_value: None,
|
||||||
is_vararg: false,
|
|
||||||
}],
|
}],
|
||||||
ret: primitive.none,
|
ret: primitive.none,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
|
@ -1,30 +1,14 @@
|
|||||||
use std::{
|
use inkwell::{
|
||||||
collections::{HashMap, HashSet},
|
types::{BasicType, BasicTypeEnum},
|
||||||
sync::{
|
values::BasicValueEnum,
|
||||||
atomic::{AtomicBool, Ordering::Relaxed},
|
AddressSpace,
|
||||||
Arc,
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use itertools::Itertools;
|
use itertools::Itertools;
|
||||||
use parking_lot::RwLock;
|
|
||||||
use pyo3::{
|
|
||||||
types::{PyDict, PyTuple},
|
|
||||||
PyAny, PyObject, PyResult, Python,
|
|
||||||
};
|
|
||||||
|
|
||||||
use nac3core::{
|
use nac3core::{
|
||||||
codegen::{
|
codegen::{
|
||||||
classes::{NDArrayType, ProxyType},
|
classes::{NDArrayType, ProxyType},
|
||||||
CodeGenContext, CodeGenerator,
|
CodeGenContext, CodeGenerator,
|
||||||
},
|
},
|
||||||
inkwell::{
|
|
||||||
module::Linkage,
|
|
||||||
types::{BasicType, BasicTypeEnum},
|
|
||||||
values::BasicValueEnum,
|
|
||||||
AddressSpace,
|
|
||||||
},
|
|
||||||
nac3parser::ast::{self, StrRef},
|
|
||||||
symbol_resolver::{StaticValue, SymbolResolver, SymbolValue, ValueEnum},
|
symbol_resolver::{StaticValue, SymbolResolver, SymbolValue, ValueEnum},
|
||||||
toplevel::{
|
toplevel::{
|
||||||
helper::PrimDef,
|
helper::PrimDef,
|
||||||
@ -36,6 +20,19 @@ use nac3core::{
|
|||||||
typedef::{into_var_map, iter_type_vars, Type, TypeEnum, TypeVar, Unifier, VarMap},
|
typedef::{into_var_map, iter_type_vars, Type, TypeEnum, TypeVar, Unifier, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use nac3parser::ast::{self, StrRef};
|
||||||
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
use pyo3::{
|
||||||
|
types::{PyDict, PyTuple},
|
||||||
|
PyAny, PyObject, PyResult, Python,
|
||||||
|
};
|
||||||
|
use std::{
|
||||||
|
collections::{HashMap, HashSet},
|
||||||
|
sync::{
|
||||||
|
atomic::{AtomicBool, Ordering::Relaxed},
|
||||||
|
Arc,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
use crate::PrimitivePythonId;
|
use crate::PrimitivePythonId;
|
||||||
|
|
||||||
@ -82,6 +79,7 @@ pub struct InnerResolver {
|
|||||||
pub id_to_primitive: RwLock<HashMap<u64, PrimitiveValue>>,
|
pub id_to_primitive: RwLock<HashMap<u64, PrimitiveValue>>,
|
||||||
pub field_to_val: RwLock<HashMap<ResolverField, Option<PyFieldHandle>>>,
|
pub field_to_val: RwLock<HashMap<ResolverField, Option<PyFieldHandle>>>,
|
||||||
pub global_value_ids: Arc<RwLock<HashMap<u64, PyObject>>>,
|
pub global_value_ids: Arc<RwLock<HashMap<u64, PyObject>>>,
|
||||||
|
pub class_names: Mutex<HashMap<StrRef, Type>>,
|
||||||
pub pyid_to_def: Arc<RwLock<HashMap<u64, DefinitionId>>>,
|
pub pyid_to_def: Arc<RwLock<HashMap<u64, DefinitionId>>>,
|
||||||
pub pyid_to_type: Arc<RwLock<HashMap<u64, Type>>>,
|
pub pyid_to_type: Arc<RwLock<HashMap<u64, Type>>>,
|
||||||
pub primitive_ids: PrimitivePythonId,
|
pub primitive_ids: PrimitivePythonId,
|
||||||
@ -135,8 +133,6 @@ impl StaticValue for PythonValue {
|
|||||||
format!("{}_const", self.id).as_str(),
|
format!("{}_const", self.id).as_str(),
|
||||||
);
|
);
|
||||||
global.set_constant(true);
|
global.set_constant(true);
|
||||||
// Set linkage of global to private to avoid name collisions
|
|
||||||
global.set_linkage(Linkage::Private);
|
|
||||||
global.set_initializer(&ctx.ctx.const_struct(
|
global.set_initializer(&ctx.ctx.const_struct(
|
||||||
&[ctx.ctx.i32_type().const_int(u64::from(id), false).into()],
|
&[ctx.ctx.i32_type().const_int(u64::from(id), false).into()],
|
||||||
false,
|
false,
|
||||||
@ -167,7 +163,7 @@ impl StaticValue for PythonValue {
|
|||||||
PrimitiveValue::Bool(val) => {
|
PrimitiveValue::Bool(val) => {
|
||||||
ctx.ctx.i8_type().const_int(u64::from(*val), false).into()
|
ctx.ctx.i8_type().const_int(u64::from(*val), false).into()
|
||||||
}
|
}
|
||||||
PrimitiveValue::Str(val) => ctx.gen_string(generator, val).into(),
|
PrimitiveValue::Str(val) => ctx.ctx.const_string(val.as_bytes(), true).into(),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
if let Some(global) = ctx.module.get_global(&self.id.to_string()) {
|
if let Some(global) = ctx.module.get_global(&self.id.to_string()) {
|
||||||
@ -333,19 +329,8 @@ impl InnerResolver {
|
|||||||
Ok(Ok((primitives.exception, true)))
|
Ok(Ok((primitives.exception, true)))
|
||||||
} else if ty_id == self.primitive_ids.list {
|
} else if ty_id == self.primitive_ids.list {
|
||||||
// do not handle type var param and concrete check here
|
// do not handle type var param and concrete check here
|
||||||
let list_tvar = if let TypeEnum::TObj { obj_id, params, .. } =
|
|
||||||
&*unifier.get_ty_immutable(primitives.list)
|
|
||||||
{
|
|
||||||
assert_eq!(*obj_id, PrimDef::List.id());
|
|
||||||
iter_type_vars(params).nth(0).unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
let var = unifier.get_dummy_var().ty;
|
let var = unifier.get_dummy_var().ty;
|
||||||
let list = unifier
|
let list = unifier.add_ty(TypeEnum::TList { ty: var });
|
||||||
.subst(primitives.list, &into_var_map([TypeVar { id: list_tvar.id, ty: var }]))
|
|
||||||
.unwrap();
|
|
||||||
Ok(Ok((list, false)))
|
Ok(Ok((list, false)))
|
||||||
} else if ty_id == self.primitive_ids.ndarray {
|
} else if ty_id == self.primitive_ids.ndarray {
|
||||||
// do not handle type var param and concrete check here
|
// do not handle type var param and concrete check here
|
||||||
@ -355,7 +340,7 @@ impl InnerResolver {
|
|||||||
Ok(Ok((ndarray, false)))
|
Ok(Ok((ndarray, false)))
|
||||||
} else if ty_id == self.primitive_ids.tuple {
|
} else if ty_id == self.primitive_ids.tuple {
|
||||||
// do not handle type var param and concrete check here
|
// do not handle type var param and concrete check here
|
||||||
Ok(Ok((unifier.add_ty(TypeEnum::TTuple { ty: vec![], is_vararg_ctx: false }), false)))
|
Ok(Ok((unifier.add_ty(TypeEnum::TTuple { ty: vec![] }), false)))
|
||||||
} else if ty_id == self.primitive_ids.option {
|
} else if ty_id == self.primitive_ids.option {
|
||||||
Ok(Ok((primitives.option, false)))
|
Ok(Ok((primitives.option, false)))
|
||||||
} else if ty_id == self.primitive_ids.none {
|
} else if ty_id == self.primitive_ids.none {
|
||||||
@ -475,7 +460,7 @@ impl InnerResolver {
|
|||||||
};
|
};
|
||||||
|
|
||||||
match &*unifier.get_ty(origin_ty) {
|
match &*unifier.get_ty(origin_ty) {
|
||||||
TypeEnum::TObj { obj_id, .. } if *obj_id == PrimDef::List.id() => {
|
TypeEnum::TList { .. } => {
|
||||||
if args.len() == 1 {
|
if args.len() == 1 {
|
||||||
let ty = match self.get_pyty_obj_type(
|
let ty = match self.get_pyty_obj_type(
|
||||||
py,
|
py,
|
||||||
@ -492,21 +477,7 @@ impl InnerResolver {
|
|||||||
"type list should take concrete parameters in typevar range".into(),
|
"type list should take concrete parameters in typevar range".into(),
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
let list_tvar = if let TypeEnum::TObj { obj_id, params, .. } =
|
Ok(Ok((unifier.add_ty(TypeEnum::TList { ty: ty.0 }), true)))
|
||||||
&*unifier.get_ty_immutable(primitives.list)
|
|
||||||
{
|
|
||||||
assert_eq!(*obj_id, PrimDef::List.id());
|
|
||||||
iter_type_vars(params).nth(0).unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
let list = unifier
|
|
||||||
.subst(
|
|
||||||
primitives.list,
|
|
||||||
&into_var_map([TypeVar { id: list_tvar.id, ty: ty.0 }]),
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
Ok(Ok((list, true)))
|
|
||||||
} else {
|
} else {
|
||||||
return Ok(Err(format!(
|
return Ok(Err(format!(
|
||||||
"type list needs exactly 1 type parameters, found {}",
|
"type list needs exactly 1 type parameters, found {}",
|
||||||
@ -559,10 +530,7 @@ impl InnerResolver {
|
|||||||
Err(err) => return Ok(Err(err)),
|
Err(err) => return Ok(Err(err)),
|
||||||
_ => return Ok(Err("tuple type needs at least 1 type parameters".to_string()))
|
_ => return Ok(Err("tuple type needs at least 1 type parameters".to_string()))
|
||||||
};
|
};
|
||||||
Ok(Ok((
|
Ok(Ok((unifier.add_ty(TypeEnum::TTuple { ty: args }), true)))
|
||||||
unifier.add_ty(TypeEnum::TTuple { ty: args, is_vararg_ctx: false }),
|
|
||||||
true,
|
|
||||||
)))
|
|
||||||
}
|
}
|
||||||
TypeEnum::TObj { params, obj_id, .. } => {
|
TypeEnum::TObj { params, obj_id, .. } => {
|
||||||
let subst = {
|
let subst = {
|
||||||
@ -659,9 +627,7 @@ impl InnerResolver {
|
|||||||
let pyid_to_def = self.pyid_to_def.read();
|
let pyid_to_def = self.pyid_to_def.read();
|
||||||
let constructor_ty = pyid_to_def.get(&py_obj_id).and_then(|def_id| {
|
let constructor_ty = pyid_to_def.get(&py_obj_id).and_then(|def_id| {
|
||||||
defs.iter().find_map(|def| {
|
defs.iter().find_map(|def| {
|
||||||
if let Some(rear_guard) = def.try_read() {
|
if let TopLevelDef::Class { object_id, methods, constructor, .. } = &*def.read() {
|
||||||
if let TopLevelDef::Class { object_id, methods, constructor, .. } = &*rear_guard
|
|
||||||
{
|
|
||||||
if object_id == def_id
|
if object_id == def_id
|
||||||
&& constructor.is_some()
|
&& constructor.is_some()
|
||||||
&& methods.iter().any(|(s, _, _)| s == &"__init__".into())
|
&& methods.iter().any(|(s, _, _)| s == &"__init__".into())
|
||||||
@ -669,7 +635,6 @@ impl InnerResolver {
|
|||||||
return *constructor;
|
return *constructor;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
None
|
None
|
||||||
})
|
})
|
||||||
});
|
});
|
||||||
@ -699,38 +664,15 @@ impl InnerResolver {
|
|||||||
primitives,
|
primitives,
|
||||||
)? {
|
)? {
|
||||||
Ok(s) => s,
|
Ok(s) => s,
|
||||||
Err(e) => {
|
Err(e) => return Ok(Err(e)),
|
||||||
// Allow access to Class Attributes of Classes without having to initialize Objects
|
|
||||||
if self.pyid_to_def.read().contains_key(&py_obj_id) {
|
|
||||||
if let Some(def_id) = self.pyid_to_def.read().get(&py_obj_id).copied() {
|
|
||||||
let def = defs[def_id.0].read();
|
|
||||||
let TopLevelDef::Class { object_id, .. } = &*def else {
|
|
||||||
// only object is supported, functions are not supported
|
|
||||||
unreachable!("function type is not supported, should not be queried")
|
|
||||||
};
|
|
||||||
|
|
||||||
let ty = TypeEnum::TObj {
|
|
||||||
obj_id: *object_id,
|
|
||||||
params: VarMap::new(),
|
|
||||||
fields: HashMap::new(),
|
|
||||||
};
|
|
||||||
(unifier.add_ty(ty), true)
|
|
||||||
} else {
|
|
||||||
return Ok(Err(e));
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
return Ok(Err(e));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
match (&*unifier.get_ty(extracted_ty), inst_check) {
|
match (&*unifier.get_ty(extracted_ty), inst_check) {
|
||||||
// do the instantiation for these four types
|
// do the instantiation for these four types
|
||||||
(TypeEnum::TObj { obj_id, params, .. }, false) if *obj_id == PrimDef::List.id() => {
|
(TypeEnum::TList { ty }, false) => {
|
||||||
let ty = iter_type_vars(params).nth(0).unwrap().ty;
|
|
||||||
let len: usize = self.helper.len_fn.call1(py, (obj,))?.extract(py)?;
|
let len: usize = self.helper.len_fn.call1(py, (obj,))?.extract(py)?;
|
||||||
if len == 0 {
|
if len == 0 {
|
||||||
assert!(matches!(
|
assert!(matches!(
|
||||||
&*unifier.get_ty(ty),
|
&*unifier.get_ty(*ty),
|
||||||
TypeEnum::TVar { fields: None, range, .. }
|
TypeEnum::TVar { fields: None, range, .. }
|
||||||
if range.is_empty()
|
if range.is_empty()
|
||||||
));
|
));
|
||||||
@ -739,25 +681,8 @@ impl InnerResolver {
|
|||||||
let actual_ty =
|
let actual_ty =
|
||||||
self.get_list_elem_type(py, obj, len, unifier, defs, primitives)?;
|
self.get_list_elem_type(py, obj, len, unifier, defs, primitives)?;
|
||||||
match actual_ty {
|
match actual_ty {
|
||||||
Ok(t) => match unifier.unify(ty, t) {
|
Ok(t) => match unifier.unify(*ty, t) {
|
||||||
Ok(()) => {
|
Ok(()) => Ok(Ok(unifier.add_ty(TypeEnum::TList { ty: *ty }))),
|
||||||
let list_tvar = if let TypeEnum::TObj { obj_id, params, .. } =
|
|
||||||
&*unifier.get_ty_immutable(primitives.list)
|
|
||||||
{
|
|
||||||
assert_eq!(*obj_id, PrimDef::List.id());
|
|
||||||
iter_type_vars(params).nth(0).unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
let list = unifier
|
|
||||||
.subst(
|
|
||||||
primitives.list,
|
|
||||||
&into_var_map([TypeVar { id: list_tvar.id, ty }]),
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
Ok(Ok(list))
|
|
||||||
}
|
|
||||||
|
|
||||||
Err(e) => Ok(Err(format!(
|
Err(e) => Ok(Err(format!(
|
||||||
"type error ({}) for the list",
|
"type error ({}) for the list",
|
||||||
e.to_display(unifier)
|
e.to_display(unifier)
|
||||||
@ -804,9 +729,7 @@ impl InnerResolver {
|
|||||||
.map(|elem| self.get_obj_type(py, elem, unifier, defs, primitives))
|
.map(|elem| self.get_obj_type(py, elem, unifier, defs, primitives))
|
||||||
.collect();
|
.collect();
|
||||||
let types = types?;
|
let types = types?;
|
||||||
Ok(types.map(|types| {
|
Ok(types.map(|types| unifier.add_ty(TypeEnum::TTuple { ty: types })))
|
||||||
unifier.add_ty(TypeEnum::TTuple { ty: types, is_vararg_ctx: false })
|
|
||||||
}))
|
|
||||||
}
|
}
|
||||||
// special handling for option type since its class member layout in python side
|
// special handling for option type since its class member layout in python side
|
||||||
// is special and cannot be mapped directly to a nac3 type as below
|
// is special and cannot be mapped directly to a nac3 type as below
|
||||||
@ -981,7 +904,7 @@ impl InnerResolver {
|
|||||||
} else if ty_id == self.primitive_ids.string || ty_id == self.primitive_ids.np_str_ {
|
} else if ty_id == self.primitive_ids.string || ty_id == self.primitive_ids.np_str_ {
|
||||||
let val: String = obj.extract().unwrap();
|
let val: String = obj.extract().unwrap();
|
||||||
self.id_to_primitive.write().insert(id, PrimitiveValue::Str(val.clone()));
|
self.id_to_primitive.write().insert(id, PrimitiveValue::Str(val.clone()));
|
||||||
Ok(Some(ctx.gen_string(generator, val).into()))
|
Ok(Some(ctx.ctx.const_string(val.as_bytes(), true).into()))
|
||||||
} else if ty_id == self.primitive_ids.float || ty_id == self.primitive_ids.float64 {
|
} else if ty_id == self.primitive_ids.float || ty_id == self.primitive_ids.float64 {
|
||||||
let val: f64 = obj.extract().unwrap();
|
let val: f64 = obj.extract().unwrap();
|
||||||
self.id_to_primitive.write().insert(id, PrimitiveValue::F64(val));
|
self.id_to_primitive.write().insert(id, PrimitiveValue::F64(val));
|
||||||
@ -994,21 +917,15 @@ impl InnerResolver {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let len: usize = self.helper.len_fn.call1(py, (obj,))?.extract(py)?;
|
let len: usize = self.helper.len_fn.call1(py, (obj,))?.extract(py)?;
|
||||||
let elem_ty = match ctx.unifier.get_ty_immutable(expected_ty).as_ref() {
|
let elem_ty = if let TypeEnum::TList { ty } =
|
||||||
TypeEnum::TObj { obj_id, params, .. } if *obj_id == PrimDef::List.id() => {
|
ctx.unifier.get_ty_immutable(expected_ty).as_ref()
|
||||||
iter_type_vars(params).nth(0).unwrap().ty
|
|
||||||
}
|
|
||||||
_ => unreachable!("must be list"),
|
|
||||||
};
|
|
||||||
let size_t = generator.get_size_type(ctx.ctx);
|
|
||||||
let ty = if len == 0
|
|
||||||
&& matches!(&*ctx.unifier.get_ty_immutable(elem_ty), TypeEnum::TVar { .. })
|
|
||||||
{
|
{
|
||||||
// The default type for zero-length lists of unknown element type is size_t
|
*ty
|
||||||
size_t.into()
|
|
||||||
} else {
|
} else {
|
||||||
ctx.get_llvm_type(generator, elem_ty)
|
unreachable!("must be list")
|
||||||
};
|
};
|
||||||
|
let ty = ctx.get_llvm_type(generator, elem_ty);
|
||||||
|
let size_t = generator.get_size_type(ctx.ctx);
|
||||||
let arr_ty = ctx
|
let arr_ty = ctx
|
||||||
.ctx
|
.ctx
|
||||||
.struct_type(&[ty.ptr_type(AddressSpace::default()).into(), size_t.into()], false);
|
.struct_type(&[ty.ptr_type(AddressSpace::default()).into(), size_t.into()], false);
|
||||||
@ -1212,9 +1129,7 @@ impl InnerResolver {
|
|||||||
Ok(Some(ndarray.as_pointer_value().into()))
|
Ok(Some(ndarray.as_pointer_value().into()))
|
||||||
} else if ty_id == self.primitive_ids.tuple {
|
} else if ty_id == self.primitive_ids.tuple {
|
||||||
let expected_ty_enum = ctx.unifier.get_ty_immutable(expected_ty);
|
let expected_ty_enum = ctx.unifier.get_ty_immutable(expected_ty);
|
||||||
let TypeEnum::TTuple { ty, is_vararg_ctx: false } = expected_ty_enum.as_ref() else {
|
let TypeEnum::TTuple { ty } = expected_ty_enum.as_ref() else { unreachable!() };
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
let tup_tys = ty.iter();
|
let tup_tys = ty.iter();
|
||||||
let elements: &PyTuple = obj.downcast()?;
|
let elements: &PyTuple = obj.downcast()?;
|
||||||
@ -1470,7 +1385,6 @@ impl SymbolResolver for Resolver {
|
|||||||
&self,
|
&self,
|
||||||
id: StrRef,
|
id: StrRef,
|
||||||
_: &mut CodeGenContext<'ctx, '_>,
|
_: &mut CodeGenContext<'ctx, '_>,
|
||||||
_: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>> {
|
) -> Option<ValueEnum<'ctx>> {
|
||||||
let sym_value = {
|
let sym_value = {
|
||||||
let id_to_val = self.0.id_to_pyval.read();
|
let id_to_val = self.0.id_to_pyval.read();
|
||||||
|
@ -1,12 +1,9 @@
|
|||||||
use itertools::Either;
|
use inkwell::{
|
||||||
|
|
||||||
use nac3core::{
|
|
||||||
codegen::CodeGenContext,
|
|
||||||
inkwell::{
|
|
||||||
values::{BasicValueEnum, CallSiteValue},
|
values::{BasicValueEnum, CallSiteValue},
|
||||||
AddressSpace, AtomicOrdering,
|
AddressSpace, AtomicOrdering,
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
use itertools::Either;
|
||||||
|
use nac3core::codegen::CodeGenContext;
|
||||||
|
|
||||||
/// Functions for manipulating the timeline.
|
/// Functions for manipulating the timeline.
|
||||||
pub trait TimeFns {
|
pub trait TimeFns {
|
||||||
@ -34,7 +31,7 @@ impl TimeFns for NowPinningTimeFns64 {
|
|||||||
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
||||||
let now_hiptr = ctx
|
let now_hiptr = ctx
|
||||||
.builder
|
.builder
|
||||||
.build_bit_cast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
.build_bitcast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
@ -83,7 +80,7 @@ impl TimeFns for NowPinningTimeFns64 {
|
|||||||
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
||||||
let now_hiptr = ctx
|
let now_hiptr = ctx
|
||||||
.builder
|
.builder
|
||||||
.build_bit_cast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
.build_bitcast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
@ -112,7 +109,7 @@ impl TimeFns for NowPinningTimeFns64 {
|
|||||||
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
||||||
let now_hiptr = ctx
|
let now_hiptr = ctx
|
||||||
.builder
|
.builder
|
||||||
.build_bit_cast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
.build_bitcast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
@ -210,7 +207,7 @@ impl TimeFns for NowPinningTimeFns {
|
|||||||
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
.unwrap_or_else(|| ctx.module.add_global(i64_type, None, "now"));
|
||||||
let now_hiptr = ctx
|
let now_hiptr = ctx
|
||||||
.builder
|
.builder
|
||||||
.build_bit_cast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
.build_bitcast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
@ -261,7 +258,7 @@ impl TimeFns for NowPinningTimeFns {
|
|||||||
let time_lo = ctx.builder.build_int_truncate(time, i32_type, "time.lo").unwrap();
|
let time_lo = ctx.builder.build_int_truncate(time, i32_type, "time.lo").unwrap();
|
||||||
let now_hiptr = ctx
|
let now_hiptr = ctx
|
||||||
.builder
|
.builder
|
||||||
.build_bit_cast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
.build_bitcast(now, i32_type.ptr_type(AddressSpace::default()), "now.hi.addr")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
|
@ -10,7 +10,7 @@ constant-optimization = ["fold"]
|
|||||||
fold = []
|
fold = []
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
lazy_static = "1.5"
|
lazy_static = "1.4"
|
||||||
parking_lot = "0.12"
|
parking_lot = "0.12"
|
||||||
string-interner = "0.17"
|
string-interner = "0.17"
|
||||||
fxhash = "0.2"
|
fxhash = "0.2"
|
||||||
|
@ -4,23 +4,20 @@ version = "0.1.0"
|
|||||||
authors = ["M-Labs"]
|
authors = ["M-Labs"]
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[features]
|
|
||||||
no-escape-analysis = []
|
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
itertools = "0.13"
|
itertools = "0.13"
|
||||||
crossbeam = "0.8"
|
crossbeam = "0.8"
|
||||||
indexmap = "2.6"
|
indexmap = "2.2"
|
||||||
parking_lot = "0.12"
|
parking_lot = "0.12"
|
||||||
rayon = "1.10"
|
rayon = "1.8"
|
||||||
nac3parser = { path = "../nac3parser" }
|
nac3parser = { path = "../nac3parser" }
|
||||||
strum = "0.26"
|
strum = "0.26.2"
|
||||||
strum_macros = "0.26"
|
strum_macros = "0.26.4"
|
||||||
|
|
||||||
[dependencies.inkwell]
|
[dependencies.inkwell]
|
||||||
version = "0.5"
|
version = "0.4"
|
||||||
default-features = false
|
default-features = false
|
||||||
features = ["llvm14-0-prefer-dynamic", "target-x86", "target-arm", "target-riscv", "no-libffi-linking"]
|
features = ["llvm14-0", "target-x86", "target-arm", "target-riscv", "no-libffi-linking"]
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
test-case = "1.2.0"
|
test-case = "1.2.0"
|
||||||
|
@ -1,3 +1,4 @@
|
|||||||
|
use regex::Regex;
|
||||||
use std::{
|
use std::{
|
||||||
env,
|
env,
|
||||||
fs::File,
|
fs::File,
|
||||||
@ -6,53 +7,34 @@ use std::{
|
|||||||
process::{Command, Stdio},
|
process::{Command, Stdio},
|
||||||
};
|
};
|
||||||
|
|
||||||
use regex::Regex;
|
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
let out_dir = env::var("OUT_DIR").unwrap();
|
const FILE: &str = "src/codegen/irrt/irrt.c";
|
||||||
let out_dir = Path::new(&out_dir);
|
|
||||||
let irrt_dir = Path::new("irrt");
|
|
||||||
|
|
||||||
let irrt_cpp_path = irrt_dir.join("irrt.cpp");
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* HACK: Sadly, clang doesn't let us emit generic LLVM bitcode.
|
* HACK: Sadly, clang doesn't let us emit generic LLVM bitcode.
|
||||||
* Compiling for WASM32 and filtering the output with regex is the closest we can get.
|
* Compiling for WASM32 and filtering the output with regex is the closest we can get.
|
||||||
*/
|
*/
|
||||||
let mut flags: Vec<&str> = vec![
|
let flags: &[&str] = &[
|
||||||
"--target=wasm32",
|
"--target=wasm32",
|
||||||
"-x",
|
FILE,
|
||||||
"c++",
|
|
||||||
"-std=c++20",
|
|
||||||
"-fno-discard-value-names",
|
"-fno-discard-value-names",
|
||||||
"-fno-exceptions",
|
match env::var("PROFILE").as_deref() {
|
||||||
"-fno-rtti",
|
Ok("debug") => "-O0",
|
||||||
|
Ok("release") => "-O3",
|
||||||
|
flavor => panic!("Unknown or missing build flavor {flavor:?}"),
|
||||||
|
},
|
||||||
"-emit-llvm",
|
"-emit-llvm",
|
||||||
"-S",
|
"-S",
|
||||||
"-Wall",
|
"-Wall",
|
||||||
"-Wextra",
|
"-Wextra",
|
||||||
"-o",
|
"-o",
|
||||||
"-",
|
"-",
|
||||||
"-I",
|
|
||||||
irrt_dir.to_str().unwrap(),
|
|
||||||
irrt_cpp_path.to_str().unwrap(),
|
|
||||||
];
|
];
|
||||||
|
|
||||||
match env::var("PROFILE").as_deref() {
|
println!("cargo:rerun-if-changed={FILE}");
|
||||||
Ok("debug") => {
|
let out_dir = env::var("OUT_DIR").unwrap();
|
||||||
flags.push("-O0");
|
let out_path = Path::new(&out_dir);
|
||||||
flags.push("-DIRRT_DEBUG_ASSERT");
|
|
||||||
}
|
|
||||||
Ok("release") => {
|
|
||||||
flags.push("-O3");
|
|
||||||
}
|
|
||||||
flavor => panic!("Unknown or missing build flavor {flavor:?}"),
|
|
||||||
}
|
|
||||||
|
|
||||||
// Tell Cargo to rerun if any file under `irrt_dir` (recursive) changes
|
|
||||||
println!("cargo:rerun-if-changed={}", irrt_dir.to_str().unwrap());
|
|
||||||
|
|
||||||
// Compile IRRT and capture the LLVM IR output
|
|
||||||
let output = Command::new("clang-irrt")
|
let output = Command::new("clang-irrt")
|
||||||
.args(flags)
|
.args(flags)
|
||||||
.output()
|
.output()
|
||||||
@ -66,17 +48,7 @@ fn main() {
|
|||||||
let output = std::str::from_utf8(&output.stdout).unwrap().replace("\r\n", "\n");
|
let output = std::str::from_utf8(&output.stdout).unwrap().replace("\r\n", "\n");
|
||||||
let mut filtered_output = String::with_capacity(output.len());
|
let mut filtered_output = String::with_capacity(output.len());
|
||||||
|
|
||||||
// Filter out irrelevant IR
|
let regex_filter = Regex::new(r"(?ms:^define.*?\}$)|(?m:^declare.*?$)").unwrap();
|
||||||
//
|
|
||||||
// Regex:
|
|
||||||
// - `(?ms:^define.*?\}$)` captures LLVM `define` blocks
|
|
||||||
// - `(?m:^declare.*?$)` captures LLVM `declare` lines
|
|
||||||
// - `(?m:^%.+?=\s*type\s*\{.+?\}$)` captures LLVM `type` declarations
|
|
||||||
// - `(?m:^@.+?=.+$)` captures global constants
|
|
||||||
let regex_filter = Regex::new(
|
|
||||||
r"(?ms:^define.*?\}$)|(?m:^declare.*?$)|(?m:^%.+?=\s*type\s*\{.+?\}$)|(?m:^@.+?=.+$)",
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
for f in regex_filter.captures_iter(&output) {
|
for f in regex_filter.captures_iter(&output) {
|
||||||
assert_eq!(f.len(), 1);
|
assert_eq!(f.len(), 1);
|
||||||
filtered_output.push_str(&f[0]);
|
filtered_output.push_str(&f[0]);
|
||||||
@ -87,22 +59,18 @@ fn main() {
|
|||||||
.unwrap()
|
.unwrap()
|
||||||
.replace_all(&filtered_output, "");
|
.replace_all(&filtered_output, "");
|
||||||
|
|
||||||
// For debugging
|
println!("cargo:rerun-if-env-changed=DEBUG_DUMP_IRRT");
|
||||||
// Doing `DEBUG_DUMP_IRRT=1 cargo build -p nac3core` dumps the LLVM IR generated
|
if env::var("DEBUG_DUMP_IRRT").is_ok() {
|
||||||
const DEBUG_DUMP_IRRT: &str = "DEBUG_DUMP_IRRT";
|
let mut file = File::create(out_path.join("irrt.ll")).unwrap();
|
||||||
println!("cargo:rerun-if-env-changed={DEBUG_DUMP_IRRT}");
|
|
||||||
if env::var(DEBUG_DUMP_IRRT).is_ok() {
|
|
||||||
let mut file = File::create(out_dir.join("irrt.ll")).unwrap();
|
|
||||||
file.write_all(output.as_bytes()).unwrap();
|
file.write_all(output.as_bytes()).unwrap();
|
||||||
|
let mut file = File::create(out_path.join("irrt-filtered.ll")).unwrap();
|
||||||
let mut file = File::create(out_dir.join("irrt-filtered.ll")).unwrap();
|
|
||||||
file.write_all(filtered_output.as_bytes()).unwrap();
|
file.write_all(filtered_output.as_bytes()).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut llvm_as = Command::new("llvm-as-irrt")
|
let mut llvm_as = Command::new("llvm-as-irrt")
|
||||||
.stdin(Stdio::piped())
|
.stdin(Stdio::piped())
|
||||||
.arg("-o")
|
.arg("-o")
|
||||||
.arg(out_dir.join("irrt.bc"))
|
.arg(out_path.join("irrt.bc"))
|
||||||
.spawn()
|
.spawn()
|
||||||
.unwrap();
|
.unwrap();
|
||||||
llvm_as.stdin.as_mut().unwrap().write_all(filtered_output.as_bytes()).unwrap();
|
llvm_as.stdin.as_mut().unwrap().write_all(filtered_output.as_bytes()).unwrap();
|
||||||
|
@ -1,6 +0,0 @@
|
|||||||
#include "irrt/exception.hpp"
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
#include "irrt/list.hpp"
|
|
||||||
#include "irrt/math.hpp"
|
|
||||||
#include "irrt/ndarray.hpp"
|
|
||||||
#include "irrt/slice.hpp"
|
|
@ -1,9 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
struct CSlice {
|
|
||||||
uint8_t* base;
|
|
||||||
SizeT len;
|
|
||||||
};
|
|
@ -1,25 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
// Set in nac3core/build.rs
|
|
||||||
#ifdef IRRT_DEBUG_ASSERT
|
|
||||||
#define IRRT_DEBUG_ASSERT_BOOL true
|
|
||||||
#else
|
|
||||||
#define IRRT_DEBUG_ASSERT_BOOL false
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#define raise_debug_assert(SizeT, msg, param1, param2, param3) \
|
|
||||||
raise_exception(SizeT, EXN_ASSERTION_ERROR, "IRRT debug assert failed: " msg, param1, param2, param3)
|
|
||||||
|
|
||||||
#define debug_assert_eq(SizeT, lhs, rhs) \
|
|
||||||
if constexpr (IRRT_DEBUG_ASSERT_BOOL) { \
|
|
||||||
if ((lhs) != (rhs)) { \
|
|
||||||
raise_debug_assert(SizeT, "LHS = {0}. RHS = {1}", lhs, rhs, NO_PARAM); \
|
|
||||||
} \
|
|
||||||
}
|
|
||||||
|
|
||||||
#define debug_assert(SizeT, expr) \
|
|
||||||
if constexpr (IRRT_DEBUG_ASSERT_BOOL) { \
|
|
||||||
if (!(expr)) { \
|
|
||||||
raise_debug_assert(SizeT, "Got false.", NO_PARAM, NO_PARAM, NO_PARAM); \
|
|
||||||
} \
|
|
||||||
}
|
|
@ -1,82 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#include "irrt/cslice.hpp"
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @brief The int type of ARTIQ exception IDs.
|
|
||||||
*/
|
|
||||||
typedef int32_t ExceptionId;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Set of exceptions C++ IRRT can use.
|
|
||||||
* Must be synchronized with `setup_irrt_exceptions` in `nac3core/src/codegen/irrt/mod.rs`.
|
|
||||||
*/
|
|
||||||
extern "C" {
|
|
||||||
ExceptionId EXN_INDEX_ERROR;
|
|
||||||
ExceptionId EXN_VALUE_ERROR;
|
|
||||||
ExceptionId EXN_ASSERTION_ERROR;
|
|
||||||
ExceptionId EXN_TYPE_ERROR;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @brief Extern function to `__nac3_raise`
|
|
||||||
*
|
|
||||||
* The parameter `err` could be `Exception<int32_t>` or `Exception<int64_t>`. The caller
|
|
||||||
* must make sure to pass `Exception`s with the correct `SizeT` depending on the `size_t` of the runtime.
|
|
||||||
*/
|
|
||||||
extern "C" void __nac3_raise(void* err);
|
|
||||||
|
|
||||||
namespace {
|
|
||||||
/**
|
|
||||||
* @brief NAC3's Exception struct
|
|
||||||
*/
|
|
||||||
template<typename SizeT>
|
|
||||||
struct Exception {
|
|
||||||
ExceptionId id;
|
|
||||||
CSlice<SizeT> filename;
|
|
||||||
int32_t line;
|
|
||||||
int32_t column;
|
|
||||||
CSlice<SizeT> function;
|
|
||||||
CSlice<SizeT> msg;
|
|
||||||
int64_t params[3];
|
|
||||||
};
|
|
||||||
|
|
||||||
constexpr int64_t NO_PARAM = 0;
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
void _raise_exception_helper(ExceptionId id,
|
|
||||||
const char* filename,
|
|
||||||
int32_t line,
|
|
||||||
const char* function,
|
|
||||||
const char* msg,
|
|
||||||
int64_t param0,
|
|
||||||
int64_t param1,
|
|
||||||
int64_t param2) {
|
|
||||||
Exception<SizeT> e = {
|
|
||||||
.id = id,
|
|
||||||
.filename = {.base = reinterpret_cast<const uint8_t*>(filename), .len = __builtin_strlen(filename)},
|
|
||||||
.line = line,
|
|
||||||
.column = 0,
|
|
||||||
.function = {.base = reinterpret_cast<const uint8_t*>(function), .len = __builtin_strlen(function)},
|
|
||||||
.msg = {.base = reinterpret_cast<const uint8_t*>(msg), .len = __builtin_strlen(msg)},
|
|
||||||
};
|
|
||||||
e.params[0] = param0;
|
|
||||||
e.params[1] = param1;
|
|
||||||
e.params[2] = param2;
|
|
||||||
__nac3_raise(reinterpret_cast<void*>(&e));
|
|
||||||
__builtin_unreachable();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @brief Raise an exception with location details (location in the IRRT source files).
|
|
||||||
* @param SizeT The runtime `size_t` type.
|
|
||||||
* @param id The ID of the exception to raise.
|
|
||||||
* @param msg A global constant C-string of the error message.
|
|
||||||
*
|
|
||||||
* `param0` to `param2` are optional format arguments of `msg`. They should be set to
|
|
||||||
* `NO_PARAM` to indicate they are unused.
|
|
||||||
*/
|
|
||||||
#define raise_exception(SizeT, id, msg, param0, param1, param2) \
|
|
||||||
_raise_exception_helper<SizeT>(id, __FILE__, __LINE__, __FUNCTION__, msg, param0, param1, param2)
|
|
||||||
} // namespace
|
|
@ -1,22 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#if __STDC_VERSION__ >= 202000
|
|
||||||
using int8_t = _BitInt(8);
|
|
||||||
using uint8_t = unsigned _BitInt(8);
|
|
||||||
using int32_t = _BitInt(32);
|
|
||||||
using uint32_t = unsigned _BitInt(32);
|
|
||||||
using int64_t = _BitInt(64);
|
|
||||||
using uint64_t = unsigned _BitInt(64);
|
|
||||||
#else
|
|
||||||
using int8_t = _ExtInt(8);
|
|
||||||
using uint8_t = unsigned _ExtInt(8);
|
|
||||||
using int32_t = _ExtInt(32);
|
|
||||||
using uint32_t = unsigned _ExtInt(32);
|
|
||||||
using int64_t = _ExtInt(64);
|
|
||||||
using uint64_t = unsigned _ExtInt(64);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
// NDArray indices are always `uint32_t`.
|
|
||||||
using NDIndex = uint32_t;
|
|
||||||
// The type of an index or a value describing the length of a range/slice is always `int32_t`.
|
|
||||||
using SliceIndex = int32_t;
|
|
@ -1,75 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
#include "irrt/math_util.hpp"
|
|
||||||
|
|
||||||
extern "C" {
|
|
||||||
// Handle list assignment and dropping part of the list when
|
|
||||||
// both dest_step and src_step are +1.
|
|
||||||
// - All the index must *not* be out-of-bound or negative,
|
|
||||||
// - The end index is *inclusive*,
|
|
||||||
// - The length of src and dest slice size should already
|
|
||||||
// be checked: if dest.step == 1 then len(src) <= len(dest) else len(src) == len(dest)
|
|
||||||
SliceIndex __nac3_list_slice_assign_var_size(SliceIndex dest_start,
|
|
||||||
SliceIndex dest_end,
|
|
||||||
SliceIndex dest_step,
|
|
||||||
uint8_t* dest_arr,
|
|
||||||
SliceIndex dest_arr_len,
|
|
||||||
SliceIndex src_start,
|
|
||||||
SliceIndex src_end,
|
|
||||||
SliceIndex src_step,
|
|
||||||
uint8_t* src_arr,
|
|
||||||
SliceIndex src_arr_len,
|
|
||||||
const SliceIndex size) {
|
|
||||||
/* if dest_arr_len == 0, do nothing since we do not support extending list */
|
|
||||||
if (dest_arr_len == 0)
|
|
||||||
return dest_arr_len;
|
|
||||||
/* if both step is 1, memmove directly, handle the dropping of the list, and shrink size */
|
|
||||||
if (src_step == dest_step && dest_step == 1) {
|
|
||||||
const SliceIndex src_len = (src_end >= src_start) ? (src_end - src_start + 1) : 0;
|
|
||||||
const SliceIndex dest_len = (dest_end >= dest_start) ? (dest_end - dest_start + 1) : 0;
|
|
||||||
if (src_len > 0) {
|
|
||||||
__builtin_memmove(dest_arr + dest_start * size, src_arr + src_start * size, src_len * size);
|
|
||||||
}
|
|
||||||
if (dest_len > 0) {
|
|
||||||
/* dropping */
|
|
||||||
__builtin_memmove(dest_arr + (dest_start + src_len) * size, dest_arr + (dest_end + 1) * size,
|
|
||||||
(dest_arr_len - dest_end - 1) * size);
|
|
||||||
}
|
|
||||||
/* shrink size */
|
|
||||||
return dest_arr_len - (dest_len - src_len);
|
|
||||||
}
|
|
||||||
/* if two range overlaps, need alloca */
|
|
||||||
uint8_t need_alloca = (dest_arr == src_arr)
|
|
||||||
&& !(max(dest_start, dest_end) < min(src_start, src_end)
|
|
||||||
|| max(src_start, src_end) < min(dest_start, dest_end));
|
|
||||||
if (need_alloca) {
|
|
||||||
uint8_t* tmp = reinterpret_cast<uint8_t*>(__builtin_alloca(src_arr_len * size));
|
|
||||||
__builtin_memcpy(tmp, src_arr, src_arr_len * size);
|
|
||||||
src_arr = tmp;
|
|
||||||
}
|
|
||||||
SliceIndex src_ind = src_start;
|
|
||||||
SliceIndex dest_ind = dest_start;
|
|
||||||
for (; (src_step > 0) ? (src_ind <= src_end) : (src_ind >= src_end); src_ind += src_step, dest_ind += dest_step) {
|
|
||||||
/* for constant optimization */
|
|
||||||
if (size == 1) {
|
|
||||||
__builtin_memcpy(dest_arr + dest_ind, src_arr + src_ind, 1);
|
|
||||||
} else if (size == 4) {
|
|
||||||
__builtin_memcpy(dest_arr + dest_ind * 4, src_arr + src_ind * 4, 4);
|
|
||||||
} else if (size == 8) {
|
|
||||||
__builtin_memcpy(dest_arr + dest_ind * 8, src_arr + src_ind * 8, 8);
|
|
||||||
} else {
|
|
||||||
/* memcpy for var size, cannot overlap after previous alloca */
|
|
||||||
__builtin_memcpy(dest_arr + dest_ind * size, src_arr + src_ind * size, size);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
/* only dest_step == 1 can we shrink the dest list. */
|
|
||||||
/* size should be ensured prior to calling this function */
|
|
||||||
if (dest_step == 1 && dest_end >= dest_start) {
|
|
||||||
__builtin_memmove(dest_arr + dest_ind * size, dest_arr + (dest_end + 1) * size,
|
|
||||||
(dest_arr_len - dest_end - 1) * size);
|
|
||||||
return dest_arr_len - (dest_end - dest_ind) - 1;
|
|
||||||
}
|
|
||||||
return dest_arr_len;
|
|
||||||
}
|
|
||||||
} // extern "C"
|
|
@ -1,93 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
namespace {
|
|
||||||
// adapted from GNU Scientific Library: https://git.savannah.gnu.org/cgit/gsl.git/tree/sys/pow_int.c
|
|
||||||
// need to make sure `exp >= 0` before calling this function
|
|
||||||
template<typename T>
|
|
||||||
T __nac3_int_exp_impl(T base, T exp) {
|
|
||||||
T res = 1;
|
|
||||||
/* repeated squaring method */
|
|
||||||
do {
|
|
||||||
if (exp & 1) {
|
|
||||||
res *= base; /* for n odd */
|
|
||||||
}
|
|
||||||
exp >>= 1;
|
|
||||||
base *= base;
|
|
||||||
} while (exp);
|
|
||||||
return res;
|
|
||||||
}
|
|
||||||
} // namespace
|
|
||||||
|
|
||||||
#define DEF_nac3_int_exp_(T) \
|
|
||||||
T __nac3_int_exp_##T(T base, T exp) { \
|
|
||||||
return __nac3_int_exp_impl(base, exp); \
|
|
||||||
}
|
|
||||||
|
|
||||||
extern "C" {
|
|
||||||
|
|
||||||
// Putting semicolons here to make clang-format not reformat this into
|
|
||||||
// a stair shape.
|
|
||||||
DEF_nac3_int_exp_(int32_t);
|
|
||||||
DEF_nac3_int_exp_(int64_t);
|
|
||||||
DEF_nac3_int_exp_(uint32_t);
|
|
||||||
DEF_nac3_int_exp_(uint64_t);
|
|
||||||
|
|
||||||
int32_t __nac3_isinf(double x) {
|
|
||||||
return __builtin_isinf(x);
|
|
||||||
}
|
|
||||||
|
|
||||||
int32_t __nac3_isnan(double x) {
|
|
||||||
return __builtin_isnan(x);
|
|
||||||
}
|
|
||||||
|
|
||||||
double tgamma(double arg);
|
|
||||||
|
|
||||||
double __nac3_gamma(double z) {
|
|
||||||
// Handling for denormals
|
|
||||||
// | x | Python gamma(x) | C tgamma(x) |
|
|
||||||
// --- | ----------------- | --------------- | ----------- |
|
|
||||||
// (1) | nan | nan | nan |
|
|
||||||
// (2) | -inf | -inf | inf |
|
|
||||||
// (3) | inf | inf | inf |
|
|
||||||
// (4) | 0.0 | inf | inf |
|
|
||||||
// (5) | {-1.0, -2.0, ...} | inf | nan |
|
|
||||||
|
|
||||||
// (1)-(3)
|
|
||||||
if (__builtin_isinf(z) || __builtin_isnan(z)) {
|
|
||||||
return z;
|
|
||||||
}
|
|
||||||
|
|
||||||
double v = tgamma(z);
|
|
||||||
|
|
||||||
// (4)-(5)
|
|
||||||
return __builtin_isinf(v) || __builtin_isnan(v) ? __builtin_inf() : v;
|
|
||||||
}
|
|
||||||
|
|
||||||
double lgamma(double arg);
|
|
||||||
|
|
||||||
double __nac3_gammaln(double x) {
|
|
||||||
// libm's handling of value overflows differs from scipy:
|
|
||||||
// - scipy: gammaln(-inf) -> -inf
|
|
||||||
// - libm : lgamma(-inf) -> inf
|
|
||||||
|
|
||||||
if (__builtin_isinf(x)) {
|
|
||||||
return x;
|
|
||||||
}
|
|
||||||
|
|
||||||
return lgamma(x);
|
|
||||||
}
|
|
||||||
|
|
||||||
double j0(double x);
|
|
||||||
|
|
||||||
double __nac3_j0(double x) {
|
|
||||||
// libm's handling of value overflows differs from scipy:
|
|
||||||
// - scipy: j0(inf) -> nan
|
|
||||||
// - libm : j0(inf) -> 0.0
|
|
||||||
|
|
||||||
if (__builtin_isinf(x)) {
|
|
||||||
return __builtin_nan("");
|
|
||||||
}
|
|
||||||
|
|
||||||
return j0(x);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,13 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
namespace {
|
|
||||||
template<typename T>
|
|
||||||
const T& max(const T& a, const T& b) {
|
|
||||||
return a > b ? a : b;
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename T>
|
|
||||||
const T& min(const T& a, const T& b) {
|
|
||||||
return a > b ? b : a;
|
|
||||||
}
|
|
||||||
} // namespace
|
|
@ -1,144 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
|
|
||||||
namespace {
|
|
||||||
template<typename SizeT>
|
|
||||||
SizeT __nac3_ndarray_calc_size_impl(const SizeT* list_data, SizeT list_len, SizeT begin_idx, SizeT end_idx) {
|
|
||||||
__builtin_assume(end_idx <= list_len);
|
|
||||||
|
|
||||||
SizeT num_elems = 1;
|
|
||||||
for (SizeT i = begin_idx; i < end_idx; ++i) {
|
|
||||||
SizeT val = list_data[i];
|
|
||||||
__builtin_assume(val > 0);
|
|
||||||
num_elems *= val;
|
|
||||||
}
|
|
||||||
return num_elems;
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
void __nac3_ndarray_calc_nd_indices_impl(SizeT index, const SizeT* dims, SizeT num_dims, NDIndex* idxs) {
|
|
||||||
SizeT stride = 1;
|
|
||||||
for (SizeT dim = 0; dim < num_dims; dim++) {
|
|
||||||
SizeT i = num_dims - dim - 1;
|
|
||||||
__builtin_assume(dims[i] > 0);
|
|
||||||
idxs[i] = (index / stride) % dims[i];
|
|
||||||
stride *= dims[i];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
SizeT __nac3_ndarray_flatten_index_impl(const SizeT* dims, SizeT num_dims, const NDIndex* indices, SizeT num_indices) {
|
|
||||||
SizeT idx = 0;
|
|
||||||
SizeT stride = 1;
|
|
||||||
for (SizeT i = 0; i < num_dims; ++i) {
|
|
||||||
SizeT ri = num_dims - i - 1;
|
|
||||||
if (ri < num_indices) {
|
|
||||||
idx += stride * indices[ri];
|
|
||||||
}
|
|
||||||
|
|
||||||
__builtin_assume(dims[i] > 0);
|
|
||||||
stride *= dims[ri];
|
|
||||||
}
|
|
||||||
return idx;
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
void __nac3_ndarray_calc_broadcast_impl(const SizeT* lhs_dims,
|
|
||||||
SizeT lhs_ndims,
|
|
||||||
const SizeT* rhs_dims,
|
|
||||||
SizeT rhs_ndims,
|
|
||||||
SizeT* out_dims) {
|
|
||||||
SizeT max_ndims = lhs_ndims > rhs_ndims ? lhs_ndims : rhs_ndims;
|
|
||||||
|
|
||||||
for (SizeT i = 0; i < max_ndims; ++i) {
|
|
||||||
const SizeT* lhs_dim_sz = i < lhs_ndims ? &lhs_dims[lhs_ndims - i - 1] : nullptr;
|
|
||||||
const SizeT* rhs_dim_sz = i < rhs_ndims ? &rhs_dims[rhs_ndims - i - 1] : nullptr;
|
|
||||||
SizeT* out_dim = &out_dims[max_ndims - i - 1];
|
|
||||||
|
|
||||||
if (lhs_dim_sz == nullptr) {
|
|
||||||
*out_dim = *rhs_dim_sz;
|
|
||||||
} else if (rhs_dim_sz == nullptr) {
|
|
||||||
*out_dim = *lhs_dim_sz;
|
|
||||||
} else if (*lhs_dim_sz == 1) {
|
|
||||||
*out_dim = *rhs_dim_sz;
|
|
||||||
} else if (*rhs_dim_sz == 1) {
|
|
||||||
*out_dim = *lhs_dim_sz;
|
|
||||||
} else if (*lhs_dim_sz == *rhs_dim_sz) {
|
|
||||||
*out_dim = *lhs_dim_sz;
|
|
||||||
} else {
|
|
||||||
__builtin_unreachable();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
template<typename SizeT>
|
|
||||||
void __nac3_ndarray_calc_broadcast_idx_impl(const SizeT* src_dims,
|
|
||||||
SizeT src_ndims,
|
|
||||||
const NDIndex* in_idx,
|
|
||||||
NDIndex* out_idx) {
|
|
||||||
for (SizeT i = 0; i < src_ndims; ++i) {
|
|
||||||
SizeT src_i = src_ndims - i - 1;
|
|
||||||
out_idx[src_i] = src_dims[src_i] == 1 ? 0 : in_idx[src_i];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} // namespace
|
|
||||||
|
|
||||||
extern "C" {
|
|
||||||
uint32_t __nac3_ndarray_calc_size(const uint32_t* list_data, uint32_t list_len, uint32_t begin_idx, uint32_t end_idx) {
|
|
||||||
return __nac3_ndarray_calc_size_impl(list_data, list_len, begin_idx, end_idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
uint64_t
|
|
||||||
__nac3_ndarray_calc_size64(const uint64_t* list_data, uint64_t list_len, uint64_t begin_idx, uint64_t end_idx) {
|
|
||||||
return __nac3_ndarray_calc_size_impl(list_data, list_len, begin_idx, end_idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_nd_indices(uint32_t index, const uint32_t* dims, uint32_t num_dims, NDIndex* idxs) {
|
|
||||||
__nac3_ndarray_calc_nd_indices_impl(index, dims, num_dims, idxs);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_nd_indices64(uint64_t index, const uint64_t* dims, uint64_t num_dims, NDIndex* idxs) {
|
|
||||||
__nac3_ndarray_calc_nd_indices_impl(index, dims, num_dims, idxs);
|
|
||||||
}
|
|
||||||
|
|
||||||
uint32_t
|
|
||||||
__nac3_ndarray_flatten_index(const uint32_t* dims, uint32_t num_dims, const NDIndex* indices, uint32_t num_indices) {
|
|
||||||
return __nac3_ndarray_flatten_index_impl(dims, num_dims, indices, num_indices);
|
|
||||||
}
|
|
||||||
|
|
||||||
uint64_t
|
|
||||||
__nac3_ndarray_flatten_index64(const uint64_t* dims, uint64_t num_dims, const NDIndex* indices, uint64_t num_indices) {
|
|
||||||
return __nac3_ndarray_flatten_index_impl(dims, num_dims, indices, num_indices);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_broadcast(const uint32_t* lhs_dims,
|
|
||||||
uint32_t lhs_ndims,
|
|
||||||
const uint32_t* rhs_dims,
|
|
||||||
uint32_t rhs_ndims,
|
|
||||||
uint32_t* out_dims) {
|
|
||||||
return __nac3_ndarray_calc_broadcast_impl(lhs_dims, lhs_ndims, rhs_dims, rhs_ndims, out_dims);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_broadcast64(const uint64_t* lhs_dims,
|
|
||||||
uint64_t lhs_ndims,
|
|
||||||
const uint64_t* rhs_dims,
|
|
||||||
uint64_t rhs_ndims,
|
|
||||||
uint64_t* out_dims) {
|
|
||||||
return __nac3_ndarray_calc_broadcast_impl(lhs_dims, lhs_ndims, rhs_dims, rhs_ndims, out_dims);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_broadcast_idx(const uint32_t* src_dims,
|
|
||||||
uint32_t src_ndims,
|
|
||||||
const NDIndex* in_idx,
|
|
||||||
NDIndex* out_idx) {
|
|
||||||
__nac3_ndarray_calc_broadcast_idx_impl(src_dims, src_ndims, in_idx, out_idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
void __nac3_ndarray_calc_broadcast_idx64(const uint64_t* src_dims,
|
|
||||||
uint64_t src_ndims,
|
|
||||||
const NDIndex* in_idx,
|
|
||||||
NDIndex* out_idx) {
|
|
||||||
__nac3_ndarray_calc_broadcast_idx_impl(src_dims, src_ndims, in_idx, out_idx);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,28 +0,0 @@
|
|||||||
#pragma once
|
|
||||||
|
|
||||||
#include "irrt/int_types.hpp"
|
|
||||||
|
|
||||||
extern "C" {
|
|
||||||
SliceIndex __nac3_slice_index_bound(SliceIndex i, const SliceIndex len) {
|
|
||||||
if (i < 0) {
|
|
||||||
i = len + i;
|
|
||||||
}
|
|
||||||
if (i < 0) {
|
|
||||||
return 0;
|
|
||||||
} else if (i > len) {
|
|
||||||
return len;
|
|
||||||
}
|
|
||||||
return i;
|
|
||||||
}
|
|
||||||
|
|
||||||
SliceIndex __nac3_range_slice_len(const SliceIndex start, const SliceIndex end, const SliceIndex step) {
|
|
||||||
SliceIndex diff = end - start;
|
|
||||||
if (diff > 0 && step > 0) {
|
|
||||||
return ((diff - 1) / step) + 1;
|
|
||||||
} else if (diff < 0 && step < 0) {
|
|
||||||
return ((diff + 1) / step) + 1;
|
|
||||||
} else {
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,103 +1,26 @@
|
|||||||
use inkwell::{
|
use inkwell::types::BasicTypeEnum;
|
||||||
types::BasicTypeEnum,
|
use inkwell::values::BasicValueEnum;
|
||||||
values::{BasicValue, BasicValueEnum, IntValue, PointerValue},
|
use inkwell::{FloatPredicate, IntPredicate, OptimizationLevel};
|
||||||
FloatPredicate, IntPredicate, OptimizationLevel,
|
|
||||||
};
|
|
||||||
use itertools::Itertools;
|
use itertools::Itertools;
|
||||||
|
|
||||||
use crate::{
|
use crate::codegen::classes::{NDArrayValue, ProxyValue, UntypedArrayLikeAccessor};
|
||||||
codegen::{
|
use crate::codegen::numpy::ndarray_elementwise_unaryop_impl;
|
||||||
classes::{
|
use crate::codegen::stmt::gen_for_callback_incrementing;
|
||||||
ArrayLikeValue, NDArrayValue, ProxyValue, RangeValue, TypedArrayLikeAccessor,
|
use crate::codegen::{extern_fns, irrt, llvm_intrinsics, numpy, CodeGenContext, CodeGenerator};
|
||||||
UntypedArrayLikeAccessor, UntypedArrayLikeMutator,
|
use crate::toplevel::helper::PrimDef;
|
||||||
},
|
use crate::toplevel::numpy::unpack_ndarray_var_tys;
|
||||||
expr::destructure_range,
|
use crate::typecheck::typedef::Type;
|
||||||
extern_fns, irrt,
|
|
||||||
irrt::calculate_len_for_slice_range,
|
|
||||||
llvm_intrinsics,
|
|
||||||
macros::codegen_unreachable,
|
|
||||||
numpy,
|
|
||||||
numpy::ndarray_elementwise_unaryop_impl,
|
|
||||||
stmt::gen_for_callback_incrementing,
|
|
||||||
CodeGenContext, CodeGenerator,
|
|
||||||
},
|
|
||||||
toplevel::helper::PrimDef,
|
|
||||||
toplevel::numpy::unpack_ndarray_var_tys,
|
|
||||||
typecheck::typedef::{Type, TypeEnum},
|
|
||||||
};
|
|
||||||
|
|
||||||
/// Shorthand for [`unreachable!()`] when a type of argument is not supported.
|
/// Shorthand for [`unreachable!()`] when a type of argument is not supported.
|
||||||
///
|
///
|
||||||
/// The generated message will contain the function name and the name of the unsupported type.
|
/// The generated message will contain the function name and the name of the unsupported type.
|
||||||
fn unsupported_type(ctx: &CodeGenContext<'_, '_>, fn_name: &str, tys: &[Type]) -> ! {
|
fn unsupported_type(ctx: &CodeGenContext<'_, '_>, fn_name: &str, tys: &[Type]) -> ! {
|
||||||
codegen_unreachable!(
|
unreachable!(
|
||||||
ctx,
|
|
||||||
"{fn_name}() not supported for '{}'",
|
"{fn_name}() not supported for '{}'",
|
||||||
tys.iter().map(|ty| format!("'{}'", ctx.unifier.stringify(*ty))).join(", "),
|
tys.iter().map(|ty| format!("'{}'", ctx.unifier.stringify(*ty))).join(", "),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Invokes the `len` builtin function.
|
|
||||||
pub fn call_len<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
n: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<IntValue<'ctx>, String> {
|
|
||||||
let llvm_i32 = ctx.ctx.i32_type();
|
|
||||||
let range_ty = ctx.primitives.range;
|
|
||||||
let (arg_ty, arg) = n;
|
|
||||||
|
|
||||||
Ok(if ctx.unifier.unioned(arg_ty, range_ty) {
|
|
||||||
let arg = RangeValue::from_ptr_val(arg.into_pointer_value(), Some("range"));
|
|
||||||
let (start, end, step) = destructure_range(ctx, arg);
|
|
||||||
calculate_len_for_slice_range(generator, ctx, start, end, step)
|
|
||||||
} else {
|
|
||||||
match &*ctx.unifier.get_ty_immutable(arg_ty) {
|
|
||||||
TypeEnum::TTuple { ty, .. } => llvm_i32.const_int(ty.len() as u64, false),
|
|
||||||
TypeEnum::TObj { obj_id, .. } if *obj_id == PrimDef::List.id() => {
|
|
||||||
let zero = llvm_i32.const_zero();
|
|
||||||
let len = ctx
|
|
||||||
.build_gep_and_load(
|
|
||||||
arg.into_pointer_value(),
|
|
||||||
&[zero, llvm_i32.const_int(1, false)],
|
|
||||||
None,
|
|
||||||
)
|
|
||||||
.into_int_value();
|
|
||||||
ctx.builder.build_int_truncate_or_bit_cast(len, llvm_i32, "len").unwrap()
|
|
||||||
}
|
|
||||||
TypeEnum::TObj { obj_id, .. } if *obj_id == PrimDef::NDArray.id() => {
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
let arg = NDArrayValue::from_ptr_val(arg.into_pointer_value(), llvm_usize, None);
|
|
||||||
|
|
||||||
let ndims = arg.dim_sizes().size(ctx, generator);
|
|
||||||
ctx.make_assert(
|
|
||||||
generator,
|
|
||||||
ctx.builder
|
|
||||||
.build_int_compare(IntPredicate::NE, ndims, llvm_usize.const_zero(), "")
|
|
||||||
.unwrap(),
|
|
||||||
"0:TypeError",
|
|
||||||
"len() of unsized object",
|
|
||||||
[None, None, None],
|
|
||||||
ctx.current_loc,
|
|
||||||
);
|
|
||||||
|
|
||||||
let len = unsafe {
|
|
||||||
arg.dim_sizes().get_typed_unchecked(
|
|
||||||
ctx,
|
|
||||||
generator,
|
|
||||||
&llvm_usize.const_zero(),
|
|
||||||
None,
|
|
||||||
)
|
|
||||||
};
|
|
||||||
|
|
||||||
ctx.builder.build_int_truncate_or_bit_cast(len, llvm_i32, "len").unwrap()
|
|
||||||
}
|
|
||||||
_ => codegen_unreachable!(ctx),
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `int32` builtin function.
|
/// Invokes the `int32` builtin function.
|
||||||
pub fn call_int32<'ctx, G: CodeGenerator + ?Sized>(
|
pub fn call_int32<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
@ -108,6 +31,7 @@ pub fn call_int32<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
|
|
||||||
let (n_ty, n) = n;
|
let (n_ty, n) = n;
|
||||||
|
|
||||||
Ok(match n {
|
Ok(match n {
|
||||||
BasicValueEnum::IntValue(n) if matches!(n.get_type().get_bit_width(), 1 | 8) => {
|
BasicValueEnum::IntValue(n) if matches!(n.get_type().get_bit_width(), 1 | 8) => {
|
||||||
debug_assert!(ctx.unifier.unioned(n_ty, ctx.primitives.bool));
|
debug_assert!(ctx.unifier.unioned(n_ty, ctx.primitives.bool));
|
||||||
@ -495,7 +419,7 @@ pub fn call_numpy_round<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
BasicValueEnum::FloatValue(n) => {
|
BasicValueEnum::FloatValue(n) => {
|
||||||
debug_assert!(ctx.unifier.unioned(n_ty, ctx.primitives.float));
|
debug_assert!(ctx.unifier.unioned(n_ty, ctx.primitives.float));
|
||||||
|
|
||||||
llvm_intrinsics::call_float_rint(ctx, n, None).into()
|
llvm_intrinsics::call_float_roundeven(ctx, n, None).into()
|
||||||
}
|
}
|
||||||
|
|
||||||
BasicValueEnum::PointerValue(n)
|
BasicValueEnum::PointerValue(n)
|
||||||
@ -678,7 +602,7 @@ pub fn call_ceil<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
ret_elem_ty,
|
ret_elem_ty,
|
||||||
None,
|
None,
|
||||||
NDArrayValue::from_ptr_val(n, llvm_usize, None),
|
NDArrayValue::from_ptr_val(n, llvm_usize, None),
|
||||||
|generator, ctx, val| call_ceil(generator, ctx, (elem_ty, val), ret_elem_ty),
|
|generator, ctx, val| call_floor(generator, ctx, (elem_ty, val), ret_elem_ty),
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
ndarray.as_base_value().into()
|
ndarray.as_base_value().into()
|
||||||
@ -737,6 +661,90 @@ pub fn call_min<'ctx>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Invokes the `np_min` builtin function.
|
||||||
|
pub fn call_numpy_min<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
|
generator: &mut G,
|
||||||
|
ctx: &mut CodeGenContext<'ctx, '_>,
|
||||||
|
a: (Type, BasicValueEnum<'ctx>),
|
||||||
|
) -> Result<BasicValueEnum<'ctx>, String> {
|
||||||
|
const FN_NAME: &str = "np_min";
|
||||||
|
|
||||||
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
|
|
||||||
|
let (a_ty, a) = a;
|
||||||
|
|
||||||
|
Ok(match a {
|
||||||
|
BasicValueEnum::IntValue(_) | BasicValueEnum::FloatValue(_) => {
|
||||||
|
debug_assert!([
|
||||||
|
ctx.primitives.bool,
|
||||||
|
ctx.primitives.int32,
|
||||||
|
ctx.primitives.uint32,
|
||||||
|
ctx.primitives.int64,
|
||||||
|
ctx.primitives.uint64,
|
||||||
|
ctx.primitives.float,
|
||||||
|
]
|
||||||
|
.iter()
|
||||||
|
.any(|ty| ctx.unifier.unioned(a_ty, *ty)));
|
||||||
|
|
||||||
|
a
|
||||||
|
}
|
||||||
|
|
||||||
|
BasicValueEnum::PointerValue(n)
|
||||||
|
if a_ty.obj_id(&ctx.unifier).is_some_and(|id| id == PrimDef::NDArray.id()) =>
|
||||||
|
{
|
||||||
|
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, a_ty);
|
||||||
|
let llvm_ndarray_ty = ctx.get_llvm_type(generator, elem_ty);
|
||||||
|
|
||||||
|
let n = NDArrayValue::from_ptr_val(n, llvm_usize, None);
|
||||||
|
let n_sz = irrt::call_ndarray_calc_size(generator, ctx, &n.dim_sizes(), (None, None));
|
||||||
|
if ctx.registry.llvm_options.opt_level == OptimizationLevel::None {
|
||||||
|
let n_sz_eqz = ctx
|
||||||
|
.builder
|
||||||
|
.build_int_compare(IntPredicate::NE, n_sz, n_sz.get_type().const_zero(), "")
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.make_assert(
|
||||||
|
generator,
|
||||||
|
n_sz_eqz,
|
||||||
|
"0:ValueError",
|
||||||
|
"zero-size array to reduction operation minimum which has no identity",
|
||||||
|
[None, None, None],
|
||||||
|
ctx.current_loc,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
let accumulator_addr = generator.gen_var_alloc(ctx, llvm_ndarray_ty, None)?;
|
||||||
|
unsafe {
|
||||||
|
let identity =
|
||||||
|
n.data().get_unchecked(ctx, generator, &llvm_usize.const_zero(), None);
|
||||||
|
ctx.builder.build_store(accumulator_addr, identity).unwrap();
|
||||||
|
}
|
||||||
|
|
||||||
|
gen_for_callback_incrementing(
|
||||||
|
generator,
|
||||||
|
ctx,
|
||||||
|
llvm_usize.const_int(1, false),
|
||||||
|
(n_sz, false),
|
||||||
|
|generator, ctx, _, idx| {
|
||||||
|
let elem = unsafe { n.data().get_unchecked(ctx, generator, &idx, None) };
|
||||||
|
|
||||||
|
let accumulator = ctx.builder.build_load(accumulator_addr, "").unwrap();
|
||||||
|
let result = call_min(ctx, (elem_ty, accumulator), (elem_ty, elem));
|
||||||
|
ctx.builder.build_store(accumulator_addr, result).unwrap();
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
llvm_usize.const_int(1, false),
|
||||||
|
)?;
|
||||||
|
|
||||||
|
let accumulator = ctx.builder.build_load(accumulator_addr, "").unwrap();
|
||||||
|
accumulator
|
||||||
|
}
|
||||||
|
|
||||||
|
_ => unsupported_type(ctx, FN_NAME, &[a_ty]),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
/// Invokes the `np_minimum` builtin function.
|
/// Invokes the `np_minimum` builtin function.
|
||||||
pub fn call_numpy_minimum<'ctx, G: CodeGenerator + ?Sized>(
|
pub fn call_numpy_minimum<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
@ -795,7 +803,7 @@ pub fn call_numpy_minimum<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -869,20 +877,18 @@ pub fn call_max<'ctx>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Invokes the `np_max`, `np_min`, `np_argmax`, `np_argmin` functions
|
/// Invokes the `np_max` builtin function.
|
||||||
/// * `fn_name`: Can be one of `"np_argmin"`, `"np_argmax"`, `"np_max"`, `"np_min"`
|
pub fn call_numpy_max<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
pub fn call_numpy_max_min<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
ctx: &mut CodeGenContext<'ctx, '_>,
|
||||||
a: (Type, BasicValueEnum<'ctx>),
|
a: (Type, BasicValueEnum<'ctx>),
|
||||||
fn_name: &str,
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
) -> Result<BasicValueEnum<'ctx>, String> {
|
||||||
debug_assert!(["np_argmin", "np_argmax", "np_max", "np_min"].iter().any(|f| *f == fn_name));
|
const FN_NAME: &str = "np_max";
|
||||||
|
|
||||||
let llvm_int64 = ctx.ctx.i64_type();
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
|
|
||||||
let (a_ty, a) = a;
|
let (a_ty, a) = a;
|
||||||
|
|
||||||
Ok(match a {
|
Ok(match a {
|
||||||
BasicValueEnum::IntValue(_) | BasicValueEnum::FloatValue(_) => {
|
BasicValueEnum::IntValue(_) | BasicValueEnum::FloatValue(_) => {
|
||||||
debug_assert!([
|
debug_assert!([
|
||||||
@ -896,12 +902,9 @@ pub fn call_numpy_max_min<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
.iter()
|
.iter()
|
||||||
.any(|ty| ctx.unifier.unioned(a_ty, *ty)));
|
.any(|ty| ctx.unifier.unioned(a_ty, *ty)));
|
||||||
|
|
||||||
match fn_name {
|
a
|
||||||
"np_argmin" | "np_argmax" => llvm_int64.const_zero().into(),
|
|
||||||
"np_max" | "np_min" => a,
|
|
||||||
_ => codegen_unreachable!(ctx),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
BasicValueEnum::PointerValue(n)
|
BasicValueEnum::PointerValue(n)
|
||||||
if a_ty.obj_id(&ctx.unifier).is_some_and(|id| id == PrimDef::NDArray.id()) =>
|
if a_ty.obj_id(&ctx.unifier).is_some_and(|id| id == PrimDef::NDArray.id()) =>
|
||||||
{
|
{
|
||||||
@ -920,82 +923,41 @@ pub fn call_numpy_max_min<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
generator,
|
generator,
|
||||||
n_sz_eqz,
|
n_sz_eqz,
|
||||||
"0:ValueError",
|
"0:ValueError",
|
||||||
format!("zero-size array to reduction operation {fn_name}").as_str(),
|
"zero-size array to reduction operation minimum which has no identity",
|
||||||
[None, None, None],
|
[None, None, None],
|
||||||
ctx.current_loc,
|
ctx.current_loc,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
let accumulator_addr = generator.gen_var_alloc(ctx, llvm_ndarray_ty, None)?;
|
let accumulator_addr = generator.gen_var_alloc(ctx, llvm_ndarray_ty, None)?;
|
||||||
let res_idx = generator.gen_var_alloc(ctx, llvm_int64.into(), None)?;
|
|
||||||
|
|
||||||
unsafe {
|
unsafe {
|
||||||
let identity =
|
let identity =
|
||||||
n.data().get_unchecked(ctx, generator, &llvm_usize.const_zero(), None);
|
n.data().get_unchecked(ctx, generator, &llvm_usize.const_zero(), None);
|
||||||
ctx.builder.build_store(accumulator_addr, identity).unwrap();
|
ctx.builder.build_store(accumulator_addr, identity).unwrap();
|
||||||
ctx.builder.build_store(res_idx, llvm_int64.const_zero()).unwrap();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
gen_for_callback_incrementing(
|
gen_for_callback_incrementing(
|
||||||
generator,
|
generator,
|
||||||
ctx,
|
ctx,
|
||||||
None,
|
llvm_usize.const_int(1, false),
|
||||||
llvm_int64.const_int(1, false),
|
|
||||||
(n_sz, false),
|
(n_sz, false),
|
||||||
|generator, ctx, _, idx| {
|
|generator, ctx, _, idx| {
|
||||||
let elem = unsafe { n.data().get_unchecked(ctx, generator, &idx, None) };
|
let elem = unsafe { n.data().get_unchecked(ctx, generator, &idx, None) };
|
||||||
|
|
||||||
let accumulator = ctx.builder.build_load(accumulator_addr, "").unwrap();
|
let accumulator = ctx.builder.build_load(accumulator_addr, "").unwrap();
|
||||||
let cur_idx = ctx.builder.build_load(res_idx, "").unwrap();
|
let result = call_max(ctx, (elem_ty, accumulator), (elem_ty, elem));
|
||||||
|
|
||||||
let result = match fn_name {
|
|
||||||
"np_argmin" | "np_min" => {
|
|
||||||
call_min(ctx, (elem_ty, accumulator), (elem_ty, elem))
|
|
||||||
}
|
|
||||||
"np_argmax" | "np_max" => {
|
|
||||||
call_max(ctx, (elem_ty, accumulator), (elem_ty, elem))
|
|
||||||
}
|
|
||||||
_ => codegen_unreachable!(ctx),
|
|
||||||
};
|
|
||||||
|
|
||||||
let updated_idx = match (accumulator, result) {
|
|
||||||
(BasicValueEnum::IntValue(m), BasicValueEnum::IntValue(n)) => ctx
|
|
||||||
.builder
|
|
||||||
.build_select(
|
|
||||||
ctx.builder.build_int_compare(IntPredicate::NE, m, n, "").unwrap(),
|
|
||||||
idx.into(),
|
|
||||||
cur_idx,
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
(BasicValueEnum::FloatValue(m), BasicValueEnum::FloatValue(n)) => ctx
|
|
||||||
.builder
|
|
||||||
.build_select(
|
|
||||||
ctx.builder
|
|
||||||
.build_float_compare(FloatPredicate::ONE, m, n, "")
|
|
||||||
.unwrap(),
|
|
||||||
idx.into(),
|
|
||||||
cur_idx,
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
_ => unsupported_type(ctx, fn_name, &[elem_ty, elem_ty]),
|
|
||||||
};
|
|
||||||
ctx.builder.build_store(res_idx, updated_idx).unwrap();
|
|
||||||
ctx.builder.build_store(accumulator_addr, result).unwrap();
|
ctx.builder.build_store(accumulator_addr, result).unwrap();
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
},
|
||||||
llvm_int64.const_int(1, false),
|
llvm_usize.const_int(1, false),
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
match fn_name {
|
let accumulator = ctx.builder.build_load(accumulator_addr, "").unwrap();
|
||||||
"np_argmin" | "np_argmax" => ctx.builder.build_load(res_idx, "").unwrap(),
|
accumulator
|
||||||
"np_max" | "np_min" => ctx.builder.build_load(accumulator_addr, "").unwrap(),
|
|
||||||
_ => codegen_unreachable!(ctx),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
_ => unsupported_type(ctx, fn_name, &[a_ty]),
|
_ => unsupported_type(ctx, FN_NAME, &[a_ty]),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1057,7 +1019,7 @@ pub fn call_numpy_maximum<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1421,7 +1383,7 @@ create_helper_call_numpy_unary_elementwise_float_to_float!(
|
|||||||
create_helper_call_numpy_unary_elementwise_float_to_float!(
|
create_helper_call_numpy_unary_elementwise_float_to_float!(
|
||||||
call_numpy_rint,
|
call_numpy_rint,
|
||||||
"np_rint",
|
"np_rint",
|
||||||
llvm_intrinsics::call_float_rint
|
llvm_intrinsics::call_float_roundeven
|
||||||
);
|
);
|
||||||
|
|
||||||
create_helper_call_numpy_unary_elementwise_float_to_float!(
|
create_helper_call_numpy_unary_elementwise_float_to_float!(
|
||||||
@ -1497,7 +1459,7 @@ pub fn call_numpy_arctan2<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1564,7 +1526,7 @@ pub fn call_numpy_copysign<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1631,7 +1593,7 @@ pub fn call_numpy_fmax<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1698,7 +1660,7 @@ pub fn call_numpy_fmin<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1821,7 +1783,7 @@ pub fn call_numpy_hypot<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1888,7 +1850,7 @@ pub fn call_numpy_nextafter<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
} else if is_ndarray2 {
|
} else if is_ndarray2 {
|
||||||
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
unpack_ndarray_var_tys(&mut ctx.unifier, x2_ty).0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
let x1_scalar_ty = if is_ndarray1 { dtype } else { x1_ty };
|
||||||
@ -1912,501 +1874,3 @@ pub fn call_numpy_nextafter<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
_ => unsupported_type(ctx, FN_NAME, &[x1_ty, x2_ty]),
|
_ => unsupported_type(ctx, FN_NAME, &[x1_ty, x2_ty]),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Allocates a struct with the fields specified by `out_matrices` and returns a pointer to it
|
|
||||||
fn build_output_struct<'ctx>(
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
out_matrices: Vec<BasicValueEnum<'ctx>>,
|
|
||||||
) -> PointerValue<'ctx> {
|
|
||||||
let field_ty =
|
|
||||||
out_matrices.iter().map(BasicValueEnum::get_type).collect::<Vec<BasicTypeEnum>>();
|
|
||||||
let out_ty = ctx.ctx.struct_type(&field_ty, false);
|
|
||||||
let out_ptr = ctx.builder.build_alloca(out_ty, "").unwrap();
|
|
||||||
|
|
||||||
for (i, v) in out_matrices.into_iter().enumerate() {
|
|
||||||
unsafe {
|
|
||||||
let ptr = ctx
|
|
||||||
.builder
|
|
||||||
.build_in_bounds_gep(
|
|
||||||
out_ptr,
|
|
||||||
&[
|
|
||||||
ctx.ctx.i32_type().const_zero(),
|
|
||||||
ctx.ctx.i32_type().const_int(i as u64, false),
|
|
||||||
],
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
ctx.builder.build_store(ptr, v).unwrap();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
out_ptr
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_cholesky` linalg function
|
|
||||||
pub fn call_np_linalg_cholesky<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_cholesky";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
|
|
||||||
let out = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_cholesky(ctx, x1, out, None);
|
|
||||||
Ok(out)
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_qr` linalg function
|
|
||||||
pub fn call_np_linalg_qr<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_qr";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unimplemented!("{FN_NAME} operates on float type NdArrays only");
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let k = llvm_intrinsics::call_int_smin(ctx, dim0, dim1, None);
|
|
||||||
|
|
||||||
let out_q = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, k])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_r = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[k, dim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_qr(ctx, x1, out_q, out_r, None);
|
|
||||||
|
|
||||||
let out_ptr = build_output_struct(ctx, vec![out_q, out_r]);
|
|
||||||
|
|
||||||
Ok(ctx.builder.build_load(out_ptr, "QR_Factorization_result").map(Into::into).unwrap())
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_svd` linalg function
|
|
||||||
pub fn call_np_linalg_svd<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_svd";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let k = llvm_intrinsics::call_int_smin(ctx, dim0, dim1, None);
|
|
||||||
|
|
||||||
let out_u = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_s = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[k])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_vh = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim1, dim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_svd(ctx, x1, out_u, out_s, out_vh, None);
|
|
||||||
|
|
||||||
let out_ptr = build_output_struct(ctx, vec![out_u, out_s, out_vh]);
|
|
||||||
|
|
||||||
Ok(ctx.builder.build_load(out_ptr, "SVD_Factorization_result").map(Into::into).unwrap())
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_inv` linalg function
|
|
||||||
pub fn call_np_linalg_inv<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_inv";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
|
|
||||||
let out = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_inv(ctx, x1, out, None);
|
|
||||||
Ok(out)
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_pinv` linalg function
|
|
||||||
pub fn call_np_linalg_pinv<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_pinv";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
|
|
||||||
let out = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim1, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_pinv(ctx, x1, out, None);
|
|
||||||
Ok(out)
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `sp_linalg_lu` linalg function
|
|
||||||
pub fn call_sp_linalg_lu<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "sp_linalg_lu";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let dim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let k = llvm_intrinsics::call_int_smin(ctx, dim0, dim1, None);
|
|
||||||
|
|
||||||
let out_l = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, k])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_u = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[k, dim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_sp_linalg_lu(ctx, x1, out_l, out_u, None);
|
|
||||||
|
|
||||||
let out_ptr = build_output_struct(ctx, vec![out_l, out_u]);
|
|
||||||
Ok(ctx.builder.build_load(out_ptr, "LU_Factorization_result").map(Into::into).unwrap())
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_matrix_power` linalg function
|
|
||||||
pub fn call_np_linalg_matrix_power<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
x2: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_matrix_power";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let (x2_ty, x2) = x2;
|
|
||||||
let x2 = call_float(generator, ctx, (x2_ty, x2)).unwrap();
|
|
||||||
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
if let (BasicValueEnum::PointerValue(n1), BasicValueEnum::FloatValue(n2)) = (x1, x2) {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty, x2_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
// Changing second parameter to a `NDArray` for uniformity in function call
|
|
||||||
let n2_array = numpy::create_ndarray_const_shape(
|
|
||||||
generator,
|
|
||||||
ctx,
|
|
||||||
elem_ty,
|
|
||||||
&[llvm_usize.const_int(1, false)],
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
unsafe {
|
|
||||||
n2_array.data().set_unchecked(
|
|
||||||
ctx,
|
|
||||||
generator,
|
|
||||||
&llvm_usize.const_zero(),
|
|
||||||
n2.as_basic_value_enum(),
|
|
||||||
);
|
|
||||||
};
|
|
||||||
let n2_array = n2_array.as_base_value().as_basic_value_enum();
|
|
||||||
|
|
||||||
let outdim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let outdim1 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_int(1, false), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
|
|
||||||
let out = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[outdim0, outdim1])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_np_linalg_matrix_power(ctx, x1, n2_array, out, None);
|
|
||||||
Ok(out)
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty, x2_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `np_linalg_det` linalg function
|
|
||||||
pub fn call_np_linalg_det<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "np_linalg_matrix_power";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
if let BasicValueEnum::PointerValue(_) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
// Changing second parameter to a `NDArray` for uniformity in function call
|
|
||||||
let out = numpy::create_ndarray_const_shape(
|
|
||||||
generator,
|
|
||||||
ctx,
|
|
||||||
elem_ty,
|
|
||||||
&[llvm_usize.const_int(1, false)],
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
extern_fns::call_np_linalg_det(ctx, x1, out.as_base_value().as_basic_value_enum(), None);
|
|
||||||
let res =
|
|
||||||
unsafe { out.data().get_unchecked(ctx, generator, &llvm_usize.const_zero(), None) };
|
|
||||||
Ok(res)
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `sp_linalg_schur` linalg function
|
|
||||||
pub fn call_sp_linalg_schur<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "sp_linalg_schur";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let out_t = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_z = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
|
|
||||||
extern_fns::call_sp_linalg_schur(ctx, x1, out_t, out_z, None);
|
|
||||||
|
|
||||||
let out_ptr = build_output_struct(ctx, vec![out_t, out_z]);
|
|
||||||
Ok(ctx.builder.build_load(out_ptr, "Schur_Factorization_result").map(Into::into).unwrap())
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the `sp_linalg_hessenberg` linalg function
|
|
||||||
pub fn call_sp_linalg_hessenberg<'ctx, G: CodeGenerator + ?Sized>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
x1: (Type, BasicValueEnum<'ctx>),
|
|
||||||
) -> Result<BasicValueEnum<'ctx>, String> {
|
|
||||||
const FN_NAME: &str = "sp_linalg_hessenberg";
|
|
||||||
let (x1_ty, x1) = x1;
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
|
|
||||||
if let BasicValueEnum::PointerValue(n1) = x1 {
|
|
||||||
let (elem_ty, _) = unpack_ndarray_var_tys(&mut ctx.unifier, x1_ty);
|
|
||||||
let n1_elem_ty = ctx.get_llvm_type(generator, elem_ty);
|
|
||||||
|
|
||||||
let BasicTypeEnum::FloatType(_) = n1_elem_ty else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty]);
|
|
||||||
};
|
|
||||||
|
|
||||||
let n1 = NDArrayValue::from_ptr_val(n1, llvm_usize, None);
|
|
||||||
|
|
||||||
let dim0 = unsafe {
|
|
||||||
n1.dim_sizes()
|
|
||||||
.get_unchecked(ctx, generator, &llvm_usize.const_zero(), None)
|
|
||||||
.into_int_value()
|
|
||||||
};
|
|
||||||
let out_h = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
let out_q = numpy::create_ndarray_const_shape(generator, ctx, elem_ty, &[dim0, dim0])
|
|
||||||
.unwrap()
|
|
||||||
.as_base_value()
|
|
||||||
.as_basic_value_enum();
|
|
||||||
extern_fns::call_sp_linalg_hessenberg(ctx, x1, out_h, out_q, None);
|
|
||||||
|
|
||||||
let out_ptr = build_output_struct(ctx, vec![out_h, out_q]);
|
|
||||||
Ok(ctx
|
|
||||||
.builder
|
|
||||||
.build_load(out_ptr, "Hessenberg_decomposition_result")
|
|
||||||
.map(Into::into)
|
|
||||||
.unwrap())
|
|
||||||
} else {
|
|
||||||
unsupported_type(ctx, FN_NAME, &[x1_ty])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
@ -1,16 +1,17 @@
|
|||||||
use inkwell::{
|
|
||||||
context::Context,
|
|
||||||
types::{AnyTypeEnum, ArrayType, BasicType, BasicTypeEnum, IntType, PointerType, StructType},
|
|
||||||
values::{ArrayValue, BasicValue, BasicValueEnum, IntValue, PointerValue, StructValue},
|
|
||||||
AddressSpace, IntPredicate,
|
|
||||||
};
|
|
||||||
|
|
||||||
use crate::codegen::{
|
use crate::codegen::{
|
||||||
irrt::{call_ndarray_calc_size, call_ndarray_flatten_index},
|
irrt::{call_ndarray_calc_size, call_ndarray_flatten_index},
|
||||||
llvm_intrinsics::call_int_umin,
|
llvm_intrinsics::call_int_umin,
|
||||||
stmt::gen_for_callback_incrementing,
|
stmt::gen_for_callback_incrementing,
|
||||||
CodeGenContext, CodeGenerator,
|
CodeGenContext, CodeGenerator,
|
||||||
};
|
};
|
||||||
|
use inkwell::context::Context;
|
||||||
|
use inkwell::types::{ArrayType, BasicType, StructType};
|
||||||
|
use inkwell::values::{ArrayValue, BasicValue, StructValue};
|
||||||
|
use inkwell::{
|
||||||
|
types::{AnyTypeEnum, BasicTypeEnum, IntType, PointerType},
|
||||||
|
values::{BasicValueEnum, IntValue, PointerValue},
|
||||||
|
AddressSpace, IntPredicate,
|
||||||
|
};
|
||||||
|
|
||||||
/// A LLVM type that is used to represent a non-primitive type in NAC3.
|
/// A LLVM type that is used to represent a non-primitive type in NAC3.
|
||||||
pub trait ProxyType<'ctx>: Into<Self::Base> {
|
pub trait ProxyType<'ctx>: Into<Self::Base> {
|
||||||
@ -712,25 +713,12 @@ impl<'ctx> ListValue<'ctx> {
|
|||||||
/// If `size` is [None], the size stored in the field of this instance is used instead.
|
/// If `size` is [None], the size stored in the field of this instance is used instead.
|
||||||
pub fn create_data(
|
pub fn create_data(
|
||||||
&self,
|
&self,
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
elem_ty: BasicTypeEnum<'ctx>,
|
elem_ty: BasicTypeEnum<'ctx>,
|
||||||
size: Option<IntValue<'ctx>>,
|
size: Option<IntValue<'ctx>>,
|
||||||
) {
|
) {
|
||||||
let size = size.unwrap_or_else(|| self.load_size(ctx, None));
|
let size = size.unwrap_or_else(|| self.load_size(ctx, None));
|
||||||
|
self.store_data(ctx, ctx.builder.build_array_alloca(elem_ty, size, "").unwrap());
|
||||||
let data = ctx
|
|
||||||
.builder
|
|
||||||
.build_select(
|
|
||||||
ctx.builder
|
|
||||||
.build_int_compare(IntPredicate::NE, size, self.llvm_usize.const_zero(), "")
|
|
||||||
.unwrap(),
|
|
||||||
ctx.builder.build_array_alloca(elem_ty, size, "").unwrap(),
|
|
||||||
elem_ty.ptr_type(AddressSpace::default()).const_zero(),
|
|
||||||
"",
|
|
||||||
)
|
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
|
||||||
.unwrap();
|
|
||||||
self.store_data(ctx, data);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns the double-indirection pointer to the `data` array, as if by calling `getelementptr`
|
/// Returns the double-indirection pointer to the `data` array, as if by calling `getelementptr`
|
||||||
@ -1249,13 +1237,11 @@ impl<'ctx> NDArrayType<'ctx> {
|
|||||||
|
|
||||||
/// Returns the element type of this `ndarray` type.
|
/// Returns the element type of this `ndarray` type.
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn element_type(&self) -> AnyTypeEnum<'ctx> {
|
pub fn element_type(&self) -> BasicTypeEnum<'ctx> {
|
||||||
self.as_base_type()
|
self.as_base_type()
|
||||||
.get_element_type()
|
.get_element_type()
|
||||||
.into_struct_type()
|
.into_struct_type()
|
||||||
.get_field_type_at_index(2)
|
.get_field_type_at_index(2)
|
||||||
.map(BasicTypeEnum::into_pointer_type)
|
|
||||||
.map(PointerType::get_element_type)
|
|
||||||
.unwrap()
|
.unwrap()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1405,7 +1391,7 @@ impl<'ctx> NDArrayValue<'ctx> {
|
|||||||
|
|
||||||
/// Returns the double-indirection pointer to the `data` array, as if by calling `getelementptr`
|
/// Returns the double-indirection pointer to the `data` array, as if by calling `getelementptr`
|
||||||
/// on the field.
|
/// on the field.
|
||||||
pub fn ptr_to_data(&self, ctx: &CodeGenContext<'ctx, '_>) -> PointerValue<'ctx> {
|
fn ptr_to_data(&self, ctx: &CodeGenContext<'ctx, '_>) -> PointerValue<'ctx> {
|
||||||
let llvm_i32 = ctx.ctx.i32_type();
|
let llvm_i32 = ctx.ctx.i32_type();
|
||||||
let var_name = self.name.map(|v| format!("{v}.data.addr")).unwrap_or_default();
|
let var_name = self.name.map(|v| format!("{v}.data.addr")).unwrap_or_default();
|
||||||
|
|
||||||
@ -1718,7 +1704,6 @@ impl<'ctx, Index: UntypedArrayLikeAccessor<'ctx>> ArrayLikeIndexer<'ctx, Index>
|
|||||||
gen_for_callback_incrementing(
|
gen_for_callback_incrementing(
|
||||||
generator,
|
generator,
|
||||||
ctx,
|
ctx,
|
||||||
None,
|
|
||||||
llvm_usize.const_zero(),
|
llvm_usize.const_zero(),
|
||||||
(len, false),
|
(len, false),
|
||||||
|generator, ctx, _, i| {
|
|generator, ctx, _, i| {
|
||||||
|
@ -1,9 +1,3 @@
|
|||||||
use std::collections::HashMap;
|
|
||||||
|
|
||||||
use indexmap::IndexMap;
|
|
||||||
|
|
||||||
use nac3parser::ast::StrRef;
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
symbol_resolver::SymbolValue,
|
symbol_resolver::SymbolValue,
|
||||||
toplevel::DefinitionId,
|
toplevel::DefinitionId,
|
||||||
@ -15,6 +9,10 @@ use crate::{
|
|||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use indexmap::IndexMap;
|
||||||
|
use nac3parser::ast::StrRef;
|
||||||
|
use std::collections::HashMap;
|
||||||
|
|
||||||
pub struct ConcreteTypeStore {
|
pub struct ConcreteTypeStore {
|
||||||
store: Vec<ConcreteTypeEnum>,
|
store: Vec<ConcreteTypeEnum>,
|
||||||
}
|
}
|
||||||
@ -27,7 +25,6 @@ pub struct ConcreteFuncArg {
|
|||||||
pub name: StrRef,
|
pub name: StrRef,
|
||||||
pub ty: ConcreteType,
|
pub ty: ConcreteType,
|
||||||
pub default_value: Option<SymbolValue>,
|
pub default_value: Option<SymbolValue>,
|
||||||
pub is_vararg: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
@ -49,7 +46,9 @@ pub enum ConcreteTypeEnum {
|
|||||||
TPrimitive(Primitive),
|
TPrimitive(Primitive),
|
||||||
TTuple {
|
TTuple {
|
||||||
ty: Vec<ConcreteType>,
|
ty: Vec<ConcreteType>,
|
||||||
is_vararg_ctx: bool,
|
},
|
||||||
|
TList {
|
||||||
|
ty: ConcreteType,
|
||||||
},
|
},
|
||||||
TObj {
|
TObj {
|
||||||
obj_id: DefinitionId,
|
obj_id: DefinitionId,
|
||||||
@ -106,16 +105,8 @@ impl ConcreteTypeStore {
|
|||||||
.iter()
|
.iter()
|
||||||
.map(|arg| ConcreteFuncArg {
|
.map(|arg| ConcreteFuncArg {
|
||||||
name: arg.name,
|
name: arg.name,
|
||||||
ty: if arg.is_vararg {
|
ty: self.from_unifier_type(unifier, primitives, arg.ty, cache),
|
||||||
let tuple_ty = unifier
|
|
||||||
.add_ty(TypeEnum::TTuple { ty: vec![arg.ty], is_vararg_ctx: true });
|
|
||||||
|
|
||||||
self.from_unifier_type(unifier, primitives, tuple_ty, cache)
|
|
||||||
} else {
|
|
||||||
self.from_unifier_type(unifier, primitives, arg.ty, cache)
|
|
||||||
},
|
|
||||||
default_value: arg.default_value.clone(),
|
default_value: arg.default_value.clone(),
|
||||||
is_vararg: arg.is_vararg,
|
|
||||||
})
|
})
|
||||||
.collect(),
|
.collect(),
|
||||||
ret: self.from_unifier_type(unifier, primitives, signature.ret, cache),
|
ret: self.from_unifier_type(unifier, primitives, signature.ret, cache),
|
||||||
@ -170,12 +161,14 @@ impl ConcreteTypeStore {
|
|||||||
cache.insert(ty, None);
|
cache.insert(ty, None);
|
||||||
let ty_enum = unifier.get_ty(ty);
|
let ty_enum = unifier.get_ty(ty);
|
||||||
let result = match &*ty_enum {
|
let result = match &*ty_enum {
|
||||||
TypeEnum::TTuple { ty, is_vararg_ctx } => ConcreteTypeEnum::TTuple {
|
TypeEnum::TTuple { ty } => ConcreteTypeEnum::TTuple {
|
||||||
ty: ty
|
ty: ty
|
||||||
.iter()
|
.iter()
|
||||||
.map(|t| self.from_unifier_type(unifier, primitives, *t, cache))
|
.map(|t| self.from_unifier_type(unifier, primitives, *t, cache))
|
||||||
.collect(),
|
.collect(),
|
||||||
is_vararg_ctx: *is_vararg_ctx,
|
},
|
||||||
|
TypeEnum::TList { ty } => ConcreteTypeEnum::TList {
|
||||||
|
ty: self.from_unifier_type(unifier, primitives, *ty, cache),
|
||||||
},
|
},
|
||||||
TypeEnum::TObj { obj_id, fields, params } => ConcreteTypeEnum::TObj {
|
TypeEnum::TObj { obj_id, fields, params } => ConcreteTypeEnum::TObj {
|
||||||
obj_id: *obj_id,
|
obj_id: *obj_id,
|
||||||
@ -261,13 +254,15 @@ impl ConcreteTypeStore {
|
|||||||
*cache.get_mut(&cty).unwrap() = Some(ty);
|
*cache.get_mut(&cty).unwrap() = Some(ty);
|
||||||
return ty;
|
return ty;
|
||||||
}
|
}
|
||||||
ConcreteTypeEnum::TTuple { ty, is_vararg_ctx } => TypeEnum::TTuple {
|
ConcreteTypeEnum::TTuple { ty } => TypeEnum::TTuple {
|
||||||
ty: ty
|
ty: ty
|
||||||
.iter()
|
.iter()
|
||||||
.map(|cty| self.to_unifier_type(unifier, primitives, *cty, cache))
|
.map(|cty| self.to_unifier_type(unifier, primitives, *cty, cache))
|
||||||
.collect(),
|
.collect(),
|
||||||
is_vararg_ctx: *is_vararg_ctx,
|
|
||||||
},
|
},
|
||||||
|
ConcreteTypeEnum::TList { ty } => {
|
||||||
|
TypeEnum::TList { ty: self.to_unifier_type(unifier, primitives, *ty, cache) }
|
||||||
|
}
|
||||||
ConcreteTypeEnum::TVirtual { ty } => {
|
ConcreteTypeEnum::TVirtual { ty } => {
|
||||||
TypeEnum::TVirtual { ty: self.to_unifier_type(unifier, primitives, *ty, cache) }
|
TypeEnum::TVirtual { ty: self.to_unifier_type(unifier, primitives, *ty, cache) }
|
||||||
}
|
}
|
||||||
@ -291,7 +286,6 @@ impl ConcreteTypeStore {
|
|||||||
name: arg.name,
|
name: arg.name,
|
||||||
ty: self.to_unifier_type(unifier, primitives, arg.ty, cache),
|
ty: self.to_unifier_type(unifier, primitives, arg.ty, cache),
|
||||||
default_value: arg.default_value.clone(),
|
default_value: arg.default_value.clone(),
|
||||||
is_vararg: false,
|
|
||||||
})
|
})
|
||||||
.collect(),
|
.collect(),
|
||||||
ret: self.to_unifier_type(unifier, primitives, *ret, cache),
|
ret: self.to_unifier_type(unifier, primitives, *ret, cache),
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -1,102 +1,517 @@
|
|||||||
use inkwell::{
|
use inkwell::attributes::{Attribute, AttributeLoc};
|
||||||
attributes::{Attribute, AttributeLoc},
|
use inkwell::values::{BasicValueEnum, CallSiteValue, FloatValue, IntValue};
|
||||||
values::{BasicValueEnum, CallSiteValue, FloatValue, IntValue},
|
|
||||||
};
|
|
||||||
use itertools::Either;
|
use itertools::Either;
|
||||||
|
|
||||||
use crate::codegen::CodeGenContext;
|
use crate::codegen::CodeGenContext;
|
||||||
|
|
||||||
/// Macro to generate extern function
|
/// Invokes the [`tan`](https://en.cppreference.com/w/c/numeric/math/tan) function.
|
||||||
/// Both function return type and function parameter type are `FloatValue`
|
pub fn call_tan<'ctx>(
|
||||||
///
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
/// Arguments:
|
arg: FloatValue<'ctx>,
|
||||||
/// * `unary/binary`: Whether the extern function requires one (unary) or two (binary) operands
|
|
||||||
/// * `$fn_name:ident`: The identifier of the rust function to be generated
|
|
||||||
/// * `$extern_fn:literal`: Name of underlying extern function
|
|
||||||
///
|
|
||||||
/// Optional Arguments:
|
|
||||||
/// * `$(,$attributes:literal)*)`: Attributes linked with the extern function.
|
|
||||||
/// The default attributes are "mustprogress", "nofree", "nounwind", "willreturn", and "writeonly".
|
|
||||||
/// These will be used unless other attributes are specified
|
|
||||||
/// * `$(,$args:ident)*`: Operands of the extern function
|
|
||||||
/// The data type of these operands will be set to `FloatValue`
|
|
||||||
///
|
|
||||||
macro_rules! generate_extern_fn {
|
|
||||||
("unary", $fn_name:ident, $extern_fn:literal) => {
|
|
||||||
generate_extern_fn!($fn_name, $extern_fn, arg, "mustprogress", "nofree", "nounwind", "willreturn", "writeonly");
|
|
||||||
};
|
|
||||||
("unary", $fn_name:ident, $extern_fn:literal $(,$attributes:literal)*) => {
|
|
||||||
generate_extern_fn!($fn_name, $extern_fn, arg $(,$attributes)*);
|
|
||||||
};
|
|
||||||
("binary", $fn_name:ident, $extern_fn:literal) => {
|
|
||||||
generate_extern_fn!($fn_name, $extern_fn, arg1, arg2, "mustprogress", "nofree", "nounwind", "willreturn", "writeonly");
|
|
||||||
};
|
|
||||||
("binary", $fn_name:ident, $extern_fn:literal $(,$attributes:literal)*) => {
|
|
||||||
generate_extern_fn!($fn_name, $extern_fn, arg1, arg2 $(,$attributes)*);
|
|
||||||
};
|
|
||||||
($fn_name:ident, $extern_fn:literal $(,$args:ident)* $(,$attributes:literal)*) => {
|
|
||||||
#[doc = concat!("Invokes the [`", stringify!($extern_fn), "`](https://en.cppreference.com/w/c/numeric/math/", stringify!($llvm_name), ") function." )]
|
|
||||||
pub fn $fn_name<'ctx>(
|
|
||||||
ctx: &CodeGenContext<'ctx, '_>
|
|
||||||
$(,$args: FloatValue<'ctx>)*,
|
|
||||||
name: Option<&str>,
|
name: Option<&str>,
|
||||||
) -> FloatValue<'ctx> {
|
) -> FloatValue<'ctx> {
|
||||||
const FN_NAME: &str = $extern_fn;
|
const FN_NAME: &str = "tan";
|
||||||
|
|
||||||
let llvm_f64 = ctx.ctx.f64_type();
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
$(debug_assert_eq!($args.get_type(), llvm_f64);)*
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
let fn_type = llvm_f64.fn_type(&[$($args.get_type().into()),*], false);
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
for attr in [$($attributes),*] {
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
func.add_attribute(
|
func.add_attribute(
|
||||||
AttributeLoc::Function,
|
AttributeLoc::Function,
|
||||||
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
func
|
func
|
||||||
});
|
});
|
||||||
|
|
||||||
ctx.builder
|
ctx.builder
|
||||||
.build_call(extern_fn, &[$($args.into()),*], name.unwrap_or_default())
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
.map(CallSiteValue::try_as_basic_value)
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
.map(Either::unwrap_left)
|
.map(Either::unwrap_left)
|
||||||
.unwrap()
|
.unwrap()
|
||||||
}
|
}
|
||||||
};
|
|
||||||
|
/// Invokes the [`asin`](https://en.cppreference.com/w/c/numeric/math/asin) function.
|
||||||
|
pub fn call_asin<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "asin";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
generate_extern_fn!("unary", call_tan, "tan");
|
func
|
||||||
generate_extern_fn!("unary", call_asin, "asin");
|
});
|
||||||
generate_extern_fn!("unary", call_acos, "acos");
|
|
||||||
generate_extern_fn!("unary", call_atan, "atan");
|
|
||||||
generate_extern_fn!("unary", call_sinh, "sinh");
|
|
||||||
generate_extern_fn!("unary", call_cosh, "cosh");
|
|
||||||
generate_extern_fn!("unary", call_tanh, "tanh");
|
|
||||||
generate_extern_fn!("unary", call_asinh, "asinh");
|
|
||||||
generate_extern_fn!("unary", call_acosh, "acosh");
|
|
||||||
generate_extern_fn!("unary", call_atanh, "atanh");
|
|
||||||
generate_extern_fn!("unary", call_expm1, "expm1");
|
|
||||||
generate_extern_fn!(
|
|
||||||
"unary",
|
|
||||||
call_cbrt,
|
|
||||||
"cbrt",
|
|
||||||
"mustprogress",
|
|
||||||
"nofree",
|
|
||||||
"nosync",
|
|
||||||
"nounwind",
|
|
||||||
"readonly",
|
|
||||||
"willreturn"
|
|
||||||
);
|
|
||||||
generate_extern_fn!("unary", call_erf, "erf", "nounwind");
|
|
||||||
generate_extern_fn!("unary", call_erfc, "erfc", "nounwind");
|
|
||||||
generate_extern_fn!("unary", call_j1, "j1", "nounwind");
|
|
||||||
|
|
||||||
generate_extern_fn!("binary", call_atan2, "atan2");
|
ctx.builder
|
||||||
generate_extern_fn!("binary", call_hypot, "hypot", "nounwind");
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
generate_extern_fn!("binary", call_nextafter, "nextafter", "nounwind");
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`acos`](https://en.cppreference.com/w/c/numeric/math/acos) function.
|
||||||
|
pub fn call_acos<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "acos";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`atan`](https://en.cppreference.com/w/c/numeric/math/atan) function.
|
||||||
|
pub fn call_atan<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "atan";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`sinh`](https://en.cppreference.com/w/c/numeric/math/sinh) function.
|
||||||
|
pub fn call_sinh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "sinh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`cosh`](https://en.cppreference.com/w/c/numeric/math/cosh) function.
|
||||||
|
pub fn call_cosh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "cosh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`tanh`](https://en.cppreference.com/w/c/numeric/math/tanh) function.
|
||||||
|
pub fn call_tanh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "tanh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`asinh`](https://en.cppreference.com/w/c/numeric/math/asinh) function.
|
||||||
|
pub fn call_asinh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "asinh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`acosh`](https://en.cppreference.com/w/c/numeric/math/acosh) function.
|
||||||
|
pub fn call_acosh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "acosh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`atanh`](https://en.cppreference.com/w/c/numeric/math/atanh) function.
|
||||||
|
pub fn call_atanh<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "atanh";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`expm1`](https://en.cppreference.com/w/c/numeric/math/expm1) function.
|
||||||
|
pub fn call_expm1<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "expm1";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`cbrt`](https://en.cppreference.com/w/c/numeric/math/cbrt) function.
|
||||||
|
pub fn call_cbrt<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "cbrt";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nosync", "nounwind", "readonly", "willreturn"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`erf`](https://en.cppreference.com/w/c/numeric/math/erf) function.
|
||||||
|
pub fn call_erf<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "erf";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id("nounwind"), 0),
|
||||||
|
);
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`erfc`](https://en.cppreference.com/w/c/numeric/math/erfc) function.
|
||||||
|
pub fn call_erfc<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "erfc";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id("nounwind"), 0),
|
||||||
|
);
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`j1`](https://www.gnu.org/software/libc/manual/html_node/Special-Functions.html#index-j1)
|
||||||
|
/// function.
|
||||||
|
pub fn call_j1<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
arg: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "j1";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(arg.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id("nounwind"), 0),
|
||||||
|
);
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[arg.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`atan2`](https://en.cppreference.com/w/c/numeric/math/atan2) function.
|
||||||
|
pub fn call_atan2<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
y: FloatValue<'ctx>,
|
||||||
|
x: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "atan2";
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(y.get_type(), llvm_f64);
|
||||||
|
debug_assert_eq!(x.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into(), llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[y.into(), x.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
/// Invokes the [`ldexp`](https://en.cppreference.com/w/c/numeric/math/ldexp) function.
|
/// Invokes the [`ldexp`](https://en.cppreference.com/w/c/numeric/math/ldexp) function.
|
||||||
pub fn call_ldexp<'ctx>(
|
pub fn call_ldexp<'ctx>(
|
||||||
@ -133,61 +548,66 @@ pub fn call_ldexp<'ctx>(
|
|||||||
.unwrap()
|
.unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Macro to generate `np_linalg` and `sp_linalg` functions
|
/// Invokes the [`hypot`](https://en.cppreference.com/w/c/numeric/math/hypot) function.
|
||||||
/// The function takes as input `NDArray` and returns ()
|
pub fn call_hypot<'ctx>(
|
||||||
///
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
/// Arguments:
|
x: FloatValue<'ctx>,
|
||||||
/// * `$fn_name:ident`: The identifier of the rust function to be generated
|
y: FloatValue<'ctx>,
|
||||||
/// * `$extern_fn:literal`: Name of underlying extern function
|
|
||||||
/// * (2/3/4): Number of `NDArray` that function takes as input
|
|
||||||
///
|
|
||||||
/// Note:
|
|
||||||
/// The operands and resulting `NDArray` are both passed as input to the funcion
|
|
||||||
/// It is the responsibility of caller to ensure that output `NDArray` is properly allocated on stack
|
|
||||||
/// The function changes the content of the output `NDArray` in-place
|
|
||||||
macro_rules! generate_linalg_extern_fn {
|
|
||||||
($fn_name:ident, $extern_fn:literal, 2) => {
|
|
||||||
generate_linalg_extern_fn!($fn_name, $extern_fn, mat1, mat2);
|
|
||||||
};
|
|
||||||
($fn_name:ident, $extern_fn:literal, 3) => {
|
|
||||||
generate_linalg_extern_fn!($fn_name, $extern_fn, mat1, mat2, mat3);
|
|
||||||
};
|
|
||||||
($fn_name:ident, $extern_fn:literal, 4) => {
|
|
||||||
generate_linalg_extern_fn!($fn_name, $extern_fn, mat1, mat2, mat3, mat4);
|
|
||||||
};
|
|
||||||
($fn_name:ident, $extern_fn:literal $(,$input_matrix:ident)*) => {
|
|
||||||
#[doc = concat!("Invokes the linalg `", stringify!($extern_fn), " function." )]
|
|
||||||
pub fn $fn_name<'ctx>(
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>
|
|
||||||
$(,$input_matrix: BasicValueEnum<'ctx>)*,
|
|
||||||
name: Option<&str>,
|
name: Option<&str>,
|
||||||
){
|
) -> FloatValue<'ctx> {
|
||||||
const FN_NAME: &str = $extern_fn;
|
const FN_NAME: &str = "hypot";
|
||||||
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
|
||||||
let fn_type = ctx.ctx.void_type().fn_type(&[$($input_matrix.get_type().into()),*], false);
|
|
||||||
|
|
||||||
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(x.get_type(), llvm_f64);
|
||||||
|
debug_assert_eq!(y.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into(), llvm_f64.into()], false);
|
||||||
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
for attr in ["mustprogress", "nofree", "nounwind", "willreturn", "writeonly"] {
|
|
||||||
func.add_attribute(
|
func.add_attribute(
|
||||||
AttributeLoc::Function,
|
AttributeLoc::Function,
|
||||||
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id(attr), 0),
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id("nounwind"), 0),
|
||||||
);
|
);
|
||||||
}
|
|
||||||
func
|
func
|
||||||
});
|
});
|
||||||
|
|
||||||
ctx.builder.build_call(extern_fn, &[$($input_matrix.into(),)*], name.unwrap_or_default()).unwrap();
|
ctx.builder
|
||||||
}
|
.build_call(extern_fn, &[x.into(), y.into()], name.unwrap_or_default())
|
||||||
};
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
generate_linalg_extern_fn!(call_np_linalg_cholesky, "np_linalg_cholesky", 2);
|
/// Invokes the [`nextafter`](https://en.cppreference.com/w/c/numeric/math/nextafter) function.
|
||||||
generate_linalg_extern_fn!(call_np_linalg_qr, "np_linalg_qr", 3);
|
pub fn call_nextafter<'ctx>(
|
||||||
generate_linalg_extern_fn!(call_np_linalg_svd, "np_linalg_svd", 4);
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
generate_linalg_extern_fn!(call_np_linalg_inv, "np_linalg_inv", 2);
|
from: FloatValue<'ctx>,
|
||||||
generate_linalg_extern_fn!(call_np_linalg_pinv, "np_linalg_pinv", 2);
|
to: FloatValue<'ctx>,
|
||||||
generate_linalg_extern_fn!(call_np_linalg_matrix_power, "np_linalg_matrix_power", 3);
|
name: Option<&str>,
|
||||||
generate_linalg_extern_fn!(call_np_linalg_det, "np_linalg_det", 2);
|
) -> FloatValue<'ctx> {
|
||||||
generate_linalg_extern_fn!(call_sp_linalg_lu, "sp_linalg_lu", 3);
|
const FN_NAME: &str = "nextafter";
|
||||||
generate_linalg_extern_fn!(call_sp_linalg_schur, "sp_linalg_schur", 3);
|
|
||||||
generate_linalg_extern_fn!(call_sp_linalg_hessenberg, "sp_linalg_hessenberg", 3);
|
let llvm_f64 = ctx.ctx.f64_type();
|
||||||
|
debug_assert_eq!(from.get_type(), llvm_f64);
|
||||||
|
debug_assert_eq!(to.get_type(), llvm_f64);
|
||||||
|
|
||||||
|
let extern_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
||||||
|
let fn_type = llvm_f64.fn_type(&[llvm_f64.into(), llvm_f64.into()], false);
|
||||||
|
let func = ctx.module.add_function(FN_NAME, fn_type, None);
|
||||||
|
func.add_attribute(
|
||||||
|
AttributeLoc::Function,
|
||||||
|
ctx.ctx.create_enum_attribute(Attribute::get_named_enum_kind_id("nounwind"), 0),
|
||||||
|
);
|
||||||
|
|
||||||
|
func
|
||||||
|
});
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(extern_fn, &[from.into(), to.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
@ -1,17 +1,15 @@
|
|||||||
use inkwell::{
|
|
||||||
context::Context,
|
|
||||||
types::{BasicTypeEnum, IntType},
|
|
||||||
values::{BasicValueEnum, IntValue, PointerValue},
|
|
||||||
};
|
|
||||||
|
|
||||||
use nac3parser::ast::{Expr, Stmt, StrRef};
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{bool_to_i1, bool_to_i8, classes::ArraySliceValue, expr::*, stmt::*, CodeGenContext},
|
codegen::{bool_to_i1, bool_to_i8, classes::ArraySliceValue, expr::*, stmt::*, CodeGenContext},
|
||||||
symbol_resolver::ValueEnum,
|
symbol_resolver::ValueEnum,
|
||||||
toplevel::{DefinitionId, TopLevelDef},
|
toplevel::{DefinitionId, TopLevelDef},
|
||||||
typecheck::typedef::{FunSignature, Type},
|
typecheck::typedef::{FunSignature, Type},
|
||||||
};
|
};
|
||||||
|
use inkwell::{
|
||||||
|
context::Context,
|
||||||
|
types::{BasicTypeEnum, IntType},
|
||||||
|
values::{BasicValueEnum, IntValue, PointerValue},
|
||||||
|
};
|
||||||
|
use nac3parser::ast::{Expr, Stmt, StrRef};
|
||||||
|
|
||||||
pub trait CodeGenerator {
|
pub trait CodeGenerator {
|
||||||
/// Return the module name for the code generator.
|
/// Return the module name for the code generator.
|
||||||
@ -59,7 +57,6 @@ pub trait CodeGenerator {
|
|||||||
/// - fun: Function signature, definition ID and the substitution key.
|
/// - fun: Function signature, definition ID and the substitution key.
|
||||||
/// - params: Function parameters. Note that this does not include the object even if the
|
/// - params: Function parameters. Note that this does not include the object even if the
|
||||||
/// function is a class method.
|
/// function is a class method.
|
||||||
///
|
|
||||||
/// Note that this function should check if the function is generated in another thread (due to
|
/// Note that this function should check if the function is generated in another thread (due to
|
||||||
/// possible race condition), see the default implementation for an example.
|
/// possible race condition), see the default implementation for an example.
|
||||||
fn gen_func_instance<'ctx>(
|
fn gen_func_instance<'ctx>(
|
||||||
@ -126,45 +123,11 @@ pub trait CodeGenerator {
|
|||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
ctx: &mut CodeGenContext<'ctx, '_>,
|
||||||
target: &Expr<Option<Type>>,
|
target: &Expr<Option<Type>>,
|
||||||
value: ValueEnum<'ctx>,
|
value: ValueEnum<'ctx>,
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String>
|
) -> Result<(), String>
|
||||||
where
|
where
|
||||||
Self: Sized,
|
Self: Sized,
|
||||||
{
|
{
|
||||||
gen_assign(self, ctx, target, value, value_ty)
|
gen_assign(self, ctx, target, value)
|
||||||
}
|
|
||||||
|
|
||||||
/// Generate code for an assignment expression where LHS is a `"target_list"`.
|
|
||||||
///
|
|
||||||
/// See <https://docs.python.org/3/reference/simple_stmts.html#assignment-statements>.
|
|
||||||
fn gen_assign_target_list<'ctx>(
|
|
||||||
&mut self,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
targets: &Vec<Expr<Option<Type>>>,
|
|
||||||
value: ValueEnum<'ctx>,
|
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String>
|
|
||||||
where
|
|
||||||
Self: Sized,
|
|
||||||
{
|
|
||||||
gen_assign_target_list(self, ctx, targets, value, value_ty)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Generate code for an item assignment.
|
|
||||||
///
|
|
||||||
/// i.e., `target[key] = value`
|
|
||||||
fn gen_setitem<'ctx>(
|
|
||||||
&mut self,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
target: &Expr<Option<Type>>,
|
|
||||||
key: &Expr<Option<Type>>,
|
|
||||||
value: ValueEnum<'ctx>,
|
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String>
|
|
||||||
where
|
|
||||||
Self: Sized,
|
|
||||||
{
|
|
||||||
gen_setitem(self, ctx, target, key, value, value_ty)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Generate code for a while expression.
|
/// Generate code for a while expression.
|
||||||
|
389
nac3core/src/codegen/irrt/irrt.c
Normal file
389
nac3core/src/codegen/irrt/irrt.c
Normal file
@ -0,0 +1,389 @@
|
|||||||
|
typedef _BitInt(8) int8_t;
|
||||||
|
typedef unsigned _BitInt(8) uint8_t;
|
||||||
|
typedef _BitInt(32) int32_t;
|
||||||
|
typedef unsigned _BitInt(32) uint32_t;
|
||||||
|
typedef _BitInt(64) int64_t;
|
||||||
|
typedef unsigned _BitInt(64) uint64_t;
|
||||||
|
|
||||||
|
# define MAX(a, b) (a > b ? a : b)
|
||||||
|
# define MIN(a, b) (a > b ? b : a)
|
||||||
|
|
||||||
|
# define NULL ((void *) 0)
|
||||||
|
|
||||||
|
// adapted from GNU Scientific Library: https://git.savannah.gnu.org/cgit/gsl.git/tree/sys/pow_int.c
|
||||||
|
// need to make sure `exp >= 0` before calling this function
|
||||||
|
#define DEF_INT_EXP(T) T __nac3_int_exp_##T( \
|
||||||
|
T base, \
|
||||||
|
T exp \
|
||||||
|
) { \
|
||||||
|
T res = (T)1; \
|
||||||
|
/* repeated squaring method */ \
|
||||||
|
do { \
|
||||||
|
if (exp & 1) res *= base; /* for n odd */ \
|
||||||
|
exp >>= 1; \
|
||||||
|
base *= base; \
|
||||||
|
} while (exp); \
|
||||||
|
return res; \
|
||||||
|
} \
|
||||||
|
|
||||||
|
DEF_INT_EXP(int32_t)
|
||||||
|
DEF_INT_EXP(int64_t)
|
||||||
|
DEF_INT_EXP(uint32_t)
|
||||||
|
DEF_INT_EXP(uint64_t)
|
||||||
|
|
||||||
|
|
||||||
|
int32_t __nac3_slice_index_bound(int32_t i, const int32_t len) {
|
||||||
|
if (i < 0) {
|
||||||
|
i = len + i;
|
||||||
|
}
|
||||||
|
if (i < 0) {
|
||||||
|
return 0;
|
||||||
|
} else if (i > len) {
|
||||||
|
return len;
|
||||||
|
}
|
||||||
|
return i;
|
||||||
|
}
|
||||||
|
|
||||||
|
int32_t __nac3_range_slice_len(const int32_t start, const int32_t end, const int32_t step) {
|
||||||
|
int32_t diff = end - start;
|
||||||
|
if (diff > 0 && step > 0) {
|
||||||
|
return ((diff - 1) / step) + 1;
|
||||||
|
} else if (diff < 0 && step < 0) {
|
||||||
|
return ((diff + 1) / step) + 1;
|
||||||
|
} else {
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handle list assignment and dropping part of the list when
|
||||||
|
// both dest_step and src_step are +1.
|
||||||
|
// - All the index must *not* be out-of-bound or negative,
|
||||||
|
// - The end index is *inclusive*,
|
||||||
|
// - The length of src and dest slice size should already
|
||||||
|
// be checked: if dest.step == 1 then len(src) <= len(dest) else len(src) == len(dest)
|
||||||
|
int32_t __nac3_list_slice_assign_var_size(
|
||||||
|
int32_t dest_start,
|
||||||
|
int32_t dest_end,
|
||||||
|
int32_t dest_step,
|
||||||
|
uint8_t *dest_arr,
|
||||||
|
int32_t dest_arr_len,
|
||||||
|
int32_t src_start,
|
||||||
|
int32_t src_end,
|
||||||
|
int32_t src_step,
|
||||||
|
uint8_t *src_arr,
|
||||||
|
int32_t src_arr_len,
|
||||||
|
const int32_t size
|
||||||
|
) {
|
||||||
|
/* if dest_arr_len == 0, do nothing since we do not support extending list */
|
||||||
|
if (dest_arr_len == 0) return dest_arr_len;
|
||||||
|
/* if both step is 1, memmove directly, handle the dropping of the list, and shrink size */
|
||||||
|
if (src_step == dest_step && dest_step == 1) {
|
||||||
|
const int32_t src_len = (src_end >= src_start) ? (src_end - src_start + 1) : 0;
|
||||||
|
const int32_t dest_len = (dest_end >= dest_start) ? (dest_end - dest_start + 1) : 0;
|
||||||
|
if (src_len > 0) {
|
||||||
|
__builtin_memmove(
|
||||||
|
dest_arr + dest_start * size,
|
||||||
|
src_arr + src_start * size,
|
||||||
|
src_len * size
|
||||||
|
);
|
||||||
|
}
|
||||||
|
if (dest_len > 0) {
|
||||||
|
/* dropping */
|
||||||
|
__builtin_memmove(
|
||||||
|
dest_arr + (dest_start + src_len) * size,
|
||||||
|
dest_arr + (dest_end + 1) * size,
|
||||||
|
(dest_arr_len - dest_end - 1) * size
|
||||||
|
);
|
||||||
|
}
|
||||||
|
/* shrink size */
|
||||||
|
return dest_arr_len - (dest_len - src_len);
|
||||||
|
}
|
||||||
|
/* if two range overlaps, need alloca */
|
||||||
|
uint8_t need_alloca =
|
||||||
|
(dest_arr == src_arr)
|
||||||
|
&& !(
|
||||||
|
MAX(dest_start, dest_end) < MIN(src_start, src_end)
|
||||||
|
|| MAX(src_start, src_end) < MIN(dest_start, dest_end)
|
||||||
|
);
|
||||||
|
if (need_alloca) {
|
||||||
|
uint8_t *tmp = __builtin_alloca(src_arr_len * size);
|
||||||
|
__builtin_memcpy(tmp, src_arr, src_arr_len * size);
|
||||||
|
src_arr = tmp;
|
||||||
|
}
|
||||||
|
int32_t src_ind = src_start;
|
||||||
|
int32_t dest_ind = dest_start;
|
||||||
|
for (;
|
||||||
|
(src_step > 0) ? (src_ind <= src_end) : (src_ind >= src_end);
|
||||||
|
src_ind += src_step, dest_ind += dest_step
|
||||||
|
) {
|
||||||
|
/* for constant optimization */
|
||||||
|
if (size == 1) {
|
||||||
|
__builtin_memcpy(dest_arr + dest_ind, src_arr + src_ind, 1);
|
||||||
|
} else if (size == 4) {
|
||||||
|
__builtin_memcpy(dest_arr + dest_ind * 4, src_arr + src_ind * 4, 4);
|
||||||
|
} else if (size == 8) {
|
||||||
|
__builtin_memcpy(dest_arr + dest_ind * 8, src_arr + src_ind * 8, 8);
|
||||||
|
} else {
|
||||||
|
/* memcpy for var size, cannot overlap after previous alloca */
|
||||||
|
__builtin_memcpy(dest_arr + dest_ind * size, src_arr + src_ind * size, size);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/* only dest_step == 1 can we shrink the dest list. */
|
||||||
|
/* size should be ensured prior to calling this function */
|
||||||
|
if (dest_step == 1 && dest_end >= dest_start) {
|
||||||
|
__builtin_memmove(
|
||||||
|
dest_arr + dest_ind * size,
|
||||||
|
dest_arr + (dest_end + 1) * size,
|
||||||
|
(dest_arr_len - dest_end - 1) * size
|
||||||
|
);
|
||||||
|
return dest_arr_len - (dest_end - dest_ind) - 1;
|
||||||
|
}
|
||||||
|
return dest_arr_len;
|
||||||
|
}
|
||||||
|
|
||||||
|
int32_t __nac3_isinf(double x) {
|
||||||
|
return __builtin_isinf(x);
|
||||||
|
}
|
||||||
|
|
||||||
|
int32_t __nac3_isnan(double x) {
|
||||||
|
return __builtin_isnan(x);
|
||||||
|
}
|
||||||
|
|
||||||
|
double tgamma(double arg);
|
||||||
|
|
||||||
|
double __nac3_gamma(double z) {
|
||||||
|
// Handling for denormals
|
||||||
|
// | x | Python gamma(x) | C tgamma(x) |
|
||||||
|
// --- | ----------------- | --------------- | ----------- |
|
||||||
|
// (1) | nan | nan | nan |
|
||||||
|
// (2) | -inf | -inf | inf |
|
||||||
|
// (3) | inf | inf | inf |
|
||||||
|
// (4) | 0.0 | inf | inf |
|
||||||
|
// (5) | {-1.0, -2.0, ...} | inf | nan |
|
||||||
|
|
||||||
|
// (1)-(3)
|
||||||
|
if (__builtin_isinf(z) || __builtin_isnan(z)) {
|
||||||
|
return z;
|
||||||
|
}
|
||||||
|
|
||||||
|
double v = tgamma(z);
|
||||||
|
|
||||||
|
// (4)-(5)
|
||||||
|
return __builtin_isinf(v) || __builtin_isnan(v) ? __builtin_inf() : v;
|
||||||
|
}
|
||||||
|
|
||||||
|
double lgamma(double arg);
|
||||||
|
|
||||||
|
double __nac3_gammaln(double x) {
|
||||||
|
// libm's handling of value overflows differs from scipy:
|
||||||
|
// - scipy: gammaln(-inf) -> -inf
|
||||||
|
// - libm : lgamma(-inf) -> inf
|
||||||
|
|
||||||
|
if (__builtin_isinf(x)) {
|
||||||
|
return x;
|
||||||
|
}
|
||||||
|
|
||||||
|
return lgamma(x);
|
||||||
|
}
|
||||||
|
|
||||||
|
double j0(double x);
|
||||||
|
|
||||||
|
double __nac3_j0(double x) {
|
||||||
|
// libm's handling of value overflows differs from scipy:
|
||||||
|
// - scipy: j0(inf) -> nan
|
||||||
|
// - libm : j0(inf) -> 0.0
|
||||||
|
|
||||||
|
if (__builtin_isinf(x)) {
|
||||||
|
return __builtin_nan("");
|
||||||
|
}
|
||||||
|
|
||||||
|
return j0(x);
|
||||||
|
}
|
||||||
|
|
||||||
|
uint32_t __nac3_ndarray_calc_size(
|
||||||
|
const uint64_t *list_data,
|
||||||
|
uint32_t list_len,
|
||||||
|
uint32_t begin_idx,
|
||||||
|
uint32_t end_idx
|
||||||
|
) {
|
||||||
|
__builtin_assume(end_idx <= list_len);
|
||||||
|
|
||||||
|
uint32_t num_elems = 1;
|
||||||
|
for (uint32_t i = begin_idx; i < end_idx; ++i) {
|
||||||
|
uint64_t val = list_data[i];
|
||||||
|
__builtin_assume(val > 0);
|
||||||
|
num_elems *= val;
|
||||||
|
}
|
||||||
|
return num_elems;
|
||||||
|
}
|
||||||
|
|
||||||
|
uint64_t __nac3_ndarray_calc_size64(
|
||||||
|
const uint64_t *list_data,
|
||||||
|
uint64_t list_len,
|
||||||
|
uint64_t begin_idx,
|
||||||
|
uint64_t end_idx
|
||||||
|
) {
|
||||||
|
__builtin_assume(end_idx <= list_len);
|
||||||
|
|
||||||
|
uint64_t num_elems = 1;
|
||||||
|
for (uint64_t i = begin_idx; i < end_idx; ++i) {
|
||||||
|
uint64_t val = list_data[i];
|
||||||
|
__builtin_assume(val > 0);
|
||||||
|
num_elems *= val;
|
||||||
|
}
|
||||||
|
return num_elems;
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_nd_indices(
|
||||||
|
uint32_t index,
|
||||||
|
const uint32_t* dims,
|
||||||
|
uint32_t num_dims,
|
||||||
|
uint32_t* idxs
|
||||||
|
) {
|
||||||
|
uint32_t stride = 1;
|
||||||
|
for (uint32_t dim = 0; dim < num_dims; dim++) {
|
||||||
|
uint32_t i = num_dims - dim - 1;
|
||||||
|
__builtin_assume(dims[i] > 0);
|
||||||
|
idxs[i] = (index / stride) % dims[i];
|
||||||
|
stride *= dims[i];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_nd_indices64(
|
||||||
|
uint64_t index,
|
||||||
|
const uint64_t* dims,
|
||||||
|
uint64_t num_dims,
|
||||||
|
uint32_t* idxs
|
||||||
|
) {
|
||||||
|
uint64_t stride = 1;
|
||||||
|
for (uint64_t dim = 0; dim < num_dims; dim++) {
|
||||||
|
uint64_t i = num_dims - dim - 1;
|
||||||
|
__builtin_assume(dims[i] > 0);
|
||||||
|
idxs[i] = (uint32_t) ((index / stride) % dims[i]);
|
||||||
|
stride *= dims[i];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
uint32_t __nac3_ndarray_flatten_index(
|
||||||
|
const uint32_t* dims,
|
||||||
|
uint32_t num_dims,
|
||||||
|
const uint32_t* indices,
|
||||||
|
uint32_t num_indices
|
||||||
|
) {
|
||||||
|
uint32_t idx = 0;
|
||||||
|
uint32_t stride = 1;
|
||||||
|
for (uint32_t i = 0; i < num_dims; ++i) {
|
||||||
|
uint32_t ri = num_dims - i - 1;
|
||||||
|
if (ri < num_indices) {
|
||||||
|
idx += (stride * indices[ri]);
|
||||||
|
}
|
||||||
|
|
||||||
|
__builtin_assume(dims[i] > 0);
|
||||||
|
stride *= dims[ri];
|
||||||
|
}
|
||||||
|
return idx;
|
||||||
|
}
|
||||||
|
|
||||||
|
uint64_t __nac3_ndarray_flatten_index64(
|
||||||
|
const uint64_t* dims,
|
||||||
|
uint64_t num_dims,
|
||||||
|
const uint32_t* indices,
|
||||||
|
uint64_t num_indices
|
||||||
|
) {
|
||||||
|
uint64_t idx = 0;
|
||||||
|
uint64_t stride = 1;
|
||||||
|
for (uint64_t i = 0; i < num_dims; ++i) {
|
||||||
|
uint64_t ri = num_dims - i - 1;
|
||||||
|
if (ri < num_indices) {
|
||||||
|
idx += (stride * indices[ri]);
|
||||||
|
}
|
||||||
|
|
||||||
|
__builtin_assume(dims[i] > 0);
|
||||||
|
stride *= dims[ri];
|
||||||
|
}
|
||||||
|
return idx;
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_broadcast(
|
||||||
|
const uint32_t *lhs_dims,
|
||||||
|
uint32_t lhs_ndims,
|
||||||
|
const uint32_t *rhs_dims,
|
||||||
|
uint32_t rhs_ndims,
|
||||||
|
uint32_t *out_dims
|
||||||
|
) {
|
||||||
|
uint32_t max_ndims = lhs_ndims > rhs_ndims ? lhs_ndims : rhs_ndims;
|
||||||
|
|
||||||
|
for (uint32_t i = 0; i < max_ndims; ++i) {
|
||||||
|
uint32_t *lhs_dim_sz = i < lhs_ndims ? &lhs_dims[lhs_ndims - i - 1] : NULL;
|
||||||
|
uint32_t *rhs_dim_sz = i < rhs_ndims ? &rhs_dims[rhs_ndims - i - 1] : NULL;
|
||||||
|
uint32_t *out_dim = &out_dims[max_ndims - i - 1];
|
||||||
|
|
||||||
|
if (lhs_dim_sz == NULL) {
|
||||||
|
*out_dim = *rhs_dim_sz;
|
||||||
|
} else if (rhs_dim_sz == NULL) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else if (*lhs_dim_sz == 1) {
|
||||||
|
*out_dim = *rhs_dim_sz;
|
||||||
|
} else if (*rhs_dim_sz == 1) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else if (*lhs_dim_sz == *rhs_dim_sz) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else {
|
||||||
|
__builtin_unreachable();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_broadcast64(
|
||||||
|
const uint64_t *lhs_dims,
|
||||||
|
uint64_t lhs_ndims,
|
||||||
|
const uint64_t *rhs_dims,
|
||||||
|
uint64_t rhs_ndims,
|
||||||
|
uint64_t *out_dims
|
||||||
|
) {
|
||||||
|
uint64_t max_ndims = lhs_ndims > rhs_ndims ? lhs_ndims : rhs_ndims;
|
||||||
|
|
||||||
|
for (uint64_t i = 0; i < max_ndims; ++i) {
|
||||||
|
uint64_t *lhs_dim_sz = i < lhs_ndims ? &lhs_dims[lhs_ndims - i - 1] : NULL;
|
||||||
|
uint64_t *rhs_dim_sz = i < rhs_ndims ? &rhs_dims[rhs_ndims - i - 1] : NULL;
|
||||||
|
uint64_t *out_dim = &out_dims[max_ndims - i - 1];
|
||||||
|
|
||||||
|
if (lhs_dim_sz == NULL) {
|
||||||
|
*out_dim = *rhs_dim_sz;
|
||||||
|
} else if (rhs_dim_sz == NULL) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else if (*lhs_dim_sz == 1) {
|
||||||
|
*out_dim = *rhs_dim_sz;
|
||||||
|
} else if (*rhs_dim_sz == 1) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else if (*lhs_dim_sz == *rhs_dim_sz) {
|
||||||
|
*out_dim = *lhs_dim_sz;
|
||||||
|
} else {
|
||||||
|
__builtin_unreachable();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_broadcast_idx(
|
||||||
|
const uint32_t *src_dims,
|
||||||
|
uint32_t src_ndims,
|
||||||
|
const uint32_t *in_idx,
|
||||||
|
uint32_t *out_idx
|
||||||
|
) {
|
||||||
|
for (uint32_t i = 0; i < src_ndims; ++i) {
|
||||||
|
uint32_t src_i = src_ndims - i - 1;
|
||||||
|
out_idx[src_i] = src_dims[src_i] == 1 ? 0 : in_idx[src_i];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void __nac3_ndarray_calc_broadcast_idx64(
|
||||||
|
const uint64_t *src_dims,
|
||||||
|
uint64_t src_ndims,
|
||||||
|
const uint32_t *in_idx,
|
||||||
|
uint32_t *out_idx
|
||||||
|
) {
|
||||||
|
for (uint64_t i = 0; i < src_ndims; ++i) {
|
||||||
|
uint64_t src_i = src_ndims - i - 1;
|
||||||
|
out_idx[src_i] = src_dims[src_i] == 1 ? 0 : (uint32_t) in_idx[src_i];
|
||||||
|
}
|
||||||
|
}
|
@ -1,30 +1,28 @@
|
|||||||
|
use crate::typecheck::typedef::Type;
|
||||||
|
|
||||||
|
use super::{
|
||||||
|
classes::{
|
||||||
|
ArrayLikeIndexer, ArrayLikeValue, ArraySliceValue, ListValue, NDArrayValue,
|
||||||
|
TypedArrayLikeAdapter, UntypedArrayLikeAccessor,
|
||||||
|
},
|
||||||
|
llvm_intrinsics, CodeGenContext, CodeGenerator,
|
||||||
|
};
|
||||||
|
use crate::codegen::classes::TypedArrayLikeAccessor;
|
||||||
|
use crate::codegen::stmt::gen_for_callback_incrementing;
|
||||||
use inkwell::{
|
use inkwell::{
|
||||||
attributes::{Attribute, AttributeLoc},
|
attributes::{Attribute, AttributeLoc},
|
||||||
context::Context,
|
context::Context,
|
||||||
memory_buffer::MemoryBuffer,
|
memory_buffer::MemoryBuffer,
|
||||||
module::Module,
|
module::Module,
|
||||||
types::{BasicTypeEnum, IntType},
|
types::{BasicTypeEnum, IntType},
|
||||||
values::{BasicValue, BasicValueEnum, CallSiteValue, FloatValue, IntValue},
|
values::{BasicValueEnum, CallSiteValue, FloatValue, IntValue},
|
||||||
AddressSpace, IntPredicate,
|
AddressSpace, IntPredicate,
|
||||||
};
|
};
|
||||||
use itertools::Either;
|
use itertools::Either;
|
||||||
|
|
||||||
use nac3parser::ast::Expr;
|
use nac3parser::ast::Expr;
|
||||||
|
|
||||||
use super::{
|
|
||||||
classes::{
|
|
||||||
ArrayLikeIndexer, ArrayLikeValue, ArraySliceValue, ListValue, NDArrayValue,
|
|
||||||
TypedArrayLikeAccessor, TypedArrayLikeAdapter, UntypedArrayLikeAccessor,
|
|
||||||
},
|
|
||||||
llvm_intrinsics,
|
|
||||||
macros::codegen_unreachable,
|
|
||||||
stmt::gen_for_callback_incrementing,
|
|
||||||
CodeGenContext, CodeGenerator,
|
|
||||||
};
|
|
||||||
use crate::{symbol_resolver::SymbolResolver, typecheck::typedef::Type};
|
|
||||||
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn load_irrt<'ctx>(ctx: &'ctx Context, symbol_resolver: &dyn SymbolResolver) -> Module<'ctx> {
|
pub fn load_irrt(ctx: &Context) -> Module {
|
||||||
let bitcode_buf = MemoryBuffer::create_from_memory_range(
|
let bitcode_buf = MemoryBuffer::create_from_memory_range(
|
||||||
include_bytes!(concat!(env!("OUT_DIR"), "/irrt.bc")),
|
include_bytes!(concat!(env!("OUT_DIR"), "/irrt.bc")),
|
||||||
"irrt_bitcode_buffer",
|
"irrt_bitcode_buffer",
|
||||||
@ -40,25 +38,6 @@ pub fn load_irrt<'ctx>(ctx: &'ctx Context, symbol_resolver: &dyn SymbolResolver)
|
|||||||
let function = irrt_mod.get_function(symbol).unwrap();
|
let function = irrt_mod.get_function(symbol).unwrap();
|
||||||
function.add_attribute(AttributeLoc::Function, ctx.create_enum_attribute(inline_attr, 0));
|
function.add_attribute(AttributeLoc::Function, ctx.create_enum_attribute(inline_attr, 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize all global `EXN_*` exception IDs in IRRT with the [`SymbolResolver`].
|
|
||||||
let exn_id_type = ctx.i32_type();
|
|
||||||
let errors = &[
|
|
||||||
("EXN_INDEX_ERROR", "0:IndexError"),
|
|
||||||
("EXN_VALUE_ERROR", "0:ValueError"),
|
|
||||||
("EXN_ASSERTION_ERROR", "0:AssertionError"),
|
|
||||||
("EXN_TYPE_ERROR", "0:TypeError"),
|
|
||||||
];
|
|
||||||
for (irrt_name, symbol_name) in errors {
|
|
||||||
let exn_id = symbol_resolver.get_string_id(symbol_name);
|
|
||||||
let exn_id = exn_id_type.const_int(exn_id as u64, false).as_basic_value_enum();
|
|
||||||
|
|
||||||
let global = irrt_mod.get_global(irrt_name).unwrap_or_else(|| {
|
|
||||||
panic!("Exception symbol name '{irrt_name}' should exist in the IRRT LLVM module")
|
|
||||||
});
|
|
||||||
global.set_initializer(&exn_id);
|
|
||||||
}
|
|
||||||
|
|
||||||
irrt_mod
|
irrt_mod
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -76,7 +55,7 @@ pub fn integer_power<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
(64, 64, true) => "__nac3_int_exp_int64_t",
|
(64, 64, true) => "__nac3_int_exp_int64_t",
|
||||||
(32, 32, false) => "__nac3_int_exp_uint32_t",
|
(32, 32, false) => "__nac3_int_exp_uint32_t",
|
||||||
(64, 64, false) => "__nac3_int_exp_uint64_t",
|
(64, 64, false) => "__nac3_int_exp_uint64_t",
|
||||||
_ => codegen_unreachable!(ctx),
|
_ => unreachable!(),
|
||||||
};
|
};
|
||||||
let base_type = base.get_type();
|
let base_type = base.get_type();
|
||||||
let pow_fun = ctx.module.get_function(symbol).unwrap_or_else(|| {
|
let pow_fun = ctx.module.get_function(symbol).unwrap_or_else(|| {
|
||||||
@ -462,7 +441,7 @@ pub fn list_slice_assignment<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
BasicTypeEnum::IntType(t) => t.size_of(),
|
BasicTypeEnum::IntType(t) => t.size_of(),
|
||||||
BasicTypeEnum::PointerType(t) => t.size_of(),
|
BasicTypeEnum::PointerType(t) => t.size_of(),
|
||||||
BasicTypeEnum::StructType(t) => t.size_of().unwrap(),
|
BasicTypeEnum::StructType(t) => t.size_of().unwrap(),
|
||||||
_ => codegen_unreachable!(ctx),
|
_ => unreachable!(),
|
||||||
};
|
};
|
||||||
ctx.builder.build_int_truncate_or_bit_cast(s, int32, "size").unwrap()
|
ctx.builder.build_int_truncate_or_bit_cast(s, int32, "size").unwrap()
|
||||||
}
|
}
|
||||||
@ -589,8 +568,7 @@ pub fn call_j0<'ctx>(ctx: &CodeGenContext<'ctx, '_>, v: FloatValue<'ctx>) -> Flo
|
|||||||
///
|
///
|
||||||
/// * `dims` - An [`ArrayLikeIndexer`] containing the size of each dimension.
|
/// * `dims` - An [`ArrayLikeIndexer`] containing the size of each dimension.
|
||||||
/// * `range` - The dimension index to begin and end (exclusively) calculating the dimensions for,
|
/// * `range` - The dimension index to begin and end (exclusively) calculating the dimensions for,
|
||||||
/// or [`None`] if starting from the first dimension and ending at the last dimension
|
/// or [`None`] if starting from the first dimension and ending at the last dimension respectively.
|
||||||
/// respectively.
|
|
||||||
pub fn call_ndarray_calc_size<'ctx, G, Dims>(
|
pub fn call_ndarray_calc_size<'ctx, G, Dims>(
|
||||||
generator: &G,
|
generator: &G,
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
@ -601,16 +579,18 @@ where
|
|||||||
G: CodeGenerator + ?Sized,
|
G: CodeGenerator + ?Sized,
|
||||||
Dims: ArrayLikeIndexer<'ctx>,
|
Dims: ArrayLikeIndexer<'ctx>,
|
||||||
{
|
{
|
||||||
|
let llvm_i64 = ctx.ctx.i64_type();
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
let llvm_pusize = llvm_usize.ptr_type(AddressSpace::default());
|
|
||||||
|
let llvm_pi64 = llvm_i64.ptr_type(AddressSpace::default());
|
||||||
|
|
||||||
let ndarray_calc_size_fn_name = match llvm_usize.get_bit_width() {
|
let ndarray_calc_size_fn_name = match llvm_usize.get_bit_width() {
|
||||||
32 => "__nac3_ndarray_calc_size",
|
32 => "__nac3_ndarray_calc_size",
|
||||||
64 => "__nac3_ndarray_calc_size64",
|
64 => "__nac3_ndarray_calc_size64",
|
||||||
bw => codegen_unreachable!(ctx, "Unsupported size type bit width: {}", bw),
|
bw => unreachable!("Unsupported size type bit width: {}", bw),
|
||||||
};
|
};
|
||||||
let ndarray_calc_size_fn_t = llvm_usize.fn_type(
|
let ndarray_calc_size_fn_t = llvm_usize.fn_type(
|
||||||
&[llvm_pusize.into(), llvm_usize.into(), llvm_usize.into(), llvm_usize.into()],
|
&[llvm_pi64.into(), llvm_usize.into(), llvm_usize.into(), llvm_usize.into()],
|
||||||
false,
|
false,
|
||||||
);
|
);
|
||||||
let ndarray_calc_size_fn =
|
let ndarray_calc_size_fn =
|
||||||
@ -658,7 +638,7 @@ pub fn call_ndarray_calc_nd_indices<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
let ndarray_calc_nd_indices_fn_name = match llvm_usize.get_bit_width() {
|
let ndarray_calc_nd_indices_fn_name = match llvm_usize.get_bit_width() {
|
||||||
32 => "__nac3_ndarray_calc_nd_indices",
|
32 => "__nac3_ndarray_calc_nd_indices",
|
||||||
64 => "__nac3_ndarray_calc_nd_indices64",
|
64 => "__nac3_ndarray_calc_nd_indices64",
|
||||||
bw => codegen_unreachable!(ctx, "Unsupported size type bit width: {}", bw),
|
bw => unreachable!("Unsupported size type bit width: {}", bw),
|
||||||
};
|
};
|
||||||
let ndarray_calc_nd_indices_fn =
|
let ndarray_calc_nd_indices_fn =
|
||||||
ctx.module.get_function(ndarray_calc_nd_indices_fn_name).unwrap_or_else(|| {
|
ctx.module.get_function(ndarray_calc_nd_indices_fn_name).unwrap_or_else(|| {
|
||||||
@ -727,7 +707,7 @@ where
|
|||||||
let ndarray_flatten_index_fn_name = match llvm_usize.get_bit_width() {
|
let ndarray_flatten_index_fn_name = match llvm_usize.get_bit_width() {
|
||||||
32 => "__nac3_ndarray_flatten_index",
|
32 => "__nac3_ndarray_flatten_index",
|
||||||
64 => "__nac3_ndarray_flatten_index64",
|
64 => "__nac3_ndarray_flatten_index64",
|
||||||
bw => codegen_unreachable!(ctx, "Unsupported size type bit width: {}", bw),
|
bw => unreachable!("Unsupported size type bit width: {}", bw),
|
||||||
};
|
};
|
||||||
let ndarray_flatten_index_fn =
|
let ndarray_flatten_index_fn =
|
||||||
ctx.module.get_function(ndarray_flatten_index_fn_name).unwrap_or_else(|| {
|
ctx.module.get_function(ndarray_flatten_index_fn_name).unwrap_or_else(|| {
|
||||||
@ -795,7 +775,7 @@ pub fn call_ndarray_calc_broadcast<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
let ndarray_calc_broadcast_fn_name = match llvm_usize.get_bit_width() {
|
let ndarray_calc_broadcast_fn_name = match llvm_usize.get_bit_width() {
|
||||||
32 => "__nac3_ndarray_calc_broadcast",
|
32 => "__nac3_ndarray_calc_broadcast",
|
||||||
64 => "__nac3_ndarray_calc_broadcast64",
|
64 => "__nac3_ndarray_calc_broadcast64",
|
||||||
bw => codegen_unreachable!(ctx, "Unsupported size type bit width: {}", bw),
|
bw => unreachable!("Unsupported size type bit width: {}", bw),
|
||||||
};
|
};
|
||||||
let ndarray_calc_broadcast_fn =
|
let ndarray_calc_broadcast_fn =
|
||||||
ctx.module.get_function(ndarray_calc_broadcast_fn_name).unwrap_or_else(|| {
|
ctx.module.get_function(ndarray_calc_broadcast_fn_name).unwrap_or_else(|| {
|
||||||
@ -820,7 +800,6 @@ pub fn call_ndarray_calc_broadcast<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
gen_for_callback_incrementing(
|
gen_for_callback_incrementing(
|
||||||
generator,
|
generator,
|
||||||
ctx,
|
ctx,
|
||||||
None,
|
|
||||||
llvm_usize.const_zero(),
|
llvm_usize.const_zero(),
|
||||||
(min_ndims, false),
|
(min_ndims, false),
|
||||||
|generator, ctx, _, idx| {
|
|generator, ctx, _, idx| {
|
||||||
@ -915,7 +894,7 @@ pub fn call_ndarray_calc_broadcast_index<
|
|||||||
let ndarray_calc_broadcast_fn_name = match llvm_usize.get_bit_width() {
|
let ndarray_calc_broadcast_fn_name = match llvm_usize.get_bit_width() {
|
||||||
32 => "__nac3_ndarray_calc_broadcast_idx",
|
32 => "__nac3_ndarray_calc_broadcast_idx",
|
||||||
64 => "__nac3_ndarray_calc_broadcast_idx64",
|
64 => "__nac3_ndarray_calc_broadcast_idx64",
|
||||||
bw => codegen_unreachable!(ctx, "Unsupported size type bit width: {}", bw),
|
bw => unreachable!("Unsupported size type bit width: {}", bw),
|
||||||
};
|
};
|
||||||
let ndarray_calc_broadcast_fn =
|
let ndarray_calc_broadcast_fn =
|
||||||
ctx.module.get_function(ndarray_calc_broadcast_fn_name).unwrap_or_else(|| {
|
ctx.module.get_function(ndarray_calc_broadcast_fn_name).unwrap_or_else(|| {
|
||||||
|
@ -1,13 +1,11 @@
|
|||||||
use inkwell::{
|
|
||||||
context::Context,
|
|
||||||
intrinsics::Intrinsic,
|
|
||||||
types::{AnyTypeEnum::IntType, FloatType},
|
|
||||||
values::{BasicValueEnum, CallSiteValue, FloatValue, IntValue, PointerValue},
|
|
||||||
AddressSpace,
|
|
||||||
};
|
|
||||||
use itertools::Either;
|
|
||||||
|
|
||||||
use crate::codegen::CodeGenContext;
|
use crate::codegen::CodeGenContext;
|
||||||
|
use inkwell::context::Context;
|
||||||
|
use inkwell::intrinsics::Intrinsic;
|
||||||
|
use inkwell::types::AnyTypeEnum::IntType;
|
||||||
|
use inkwell::types::FloatType;
|
||||||
|
use inkwell::values::{BasicValueEnum, CallSiteValue, FloatValue, IntValue, PointerValue};
|
||||||
|
use inkwell::AddressSpace;
|
||||||
|
use itertools::Either;
|
||||||
|
|
||||||
/// Returns the string representation for the floating-point type `ft` when used in intrinsic
|
/// Returns the string representation for the floating-point type `ft` when used in intrinsic
|
||||||
/// functions.
|
/// functions.
|
||||||
@ -37,40 +35,6 @@ fn get_float_intrinsic_repr(ctx: &Context, ft: FloatType) -> &'static str {
|
|||||||
unreachable!()
|
unreachable!()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Invokes the [`llvm.va_start`](https://llvm.org/docs/LangRef.html#llvm-va-start-intrinsic)
|
|
||||||
/// intrinsic.
|
|
||||||
pub fn call_va_start<'ctx>(ctx: &CodeGenContext<'ctx, '_>, arglist: PointerValue<'ctx>) {
|
|
||||||
const FN_NAME: &str = "llvm.va_start";
|
|
||||||
|
|
||||||
let intrinsic_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
|
||||||
let llvm_void = ctx.ctx.void_type();
|
|
||||||
let llvm_i8 = ctx.ctx.i8_type();
|
|
||||||
let llvm_p0i8 = llvm_i8.ptr_type(AddressSpace::default());
|
|
||||||
let fn_type = llvm_void.fn_type(&[llvm_p0i8.into()], false);
|
|
||||||
|
|
||||||
ctx.module.add_function(FN_NAME, fn_type, None)
|
|
||||||
});
|
|
||||||
|
|
||||||
ctx.builder.build_call(intrinsic_fn, &[arglist.into()], "").unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the [`llvm.va_start`](https://llvm.org/docs/LangRef.html#llvm-va-start-intrinsic)
|
|
||||||
/// intrinsic.
|
|
||||||
pub fn call_va_end<'ctx>(ctx: &CodeGenContext<'ctx, '_>, arglist: PointerValue<'ctx>) {
|
|
||||||
const FN_NAME: &str = "llvm.va_end";
|
|
||||||
|
|
||||||
let intrinsic_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
|
||||||
let llvm_void = ctx.ctx.void_type();
|
|
||||||
let llvm_i8 = ctx.ctx.i8_type();
|
|
||||||
let llvm_p0i8 = llvm_i8.ptr_type(AddressSpace::default());
|
|
||||||
let fn_type = llvm_void.fn_type(&[llvm_p0i8.into()], false);
|
|
||||||
|
|
||||||
ctx.module.add_function(FN_NAME, fn_type, None)
|
|
||||||
});
|
|
||||||
|
|
||||||
ctx.builder.build_call(intrinsic_fn, &[arglist.into()], "").unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the [`llvm.stacksave`](https://llvm.org/docs/LangRef.html#llvm-stacksave-intrinsic)
|
/// Invokes the [`llvm.stacksave`](https://llvm.org/docs/LangRef.html#llvm-stacksave-intrinsic)
|
||||||
/// intrinsic.
|
/// intrinsic.
|
||||||
pub fn call_stacksave<'ctx>(
|
pub fn call_stacksave<'ctx>(
|
||||||
@ -98,30 +62,145 @@ pub fn call_stacksave<'ctx>(
|
|||||||
pub fn call_stackrestore<'ctx>(ctx: &CodeGenContext<'ctx, '_>, ptr: PointerValue<'ctx>) {
|
pub fn call_stackrestore<'ctx>(ctx: &CodeGenContext<'ctx, '_>, ptr: PointerValue<'ctx>) {
|
||||||
const FN_NAME: &str = "llvm.stackrestore";
|
const FN_NAME: &str = "llvm.stackrestore";
|
||||||
|
|
||||||
/*
|
let llvm_i8 = ctx.ctx.i8_type();
|
||||||
SEE https://github.com/TheDan64/inkwell/issues/496
|
let llvm_p0i8 = llvm_i8.ptr_type(AddressSpace::default());
|
||||||
|
|
||||||
We want `llvm.stackrestore`, but the following would generate `llvm.stackrestore.p0i8`.
|
|
||||||
```ignore
|
|
||||||
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_p0i8.into()]))
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_p0i8.into()]))
|
||||||
.unwrap();
|
.unwrap();
|
||||||
```
|
|
||||||
|
|
||||||
Temp workaround by manually declaring the intrinsic with the correct function name instead.
|
|
||||||
*/
|
|
||||||
let intrinsic_fn = ctx.module.get_function(FN_NAME).unwrap_or_else(|| {
|
|
||||||
let llvm_void = ctx.ctx.void_type();
|
|
||||||
let llvm_i8 = ctx.ctx.i8_type();
|
|
||||||
let llvm_p0i8 = llvm_i8.ptr_type(AddressSpace::default());
|
|
||||||
let fn_type = llvm_void.fn_type(&[llvm_p0i8.into()], false);
|
|
||||||
|
|
||||||
ctx.module.add_function(FN_NAME, fn_type, None)
|
|
||||||
});
|
|
||||||
|
|
||||||
ctx.builder.build_call(intrinsic_fn, &[ptr.into()], "").unwrap();
|
ctx.builder.build_call(intrinsic_fn, &[ptr.into()], "").unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.abs`](https://llvm.org/docs/LangRef.html#llvm-abs-intrinsic) intrinsic.
|
||||||
|
///
|
||||||
|
/// * `src` - The value for which the absolute value is to be returned.
|
||||||
|
/// * `is_int_min_poison` - Whether `poison` is to be returned if `src` is `INT_MIN`.
|
||||||
|
pub fn call_int_abs<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
src: IntValue<'ctx>,
|
||||||
|
is_int_min_poison: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.abs";
|
||||||
|
|
||||||
|
debug_assert_eq!(is_int_min_poison.get_type().get_bit_width(), 1);
|
||||||
|
debug_assert!(is_int_min_poison.is_const());
|
||||||
|
|
||||||
|
let llvm_src_t = src.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_src_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[src.into(), is_int_min_poison.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.smax`](https://llvm.org/docs/LangRef.html#llvm-smax-intrinsic) intrinsic.
|
||||||
|
pub fn call_int_smax<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
a: IntValue<'ctx>,
|
||||||
|
b: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.smax";
|
||||||
|
|
||||||
|
debug_assert_eq!(a.get_type().get_bit_width(), b.get_type().get_bit_width());
|
||||||
|
|
||||||
|
let llvm_int_t = a.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_int_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[a.into(), b.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.smin`](https://llvm.org/docs/LangRef.html#llvm-smin-intrinsic) intrinsic.
|
||||||
|
pub fn call_int_smin<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
a: IntValue<'ctx>,
|
||||||
|
b: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.smin";
|
||||||
|
|
||||||
|
debug_assert_eq!(a.get_type().get_bit_width(), b.get_type().get_bit_width());
|
||||||
|
|
||||||
|
let llvm_int_t = a.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_int_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[a.into(), b.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.umax`](https://llvm.org/docs/LangRef.html#llvm-umax-intrinsic) intrinsic.
|
||||||
|
pub fn call_int_umax<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
a: IntValue<'ctx>,
|
||||||
|
b: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.umax";
|
||||||
|
|
||||||
|
debug_assert_eq!(a.get_type().get_bit_width(), b.get_type().get_bit_width());
|
||||||
|
|
||||||
|
let llvm_int_t = a.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_int_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[a.into(), b.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.umin`](https://llvm.org/docs/LangRef.html#llvm-umin-intrinsic) intrinsic.
|
||||||
|
pub fn call_int_umin<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
a: IntValue<'ctx>,
|
||||||
|
b: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.umin";
|
||||||
|
|
||||||
|
debug_assert_eq!(a.get_type().get_bit_width(), b.get_type().get_bit_width());
|
||||||
|
|
||||||
|
let llvm_int_t = a.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_int_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[a.into(), b.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
/// Invokes the [`llvm.memcpy`](https://llvm.org/docs/LangRef.html#llvm-memcpy-intrinsic) intrinsic.
|
/// Invokes the [`llvm.memcpy`](https://llvm.org/docs/LangRef.html#llvm-memcpy-intrinsic) intrinsic.
|
||||||
///
|
///
|
||||||
/// * `dest` - The pointer to the destination. Must be a pointer to an integer type.
|
/// * `dest` - The pointer to the destination. Must be a pointer to an integer type.
|
||||||
@ -185,7 +264,7 @@ pub fn call_memcpy_generic<'ctx>(
|
|||||||
dest
|
dest
|
||||||
} else {
|
} else {
|
||||||
ctx.builder
|
ctx.builder
|
||||||
.build_bit_cast(dest, llvm_p0i8, "")
|
.build_bitcast(dest, llvm_p0i8, "")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap()
|
.unwrap()
|
||||||
};
|
};
|
||||||
@ -193,7 +272,7 @@ pub fn call_memcpy_generic<'ctx>(
|
|||||||
src
|
src
|
||||||
} else {
|
} else {
|
||||||
ctx.builder
|
ctx.builder
|
||||||
.build_bit_cast(src, llvm_p0i8, "")
|
.build_bitcast(src, llvm_p0i8, "")
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
.map(BasicValueEnum::into_pointer_value)
|
||||||
.unwrap()
|
.unwrap()
|
||||||
};
|
};
|
||||||
@ -201,123 +280,28 @@ pub fn call_memcpy_generic<'ctx>(
|
|||||||
call_memcpy(ctx, dest, src, len, is_volatile);
|
call_memcpy(ctx, dest, src, len, is_volatile);
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Macro to find and generate build call for llvm intrinsic (body of llvm intrinsic function)
|
/// Invokes the [`llvm.sqrt`](https://llvm.org/docs/LangRef.html#llvm-sqrt-intrinsic) intrinsic.
|
||||||
///
|
pub fn call_float_sqrt<'ctx>(
|
||||||
/// Arguments:
|
|
||||||
/// * `$ctx:ident`: Reference to the current Code Generation Context
|
|
||||||
/// * `$name:ident`: Optional name to be assigned to the llvm build call (Option<&str>)
|
|
||||||
/// * `$llvm_name:literal`: Name of underlying llvm intrinsic function
|
|
||||||
/// * `$map_fn:ident`: Mapping function to be applied on `BasicValue` (`BasicValue` -> Function Return Type).
|
|
||||||
/// Use `BasicValueEnum::into_int_value` for Integer return type and
|
|
||||||
/// `BasicValueEnum::into_float_value` for Float return type
|
|
||||||
/// * `$llvm_ty:ident`: Type of first operand
|
|
||||||
/// * `,($val:ident)*`: Comma separated list of operands
|
|
||||||
macro_rules! generate_llvm_intrinsic_fn_body {
|
|
||||||
($ctx:ident, $name:ident, $llvm_name:literal, $map_fn:expr, $llvm_ty:ident $(,$val:ident)*) => {{
|
|
||||||
const FN_NAME: &str = concat!("llvm.", $llvm_name);
|
|
||||||
let intrinsic_fn = Intrinsic::find(FN_NAME).and_then(|intrinsic| intrinsic.get_declaration(&$ctx.module, &[$llvm_ty.into()])).unwrap();
|
|
||||||
$ctx.builder.build_call(intrinsic_fn, &[$($val.into()),*], $name.unwrap_or_default()).map(CallSiteValue::try_as_basic_value).map(|v| v.map_left($map_fn)).map(Either::unwrap_left).unwrap()
|
|
||||||
}};
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Macro to generate the llvm intrinsic function using [`generate_llvm_intrinsic_fn_body`].
|
|
||||||
///
|
|
||||||
/// Arguments:
|
|
||||||
/// * `float/int`: Indicates the return and argument type of the function
|
|
||||||
/// * `$fn_name:ident`: The identifier of the rust function to be generated
|
|
||||||
/// * `$llvm_name:literal`: Name of underlying llvm intrinsic function.
|
|
||||||
/// Omit "llvm." prefix from the function name i.e. use "ceil" instead of "llvm.ceil"
|
|
||||||
/// * `$val:ident`: The operand for unary operations
|
|
||||||
/// * `$val1:ident`, `$val2:ident`: The operands for binary operations
|
|
||||||
macro_rules! generate_llvm_intrinsic_fn {
|
|
||||||
("float", $fn_name:ident, $llvm_name:literal, $val:ident) => {
|
|
||||||
#[doc = concat!("Invokes the [`", stringify!($llvm_name), "`](https://llvm.org/docs/LangRef.html#llvm-", stringify!($llvm_name), "-intrinsic) intrinsic." )]
|
|
||||||
pub fn $fn_name<'ctx> (
|
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
$val: FloatValue<'ctx>,
|
val: FloatValue<'ctx>,
|
||||||
name: Option<&str>,
|
name: Option<&str>,
|
||||||
) -> FloatValue<'ctx> {
|
) -> FloatValue<'ctx> {
|
||||||
let llvm_ty = $val.get_type();
|
const FN_NAME: &str = "llvm.sqrt";
|
||||||
generate_llvm_intrinsic_fn_body!(ctx, name, $llvm_name, BasicValueEnum::into_float_value, llvm_ty, $val)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
("float", $fn_name:ident, $llvm_name:literal, $val1:ident, $val2:ident) => {
|
|
||||||
#[doc = concat!("Invokes the [`", stringify!($llvm_name), "`](https://llvm.org/docs/LangRef.html#llvm-", stringify!($llvm_name), "-intrinsic) intrinsic." )]
|
|
||||||
pub fn $fn_name<'ctx> (
|
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
|
||||||
$val1: FloatValue<'ctx>,
|
|
||||||
$val2: FloatValue<'ctx>,
|
|
||||||
name: Option<&str>,
|
|
||||||
) -> FloatValue<'ctx> {
|
|
||||||
debug_assert_eq!($val1.get_type(), $val2.get_type());
|
|
||||||
let llvm_ty = $val1.get_type();
|
|
||||||
generate_llvm_intrinsic_fn_body!(ctx, name, $llvm_name, BasicValueEnum::into_float_value, llvm_ty, $val1, $val2)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
("int", $fn_name:ident, $llvm_name:literal, $val1:ident, $val2:ident) => {
|
|
||||||
#[doc = concat!("Invokes the [`", stringify!($llvm_name), "`](https://llvm.org/docs/LangRef.html#llvm-", stringify!($llvm_name), "-intrinsic) intrinsic." )]
|
|
||||||
pub fn $fn_name<'ctx> (
|
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
|
||||||
$val1: IntValue<'ctx>,
|
|
||||||
$val2: IntValue<'ctx>,
|
|
||||||
name: Option<&str>,
|
|
||||||
) -> IntValue<'ctx> {
|
|
||||||
debug_assert_eq!($val1.get_type().get_bit_width(), $val2.get_type().get_bit_width());
|
|
||||||
let llvm_ty = $val1.get_type();
|
|
||||||
generate_llvm_intrinsic_fn_body!(ctx, name, $llvm_name, BasicValueEnum::into_int_value, llvm_ty, $val1, $val2)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Invokes the [`llvm.abs`](https://llvm.org/docs/LangRef.html#llvm-abs-intrinsic) intrinsic.
|
let llvm_float_t = val.get_type();
|
||||||
///
|
|
||||||
/// * `src` - The value for which the absolute value is to be returned.
|
|
||||||
/// * `is_int_min_poison` - Whether `poison` is to be returned if `src` is `INT_MIN`.
|
|
||||||
pub fn call_int_abs<'ctx>(
|
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
|
||||||
src: IntValue<'ctx>,
|
|
||||||
is_int_min_poison: IntValue<'ctx>,
|
|
||||||
name: Option<&str>,
|
|
||||||
) -> IntValue<'ctx> {
|
|
||||||
debug_assert_eq!(is_int_min_poison.get_type().get_bit_width(), 1);
|
|
||||||
debug_assert!(is_int_min_poison.is_const());
|
|
||||||
|
|
||||||
let src_type = src.get_type();
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
generate_llvm_intrinsic_fn_body!(
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
ctx,
|
.unwrap();
|
||||||
name,
|
|
||||||
"abs",
|
ctx.builder
|
||||||
BasicValueEnum::into_int_value,
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
src_type,
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
src,
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
is_int_min_poison
|
.map(Either::unwrap_left)
|
||||||
)
|
.unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
generate_llvm_intrinsic_fn!("int", call_int_smax, "smax", a, b);
|
|
||||||
generate_llvm_intrinsic_fn!("int", call_int_smin, "smin", a, b);
|
|
||||||
generate_llvm_intrinsic_fn!("int", call_int_umax, "umax", a, b);
|
|
||||||
generate_llvm_intrinsic_fn!("int", call_int_umin, "umin", a, b);
|
|
||||||
generate_llvm_intrinsic_fn!("int", call_expect, "expect", val, expected_val);
|
|
||||||
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_sqrt, "sqrt", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_sin, "sin", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_cos, "cos", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_pow, "pow", val, power);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_exp, "exp", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_exp2, "exp2", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_log, "log", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_log10, "log10", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_log2, "log2", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_fabs, "fabs", src);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_minnum, "minnum", val, power);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_maxnum, "maxnum", val, power);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_copysign, "copysign", mag, sgn);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_floor, "floor", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_ceil, "ceil", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_round, "round", val);
|
|
||||||
generate_llvm_intrinsic_fn!("float", call_float_rint, "rint", val);
|
|
||||||
|
|
||||||
/// Invokes the [`llvm.powi`](https://llvm.org/docs/LangRef.html#llvm-powi-intrinsic) intrinsic.
|
/// Invokes the [`llvm.powi`](https://llvm.org/docs/LangRef.html#llvm-powi-intrinsic) intrinsic.
|
||||||
pub fn call_float_powi<'ctx>(
|
pub fn call_float_powi<'ctx>(
|
||||||
ctx: &CodeGenContext<'ctx, '_>,
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
@ -343,3 +327,393 @@ pub fn call_float_powi<'ctx>(
|
|||||||
.map(Either::unwrap_left)
|
.map(Either::unwrap_left)
|
||||||
.unwrap()
|
.unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.sin`](https://llvm.org/docs/LangRef.html#llvm-sin-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_sin<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.sin";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.cos`](https://llvm.org/docs/LangRef.html#llvm-cos-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_cos<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.cos";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.pow`](https://llvm.org/docs/LangRef.html#llvm-pow-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_pow<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
power: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.pow";
|
||||||
|
|
||||||
|
debug_assert_eq!(val.get_type(), power.get_type());
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into(), power.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.exp`](https://llvm.org/docs/LangRef.html#llvm-exp-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_exp<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.exp";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.exp2`](https://llvm.org/docs/LangRef.html#llvm-exp2-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_exp2<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.exp2";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.log`](https://llvm.org/docs/LangRef.html#llvm-log-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_log<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.log";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.log10`](https://llvm.org/docs/LangRef.html#llvm-log10-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_log10<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.log10";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.log2`](https://llvm.org/docs/LangRef.html#llvm-log2-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_log2<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.log2";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.fabs`](https://llvm.org/docs/LangRef.html#llvm-fabs-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_fabs<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
src: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.fabs";
|
||||||
|
|
||||||
|
let llvm_src_t = src.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_src_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[src.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.minnum`](https://llvm.org/docs/LangRef.html#llvm-minnum-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_minnum<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val1: FloatValue<'ctx>,
|
||||||
|
val2: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.minnum";
|
||||||
|
|
||||||
|
debug_assert_eq!(val1.get_type(), val2.get_type());
|
||||||
|
|
||||||
|
let llvm_float_t = val1.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val1.into(), val2.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.maxnum`](https://llvm.org/docs/LangRef.html#llvm-maxnum-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_maxnum<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val1: FloatValue<'ctx>,
|
||||||
|
val2: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.maxnum";
|
||||||
|
|
||||||
|
debug_assert_eq!(val1.get_type(), val2.get_type());
|
||||||
|
|
||||||
|
let llvm_float_t = val1.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val1.into(), val2.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.copysign`](https://llvm.org/docs/LangRef.html#llvm-copysign-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_copysign<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
mag: FloatValue<'ctx>,
|
||||||
|
sgn: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.copysign";
|
||||||
|
|
||||||
|
debug_assert_eq!(mag.get_type(), sgn.get_type());
|
||||||
|
|
||||||
|
let llvm_float_t = mag.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[mag.into(), sgn.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.floor`](https://llvm.org/docs/LangRef.html#llvm-floor-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_floor<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.floor";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.ceil`](https://llvm.org/docs/LangRef.html#llvm-ceil-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_ceil<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.ceil";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.round`](https://llvm.org/docs/LangRef.html#llvm-round-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_round<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.round";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the
|
||||||
|
/// [`llvm.roundeven`](https://llvm.org/docs/LangRef.html#llvm-roundeven-intrinsic) intrinsic.
|
||||||
|
pub fn call_float_roundeven<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: FloatValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> FloatValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.roundeven";
|
||||||
|
|
||||||
|
let llvm_float_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_float_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_float_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Invokes the [`llvm.expect`](https://llvm.org/docs/LangRef.html#llvm-expect-intrinsic) intrinsic.
|
||||||
|
pub fn call_expect<'ctx>(
|
||||||
|
ctx: &CodeGenContext<'ctx, '_>,
|
||||||
|
val: IntValue<'ctx>,
|
||||||
|
expected_val: IntValue<'ctx>,
|
||||||
|
name: Option<&str>,
|
||||||
|
) -> IntValue<'ctx> {
|
||||||
|
const FN_NAME: &str = "llvm.expect";
|
||||||
|
|
||||||
|
debug_assert_eq!(val.get_type().get_bit_width(), expected_val.get_type().get_bit_width());
|
||||||
|
|
||||||
|
let llvm_int_t = val.get_type();
|
||||||
|
|
||||||
|
let intrinsic_fn = Intrinsic::find(FN_NAME)
|
||||||
|
.and_then(|intrinsic| intrinsic.get_declaration(&ctx.module, &[llvm_int_t.into()]))
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
ctx.builder
|
||||||
|
.build_call(intrinsic_fn, &[val.into(), expected_val.into()], name.unwrap_or_default())
|
||||||
|
.map(CallSiteValue::try_as_basic_value)
|
||||||
|
.map(|v| v.map_left(BasicValueEnum::into_int_value))
|
||||||
|
.map(Either::unwrap_left)
|
||||||
|
.unwrap()
|
||||||
|
}
|
||||||
|
@ -1,12 +1,12 @@
|
|||||||
use std::{
|
use crate::{
|
||||||
collections::{HashMap, HashSet},
|
codegen::classes::{ListType, NDArrayType, ProxyType, RangeType},
|
||||||
sync::{
|
symbol_resolver::{StaticValue, SymbolResolver},
|
||||||
atomic::{AtomicBool, Ordering},
|
toplevel::{helper::PrimDef, numpy::unpack_ndarray_var_tys, TopLevelContext, TopLevelDef},
|
||||||
Arc,
|
typecheck::{
|
||||||
|
type_inferencer::{CodeLocation, PrimitiveStore},
|
||||||
|
typedef::{CallId, FuncArg, Type, TypeEnum, Unifier},
|
||||||
},
|
},
|
||||||
thread,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use crossbeam::channel::{unbounded, Receiver, Sender};
|
use crossbeam::channel::{unbounded, Receiver, Sender};
|
||||||
use inkwell::{
|
use inkwell::{
|
||||||
attributes::{Attribute, AttributeLoc},
|
attributes::{Attribute, AttributeLoc},
|
||||||
@ -24,19 +24,14 @@ use inkwell::{
|
|||||||
AddressSpace, IntPredicate, OptimizationLevel,
|
AddressSpace, IntPredicate, OptimizationLevel,
|
||||||
};
|
};
|
||||||
use itertools::Itertools;
|
use itertools::Itertools;
|
||||||
use parking_lot::{Condvar, Mutex};
|
|
||||||
|
|
||||||
use nac3parser::ast::{Location, Stmt, StrRef};
|
use nac3parser::ast::{Location, Stmt, StrRef};
|
||||||
|
use parking_lot::{Condvar, Mutex};
|
||||||
use crate::{
|
use std::collections::{HashMap, HashSet};
|
||||||
codegen::classes::{ListType, NDArrayType, ProxyType, RangeType},
|
use std::sync::{
|
||||||
symbol_resolver::{StaticValue, SymbolResolver},
|
atomic::{AtomicBool, Ordering},
|
||||||
toplevel::{helper::PrimDef, numpy::unpack_ndarray_var_tys, TopLevelContext, TopLevelDef},
|
Arc,
|
||||||
typecheck::{
|
|
||||||
type_inferencer::{CodeLocation, PrimitiveStore},
|
|
||||||
typedef::{CallId, FuncArg, Type, TypeEnum, Unifier},
|
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
use std::thread;
|
||||||
|
|
||||||
pub mod builtin_fns;
|
pub mod builtin_fns;
|
||||||
pub mod classes;
|
pub mod classes;
|
||||||
@ -55,22 +50,6 @@ mod test;
|
|||||||
use concrete_type::{ConcreteType, ConcreteTypeEnum, ConcreteTypeStore};
|
use concrete_type::{ConcreteType, ConcreteTypeEnum, ConcreteTypeStore};
|
||||||
pub use generator::{CodeGenerator, DefaultCodeGenerator};
|
pub use generator::{CodeGenerator, DefaultCodeGenerator};
|
||||||
|
|
||||||
mod macros {
|
|
||||||
/// Codegen-variant of [`std::unreachable`] which accepts an instance of [`CodeGenContext`] as
|
|
||||||
/// its first argument to provide Python source information to indicate the codegen location
|
|
||||||
/// causing the assertion.
|
|
||||||
macro_rules! codegen_unreachable {
|
|
||||||
($ctx:expr $(,)?) => {
|
|
||||||
std::unreachable!("unreachable code while processing {}", &$ctx.current_loc)
|
|
||||||
};
|
|
||||||
($ctx:expr, $($arg:tt)*) => {
|
|
||||||
std::unreachable!("unreachable code while processing {}: {}", &$ctx.current_loc, std::format!("{}", std::format_args!($($arg)+)))
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
pub(crate) use codegen_unreachable;
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Default)]
|
#[derive(Default)]
|
||||||
pub struct StaticValueStore {
|
pub struct StaticValueStore {
|
||||||
pub lookup: HashMap<Vec<(usize, u64)>, usize>,
|
pub lookup: HashMap<Vec<(usize, u64)>, usize>,
|
||||||
@ -89,16 +68,6 @@ pub struct CodeGenLLVMOptions {
|
|||||||
pub target: CodeGenTargetMachineOptions,
|
pub target: CodeGenTargetMachineOptions,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl CodeGenLLVMOptions {
|
|
||||||
/// Creates a [`TargetMachine`] using the target options specified by this struct.
|
|
||||||
///
|
|
||||||
/// See [`Target::create_target_machine`].
|
|
||||||
#[must_use]
|
|
||||||
pub fn create_target_machine(&self) -> Option<TargetMachine> {
|
|
||||||
self.target.create_target_machine(self.opt_level)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Additional options for code generation for the target machine.
|
/// Additional options for code generation for the target machine.
|
||||||
#[derive(Clone, Debug, Eq, PartialEq)]
|
#[derive(Clone, Debug, Eq, PartialEq)]
|
||||||
pub struct CodeGenTargetMachineOptions {
|
pub struct CodeGenTargetMachineOptions {
|
||||||
@ -369,10 +338,6 @@ impl WorkerRegistry {
|
|||||||
let mut builder = context.create_builder();
|
let mut builder = context.create_builder();
|
||||||
let mut module = context.create_module(generator.get_name());
|
let mut module = context.create_module(generator.get_name());
|
||||||
|
|
||||||
let target_machine = self.llvm_options.create_target_machine().unwrap();
|
|
||||||
module.set_data_layout(&target_machine.get_target_data().get_data_layout());
|
|
||||||
module.set_triple(&target_machine.get_triple());
|
|
||||||
|
|
||||||
module.add_basic_value_flag(
|
module.add_basic_value_flag(
|
||||||
"Debug Info Version",
|
"Debug Info Version",
|
||||||
inkwell::module::FlagBehavior::Warning,
|
inkwell::module::FlagBehavior::Warning,
|
||||||
@ -396,10 +361,6 @@ impl WorkerRegistry {
|
|||||||
errors.insert(e);
|
errors.insert(e);
|
||||||
// create a new empty module just to continue codegen and collect errors
|
// create a new empty module just to continue codegen and collect errors
|
||||||
module = context.create_module(&format!("{}_recover", generator.get_name()));
|
module = context.create_module(&format!("{}_recover", generator.get_name()));
|
||||||
|
|
||||||
let target_machine = self.llvm_options.create_target_machine().unwrap();
|
|
||||||
module.set_data_layout(&target_machine.get_target_data().get_data_layout());
|
|
||||||
module.set_triple(&target_machine.get_triple());
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
*self.task_count.lock() -= 1;
|
*self.task_count.lock() -= 1;
|
||||||
@ -465,7 +426,7 @@ pub struct CodeGenTask {
|
|||||||
fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
ctx: &'ctx Context,
|
ctx: &'ctx Context,
|
||||||
module: &Module<'ctx>,
|
module: &Module<'ctx>,
|
||||||
generator: &G,
|
generator: &mut G,
|
||||||
unifier: &mut Unifier,
|
unifier: &mut Unifier,
|
||||||
top_level: &TopLevelContext,
|
top_level: &TopLevelContext,
|
||||||
type_cache: &mut HashMap<Type, BasicTypeEnum<'ctx>>,
|
type_cache: &mut HashMap<Type, BasicTypeEnum<'ctx>>,
|
||||||
@ -495,20 +456,6 @@ fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
.into()
|
.into()
|
||||||
}
|
}
|
||||||
|
|
||||||
TObj { obj_id, params, .. } if *obj_id == PrimDef::List.id() => {
|
|
||||||
let element_type = get_llvm_type(
|
|
||||||
ctx,
|
|
||||||
module,
|
|
||||||
generator,
|
|
||||||
unifier,
|
|
||||||
top_level,
|
|
||||||
type_cache,
|
|
||||||
*params.iter().next().unwrap().1,
|
|
||||||
);
|
|
||||||
|
|
||||||
ListType::new(generator, ctx, element_type).as_base_type().into()
|
|
||||||
}
|
|
||||||
|
|
||||||
TObj { obj_id, .. } if *obj_id == PrimDef::NDArray.id() => {
|
TObj { obj_id, .. } if *obj_id == PrimDef::NDArray.id() => {
|
||||||
let (dtype, _) = unpack_ndarray_var_tys(unifier, ty);
|
let (dtype, _) = unpack_ndarray_var_tys(unifier, ty);
|
||||||
let element_type = get_llvm_type(
|
let element_type = get_llvm_type(
|
||||||
@ -559,10 +506,8 @@ fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
};
|
};
|
||||||
return ty;
|
return ty;
|
||||||
}
|
}
|
||||||
TTuple { ty, is_vararg_ctx } => {
|
TTuple { ty } => {
|
||||||
// a struct with fields in the order present in the tuple
|
// a struct with fields in the order present in the tuple
|
||||||
assert!(!is_vararg_ctx, "Tuples in vararg context must be instantiated with the correct number of arguments before calling get_llvm_type");
|
|
||||||
|
|
||||||
let fields = ty
|
let fields = ty
|
||||||
.iter()
|
.iter()
|
||||||
.map(|ty| {
|
.map(|ty| {
|
||||||
@ -571,6 +516,12 @@ fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
.collect_vec();
|
.collect_vec();
|
||||||
ctx.struct_type(&fields, false).into()
|
ctx.struct_type(&fields, false).into()
|
||||||
}
|
}
|
||||||
|
TList { ty } => {
|
||||||
|
let element_type =
|
||||||
|
get_llvm_type(ctx, module, generator, unifier, top_level, type_cache, *ty);
|
||||||
|
|
||||||
|
ListType::new(generator, ctx, element_type).as_base_type().into()
|
||||||
|
}
|
||||||
TVirtual { .. } => unimplemented!(),
|
TVirtual { .. } => unimplemented!(),
|
||||||
_ => unreachable!("{}", ty_enum.get_type_name()),
|
_ => unreachable!("{}", ty_enum.get_type_name()),
|
||||||
};
|
};
|
||||||
@ -592,7 +543,7 @@ fn get_llvm_type<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
fn get_llvm_abi_type<'ctx, G: CodeGenerator + ?Sized>(
|
fn get_llvm_abi_type<'ctx, G: CodeGenerator + ?Sized>(
|
||||||
ctx: &'ctx Context,
|
ctx: &'ctx Context,
|
||||||
module: &Module<'ctx>,
|
module: &Module<'ctx>,
|
||||||
generator: &G,
|
generator: &mut G,
|
||||||
unifier: &mut Unifier,
|
unifier: &mut Unifier,
|
||||||
top_level: &TopLevelContext,
|
top_level: &TopLevelContext,
|
||||||
type_cache: &mut HashMap<Type, BasicTypeEnum<'ctx>>,
|
type_cache: &mut HashMap<Type, BasicTypeEnum<'ctx>>,
|
||||||
@ -601,11 +552,11 @@ fn get_llvm_abi_type<'ctx, G: CodeGenerator + ?Sized>(
|
|||||||
) -> BasicTypeEnum<'ctx> {
|
) -> BasicTypeEnum<'ctx> {
|
||||||
// If the type is used in the definition of a function, return `i1` instead of `i8` for ABI
|
// If the type is used in the definition of a function, return `i1` instead of `i8` for ABI
|
||||||
// consistency.
|
// consistency.
|
||||||
if unifier.unioned(ty, primitives.bool) {
|
return if unifier.unioned(ty, primitives.bool) {
|
||||||
ctx.bool_type().into()
|
ctx.bool_type().into()
|
||||||
} else {
|
} else {
|
||||||
get_llvm_type(ctx, module, generator, unifier, top_level, type_cache, ty)
|
get_llvm_type(ctx, module, generator, unifier, top_level, type_cache, ty)
|
||||||
}
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Whether `sret` is needed for a return value with type `ty`.
|
/// Whether `sret` is needed for a return value with type `ty`.
|
||||||
@ -630,40 +581,6 @@ fn need_sret(ty: BasicTypeEnum) -> bool {
|
|||||||
need_sret_impl(ty, true)
|
need_sret_impl(ty, true)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns the [`BasicTypeEnum`] representing a `va_list` struct for variadic arguments.
|
|
||||||
fn get_llvm_valist_type<'ctx>(ctx: &'ctx Context, triple: &TargetTriple) -> BasicTypeEnum<'ctx> {
|
|
||||||
let triple = TargetMachine::normalize_triple(triple);
|
|
||||||
let triple = triple.as_str().to_str().unwrap();
|
|
||||||
let arch = triple.split('-').next().unwrap();
|
|
||||||
|
|
||||||
let llvm_pi8 = ctx.i8_type().ptr_type(AddressSpace::default());
|
|
||||||
|
|
||||||
// Referenced from parseArch() in llvm/lib/Support/Triple.cpp
|
|
||||||
match arch {
|
|
||||||
"i386" | "i486" | "i586" | "i686" | "riscv32" => {
|
|
||||||
ctx.i8_type().ptr_type(AddressSpace::default()).into()
|
|
||||||
}
|
|
||||||
"amd64" | "x86_64" | "x86_64h" => {
|
|
||||||
let llvm_i32 = ctx.i32_type();
|
|
||||||
|
|
||||||
let va_list_tag = ctx.opaque_struct_type("struct.__va_list_tag");
|
|
||||||
va_list_tag.set_body(
|
|
||||||
&[llvm_i32.into(), llvm_i32.into(), llvm_pi8.into(), llvm_pi8.into()],
|
|
||||||
false,
|
|
||||||
);
|
|
||||||
va_list_tag.into()
|
|
||||||
}
|
|
||||||
"armv7" => {
|
|
||||||
let va_list = ctx.opaque_struct_type("struct.__va_list");
|
|
||||||
va_list.set_body(&[llvm_pi8.into()], false);
|
|
||||||
va_list.into()
|
|
||||||
}
|
|
||||||
triple => {
|
|
||||||
todo!("Unsupported platform for varargs: {triple}")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Implementation for generating LLVM IR for a function.
|
/// Implementation for generating LLVM IR for a function.
|
||||||
pub fn gen_func_impl<
|
pub fn gen_func_impl<
|
||||||
'ctx,
|
'ctx,
|
||||||
@ -775,7 +692,6 @@ pub fn gen_func_impl<
|
|||||||
name: arg.name,
|
name: arg.name,
|
||||||
ty: task.store.to_unifier_type(&mut unifier, &primitives, arg.ty, &mut cache),
|
ty: task.store.to_unifier_type(&mut unifier, &primitives, arg.ty, &mut cache),
|
||||||
default_value: arg.default_value.clone(),
|
default_value: arg.default_value.clone(),
|
||||||
is_vararg: arg.is_vararg,
|
|
||||||
})
|
})
|
||||||
.collect_vec(),
|
.collect_vec(),
|
||||||
task.store.to_unifier_type(&mut unifier, &primitives, *ret, &mut cache),
|
task.store.to_unifier_type(&mut unifier, &primitives, *ret, &mut cache),
|
||||||
@ -798,10 +714,7 @@ pub fn gen_func_impl<
|
|||||||
let has_sret = ret_type.map_or(false, |ty| need_sret(ty));
|
let has_sret = ret_type.map_or(false, |ty| need_sret(ty));
|
||||||
let mut params = args
|
let mut params = args
|
||||||
.iter()
|
.iter()
|
||||||
.filter(|arg| !arg.is_vararg)
|
|
||||||
.map(|arg| {
|
.map(|arg| {
|
||||||
debug_assert!(!arg.is_vararg);
|
|
||||||
|
|
||||||
get_llvm_abi_type(
|
get_llvm_abi_type(
|
||||||
context,
|
context,
|
||||||
&module,
|
&module,
|
||||||
@ -820,12 +733,9 @@ pub fn gen_func_impl<
|
|||||||
params.insert(0, ret_type.unwrap().ptr_type(AddressSpace::default()).into());
|
params.insert(0, ret_type.unwrap().ptr_type(AddressSpace::default()).into());
|
||||||
}
|
}
|
||||||
|
|
||||||
debug_assert!(matches!(args.iter().filter(|arg| arg.is_vararg).count(), 0..=1));
|
|
||||||
let vararg_arg = args.iter().find(|arg| arg.is_vararg);
|
|
||||||
|
|
||||||
let fn_type = match ret_type {
|
let fn_type = match ret_type {
|
||||||
Some(ret_type) if !has_sret => ret_type.fn_type(¶ms, vararg_arg.is_some()),
|
Some(ret_type) if !has_sret => ret_type.fn_type(¶ms, false),
|
||||||
_ => context.void_type().fn_type(¶ms, vararg_arg.is_some()),
|
_ => context.void_type().fn_type(¶ms, false),
|
||||||
};
|
};
|
||||||
|
|
||||||
let symbol = &task.symbol_name;
|
let symbol = &task.symbol_name;
|
||||||
@ -853,10 +763,9 @@ pub fn gen_func_impl<
|
|||||||
builder.position_at_end(init_bb);
|
builder.position_at_end(init_bb);
|
||||||
let body_bb = context.append_basic_block(fn_val, "body");
|
let body_bb = context.append_basic_block(fn_val, "body");
|
||||||
|
|
||||||
// Store non-vararg argument values into local variables
|
|
||||||
let mut var_assignment = HashMap::new();
|
let mut var_assignment = HashMap::new();
|
||||||
let offset = u32::from(has_sret);
|
let offset = u32::from(has_sret);
|
||||||
for (n, arg) in args.iter().enumerate().filter(|(_, arg)| !arg.is_vararg) {
|
for (n, arg) in args.iter().enumerate() {
|
||||||
let param = fn_val.get_nth_param((n as u32) + offset).unwrap();
|
let param = fn_val.get_nth_param((n as u32) + offset).unwrap();
|
||||||
let local_type = get_llvm_type(
|
let local_type = get_llvm_type(
|
||||||
context,
|
context,
|
||||||
@ -889,8 +798,6 @@ pub fn gen_func_impl<
|
|||||||
var_assignment.insert(arg.name, (alloca, None, 0));
|
var_assignment.insert(arg.name, (alloca, None, 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Save vararg parameters as list
|
|
||||||
|
|
||||||
let return_buffer = if has_sret {
|
let return_buffer = if has_sret {
|
||||||
Some(fn_val.get_nth_param(0).unwrap().into_pointer_value())
|
Some(fn_val.get_nth_param(0).unwrap().into_pointer_value())
|
||||||
} else {
|
} else {
|
||||||
@ -1113,9 +1020,3 @@ fn gen_in_range_check<'ctx>(
|
|||||||
|
|
||||||
ctx.builder.build_int_compare(IntPredicate::SLT, lo, hi, "cmp").unwrap()
|
ctx.builder.build_int_compare(IntPredicate::SLT, lo, hi, "cmp").unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns the internal name for the `va_count` argument, used to indicate the number of arguments
|
|
||||||
/// passed to the variadic function.
|
|
||||||
fn get_va_count_arg_name(arg_name: StrRef) -> StrRef {
|
|
||||||
format!("__{}_va_count", &arg_name).into()
|
|
||||||
}
|
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -1,3 +1,18 @@
|
|||||||
|
use super::{
|
||||||
|
super::symbol_resolver::ValueEnum,
|
||||||
|
expr::destructure_range,
|
||||||
|
irrt::{handle_slice_indices, list_slice_assignment},
|
||||||
|
CodeGenContext, CodeGenerator,
|
||||||
|
};
|
||||||
|
use crate::{
|
||||||
|
codegen::{
|
||||||
|
classes::{ArrayLikeIndexer, ArraySliceValue, ListValue, RangeValue},
|
||||||
|
expr::gen_binop_expr,
|
||||||
|
gen_in_range_check,
|
||||||
|
},
|
||||||
|
toplevel::{helper::PrimDef, numpy::unpack_ndarray_var_tys, DefinitionId, TopLevelDef},
|
||||||
|
typecheck::typedef::{FunSignature, Type, TypeEnum},
|
||||||
|
};
|
||||||
use inkwell::{
|
use inkwell::{
|
||||||
attributes::{Attribute, AttributeLoc},
|
attributes::{Attribute, AttributeLoc},
|
||||||
basic_block::BasicBlock,
|
basic_block::BasicBlock,
|
||||||
@ -5,28 +20,10 @@ use inkwell::{
|
|||||||
values::{BasicValue, BasicValueEnum, FunctionValue, IntValue, PointerValue},
|
values::{BasicValue, BasicValueEnum, FunctionValue, IntValue, PointerValue},
|
||||||
IntPredicate,
|
IntPredicate,
|
||||||
};
|
};
|
||||||
use itertools::{izip, Itertools};
|
|
||||||
|
|
||||||
use nac3parser::ast::{
|
use nac3parser::ast::{
|
||||||
Constant, ExcepthandlerKind, Expr, ExprKind, Location, Stmt, StmtKind, StrRef,
|
Constant, ExcepthandlerKind, Expr, ExprKind, Location, Stmt, StmtKind, StrRef,
|
||||||
};
|
};
|
||||||
|
use std::convert::TryFrom;
|
||||||
use super::{
|
|
||||||
classes::{ArrayLikeIndexer, ArraySliceValue, ListValue, RangeValue},
|
|
||||||
expr::{destructure_range, gen_binop_expr},
|
|
||||||
gen_in_range_check,
|
|
||||||
irrt::{handle_slice_indices, list_slice_assignment},
|
|
||||||
macros::codegen_unreachable,
|
|
||||||
CodeGenContext, CodeGenerator,
|
|
||||||
};
|
|
||||||
use crate::{
|
|
||||||
symbol_resolver::ValueEnum,
|
|
||||||
toplevel::{DefinitionId, TopLevelDef},
|
|
||||||
typecheck::{
|
|
||||||
magic_methods::Binop,
|
|
||||||
typedef::{iter_type_vars, FunSignature, Type, TypeEnum},
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
/// See [`CodeGenerator::gen_var_alloc`].
|
/// See [`CodeGenerator::gen_var_alloc`].
|
||||||
pub fn gen_var<'ctx>(
|
pub fn gen_var<'ctx>(
|
||||||
@ -97,6 +94,8 @@ pub fn gen_store_target<'ctx, G: CodeGenerator>(
|
|||||||
pattern: &Expr<Option<Type>>,
|
pattern: &Expr<Option<Type>>,
|
||||||
name: Option<&str>,
|
name: Option<&str>,
|
||||||
) -> Result<Option<PointerValue<'ctx>>, String> {
|
) -> Result<Option<PointerValue<'ctx>>, String> {
|
||||||
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
|
|
||||||
// very similar to gen_expr, but we don't do an extra load at the end
|
// very similar to gen_expr, but we don't do an extra load at the end
|
||||||
// and we flatten nested tuples
|
// and we flatten nested tuples
|
||||||
Ok(Some(match &pattern.node {
|
Ok(Some(match &pattern.node {
|
||||||
@ -114,14 +113,14 @@ pub fn gen_store_target<'ctx, G: CodeGenerator>(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
ExprKind::Attribute { value, attr, .. } => {
|
ExprKind::Attribute { value, attr, .. } => {
|
||||||
let (index, _) = ctx.get_attr_index(value.custom.unwrap(), *attr);
|
let index = ctx.get_attr_index(value.custom.unwrap(), *attr);
|
||||||
let val = if let Some(v) = generator.gen_expr(ctx, value)? {
|
let val = if let Some(v) = generator.gen_expr(ctx, value)? {
|
||||||
v.to_basic_value_enum(ctx, generator, value.custom.unwrap())?
|
v.to_basic_value_enum(ctx, generator, value.custom.unwrap())?
|
||||||
} else {
|
} else {
|
||||||
return Ok(None);
|
return Ok(None);
|
||||||
};
|
};
|
||||||
let BasicValueEnum::PointerValue(ptr) = val else {
|
let BasicValueEnum::PointerValue(ptr) = val else {
|
||||||
codegen_unreachable!(ctx);
|
unreachable!();
|
||||||
};
|
};
|
||||||
unsafe {
|
unsafe {
|
||||||
ctx.builder.build_in_bounds_gep(
|
ctx.builder.build_in_bounds_gep(
|
||||||
@ -135,7 +134,66 @@ pub fn gen_store_target<'ctx, G: CodeGenerator>(
|
|||||||
}
|
}
|
||||||
.unwrap()
|
.unwrap()
|
||||||
}
|
}
|
||||||
_ => codegen_unreachable!(ctx),
|
ExprKind::Subscript { value, slice, .. } => {
|
||||||
|
match ctx.unifier.get_ty_immutable(value.custom.unwrap()).as_ref() {
|
||||||
|
TypeEnum::TList { .. } => {
|
||||||
|
let v = generator
|
||||||
|
.gen_expr(ctx, value)?
|
||||||
|
.unwrap()
|
||||||
|
.to_basic_value_enum(ctx, generator, value.custom.unwrap())?
|
||||||
|
.into_pointer_value();
|
||||||
|
let v = ListValue::from_ptr_val(v, llvm_usize, None);
|
||||||
|
let len = v.load_size(ctx, Some("len"));
|
||||||
|
let raw_index = generator
|
||||||
|
.gen_expr(ctx, slice)?
|
||||||
|
.unwrap()
|
||||||
|
.to_basic_value_enum(ctx, generator, slice.custom.unwrap())?
|
||||||
|
.into_int_value();
|
||||||
|
let raw_index = ctx
|
||||||
|
.builder
|
||||||
|
.build_int_s_extend(raw_index, generator.get_size_type(ctx.ctx), "sext")
|
||||||
|
.unwrap();
|
||||||
|
// handle negative index
|
||||||
|
let is_negative = ctx
|
||||||
|
.builder
|
||||||
|
.build_int_compare(
|
||||||
|
IntPredicate::SLT,
|
||||||
|
raw_index,
|
||||||
|
generator.get_size_type(ctx.ctx).const_zero(),
|
||||||
|
"is_neg",
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
let adjusted = ctx.builder.build_int_add(raw_index, len, "adjusted").unwrap();
|
||||||
|
let index = ctx
|
||||||
|
.builder
|
||||||
|
.build_select(is_negative, adjusted, raw_index, "index")
|
||||||
|
.map(BasicValueEnum::into_int_value)
|
||||||
|
.unwrap();
|
||||||
|
// unsigned less than is enough, because negative index after adjustment is
|
||||||
|
// bigger than the length (for unsigned cmp)
|
||||||
|
let bound_check = ctx
|
||||||
|
.builder
|
||||||
|
.build_int_compare(IntPredicate::ULT, index, len, "inbound")
|
||||||
|
.unwrap();
|
||||||
|
ctx.make_assert(
|
||||||
|
generator,
|
||||||
|
bound_check,
|
||||||
|
"0:IndexError",
|
||||||
|
"index {0} out of bounds 0:{1}",
|
||||||
|
[Some(raw_index), Some(len), None],
|
||||||
|
slice.location,
|
||||||
|
);
|
||||||
|
v.data().ptr_offset(ctx, generator, &index, name)
|
||||||
|
}
|
||||||
|
|
||||||
|
TypeEnum::TObj { obj_id, .. } if *obj_id == PrimDef::NDArray.id() => {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -145,20 +203,68 @@ pub fn gen_assign<'ctx, G: CodeGenerator>(
|
|||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
ctx: &mut CodeGenContext<'ctx, '_>,
|
||||||
target: &Expr<Option<Type>>,
|
target: &Expr<Option<Type>>,
|
||||||
value: ValueEnum<'ctx>,
|
value: ValueEnum<'ctx>,
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
// See https://docs.python.org/3/reference/simple_stmts.html#assignment-statements.
|
let llvm_usize = generator.get_size_type(ctx.ctx);
|
||||||
|
|
||||||
match &target.node {
|
match &target.node {
|
||||||
ExprKind::Subscript { value: target, slice: key, .. } => {
|
ExprKind::Tuple { elts, .. } => {
|
||||||
// Handle "slicing" or "subscription"
|
let BasicValueEnum::StructValue(v) =
|
||||||
generator.gen_setitem(ctx, target, key, value, value_ty)?;
|
value.to_basic_value_enum(ctx, generator, target.custom.unwrap())?
|
||||||
|
else {
|
||||||
|
unreachable!()
|
||||||
|
};
|
||||||
|
|
||||||
|
for (i, elt) in elts.iter().enumerate() {
|
||||||
|
let v = ctx
|
||||||
|
.builder
|
||||||
|
.build_extract_value(v, u32::try_from(i).unwrap(), "struct_elem")
|
||||||
|
.unwrap();
|
||||||
|
generator.gen_assign(ctx, elt, v.into())?;
|
||||||
}
|
}
|
||||||
ExprKind::Tuple { elts, .. } | ExprKind::List { elts, .. } => {
|
}
|
||||||
// Fold on `"[" [target_list] "]"` and `"(" [target_list] ")"`
|
ExprKind::Subscript { value: ls, slice, .. }
|
||||||
generator.gen_assign_target_list(ctx, elts, value, value_ty)?;
|
if matches!(&slice.node, ExprKind::Slice { .. }) =>
|
||||||
|
{
|
||||||
|
let ExprKind::Slice { lower, upper, step } = &slice.node else { unreachable!() };
|
||||||
|
|
||||||
|
let ls = generator
|
||||||
|
.gen_expr(ctx, ls)?
|
||||||
|
.unwrap()
|
||||||
|
.to_basic_value_enum(ctx, generator, ls.custom.unwrap())?
|
||||||
|
.into_pointer_value();
|
||||||
|
let ls = ListValue::from_ptr_val(ls, llvm_usize, None);
|
||||||
|
let Some((start, end, step)) =
|
||||||
|
handle_slice_indices(lower, upper, step, ctx, generator, ls.load_size(ctx, None))?
|
||||||
|
else {
|
||||||
|
return Ok(());
|
||||||
|
};
|
||||||
|
let value = value
|
||||||
|
.to_basic_value_enum(ctx, generator, target.custom.unwrap())?
|
||||||
|
.into_pointer_value();
|
||||||
|
let value = ListValue::from_ptr_val(value, llvm_usize, None);
|
||||||
|
let ty = match &*ctx.unifier.get_ty_immutable(target.custom.unwrap()) {
|
||||||
|
TypeEnum::TList { ty } => *ty,
|
||||||
|
TypeEnum::TObj { obj_id, .. } if *obj_id == PrimDef::NDArray.id() => {
|
||||||
|
unpack_ndarray_var_tys(&mut ctx.unifier, target.custom.unwrap()).0
|
||||||
|
}
|
||||||
|
_ => unreachable!(),
|
||||||
|
};
|
||||||
|
|
||||||
|
let ty = ctx.get_llvm_type(generator, ty);
|
||||||
|
let Some(src_ind) = handle_slice_indices(
|
||||||
|
&None,
|
||||||
|
&None,
|
||||||
|
&None,
|
||||||
|
ctx,
|
||||||
|
generator,
|
||||||
|
value.load_size(ctx, None),
|
||||||
|
)?
|
||||||
|
else {
|
||||||
|
return Ok(());
|
||||||
|
};
|
||||||
|
list_slice_assignment(generator, ctx, ty, ls, (start, end, step), value, src_ind);
|
||||||
}
|
}
|
||||||
_ => {
|
_ => {
|
||||||
// Handle attribute and direct variable assignments.
|
|
||||||
let name = if let ExprKind::Name { id, .. } = &target.node {
|
let name = if let ExprKind::Name { id, .. } = &target.node {
|
||||||
format!("{id}.addr")
|
format!("{id}.addr")
|
||||||
} else {
|
} else {
|
||||||
@ -176,259 +282,19 @@ pub fn gen_assign<'ctx, G: CodeGenerator>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
let val = value.to_basic_value_enum(ctx, generator, target.custom.unwrap())?;
|
let val = value.to_basic_value_enum(ctx, generator, target.custom.unwrap())?;
|
||||||
|
|
||||||
// Perform i1 <-> i8 conversion as needed
|
|
||||||
let val = if ctx.unifier.unioned(target.custom.unwrap(), ctx.primitives.bool) {
|
|
||||||
generator.bool_to_i8(ctx, val.into_int_value()).into()
|
|
||||||
} else {
|
|
||||||
val
|
|
||||||
};
|
|
||||||
|
|
||||||
ctx.builder.build_store(ptr, val).unwrap();
|
ctx.builder.build_store(ptr, val).unwrap();
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// See [`CodeGenerator::gen_assign_target_list`].
|
|
||||||
pub fn gen_assign_target_list<'ctx, G: CodeGenerator>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
targets: &Vec<Expr<Option<Type>>>,
|
|
||||||
value: ValueEnum<'ctx>,
|
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String> {
|
|
||||||
// Deconstruct the tuple `value`
|
|
||||||
let BasicValueEnum::StructValue(tuple) = value.to_basic_value_enum(ctx, generator, value_ty)?
|
|
||||||
else {
|
|
||||||
codegen_unreachable!(ctx)
|
|
||||||
};
|
|
||||||
|
|
||||||
// NOTE: Currently, RHS's type is forced to be a Tuple by the type inferencer.
|
|
||||||
let TypeEnum::TTuple { ty: tuple_tys, .. } = &*ctx.unifier.get_ty(value_ty) else {
|
|
||||||
codegen_unreachable!(ctx);
|
|
||||||
};
|
|
||||||
|
|
||||||
assert_eq!(tuple.get_type().count_fields() as usize, tuple_tys.len());
|
|
||||||
|
|
||||||
let tuple = (0..tuple.get_type().count_fields())
|
|
||||||
.map(|i| ctx.builder.build_extract_value(tuple, i, "item").unwrap())
|
|
||||||
.collect_vec();
|
|
||||||
|
|
||||||
// Find the starred target if it exists.
|
|
||||||
let mut starred_target_index: Option<usize> = None; // Index of the "starred" target. If it exists, there may only be one.
|
|
||||||
for (i, target) in targets.iter().enumerate() {
|
|
||||||
if matches!(target.node, ExprKind::Starred { .. }) {
|
|
||||||
assert!(starred_target_index.is_none()); // The typechecker ensures this
|
|
||||||
starred_target_index = Some(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Some(starred_target_index) = starred_target_index {
|
|
||||||
assert!(tuple_tys.len() >= targets.len() - 1); // The typechecker ensures this
|
|
||||||
|
|
||||||
let a = starred_target_index; // Number of RHS values before the starred target
|
|
||||||
let b = tuple_tys.len() - (targets.len() - 1 - starred_target_index); // Number of RHS values after the starred target
|
|
||||||
// Thus `tuple[a..b]` is assigned to the starred target.
|
|
||||||
|
|
||||||
// Handle assignment before the starred target
|
|
||||||
for (target, val, val_ty) in
|
|
||||||
izip!(&targets[..starred_target_index], &tuple[..a], &tuple_tys[..a])
|
|
||||||
{
|
|
||||||
generator.gen_assign(ctx, target, ValueEnum::Dynamic(*val), *val_ty)?;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle assignment to the starred target
|
|
||||||
if let ExprKind::Starred { value: target, .. } = &targets[starred_target_index].node {
|
|
||||||
let vals = &tuple[a..b];
|
|
||||||
let val_tys = &tuple_tys[a..b];
|
|
||||||
|
|
||||||
// Create a sub-tuple from `value` for the starred target.
|
|
||||||
let sub_tuple_ty = ctx
|
|
||||||
.ctx
|
|
||||||
.struct_type(&vals.iter().map(BasicValueEnum::get_type).collect_vec(), false);
|
|
||||||
let psub_tuple_val =
|
|
||||||
ctx.builder.build_alloca(sub_tuple_ty, "starred_target_value_ptr").unwrap();
|
|
||||||
for (i, val) in vals.iter().enumerate() {
|
|
||||||
let pitem = ctx
|
|
||||||
.builder
|
|
||||||
.build_struct_gep(psub_tuple_val, i as u32, "starred_target_value_item")
|
|
||||||
.unwrap();
|
|
||||||
ctx.builder.build_store(pitem, *val).unwrap();
|
|
||||||
}
|
|
||||||
let sub_tuple_val =
|
|
||||||
ctx.builder.build_load(psub_tuple_val, "starred_target_value").unwrap();
|
|
||||||
|
|
||||||
// Create the typechecker type of the sub-tuple
|
|
||||||
let sub_tuple_ty =
|
|
||||||
ctx.unifier.add_ty(TypeEnum::TTuple { ty: val_tys.to_vec(), is_vararg_ctx: false });
|
|
||||||
|
|
||||||
// Now assign with that sub-tuple to the starred target.
|
|
||||||
generator.gen_assign(ctx, target, ValueEnum::Dynamic(sub_tuple_val), sub_tuple_ty)?;
|
|
||||||
} else {
|
|
||||||
codegen_unreachable!(ctx) // The typechecker ensures this
|
|
||||||
}
|
|
||||||
|
|
||||||
// Handle assignment after the starred target
|
|
||||||
for (target, val, val_ty) in
|
|
||||||
izip!(&targets[starred_target_index + 1..], &tuple[b..], &tuple_tys[b..])
|
|
||||||
{
|
|
||||||
generator.gen_assign(ctx, target, ValueEnum::Dynamic(*val), *val_ty)?;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
assert_eq!(tuple_tys.len(), targets.len()); // The typechecker ensures this
|
|
||||||
|
|
||||||
for (target, val, val_ty) in izip!(targets, tuple, tuple_tys) {
|
|
||||||
generator.gen_assign(ctx, target, ValueEnum::Dynamic(val), *val_ty)?;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// See [`CodeGenerator::gen_setitem`].
|
|
||||||
pub fn gen_setitem<'ctx, G: CodeGenerator>(
|
|
||||||
generator: &mut G,
|
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
|
||||||
target: &Expr<Option<Type>>,
|
|
||||||
key: &Expr<Option<Type>>,
|
|
||||||
value: ValueEnum<'ctx>,
|
|
||||||
value_ty: Type,
|
|
||||||
) -> Result<(), String> {
|
|
||||||
let target_ty = target.custom.unwrap();
|
|
||||||
let key_ty = key.custom.unwrap();
|
|
||||||
|
|
||||||
match &*ctx.unifier.get_ty(target_ty) {
|
|
||||||
TypeEnum::TObj { obj_id, params: list_params, .. }
|
|
||||||
if *obj_id == ctx.primitives.list.obj_id(&ctx.unifier).unwrap() =>
|
|
||||||
{
|
|
||||||
// Handle list item assignment
|
|
||||||
let llvm_usize = generator.get_size_type(ctx.ctx);
|
|
||||||
let target_item_ty = iter_type_vars(list_params).next().unwrap().ty;
|
|
||||||
|
|
||||||
let target = generator
|
|
||||||
.gen_expr(ctx, target)?
|
|
||||||
.unwrap()
|
|
||||||
.to_basic_value_enum(ctx, generator, target_ty)?
|
|
||||||
.into_pointer_value();
|
|
||||||
let target = ListValue::from_ptr_val(target, llvm_usize, None);
|
|
||||||
|
|
||||||
if let ExprKind::Slice { .. } = &key.node {
|
|
||||||
// Handle assigning to a slice
|
|
||||||
let ExprKind::Slice { lower, upper, step } = &key.node else {
|
|
||||||
codegen_unreachable!(ctx)
|
|
||||||
};
|
|
||||||
let Some((start, end, step)) = handle_slice_indices(
|
|
||||||
lower,
|
|
||||||
upper,
|
|
||||||
step,
|
|
||||||
ctx,
|
|
||||||
generator,
|
|
||||||
target.load_size(ctx, None),
|
|
||||||
)?
|
|
||||||
else {
|
|
||||||
return Ok(());
|
|
||||||
};
|
|
||||||
|
|
||||||
let value =
|
|
||||||
value.to_basic_value_enum(ctx, generator, value_ty)?.into_pointer_value();
|
|
||||||
let value = ListValue::from_ptr_val(value, llvm_usize, None);
|
|
||||||
|
|
||||||
let target_item_ty = ctx.get_llvm_type(generator, target_item_ty);
|
|
||||||
let Some(src_ind) = handle_slice_indices(
|
|
||||||
&None,
|
|
||||||
&None,
|
|
||||||
&None,
|
|
||||||
ctx,
|
|
||||||
generator,
|
|
||||||
value.load_size(ctx, None),
|
|
||||||
)?
|
|
||||||
else {
|
|
||||||
return Ok(());
|
|
||||||
};
|
|
||||||
list_slice_assignment(
|
|
||||||
generator,
|
|
||||||
ctx,
|
|
||||||
target_item_ty,
|
|
||||||
target,
|
|
||||||
(start, end, step),
|
|
||||||
value,
|
|
||||||
src_ind,
|
|
||||||
);
|
|
||||||
} else {
|
|
||||||
// Handle assigning to an index
|
|
||||||
let len = target.load_size(ctx, Some("len"));
|
|
||||||
|
|
||||||
let index = generator
|
|
||||||
.gen_expr(ctx, key)?
|
|
||||||
.unwrap()
|
|
||||||
.to_basic_value_enum(ctx, generator, key_ty)?
|
|
||||||
.into_int_value();
|
|
||||||
let index = ctx
|
|
||||||
.builder
|
|
||||||
.build_int_s_extend(index, generator.get_size_type(ctx.ctx), "sext")
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
// handle negative index
|
|
||||||
let is_negative = ctx
|
|
||||||
.builder
|
|
||||||
.build_int_compare(
|
|
||||||
IntPredicate::SLT,
|
|
||||||
index,
|
|
||||||
generator.get_size_type(ctx.ctx).const_zero(),
|
|
||||||
"is_neg",
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
let adjusted = ctx.builder.build_int_add(index, len, "adjusted").unwrap();
|
|
||||||
let index = ctx
|
|
||||||
.builder
|
|
||||||
.build_select(is_negative, adjusted, index, "index")
|
|
||||||
.map(BasicValueEnum::into_int_value)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
// unsigned less than is enough, because negative index after adjustment is
|
|
||||||
// bigger than the length (for unsigned cmp)
|
|
||||||
let bound_check = ctx
|
|
||||||
.builder
|
|
||||||
.build_int_compare(IntPredicate::ULT, index, len, "inbound")
|
|
||||||
.unwrap();
|
|
||||||
ctx.make_assert(
|
|
||||||
generator,
|
|
||||||
bound_check,
|
|
||||||
"0:IndexError",
|
|
||||||
"index {0} out of bounds 0:{1}",
|
|
||||||
[Some(index), Some(len), None],
|
|
||||||
key.location,
|
|
||||||
);
|
|
||||||
|
|
||||||
// Write value to index on list
|
|
||||||
let item_ptr =
|
|
||||||
target.data().ptr_offset(ctx, generator, &index, Some("list_item_ptr"));
|
|
||||||
let value = value.to_basic_value_enum(ctx, generator, value_ty)?;
|
|
||||||
ctx.builder.build_store(item_ptr, value).unwrap();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
TypeEnum::TObj { obj_id, .. }
|
|
||||||
if *obj_id == ctx.primitives.ndarray.obj_id(&ctx.unifier).unwrap() =>
|
|
||||||
{
|
|
||||||
// Handle NDArray item assignment
|
|
||||||
todo!("ndarray subscript assignment is not yet implemented");
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
panic!("encountered unknown target type: {}", ctx.unifier.stringify(target_ty));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
/// See [`CodeGenerator::gen_for`].
|
/// See [`CodeGenerator::gen_for`].
|
||||||
pub fn gen_for<G: CodeGenerator>(
|
pub fn gen_for<G: CodeGenerator>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
ctx: &mut CodeGenContext<'_, '_>,
|
ctx: &mut CodeGenContext<'_, '_>,
|
||||||
stmt: &Stmt<Option<Type>>,
|
stmt: &Stmt<Option<Type>>,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
let StmtKind::For { iter, target, body, orelse, .. } = &stmt.node else {
|
let StmtKind::For { iter, target, body, orelse, .. } = &stmt.node else { unreachable!() };
|
||||||
codegen_unreachable!(ctx)
|
|
||||||
};
|
|
||||||
|
|
||||||
// var_assignment static values may be changed in another branch
|
// var_assignment static values may be changed in another branch
|
||||||
// if so, remove the static value as it may not be correct in this branch
|
// if so, remove the static value as it may not be correct in this branch
|
||||||
@ -444,6 +310,9 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
let orelse_bb =
|
let orelse_bb =
|
||||||
if orelse.is_empty() { cont_bb } else { ctx.ctx.append_basic_block(current, "for.orelse") };
|
if orelse.is_empty() { cont_bb } else { ctx.ctx.append_basic_block(current, "for.orelse") };
|
||||||
|
|
||||||
|
// Whether the iterable is a range() expression
|
||||||
|
let is_iterable_range_expr = ctx.unifier.unioned(iter.custom.unwrap(), ctx.primitives.range);
|
||||||
|
|
||||||
// The BB containing the increment expression
|
// The BB containing the increment expression
|
||||||
let incr_bb = ctx.ctx.append_basic_block(current, "for.incr");
|
let incr_bb = ctx.ctx.append_basic_block(current, "for.incr");
|
||||||
// The BB containing the loop condition check
|
// The BB containing the loop condition check
|
||||||
@ -452,35 +321,27 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
// store loop bb information and restore it later
|
// store loop bb information and restore it later
|
||||||
let loop_bb = ctx.loop_target.replace((incr_bb, cont_bb));
|
let loop_bb = ctx.loop_target.replace((incr_bb, cont_bb));
|
||||||
|
|
||||||
let iter_ty = iter.custom.unwrap();
|
|
||||||
let iter_val = if let Some(v) = generator.gen_expr(ctx, iter)? {
|
let iter_val = if let Some(v) = generator.gen_expr(ctx, iter)? {
|
||||||
v.to_basic_value_enum(ctx, generator, iter_ty)?
|
v.to_basic_value_enum(ctx, generator, iter.custom.unwrap())?
|
||||||
} else {
|
} else {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
};
|
};
|
||||||
|
if is_iterable_range_expr {
|
||||||
match &*ctx.unifier.get_ty(iter_ty) {
|
|
||||||
TypeEnum::TObj { obj_id, .. }
|
|
||||||
if *obj_id == ctx.primitives.range.obj_id(&ctx.unifier).unwrap() =>
|
|
||||||
{
|
|
||||||
let iter_val = RangeValue::from_ptr_val(iter_val.into_pointer_value(), Some("range"));
|
let iter_val = RangeValue::from_ptr_val(iter_val.into_pointer_value(), Some("range"));
|
||||||
// Internal variable for loop; Cannot be assigned
|
// Internal variable for loop; Cannot be assigned
|
||||||
let i = generator.gen_var_alloc(ctx, int32.into(), Some("for.i.addr"))?;
|
let i = generator.gen_var_alloc(ctx, int32.into(), Some("for.i.addr"))?;
|
||||||
// Variable declared in "target" expression of the loop; Can be reassigned *or* shadowed
|
// Variable declared in "target" expression of the loop; Can be reassigned *or* shadowed
|
||||||
let Some(target_i) =
|
let Some(target_i) = generator.gen_store_target(ctx, target, Some("for.target.addr"))?
|
||||||
generator.gen_store_target(ctx, target, Some("for.target.addr"))?
|
|
||||||
else {
|
else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
let (start, stop, step) = destructure_range(ctx, iter_val);
|
let (start, stop, step) = destructure_range(ctx, iter_val);
|
||||||
|
|
||||||
ctx.builder.build_store(i, start).unwrap();
|
ctx.builder.build_store(i, start).unwrap();
|
||||||
|
|
||||||
// Check "If step is zero, ValueError is raised."
|
// Check "If step is zero, ValueError is raised."
|
||||||
let rangenez = ctx
|
let rangenez =
|
||||||
.builder
|
ctx.builder.build_int_compare(IntPredicate::NE, step, int32.const_zero(), "").unwrap();
|
||||||
.build_int_compare(IntPredicate::NE, step, int32.const_zero(), "")
|
|
||||||
.unwrap();
|
|
||||||
ctx.make_assert(
|
ctx.make_assert(
|
||||||
generator,
|
generator,
|
||||||
rangenez,
|
rangenez,
|
||||||
@ -497,10 +358,7 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
.build_conditional_branch(
|
.build_conditional_branch(
|
||||||
gen_in_range_check(
|
gen_in_range_check(
|
||||||
ctx,
|
ctx,
|
||||||
ctx.builder
|
ctx.builder.build_load(i, "").map(BasicValueEnum::into_int_value).unwrap(),
|
||||||
.build_load(i, "")
|
|
||||||
.map(BasicValueEnum::into_int_value)
|
|
||||||
.unwrap(),
|
|
||||||
stop,
|
stop,
|
||||||
step,
|
step,
|
||||||
),
|
),
|
||||||
@ -530,10 +388,7 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
)
|
)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
generator.gen_block(ctx, body.iter())?;
|
generator.gen_block(ctx, body.iter())?;
|
||||||
}
|
} else {
|
||||||
TypeEnum::TObj { obj_id, params: list_params, .. }
|
|
||||||
if *obj_id == ctx.primitives.list.obj_id(&ctx.unifier).unwrap() =>
|
|
||||||
{
|
|
||||||
let index_addr = generator.gen_var_alloc(ctx, size_t.into(), Some("for.index.addr"))?;
|
let index_addr = generator.gen_var_alloc(ctx, size_t.into(), Some("for.index.addr"))?;
|
||||||
ctx.builder.build_store(index_addr, size_t.const_zero()).unwrap();
|
ctx.builder.build_store(index_addr, size_t.const_zero()).unwrap();
|
||||||
let len = ctx
|
let len = ctx
|
||||||
@ -571,14 +426,9 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
.map(BasicValueEnum::into_int_value)
|
.map(BasicValueEnum::into_int_value)
|
||||||
.unwrap();
|
.unwrap();
|
||||||
let val = ctx.build_gep_and_load(arr_ptr, &[index], Some("val"));
|
let val = ctx.build_gep_and_load(arr_ptr, &[index], Some("val"));
|
||||||
let val_ty = iter_type_vars(list_params).next().unwrap().ty;
|
generator.gen_assign(ctx, target, val.into())?;
|
||||||
generator.gen_assign(ctx, target, val.into(), val_ty)?;
|
|
||||||
generator.gen_block(ctx, body.iter())?;
|
generator.gen_block(ctx, body.iter())?;
|
||||||
}
|
}
|
||||||
_ => {
|
|
||||||
panic!("unsupported for loop iterator type: {}", ctx.unifier.stringify(iter_ty));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for (k, (_, _, counter)) in &var_assignment {
|
for (k, (_, _, counter)) in &var_assignment {
|
||||||
let (_, static_val, counter2) = ctx.var_assignment.get_mut(k).unwrap();
|
let (_, static_val, counter2) = ctx.var_assignment.get_mut(k).unwrap();
|
||||||
@ -612,14 +462,12 @@ pub fn gen_for<G: CodeGenerator>(
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(PartialEq, Eq, Debug, Clone, Copy, Hash)]
|
#[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)]
|
||||||
pub struct BreakContinueHooks<'ctx> {
|
pub struct BreakContinueHooks<'ctx> {
|
||||||
/// The [exit block][`BasicBlock`] to branch to when `break`-ing out of a loop.
|
/// [`BasicBlock`] to branch to for `break`-ing out of the loop.
|
||||||
pub exit_bb: BasicBlock<'ctx>,
|
pub break_bb: BasicBlock<'ctx>,
|
||||||
|
/// [`BasicBlock`] to branch to for `continue`-ing to the next iteration in the loop.
|
||||||
/// The [latch basic block][`BasicBlock`] to branch to for `continue`-ing to the next iteration
|
pub continue_bb: BasicBlock<'ctx>,
|
||||||
/// of the loop.
|
|
||||||
pub latch_bb: BasicBlock<'ctx>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Generates a C-style `for` construct using lambdas, similar to the following C code:
|
/// Generates a C-style `for` construct using lambdas, similar to the following C code:
|
||||||
@ -639,7 +487,6 @@ pub struct BreakContinueHooks<'ctx> {
|
|||||||
pub fn gen_for_callback<'ctx, 'a, G, I, InitFn, CondFn, BodyFn, UpdateFn>(
|
pub fn gen_for_callback<'ctx, 'a, G, I, InitFn, CondFn, BodyFn, UpdateFn>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
ctx: &mut CodeGenContext<'ctx, 'a>,
|
ctx: &mut CodeGenContext<'ctx, 'a>,
|
||||||
label: Option<&str>,
|
|
||||||
init: InitFn,
|
init: InitFn,
|
||||||
cond: CondFn,
|
cond: CondFn,
|
||||||
body: BodyFn,
|
body: BodyFn,
|
||||||
@ -650,24 +497,18 @@ where
|
|||||||
I: Clone,
|
I: Clone,
|
||||||
InitFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<I, String>,
|
InitFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<I, String>,
|
||||||
CondFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, I) -> Result<IntValue<'ctx>, String>,
|
CondFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, I) -> Result<IntValue<'ctx>, String>,
|
||||||
BodyFn: FnOnce(
|
BodyFn:
|
||||||
&mut G,
|
FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, BreakContinueHooks, I) -> Result<(), String>,
|
||||||
&mut CodeGenContext<'ctx, 'a>,
|
|
||||||
BreakContinueHooks<'ctx>,
|
|
||||||
I,
|
|
||||||
) -> Result<(), String>,
|
|
||||||
UpdateFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, I) -> Result<(), String>,
|
UpdateFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, I) -> Result<(), String>,
|
||||||
{
|
{
|
||||||
let label = label.unwrap_or("for");
|
|
||||||
|
|
||||||
let current_bb = ctx.builder.get_insert_block().unwrap();
|
let current_bb = ctx.builder.get_insert_block().unwrap();
|
||||||
let init_bb = ctx.ctx.insert_basic_block_after(current_bb, &format!("{label}.init"));
|
let init_bb = ctx.ctx.insert_basic_block_after(current_bb, "for.init");
|
||||||
// The BB containing the loop condition check
|
// The BB containing the loop condition check
|
||||||
let cond_bb = ctx.ctx.insert_basic_block_after(init_bb, &format!("{label}.cond"));
|
let cond_bb = ctx.ctx.insert_basic_block_after(init_bb, "for.cond");
|
||||||
let body_bb = ctx.ctx.insert_basic_block_after(cond_bb, &format!("{label}.body"));
|
let body_bb = ctx.ctx.insert_basic_block_after(cond_bb, "for.body");
|
||||||
// The BB containing the increment expression
|
// The BB containing the increment expression
|
||||||
let update_bb = ctx.ctx.insert_basic_block_after(body_bb, &format!("{label}.update"));
|
let update_bb = ctx.ctx.insert_basic_block_after(body_bb, "for.update");
|
||||||
let cont_bb = ctx.ctx.insert_basic_block_after(update_bb, &format!("{label}.end"));
|
let cont_bb = ctx.ctx.insert_basic_block_after(update_bb, "for.end");
|
||||||
|
|
||||||
// store loop bb information and restore it later
|
// store loop bb information and restore it later
|
||||||
let loop_bb = ctx.loop_target.replace((update_bb, cont_bb));
|
let loop_bb = ctx.loop_target.replace((update_bb, cont_bb));
|
||||||
@ -688,7 +529,7 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
ctx.builder.position_at_end(body_bb);
|
ctx.builder.position_at_end(body_bb);
|
||||||
let hooks = BreakContinueHooks { exit_bb: cont_bb, latch_bb: update_bb };
|
let hooks = BreakContinueHooks { break_bb: cont_bb, continue_bb: update_bb };
|
||||||
body(generator, ctx, hooks, loop_var.clone())?;
|
body(generator, ctx, hooks, loop_var.clone())?;
|
||||||
if !ctx.is_terminated() {
|
if !ctx.is_terminated() {
|
||||||
ctx.builder.build_unconditional_branch(update_bb).unwrap();
|
ctx.builder.build_unconditional_branch(update_bb).unwrap();
|
||||||
@ -724,7 +565,6 @@ where
|
|||||||
pub fn gen_for_callback_incrementing<'ctx, 'a, G, BodyFn>(
|
pub fn gen_for_callback_incrementing<'ctx, 'a, G, BodyFn>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
ctx: &mut CodeGenContext<'ctx, 'a>,
|
ctx: &mut CodeGenContext<'ctx, 'a>,
|
||||||
label: Option<&str>,
|
|
||||||
init_val: IntValue<'ctx>,
|
init_val: IntValue<'ctx>,
|
||||||
max_val: (IntValue<'ctx>, bool),
|
max_val: (IntValue<'ctx>, bool),
|
||||||
body: BodyFn,
|
body: BodyFn,
|
||||||
@ -735,7 +575,7 @@ where
|
|||||||
BodyFn: FnOnce(
|
BodyFn: FnOnce(
|
||||||
&mut G,
|
&mut G,
|
||||||
&mut CodeGenContext<'ctx, 'a>,
|
&mut CodeGenContext<'ctx, 'a>,
|
||||||
BreakContinueHooks<'ctx>,
|
BreakContinueHooks,
|
||||||
IntValue<'ctx>,
|
IntValue<'ctx>,
|
||||||
) -> Result<(), String>,
|
) -> Result<(), String>,
|
||||||
{
|
{
|
||||||
@ -744,7 +584,6 @@ where
|
|||||||
gen_for_callback(
|
gen_for_callback(
|
||||||
generator,
|
generator,
|
||||||
ctx,
|
ctx,
|
||||||
label,
|
|
||||||
|generator, ctx| {
|
|generator, ctx| {
|
||||||
let i_addr = generator.gen_var_alloc(ctx, init_val_t.into(), None)?;
|
let i_addr = generator.gen_var_alloc(ctx, init_val_t.into(), None)?;
|
||||||
ctx.builder.build_store(i_addr, init_val).unwrap();
|
ctx.builder.build_store(i_addr, init_val).unwrap();
|
||||||
@ -796,11 +635,9 @@ where
|
|||||||
/// - `step_fn`: A lambda of IR statements that retrieves the `step` value of the `range`-like
|
/// - `step_fn`: A lambda of IR statements that retrieves the `step` value of the `range`-like
|
||||||
/// iterable. This value will be extended to the size of `start`.
|
/// iterable. This value will be extended to the size of `start`.
|
||||||
/// - `body_fn`: A lambda of IR statements within the loop body.
|
/// - `body_fn`: A lambda of IR statements within the loop body.
|
||||||
#[allow(clippy::too_many_arguments)]
|
|
||||||
pub fn gen_for_range_callback<'ctx, 'a, G, StartFn, StopFn, StepFn, BodyFn>(
|
pub fn gen_for_range_callback<'ctx, 'a, G, StartFn, StopFn, StepFn, BodyFn>(
|
||||||
generator: &mut G,
|
generator: &mut G,
|
||||||
ctx: &mut CodeGenContext<'ctx, 'a>,
|
ctx: &mut CodeGenContext<'ctx, 'a>,
|
||||||
label: Option<&str>,
|
|
||||||
is_unsigned: bool,
|
is_unsigned: bool,
|
||||||
start_fn: StartFn,
|
start_fn: StartFn,
|
||||||
(stop_fn, stop_inclusive): (StopFn, bool),
|
(stop_fn, stop_inclusive): (StopFn, bool),
|
||||||
@ -812,19 +649,13 @@ where
|
|||||||
StartFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
StartFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
||||||
StopFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
StopFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
||||||
StepFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
StepFn: Fn(&mut G, &mut CodeGenContext<'ctx, 'a>) -> Result<IntValue<'ctx>, String>,
|
||||||
BodyFn: FnOnce(
|
BodyFn: FnOnce(&mut G, &mut CodeGenContext<'ctx, 'a>, IntValue<'ctx>) -> Result<(), String>,
|
||||||
&mut G,
|
|
||||||
&mut CodeGenContext<'ctx, 'a>,
|
|
||||||
BreakContinueHooks<'ctx>,
|
|
||||||
IntValue<'ctx>,
|
|
||||||
) -> Result<(), String>,
|
|
||||||
{
|
{
|
||||||
let init_val_t = start_fn(generator, ctx).map(IntValue::get_type).unwrap();
|
let init_val_t = start_fn(generator, ctx).map(IntValue::get_type).unwrap();
|
||||||
|
|
||||||
gen_for_callback(
|
gen_for_callback(
|
||||||
generator,
|
generator,
|
||||||
ctx,
|
ctx,
|
||||||
label,
|
|
||||||
|generator, ctx| {
|
|generator, ctx| {
|
||||||
let i_addr = generator.gen_var_alloc(ctx, init_val_t.into(), None)?;
|
let i_addr = generator.gen_var_alloc(ctx, init_val_t.into(), None)?;
|
||||||
|
|
||||||
@ -882,10 +713,10 @@ where
|
|||||||
|
|
||||||
Ok(cond)
|
Ok(cond)
|
||||||
},
|
},
|
||||||
|generator, ctx, hooks, (i_addr, _)| {
|
|generator, ctx, _, (i_addr, _)| {
|
||||||
let i = ctx.builder.build_load(i_addr, "").map(BasicValueEnum::into_int_value).unwrap();
|
let i = ctx.builder.build_load(i_addr, "").map(BasicValueEnum::into_int_value).unwrap();
|
||||||
|
|
||||||
body_fn(generator, ctx, hooks, i)
|
body_fn(generator, ctx, i)
|
||||||
},
|
},
|
||||||
|generator, ctx, (i_addr, _)| {
|
|generator, ctx, (i_addr, _)| {
|
||||||
let i = ctx.builder.build_load(i_addr, "").map(BasicValueEnum::into_int_value).unwrap();
|
let i = ctx.builder.build_load(i_addr, "").map(BasicValueEnum::into_int_value).unwrap();
|
||||||
@ -913,7 +744,7 @@ pub fn gen_while<G: CodeGenerator>(
|
|||||||
ctx: &mut CodeGenContext<'_, '_>,
|
ctx: &mut CodeGenContext<'_, '_>,
|
||||||
stmt: &Stmt<Option<Type>>,
|
stmt: &Stmt<Option<Type>>,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
let StmtKind::While { test, body, orelse, .. } = &stmt.node else { codegen_unreachable!(ctx) };
|
let StmtKind::While { test, body, orelse, .. } = &stmt.node else { unreachable!() };
|
||||||
|
|
||||||
// var_assignment static values may be changed in another branch
|
// var_assignment static values may be changed in another branch
|
||||||
// if so, remove the static value as it may not be correct in this branch
|
// if so, remove the static value as it may not be correct in this branch
|
||||||
@ -943,7 +774,7 @@ pub fn gen_while<G: CodeGenerator>(
|
|||||||
|
|
||||||
return Ok(());
|
return Ok(());
|
||||||
};
|
};
|
||||||
let BasicValueEnum::IntValue(test) = test else { codegen_unreachable!(ctx) };
|
let BasicValueEnum::IntValue(test) = test else { unreachable!() };
|
||||||
|
|
||||||
ctx.builder
|
ctx.builder
|
||||||
.build_conditional_branch(generator.bool_to_i1(ctx, test), body_bb, orelse_bb)
|
.build_conditional_branch(generator.bool_to_i1(ctx, test), body_bb, orelse_bb)
|
||||||
@ -1091,7 +922,7 @@ pub fn gen_if<G: CodeGenerator>(
|
|||||||
ctx: &mut CodeGenContext<'_, '_>,
|
ctx: &mut CodeGenContext<'_, '_>,
|
||||||
stmt: &Stmt<Option<Type>>,
|
stmt: &Stmt<Option<Type>>,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
let StmtKind::If { test, body, orelse, .. } = &stmt.node else { codegen_unreachable!(ctx) };
|
let StmtKind::If { test, body, orelse, .. } = &stmt.node else { unreachable!() };
|
||||||
|
|
||||||
// var_assignment static values may be changed in another branch
|
// var_assignment static values may be changed in another branch
|
||||||
// if so, remove the static value as it may not be correct in this branch
|
// if so, remove the static value as it may not be correct in this branch
|
||||||
@ -1214,11 +1045,11 @@ pub fn exn_constructor<'ctx>(
|
|||||||
let zelf_id = if let TypeEnum::TObj { obj_id, .. } = &*ctx.unifier.get_ty(zelf_ty) {
|
let zelf_id = if let TypeEnum::TObj { obj_id, .. } = &*ctx.unifier.get_ty(zelf_ty) {
|
||||||
obj_id.0
|
obj_id.0
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
let defs = ctx.top_level.definitions.read();
|
let defs = ctx.top_level.definitions.read();
|
||||||
let def = defs[zelf_id].read();
|
let def = defs[zelf_id].read();
|
||||||
let TopLevelDef::Class { name: zelf_name, .. } = &*def else { codegen_unreachable!(ctx) };
|
let TopLevelDef::Class { name: zelf_name, .. } = &*def else { unreachable!() };
|
||||||
let exception_name = format!("{}:{}", ctx.resolver.get_exception_id(zelf_id), zelf_name);
|
let exception_name = format!("{}:{}", ctx.resolver.get_exception_id(zelf_id), zelf_name);
|
||||||
unsafe {
|
unsafe {
|
||||||
let id_ptr = ctx.builder.build_in_bounds_gep(zelf, &[zero, zero], "exn.id").unwrap();
|
let id_ptr = ctx.builder.build_in_bounds_gep(zelf, &[zero, zero], "exn.id").unwrap();
|
||||||
@ -1326,7 +1157,7 @@ pub fn gen_try<'ctx, 'a, G: CodeGenerator>(
|
|||||||
target: &Stmt<Option<Type>>,
|
target: &Stmt<Option<Type>>,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
let StmtKind::Try { body, handlers, orelse, finalbody, .. } = &target.node else {
|
let StmtKind::Try { body, handlers, orelse, finalbody, .. } = &target.node else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
|
|
||||||
// if we need to generate anything related to exception, we must have personality defined
|
// if we need to generate anything related to exception, we must have personality defined
|
||||||
@ -1403,7 +1234,7 @@ pub fn gen_try<'ctx, 'a, G: CodeGenerator>(
|
|||||||
if let TypeEnum::TObj { obj_id, .. } = &*ctx.unifier.get_ty(type_.custom.unwrap()) {
|
if let TypeEnum::TObj { obj_id, .. } = &*ctx.unifier.get_ty(type_.custom.unwrap()) {
|
||||||
*obj_id
|
*obj_id
|
||||||
} else {
|
} else {
|
||||||
codegen_unreachable!(ctx)
|
unreachable!()
|
||||||
};
|
};
|
||||||
let exception_name = format!("{}:{}", ctx.resolver.get_exception_id(obj_id.0), exn_name);
|
let exception_name = format!("{}:{}", ctx.resolver.get_exception_id(obj_id.0), exn_name);
|
||||||
let exn_id = ctx.resolver.get_string_id(&exception_name);
|
let exn_id = ctx.resolver.get_string_id(&exception_name);
|
||||||
@ -1675,23 +1506,6 @@ pub fn gen_return<G: CodeGenerator>(
|
|||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
};
|
};
|
||||||
|
|
||||||
// Remap boolean return type into i1
|
|
||||||
let value = value.map(|ret_val| {
|
|
||||||
// The "return type" of a sret function is in the first parameter
|
|
||||||
let expected_ty = if ctx.need_sret {
|
|
||||||
func.get_type().get_param_types()[0]
|
|
||||||
} else {
|
|
||||||
func.get_type().get_return_type().unwrap()
|
|
||||||
};
|
|
||||||
|
|
||||||
if matches!(expected_ty, BasicTypeEnum::IntType(ty) if ty.get_bit_width() == 1) {
|
|
||||||
generator.bool_to_i1(ctx, ret_val.into_int_value()).into()
|
|
||||||
} else {
|
|
||||||
ret_val
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
if let Some(return_target) = ctx.return_target {
|
if let Some(return_target) = ctx.return_target {
|
||||||
if let Some(value) = value {
|
if let Some(value) = value {
|
||||||
ctx.builder.build_store(ctx.return_buffer.unwrap(), value).unwrap();
|
ctx.builder.build_store(ctx.return_buffer.unwrap(), value).unwrap();
|
||||||
@ -1702,6 +1516,25 @@ pub fn gen_return<G: CodeGenerator>(
|
|||||||
ctx.builder.build_store(ctx.return_buffer.unwrap(), value.unwrap()).unwrap();
|
ctx.builder.build_store(ctx.return_buffer.unwrap(), value.unwrap()).unwrap();
|
||||||
ctx.builder.build_return(None).unwrap();
|
ctx.builder.build_return(None).unwrap();
|
||||||
} else {
|
} else {
|
||||||
|
// Remap boolean return type into i1
|
||||||
|
let value = value.map(|v| {
|
||||||
|
let expected_ty = func.get_type().get_return_type().unwrap();
|
||||||
|
let ret_val = v.as_basic_value_enum();
|
||||||
|
|
||||||
|
if expected_ty.is_int_type() && ret_val.is_int_value() {
|
||||||
|
let ret_type = expected_ty.into_int_type();
|
||||||
|
let ret_val = ret_val.into_int_value();
|
||||||
|
|
||||||
|
if ret_type.get_bit_width() == 1 && ret_val.get_type().get_bit_width() != 1 {
|
||||||
|
generator.bool_to_i1(ctx, ret_val)
|
||||||
|
} else {
|
||||||
|
ret_val
|
||||||
|
}
|
||||||
|
.into()
|
||||||
|
} else {
|
||||||
|
ret_val
|
||||||
|
}
|
||||||
|
});
|
||||||
let value = value.as_ref().map(|v| v as &dyn BasicValue);
|
let value = value.as_ref().map(|v| v as &dyn BasicValue);
|
||||||
ctx.builder.build_return(value).unwrap();
|
ctx.builder.build_return(value).unwrap();
|
||||||
}
|
}
|
||||||
@ -1735,14 +1568,14 @@ pub fn gen_stmt<G: CodeGenerator>(
|
|||||||
}
|
}
|
||||||
StmtKind::AnnAssign { target, value, .. } => {
|
StmtKind::AnnAssign { target, value, .. } => {
|
||||||
if let Some(value) = value {
|
if let Some(value) = value {
|
||||||
let Some(value_enum) = generator.gen_expr(ctx, value)? else { return Ok(()) };
|
let Some(value) = generator.gen_expr(ctx, value)? else { return Ok(()) };
|
||||||
generator.gen_assign(ctx, target, value_enum, value.custom.unwrap())?;
|
generator.gen_assign(ctx, target, value)?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
StmtKind::Assign { targets, value, .. } => {
|
StmtKind::Assign { targets, value, .. } => {
|
||||||
let Some(value_enum) = generator.gen_expr(ctx, value)? else { return Ok(()) };
|
let Some(value) = generator.gen_expr(ctx, value)? else { return Ok(()) };
|
||||||
for target in targets {
|
for target in targets {
|
||||||
generator.gen_assign(ctx, target, value_enum.clone(), value.custom.unwrap())?;
|
generator.gen_assign(ctx, target, value.clone())?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
StmtKind::Continue { .. } => {
|
StmtKind::Continue { .. } => {
|
||||||
@ -1756,44 +1589,13 @@ pub fn gen_stmt<G: CodeGenerator>(
|
|||||||
StmtKind::For { .. } => generator.gen_for(ctx, stmt)?,
|
StmtKind::For { .. } => generator.gen_for(ctx, stmt)?,
|
||||||
StmtKind::With { .. } => generator.gen_with(ctx, stmt)?,
|
StmtKind::With { .. } => generator.gen_with(ctx, stmt)?,
|
||||||
StmtKind::AugAssign { target, op, value, .. } => {
|
StmtKind::AugAssign { target, op, value, .. } => {
|
||||||
let value_enum = gen_binop_expr(
|
let value = gen_binop_expr(generator, ctx, target, *op, value, stmt.location, true)?;
|
||||||
generator,
|
generator.gen_assign(ctx, target, value.unwrap())?;
|
||||||
ctx,
|
|
||||||
target,
|
|
||||||
Binop::aug_assign(*op),
|
|
||||||
value,
|
|
||||||
stmt.location,
|
|
||||||
)?
|
|
||||||
.unwrap();
|
|
||||||
generator.gen_assign(ctx, target, value_enum, value.custom.unwrap())?;
|
|
||||||
}
|
}
|
||||||
StmtKind::Try { .. } => gen_try(generator, ctx, stmt)?,
|
StmtKind::Try { .. } => gen_try(generator, ctx, stmt)?,
|
||||||
StmtKind::Raise { exc, .. } => {
|
StmtKind::Raise { exc, .. } => {
|
||||||
if let Some(exc) = exc {
|
if let Some(exc) = exc {
|
||||||
let exn = if let ExprKind::Name { id, .. } = &exc.node {
|
let exc = if let Some(v) = generator.gen_expr(ctx, exc)? {
|
||||||
// Handle "raise Exception" short form
|
|
||||||
let def_id = ctx.resolver.get_identifier_def(*id).map_err(|e| {
|
|
||||||
format!("{} (at {})", e.iter().next().unwrap(), exc.location)
|
|
||||||
})?;
|
|
||||||
let def = ctx.top_level.definitions.read();
|
|
||||||
let TopLevelDef::Class { constructor, .. } = *def[def_id.0].read() else {
|
|
||||||
return Err(format!("Failed to resolve symbol {id} (at {})", exc.location));
|
|
||||||
};
|
|
||||||
|
|
||||||
let TypeEnum::TFunc(signature) =
|
|
||||||
ctx.unifier.get_ty(constructor.unwrap()).as_ref().clone()
|
|
||||||
else {
|
|
||||||
return Err(format!("Failed to resolve symbol {id} (at {})", exc.location));
|
|
||||||
};
|
|
||||||
|
|
||||||
generator
|
|
||||||
.gen_call(ctx, None, (&signature, def_id), Vec::default())?
|
|
||||||
.map(Into::into)
|
|
||||||
} else {
|
|
||||||
generator.gen_expr(ctx, exc)?
|
|
||||||
};
|
|
||||||
|
|
||||||
let exc = if let Some(v) = exn {
|
|
||||||
v.to_basic_value_enum(ctx, generator, exc.custom.unwrap())?
|
v.to_basic_value_enum(ctx, generator, exc.custom.unwrap())?
|
||||||
} else {
|
} else {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
@ -1817,48 +1619,17 @@ pub fn gen_stmt<G: CodeGenerator>(
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
None => ctx.gen_string(generator, "").into(),
|
None => ctx.gen_string(generator, ""),
|
||||||
};
|
};
|
||||||
ctx.make_assert_impl(
|
ctx.make_assert_impl(
|
||||||
generator,
|
generator,
|
||||||
generator.bool_to_i1(ctx, test.into_int_value()),
|
test.into_int_value(),
|
||||||
"0:AssertionError",
|
"0:AssertionError",
|
||||||
err_msg,
|
err_msg,
|
||||||
[None, None, None],
|
[None, None, None],
|
||||||
stmt.location,
|
stmt.location,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
StmtKind::Global { names, .. } => {
|
|
||||||
let registered_globals = ctx
|
|
||||||
.top_level
|
|
||||||
.definitions
|
|
||||||
.read()
|
|
||||||
.iter()
|
|
||||||
.filter_map(|def| {
|
|
||||||
if let TopLevelDef::Variable { simple_name, ty, .. } = &*def.read() {
|
|
||||||
Some((*simple_name, *ty))
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
}
|
|
||||||
})
|
|
||||||
.collect_vec();
|
|
||||||
|
|
||||||
for id in names {
|
|
||||||
let Some((_, ty)) = registered_globals.iter().find(|(name, _)| name == id) else {
|
|
||||||
return Err(format!("{id} is not a global at {}", stmt.location));
|
|
||||||
};
|
|
||||||
|
|
||||||
let resolver = ctx.resolver.clone();
|
|
||||||
let ptr = resolver
|
|
||||||
.get_symbol_value(*id, ctx, generator)
|
|
||||||
.map(|val| val.to_basic_value_enum(ctx, generator, *ty))
|
|
||||||
.transpose()?
|
|
||||||
.map(BasicValueEnum::into_pointer_value)
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
ctx.var_assignment.insert(*id, (ptr, None, 0));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => unimplemented!(),
|
_ => unimplemented!(),
|
||||||
};
|
};
|
||||||
Ok(())
|
Ok(())
|
||||||
|
@ -1,20 +1,3 @@
|
|||||||
use std::{
|
|
||||||
collections::{HashMap, HashSet},
|
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
use indexmap::IndexMap;
|
|
||||||
use indoc::indoc;
|
|
||||||
use inkwell::{
|
|
||||||
targets::{InitializationConfig, Target},
|
|
||||||
OptimizationLevel,
|
|
||||||
};
|
|
||||||
use nac3parser::{
|
|
||||||
ast::{fold::Fold, FileName, StrRef},
|
|
||||||
parser::parse_program,
|
|
||||||
};
|
|
||||||
use parking_lot::RwLock;
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{
|
codegen::{
|
||||||
classes::{ListType, NDArrayType, ProxyType, RangeType},
|
classes::{ListType, NDArrayType, ProxyType, RangeType},
|
||||||
@ -28,10 +11,24 @@ use crate::{
|
|||||||
DefinitionId, FunInstance, TopLevelContext, TopLevelDef,
|
DefinitionId, FunInstance, TopLevelContext, TopLevelDef,
|
||||||
},
|
},
|
||||||
typecheck::{
|
typecheck::{
|
||||||
type_inferencer::{FunctionData, IdentifierInfo, Inferencer, PrimitiveStore},
|
type_inferencer::{FunctionData, Inferencer, PrimitiveStore},
|
||||||
typedef::{FunSignature, FuncArg, Type, TypeEnum, Unifier, VarMap},
|
typedef::{FunSignature, FuncArg, Type, TypeEnum, Unifier, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use indexmap::IndexMap;
|
||||||
|
use indoc::indoc;
|
||||||
|
use inkwell::{
|
||||||
|
targets::{InitializationConfig, Target},
|
||||||
|
OptimizationLevel,
|
||||||
|
};
|
||||||
|
use nac3parser::ast::FileName;
|
||||||
|
use nac3parser::{
|
||||||
|
ast::{fold::Fold, StrRef},
|
||||||
|
parser::parse_program,
|
||||||
|
};
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
use std::collections::{HashMap, HashSet};
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
struct Resolver {
|
struct Resolver {
|
||||||
id_to_type: HashMap<StrRef, Type>,
|
id_to_type: HashMap<StrRef, Type>,
|
||||||
@ -67,7 +64,6 @@ impl SymbolResolver for Resolver {
|
|||||||
&self,
|
&self,
|
||||||
_: StrRef,
|
_: StrRef,
|
||||||
_: &mut CodeGenContext<'ctx, '_>,
|
_: &mut CodeGenContext<'ctx, '_>,
|
||||||
_: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>> {
|
) -> Option<ValueEnum<'ctx>> {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
@ -98,7 +94,7 @@ fn test_primitives() {
|
|||||||
"};
|
"};
|
||||||
let statements = parse_program(source, FileName::default()).unwrap();
|
let statements = parse_program(source, FileName::default()).unwrap();
|
||||||
|
|
||||||
let composer = TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 32).0;
|
let composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 32).0;
|
||||||
let mut unifier = composer.unifier.clone();
|
let mut unifier = composer.unifier.clone();
|
||||||
let primitives = composer.primitives_ty;
|
let primitives = composer.primitives_ty;
|
||||||
let top_level = Arc::new(composer.make_top_level_context());
|
let top_level = Arc::new(composer.make_top_level_context());
|
||||||
@ -113,18 +109,8 @@ fn test_primitives() {
|
|||||||
let threads = vec![DefaultCodeGenerator::new("test".into(), 32).into()];
|
let threads = vec![DefaultCodeGenerator::new("test".into(), 32).into()];
|
||||||
let signature = FunSignature {
|
let signature = FunSignature {
|
||||||
args: vec![
|
args: vec![
|
||||||
FuncArg {
|
FuncArg { name: "a".into(), ty: primitives.int32, default_value: None },
|
||||||
name: "a".into(),
|
FuncArg { name: "b".into(), ty: primitives.int32, default_value: None },
|
||||||
ty: primitives.int32,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
},
|
|
||||||
FuncArg {
|
|
||||||
name: "b".into(),
|
|
||||||
ty: primitives.int32,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
},
|
|
||||||
],
|
],
|
||||||
ret: primitives.int32,
|
ret: primitives.int32,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
@ -142,8 +128,7 @@ fn test_primitives() {
|
|||||||
};
|
};
|
||||||
let mut virtual_checks = Vec::new();
|
let mut virtual_checks = Vec::new();
|
||||||
let mut calls = HashMap::new();
|
let mut calls = HashMap::new();
|
||||||
let mut identifiers: HashMap<_, _> =
|
let mut identifiers: HashSet<_> = ["a".into(), "b".into()].into();
|
||||||
["a".into(), "b".into()].map(|id| (id, IdentifierInfo::default())).into();
|
|
||||||
let mut inferencer = Inferencer {
|
let mut inferencer = Inferencer {
|
||||||
top_level: &top_level,
|
top_level: &top_level,
|
||||||
function_data: &mut function_data,
|
function_data: &mut function_data,
|
||||||
@ -204,8 +189,6 @@ fn test_primitives() {
|
|||||||
let expected = indoc! {"
|
let expected = indoc! {"
|
||||||
; ModuleID = 'test'
|
; ModuleID = 'test'
|
||||||
source_filename = \"test\"
|
source_filename = \"test\"
|
||||||
target datalayout = \"e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-f80:128-n8:16:32:64-S128\"
|
|
||||||
target triple = \"x86_64-unknown-linux-gnu\"
|
|
||||||
|
|
||||||
; Function Attrs: mustprogress nofree norecurse nosync nounwind readnone willreturn
|
; Function Attrs: mustprogress nofree norecurse nosync nounwind readnone willreturn
|
||||||
define i32 @testing(i32 %0, i32 %1) local_unnamed_addr #0 !dbg !4 {
|
define i32 @testing(i32 %0, i32 %1) local_unnamed_addr #0 !dbg !4 {
|
||||||
@ -263,19 +246,14 @@ fn test_simple_call() {
|
|||||||
"};
|
"};
|
||||||
let statements_2 = parse_program(source_2, FileName::default()).unwrap();
|
let statements_2 = parse_program(source_2, FileName::default()).unwrap();
|
||||||
|
|
||||||
let composer = TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 32).0;
|
let composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 32).0;
|
||||||
let mut unifier = composer.unifier.clone();
|
let mut unifier = composer.unifier.clone();
|
||||||
let primitives = composer.primitives_ty;
|
let primitives = composer.primitives_ty;
|
||||||
let top_level = Arc::new(composer.make_top_level_context());
|
let top_level = Arc::new(composer.make_top_level_context());
|
||||||
unifier.top_level = Some(top_level.clone());
|
unifier.top_level = Some(top_level.clone());
|
||||||
|
|
||||||
let signature = FunSignature {
|
let signature = FunSignature {
|
||||||
args: vec![FuncArg {
|
args: vec![FuncArg { name: "a".into(), ty: primitives.int32, default_value: None }],
|
||||||
name: "a".into(),
|
|
||||||
ty: primitives.int32,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
}],
|
|
||||||
ret: primitives.int32,
|
ret: primitives.int32,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
};
|
};
|
||||||
@ -322,8 +300,7 @@ fn test_simple_call() {
|
|||||||
};
|
};
|
||||||
let mut virtual_checks = Vec::new();
|
let mut virtual_checks = Vec::new();
|
||||||
let mut calls = HashMap::new();
|
let mut calls = HashMap::new();
|
||||||
let mut identifiers: HashMap<_, _> =
|
let mut identifiers: HashSet<_> = ["a".into(), "foo".into()].into();
|
||||||
["a".into(), "foo".into()].map(|id| (id, IdentifierInfo::default())).into();
|
|
||||||
let mut inferencer = Inferencer {
|
let mut inferencer = Inferencer {
|
||||||
top_level: &top_level,
|
top_level: &top_level,
|
||||||
function_data: &mut function_data,
|
function_data: &mut function_data,
|
||||||
@ -391,8 +368,6 @@ fn test_simple_call() {
|
|||||||
let expected = indoc! {"
|
let expected = indoc! {"
|
||||||
; ModuleID = 'test'
|
; ModuleID = 'test'
|
||||||
source_filename = \"test\"
|
source_filename = \"test\"
|
||||||
target datalayout = \"e-m:e-p270:32:32-p271:32:32-p272:64:64-i64:64-f80:128-n8:16:32:64-S128\"
|
|
||||||
target triple = \"x86_64-unknown-linux-gnu\"
|
|
||||||
|
|
||||||
; Function Attrs: mustprogress nofree norecurse nosync nounwind readnone willreturn
|
; Function Attrs: mustprogress nofree norecurse nosync nounwind readnone willreturn
|
||||||
define i32 @testing(i32 %0) local_unnamed_addr #0 !dbg !5 {
|
define i32 @testing(i32 %0) local_unnamed_addr #0 !dbg !5 {
|
||||||
|
@ -19,10 +19,6 @@
|
|||||||
clippy::wildcard_imports
|
clippy::wildcard_imports
|
||||||
)]
|
)]
|
||||||
|
|
||||||
// users of nac3core need to use the same version of these dependencies, so expose them as nac3core::*
|
|
||||||
pub use inkwell;
|
|
||||||
pub use nac3parser;
|
|
||||||
|
|
||||||
pub mod codegen;
|
pub mod codegen;
|
||||||
pub mod symbol_resolver;
|
pub mod symbol_resolver;
|
||||||
pub mod toplevel;
|
pub mod toplevel;
|
||||||
|
@ -1,15 +1,7 @@
|
|||||||
use std::{
|
use std::fmt::Debug;
|
||||||
collections::{HashMap, HashSet},
|
use std::rc::Rc;
|
||||||
fmt::{Debug, Display},
|
use std::sync::Arc;
|
||||||
rc::Rc,
|
use std::{collections::HashMap, collections::HashSet, fmt::Display};
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
use inkwell::values::{BasicValueEnum, FloatValue, IntValue, PointerValue, StructValue};
|
|
||||||
use itertools::{chain, izip, Itertools};
|
|
||||||
use parking_lot::RwLock;
|
|
||||||
|
|
||||||
use nac3parser::ast::{Constant, Expr, Location, StrRef};
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{CodeGenContext, CodeGenerator},
|
codegen::{CodeGenContext, CodeGenerator},
|
||||||
@ -19,6 +11,10 @@ use crate::{
|
|||||||
typedef::{Type, TypeEnum, Unifier, VarMap},
|
typedef::{Type, TypeEnum, Unifier, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use inkwell::values::{BasicValueEnum, FloatValue, IntValue, PointerValue, StructValue};
|
||||||
|
use itertools::{chain, izip, Itertools};
|
||||||
|
use nac3parser::ast::{Constant, Expr, Location, StrRef};
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Debug)]
|
#[derive(Clone, PartialEq, Debug)]
|
||||||
pub enum SymbolValue {
|
pub enum SymbolValue {
|
||||||
@ -82,14 +78,14 @@ impl SymbolValue {
|
|||||||
}
|
}
|
||||||
Constant::Tuple(t) => {
|
Constant::Tuple(t) => {
|
||||||
let expected_ty = unifier.get_ty(expected_ty);
|
let expected_ty = unifier.get_ty(expected_ty);
|
||||||
let TypeEnum::TTuple { ty, is_vararg_ctx } = expected_ty.as_ref() else {
|
let TypeEnum::TTuple { ty } = expected_ty.as_ref() else {
|
||||||
return Err(format!(
|
return Err(format!(
|
||||||
"Expected {:?}, but got Tuple",
|
"Expected {:?}, but got Tuple",
|
||||||
expected_ty.get_type_name()
|
expected_ty.get_type_name()
|
||||||
));
|
));
|
||||||
};
|
};
|
||||||
|
|
||||||
assert!(*is_vararg_ctx || ty.len() == t.len());
|
assert_eq!(ty.len(), t.len());
|
||||||
|
|
||||||
let elems = t
|
let elems = t
|
||||||
.iter()
|
.iter()
|
||||||
@ -159,7 +155,7 @@ impl SymbolValue {
|
|||||||
SymbolValue::Bool(_) => primitives.bool,
|
SymbolValue::Bool(_) => primitives.bool,
|
||||||
SymbolValue::Tuple(vs) => {
|
SymbolValue::Tuple(vs) => {
|
||||||
let vs_tys = vs.iter().map(|v| v.get_type(primitives, unifier)).collect::<Vec<_>>();
|
let vs_tys = vs.iter().map(|v| v.get_type(primitives, unifier)).collect::<Vec<_>>();
|
||||||
unifier.add_ty(TypeEnum::TTuple { ty: vs_tys, is_vararg_ctx: false })
|
unifier.add_ty(TypeEnum::TTuple { ty: vs_tys })
|
||||||
}
|
}
|
||||||
SymbolValue::OptionSome(_) | SymbolValue::OptionNone => primitives.option,
|
SymbolValue::OptionSome(_) | SymbolValue::OptionNone => primitives.option,
|
||||||
}
|
}
|
||||||
@ -369,7 +365,6 @@ pub trait SymbolResolver {
|
|||||||
&self,
|
&self,
|
||||||
str: StrRef,
|
str: StrRef,
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
ctx: &mut CodeGenContext<'ctx, '_>,
|
||||||
generator: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>>;
|
) -> Option<ValueEnum<'ctx>>;
|
||||||
|
|
||||||
fn get_default_param_value(&self, expr: &Expr) -> Option<SymbolValue>;
|
fn get_default_param_value(&self, expr: &Expr) -> Option<SymbolValue>;
|
||||||
@ -387,12 +382,13 @@ pub trait SymbolResolver {
|
|||||||
}
|
}
|
||||||
|
|
||||||
thread_local! {
|
thread_local! {
|
||||||
static IDENTIFIER_ID: [StrRef; 11] = [
|
static IDENTIFIER_ID: [StrRef; 12] = [
|
||||||
"int32".into(),
|
"int32".into(),
|
||||||
"int64".into(),
|
"int64".into(),
|
||||||
"float".into(),
|
"float".into(),
|
||||||
"bool".into(),
|
"bool".into(),
|
||||||
"virtual".into(),
|
"virtual".into(),
|
||||||
|
"list".into(),
|
||||||
"tuple".into(),
|
"tuple".into(),
|
||||||
"str".into(),
|
"str".into(),
|
||||||
"Exception".into(),
|
"Exception".into(),
|
||||||
@ -417,12 +413,13 @@ pub fn parse_type_annotation<T>(
|
|||||||
let float_id = ids[2];
|
let float_id = ids[2];
|
||||||
let bool_id = ids[3];
|
let bool_id = ids[3];
|
||||||
let virtual_id = ids[4];
|
let virtual_id = ids[4];
|
||||||
let tuple_id = ids[5];
|
let list_id = ids[5];
|
||||||
let str_id = ids[6];
|
let tuple_id = ids[6];
|
||||||
let exn_id = ids[7];
|
let str_id = ids[7];
|
||||||
let uint32_id = ids[8];
|
let exn_id = ids[8];
|
||||||
let uint64_id = ids[9];
|
let uint32_id = ids[9];
|
||||||
let literal_id = ids[10];
|
let uint64_id = ids[10];
|
||||||
|
let literal_id = ids[11];
|
||||||
|
|
||||||
let name_handling = |id: &StrRef, loc: Location, unifier: &mut Unifier| {
|
let name_handling = |id: &StrRef, loc: Location, unifier: &mut Unifier| {
|
||||||
if *id == int32_id {
|
if *id == int32_id {
|
||||||
@ -479,6 +476,9 @@ pub fn parse_type_annotation<T>(
|
|||||||
if *id == virtual_id {
|
if *id == virtual_id {
|
||||||
let ty = parse_type_annotation(resolver, top_level_defs, unifier, primitives, slice)?;
|
let ty = parse_type_annotation(resolver, top_level_defs, unifier, primitives, slice)?;
|
||||||
Ok(unifier.add_ty(TypeEnum::TVirtual { ty }))
|
Ok(unifier.add_ty(TypeEnum::TVirtual { ty }))
|
||||||
|
} else if *id == list_id {
|
||||||
|
let ty = parse_type_annotation(resolver, top_level_defs, unifier, primitives, slice)?;
|
||||||
|
Ok(unifier.add_ty(TypeEnum::TList { ty }))
|
||||||
} else if *id == tuple_id {
|
} else if *id == tuple_id {
|
||||||
if let Tuple { elts, .. } = &slice.node {
|
if let Tuple { elts, .. } = &slice.node {
|
||||||
let ty = elts
|
let ty = elts
|
||||||
@ -487,7 +487,7 @@ pub fn parse_type_annotation<T>(
|
|||||||
parse_type_annotation(resolver, top_level_defs, unifier, primitives, elt)
|
parse_type_annotation(resolver, top_level_defs, unifier, primitives, elt)
|
||||||
})
|
})
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
.collect::<Result<Vec<_>, _>>()?;
|
||||||
Ok(unifier.add_ty(TypeEnum::TTuple { ty, is_vararg_ctx: false }))
|
Ok(unifier.add_ty(TypeEnum::TTuple { ty }))
|
||||||
} else {
|
} else {
|
||||||
Err(HashSet::from(["Expected multiple elements for tuple".into()]))
|
Err(HashSet::from(["Expected multiple elements for tuple".into()]))
|
||||||
}
|
}
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -1,17 +1,17 @@
|
|||||||
|
use nac3parser::ast::fold::Fold;
|
||||||
use std::rc::Rc;
|
use std::rc::Rc;
|
||||||
|
|
||||||
use nac3parser::ast::{fold::Fold, ExprKind, Ident};
|
|
||||||
|
|
||||||
use super::*;
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{expr::get_subst_key, stmt::exn_constructor},
|
codegen::{expr::get_subst_key, stmt::exn_constructor},
|
||||||
symbol_resolver::SymbolValue,
|
symbol_resolver::SymbolValue,
|
||||||
typecheck::{
|
typecheck::{
|
||||||
type_inferencer::{FunctionData, IdentifierInfo, Inferencer},
|
type_inferencer::{FunctionData, Inferencer},
|
||||||
typedef::{TypeVar, VarMap},
|
typedef::{TypeVar, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
pub struct ComposerConfig {
|
pub struct ComposerConfig {
|
||||||
pub kernel_ann: Option<&'static str>,
|
pub kernel_ann: Option<&'static str>,
|
||||||
pub kernel_invariant_ann: &'static str,
|
pub kernel_invariant_ann: &'static str,
|
||||||
@ -23,7 +23,7 @@ impl Default for ComposerConfig {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub type DefAst = (Arc<RwLock<TopLevelDef>>, Option<Stmt<()>>);
|
type DefAst = (Arc<RwLock<TopLevelDef>>, Option<Stmt<()>>);
|
||||||
pub struct TopLevelComposer {
|
pub struct TopLevelComposer {
|
||||||
// list of top level definitions, same as top level context
|
// list of top level definitions, same as top level context
|
||||||
pub definition_ast_list: Vec<DefAst>,
|
pub definition_ast_list: Vec<DefAst>,
|
||||||
@ -44,27 +44,12 @@ pub struct TopLevelComposer {
|
|||||||
pub size_t: u32,
|
pub size_t: u32,
|
||||||
}
|
}
|
||||||
|
|
||||||
/// The specification for a builtin function, consisting of the function name, the function
|
|
||||||
/// signature, and a [code generation callback][`GenCall`].
|
|
||||||
pub type BuiltinFuncSpec = (StrRef, FunSignature, Arc<GenCall>);
|
|
||||||
|
|
||||||
/// A function that creates a [`BuiltinFuncSpec`] using the provided [`PrimitiveStore`] and
|
|
||||||
/// [`Unifier`].
|
|
||||||
pub type BuiltinFuncCreator = dyn Fn(&PrimitiveStore, &mut Unifier) -> BuiltinFuncSpec;
|
|
||||||
|
|
||||||
impl TopLevelComposer {
|
impl TopLevelComposer {
|
||||||
/// return a composer and things to make a "primitive" symbol resolver, so that the symbol
|
/// return a composer and things to make a "primitive" symbol resolver, so that the symbol
|
||||||
/// resolver can later figure out primitive tye definitions when passed a primitive type name
|
/// resolver can later figure out primitive type definitions when passed a primitive type name
|
||||||
///
|
|
||||||
/// `lateinit_builtins` are specifically for the ARTIQ module. Since the [`Unifier`] instance
|
|
||||||
/// used to create builtin functions do not persist until method compilation, any types
|
|
||||||
/// created (e.g. [`TypeEnum::TVar`]) also do not persist. Those functions should be instead put
|
|
||||||
/// in `lateinit_builtins`, where they will be instantiated with the [`Unifier`] instance used
|
|
||||||
/// for method compilation.
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn new(
|
pub fn new(
|
||||||
builtins: Vec<BuiltinFuncSpec>,
|
builtins: Vec<(StrRef, FunSignature, Arc<GenCall>)>,
|
||||||
lateinit_builtins: Vec<Box<BuiltinFuncCreator>>,
|
|
||||||
core_config: ComposerConfig,
|
core_config: ComposerConfig,
|
||||||
size_t: u32,
|
size_t: u32,
|
||||||
) -> (Self, HashMap<StrRef, DefinitionId>, HashMap<StrRef, Type>) {
|
) -> (Self, HashMap<StrRef, DefinitionId>, HashMap<StrRef, Type>) {
|
||||||
@ -101,8 +86,7 @@ impl TopLevelComposer {
|
|||||||
.iter()
|
.iter()
|
||||||
.map(|def_ast| match *def_ast.0.read() {
|
.map(|def_ast| match *def_ast.0.read() {
|
||||||
TopLevelDef::Class { name, .. } => name.to_string(),
|
TopLevelDef::Class { name, .. } => name.to_string(),
|
||||||
TopLevelDef::Function { simple_name, .. }
|
TopLevelDef::Function { simple_name, .. } => simple_name.to_string(),
|
||||||
| TopLevelDef::Variable { simple_name, .. } => simple_name.to_string(),
|
|
||||||
})
|
})
|
||||||
.collect_vec();
|
.collect_vec();
|
||||||
|
|
||||||
@ -135,13 +119,7 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Materialize lateinit_builtins, now that the unifier is ready
|
for (name, sig, codegen_callback) in builtins {
|
||||||
let lateinit_builtins = lateinit_builtins
|
|
||||||
.into_iter()
|
|
||||||
.map(|builtin| builtin(&primitives_ty, &mut unifier))
|
|
||||||
.collect_vec();
|
|
||||||
|
|
||||||
for (name, sig, codegen_callback) in builtins.into_iter().chain(lateinit_builtins) {
|
|
||||||
let fun_sig = unifier.add_ty(TypeEnum::TFunc(sig));
|
let fun_sig = unifier.add_ty(TypeEnum::TFunc(sig));
|
||||||
builtin_ty.insert(name, fun_sig);
|
builtin_ty.insert(name, fun_sig);
|
||||||
builtin_id.insert(name, DefinitionId(definition_ast_list.len()));
|
builtin_id.insert(name, DefinitionId(definition_ast_list.len()));
|
||||||
@ -382,87 +360,13 @@ impl TopLevelComposer {
|
|||||||
))
|
))
|
||||||
}
|
}
|
||||||
|
|
||||||
ast::StmtKind::Assign { .. } => {
|
|
||||||
// Assignment statements can assign to (and therefore create) more than one
|
|
||||||
// variable, but this function only allows returning one set of symbol information.
|
|
||||||
// We want to avoid changing this to return a `Vec` of symbol info, as this would
|
|
||||||
// require `iter().next().unwrap()` on every variable created from a non-Assign
|
|
||||||
// statement.
|
|
||||||
//
|
|
||||||
// Make callers use `register_top_level_var` instead, as it provides more
|
|
||||||
// fine-grained control over which symbols to register, while also simplifying the
|
|
||||||
// usage of this function.
|
|
||||||
panic!("Registration of top-level Assign statements must use TopLevelComposer::register_top_level_var (at {})", ast.location);
|
|
||||||
}
|
|
||||||
|
|
||||||
ast::StmtKind::AnnAssign { target, annotation, .. } => {
|
|
||||||
let ExprKind::Name { id: name, .. } = target.node else {
|
|
||||||
return Err(format!(
|
|
||||||
"global variable declaration must be an identifier (at {})",
|
|
||||||
target.location
|
|
||||||
));
|
|
||||||
};
|
|
||||||
|
|
||||||
self.register_top_level_var(
|
|
||||||
name,
|
|
||||||
Some(annotation.as_ref().clone()),
|
|
||||||
resolver,
|
|
||||||
mod_path,
|
|
||||||
target.location,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
_ => Err(format!(
|
_ => Err(format!(
|
||||||
"registrations of constructs other than top level classes/functions/variables are not supported (at {})",
|
"registrations of constructs other than top level classes/functions are not supported (at {})",
|
||||||
ast.location
|
ast.location
|
||||||
)),
|
)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Registers a top-level variable with the given `name` into the composer.
|
|
||||||
///
|
|
||||||
/// `annotation` - The type annotation of the top-level variable, or [`None`] if no type
|
|
||||||
/// annotation is provided.
|
|
||||||
/// `location` - The location of the top-level variable.
|
|
||||||
pub fn register_top_level_var(
|
|
||||||
&mut self,
|
|
||||||
name: Ident,
|
|
||||||
annotation: Option<Expr>,
|
|
||||||
resolver: Option<Arc<dyn SymbolResolver + Send + Sync>>,
|
|
||||||
mod_path: &str,
|
|
||||||
location: Location,
|
|
||||||
) -> Result<(StrRef, DefinitionId, Option<Type>), String> {
|
|
||||||
if self.keyword_list.contains(&name) {
|
|
||||||
return Err(format!("cannot use keyword `{name}` as a class name (at {location})"));
|
|
||||||
}
|
|
||||||
|
|
||||||
let global_var_name =
|
|
||||||
if mod_path.is_empty() { name.to_string() } else { format!("{mod_path}.{name}") };
|
|
||||||
|
|
||||||
if !self.defined_names.insert(global_var_name.clone()) {
|
|
||||||
return Err(format!(
|
|
||||||
"global variable `{global_var_name}` defined twice (at {location})"
|
|
||||||
));
|
|
||||||
}
|
|
||||||
|
|
||||||
let ty_to_be_unified = self.unifier.get_dummy_var().ty;
|
|
||||||
self.definition_ast_list.push((
|
|
||||||
RwLock::new(Self::make_top_level_variable_def(
|
|
||||||
global_var_name,
|
|
||||||
name,
|
|
||||||
// dummy here, unify with correct type later,
|
|
||||||
ty_to_be_unified,
|
|
||||||
annotation,
|
|
||||||
resolver,
|
|
||||||
Some(location),
|
|
||||||
))
|
|
||||||
.into(),
|
|
||||||
None,
|
|
||||||
));
|
|
||||||
|
|
||||||
Ok((name, DefinitionId(self.definition_ast_list.len() - 1), Some(ty_to_be_unified)))
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn start_analysis(&mut self, inference: bool) -> Result<(), HashSet<String>> {
|
pub fn start_analysis(&mut self, inference: bool) -> Result<(), HashSet<String>> {
|
||||||
self.analyze_top_level_class_type_var()?;
|
self.analyze_top_level_class_type_var()?;
|
||||||
self.analyze_top_level_class_bases()?;
|
self.analyze_top_level_class_bases()?;
|
||||||
@ -471,7 +375,6 @@ impl TopLevelComposer {
|
|||||||
if inference {
|
if inference {
|
||||||
self.analyze_function_instance()?;
|
self.analyze_function_instance()?;
|
||||||
}
|
}
|
||||||
self.analyze_top_level_variables()?;
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -509,7 +412,7 @@ impl TopLevelComposer {
|
|||||||
// things like `class A(Generic[T, V, ImportedModule.T])` is not supported
|
// things like `class A(Generic[T, V, ImportedModule.T])` is not supported
|
||||||
// i.e. only simple names are allowed in the subscript
|
// i.e. only simple names are allowed in the subscript
|
||||||
// should update the TopLevelDef::Class.typevars and the TypeEnum::TObj.params
|
// should update the TopLevelDef::Class.typevars and the TypeEnum::TObj.params
|
||||||
ExprKind::Subscript { value, slice, .. }
|
ast::ExprKind::Subscript { value, slice, .. }
|
||||||
if {
|
if {
|
||||||
matches!(
|
matches!(
|
||||||
&value.node,
|
&value.node,
|
||||||
@ -525,9 +428,9 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
is_generic = true;
|
is_generic = true;
|
||||||
|
|
||||||
let type_var_list: Vec<&Expr<()>>;
|
let type_var_list: Vec<&ast::Expr<()>>;
|
||||||
// if `class A(Generic[T, V, G])`
|
// if `class A(Generic[T, V, G])`
|
||||||
if let ExprKind::Tuple { elts, .. } = &slice.node {
|
if let ast::ExprKind::Tuple { elts, .. } = &slice.node {
|
||||||
type_var_list = elts.iter().collect_vec();
|
type_var_list = elts.iter().collect_vec();
|
||||||
// `class A(Generic[T])`
|
// `class A(Generic[T])`
|
||||||
} else {
|
} else {
|
||||||
@ -576,7 +479,6 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut errors = HashSet::new();
|
let mut errors = HashSet::new();
|
||||||
for (class_def, class_ast) in def_list.iter().skip(self.builtin_num) {
|
for (class_def, class_ast) in def_list.iter().skip(self.builtin_num) {
|
||||||
if class_ast.is_none() {
|
if class_ast.is_none() {
|
||||||
@ -864,7 +766,6 @@ impl TopLevelComposer {
|
|||||||
let target_ty = get_type_from_type_annotation_kinds(
|
let target_ty = get_type_from_type_annotation_kinds(
|
||||||
&temp_def_list,
|
&temp_def_list,
|
||||||
unifier,
|
unifier,
|
||||||
primitives,
|
|
||||||
&def,
|
&def,
|
||||||
&mut subst_list,
|
&mut subst_list,
|
||||||
)?;
|
)?;
|
||||||
@ -930,6 +831,7 @@ impl TopLevelComposer {
|
|||||||
let unifier = self.unifier.borrow_mut();
|
let unifier = self.unifier.borrow_mut();
|
||||||
let primitives_store = &self.primitives_ty;
|
let primitives_store = &self.primitives_ty;
|
||||||
|
|
||||||
|
let mut errors = HashSet::new();
|
||||||
let mut analyze = |function_def: &Arc<RwLock<TopLevelDef>>, function_ast: &Option<Stmt>| {
|
let mut analyze = |function_def: &Arc<RwLock<TopLevelDef>>, function_ast: &Option<Stmt>| {
|
||||||
let mut function_def = function_def.write();
|
let mut function_def = function_def.write();
|
||||||
let function_def = &mut *function_def;
|
let function_def = &mut *function_def;
|
||||||
@ -957,73 +859,7 @@ impl TopLevelComposer {
|
|||||||
let resolver = &**resolver;
|
let resolver = &**resolver;
|
||||||
|
|
||||||
let mut function_var_map = VarMap::new();
|
let mut function_var_map = VarMap::new();
|
||||||
|
let arg_types = {
|
||||||
let vararg = args
|
|
||||||
.vararg
|
|
||||||
.as_ref()
|
|
||||||
.map(|vararg| -> Result<_, HashSet<String>> {
|
|
||||||
let vararg = vararg.as_ref();
|
|
||||||
|
|
||||||
let annotation = vararg
|
|
||||||
.node
|
|
||||||
.annotation
|
|
||||||
.as_ref()
|
|
||||||
.ok_or_else(|| {
|
|
||||||
HashSet::from([format!(
|
|
||||||
"function parameter `{}` needs type annotation at {}",
|
|
||||||
vararg.node.arg, vararg.location
|
|
||||||
)])
|
|
||||||
})?
|
|
||||||
.as_ref();
|
|
||||||
|
|
||||||
let type_annotation = parse_ast_to_type_annotation_kinds(
|
|
||||||
resolver,
|
|
||||||
temp_def_list.as_slice(),
|
|
||||||
unifier,
|
|
||||||
primitives_store,
|
|
||||||
annotation,
|
|
||||||
// NOTE: since only class need this, for function
|
|
||||||
// it should be fine to be empty map
|
|
||||||
HashMap::new(),
|
|
||||||
)?;
|
|
||||||
|
|
||||||
let type_vars_within =
|
|
||||||
get_type_var_contained_in_type_annotation(&type_annotation)
|
|
||||||
.into_iter()
|
|
||||||
.map(|x| -> Result<TypeVar, HashSet<String>> {
|
|
||||||
let TypeAnnotation::TypeVar(ty) = x else {
|
|
||||||
unreachable!("must be type var annotation kind")
|
|
||||||
};
|
|
||||||
|
|
||||||
let id = Self::get_var_id(ty, unifier)?;
|
|
||||||
Ok(TypeVar { id, ty })
|
|
||||||
})
|
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
|
||||||
for var in type_vars_within {
|
|
||||||
if let Some(prev_ty) = function_var_map.insert(var.id, var.ty) {
|
|
||||||
// if already have the type inserted, make sure they are the same thing
|
|
||||||
assert_eq!(prev_ty, var.ty);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let ty = get_type_from_type_annotation_kinds(
|
|
||||||
temp_def_list.as_ref(),
|
|
||||||
unifier,
|
|
||||||
primitives_store,
|
|
||||||
&type_annotation,
|
|
||||||
&mut None,
|
|
||||||
)?;
|
|
||||||
|
|
||||||
Ok(FuncArg {
|
|
||||||
name: vararg.node.arg,
|
|
||||||
ty,
|
|
||||||
default_value: Some(SymbolValue::Tuple(Vec::default())),
|
|
||||||
is_vararg: true,
|
|
||||||
})
|
|
||||||
})
|
|
||||||
.transpose()?;
|
|
||||||
|
|
||||||
let mut arg_types = {
|
|
||||||
// make sure no duplicate parameter
|
// make sure no duplicate parameter
|
||||||
let mut defined_parameter_name: HashSet<_> = HashSet::new();
|
let mut defined_parameter_name: HashSet<_> = HashSet::new();
|
||||||
for x in &args.args {
|
for x in &args.args {
|
||||||
@ -1038,15 +874,15 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let arg_with_default: Vec<(&ast::Located<ast::ArgData<()>>, Option<&Expr>)> = args
|
let arg_with_default: Vec<(&ast::Located<ast::ArgData<()>>, Option<&ast::Expr>)> =
|
||||||
.args
|
args.args
|
||||||
.iter()
|
.iter()
|
||||||
.rev()
|
.rev()
|
||||||
.zip(
|
.zip(
|
||||||
args.defaults
|
args.defaults
|
||||||
.iter()
|
.iter()
|
||||||
.rev()
|
.rev()
|
||||||
.map(|x| -> Option<&Expr> { Some(x) })
|
.map(|x| -> Option<&ast::Expr> { Some(x) })
|
||||||
.chain(std::iter::repeat(None)),
|
.chain(std::iter::repeat(None)),
|
||||||
)
|
)
|
||||||
.collect_vec();
|
.collect_vec();
|
||||||
@ -1100,7 +936,6 @@ impl TopLevelComposer {
|
|||||||
let ty = get_type_from_type_annotation_kinds(
|
let ty = get_type_from_type_annotation_kinds(
|
||||||
temp_def_list.as_ref(),
|
temp_def_list.as_ref(),
|
||||||
unifier,
|
unifier,
|
||||||
primitives_store,
|
|
||||||
&type_annotation,
|
&type_annotation,
|
||||||
&mut None,
|
&mut None,
|
||||||
)?;
|
)?;
|
||||||
@ -1124,18 +959,11 @@ impl TopLevelComposer {
|
|||||||
v
|
v
|
||||||
}),
|
}),
|
||||||
},
|
},
|
||||||
is_vararg: false,
|
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
.collect::<Result<Vec<_>, _>>()?
|
.collect::<Result<Vec<_>, _>>()?
|
||||||
};
|
};
|
||||||
|
|
||||||
if let Some(vararg) = vararg {
|
|
||||||
arg_types.push(vararg);
|
|
||||||
};
|
|
||||||
|
|
||||||
let arg_types = arg_types;
|
|
||||||
|
|
||||||
let return_ty = {
|
let return_ty = {
|
||||||
if let Some(returns) = returns {
|
if let Some(returns) = returns {
|
||||||
let return_ty_annotation = {
|
let return_ty_annotation = {
|
||||||
@ -1174,7 +1002,6 @@ impl TopLevelComposer {
|
|||||||
get_type_from_type_annotation_kinds(
|
get_type_from_type_annotation_kinds(
|
||||||
&temp_def_list,
|
&temp_def_list,
|
||||||
unifier,
|
unifier,
|
||||||
primitives_store,
|
|
||||||
&return_ty_annotation,
|
&return_ty_annotation,
|
||||||
&mut None,
|
&mut None,
|
||||||
)?
|
)?
|
||||||
@ -1204,8 +1031,6 @@ impl TopLevelComposer {
|
|||||||
})?;
|
})?;
|
||||||
Ok(())
|
Ok(())
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut errors = HashSet::new();
|
|
||||||
for (function_def, function_ast) in def_list.iter().skip(self.builtin_num) {
|
for (function_def, function_ast) in def_list.iter().skip(self.builtin_num) {
|
||||||
if function_ast.is_none() {
|
if function_ast.is_none() {
|
||||||
continue;
|
continue;
|
||||||
@ -1232,14 +1057,7 @@ impl TopLevelComposer {
|
|||||||
let (keyword_list, core_config) = core_info;
|
let (keyword_list, core_config) = core_info;
|
||||||
let mut class_def = class_def.write();
|
let mut class_def = class_def.write();
|
||||||
let TopLevelDef::Class {
|
let TopLevelDef::Class {
|
||||||
object_id,
|
object_id, ancestors, fields, methods, resolver, type_vars, ..
|
||||||
ancestors,
|
|
||||||
fields,
|
|
||||||
attributes,
|
|
||||||
methods,
|
|
||||||
resolver,
|
|
||||||
type_vars,
|
|
||||||
..
|
|
||||||
} = &mut *class_def
|
} = &mut *class_def
|
||||||
else {
|
else {
|
||||||
unreachable!("here must be toplevel class def");
|
unreachable!("here must be toplevel class def");
|
||||||
@ -1255,14 +1073,10 @@ impl TopLevelComposer {
|
|||||||
class_body_ast,
|
class_body_ast,
|
||||||
_class_ancestor_def,
|
_class_ancestor_def,
|
||||||
class_fields_def,
|
class_fields_def,
|
||||||
class_attributes_def,
|
|
||||||
class_methods_def,
|
class_methods_def,
|
||||||
class_type_vars_def,
|
class_type_vars_def,
|
||||||
class_resolver,
|
class_resolver,
|
||||||
) = (
|
) = (*object_id, *name, bases, body, ancestors, fields, methods, type_vars, resolver);
|
||||||
*object_id, *name, bases, body, ancestors, fields, attributes, methods, type_vars,
|
|
||||||
resolver,
|
|
||||||
);
|
|
||||||
|
|
||||||
let class_resolver = class_resolver.as_ref().unwrap();
|
let class_resolver = class_resolver.as_ref().unwrap();
|
||||||
let class_resolver = class_resolver.as_ref();
|
let class_resolver = class_resolver.as_ref();
|
||||||
@ -1307,7 +1121,7 @@ impl TopLevelComposer {
|
|||||||
|
|
||||||
let arg_with_default: Vec<(
|
let arg_with_default: Vec<(
|
||||||
&ast::Located<ast::ArgData<()>>,
|
&ast::Located<ast::ArgData<()>>,
|
||||||
Option<&Expr>,
|
Option<&ast::Expr>,
|
||||||
)> = args
|
)> = args
|
||||||
.args
|
.args
|
||||||
.iter()
|
.iter()
|
||||||
@ -1316,7 +1130,7 @@ impl TopLevelComposer {
|
|||||||
args.defaults
|
args.defaults
|
||||||
.iter()
|
.iter()
|
||||||
.rev()
|
.rev()
|
||||||
.map(|x| -> Option<&Expr> { Some(x) })
|
.map(|x| -> Option<&ast::Expr> { Some(x) })
|
||||||
.chain(std::iter::repeat(None)),
|
.chain(std::iter::repeat(None)),
|
||||||
)
|
)
|
||||||
.collect_vec();
|
.collect_vec();
|
||||||
@ -1389,7 +1203,6 @@ impl TopLevelComposer {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
is_vararg: false,
|
|
||||||
};
|
};
|
||||||
// push the dummy type and the type annotation
|
// push the dummy type and the type annotation
|
||||||
// into the list for later unification
|
// into the list for later unification
|
||||||
@ -1472,16 +1285,14 @@ impl TopLevelComposer {
|
|||||||
.unify(method_dummy_ty, method_type)
|
.unify(method_dummy_ty, method_type)
|
||||||
.map_err(|e| HashSet::from([e.to_display(unifier).to_string()]))?;
|
.map_err(|e| HashSet::from([e.to_display(unifier).to_string()]))?;
|
||||||
}
|
}
|
||||||
ast::StmtKind::AnnAssign { target, annotation, value, .. } => {
|
ast::StmtKind::AnnAssign { target, annotation, value: None, .. } => {
|
||||||
if let ExprKind::Name { id: attr, .. } = &target.node {
|
if let ast::ExprKind::Name { id: attr, .. } = &target.node {
|
||||||
if defined_fields.insert(attr.to_string()) {
|
if defined_fields.insert(attr.to_string()) {
|
||||||
let dummy_field_type = unifier.get_dummy_var().ty;
|
let dummy_field_type = unifier.get_dummy_var().ty;
|
||||||
|
|
||||||
let annotation = match value {
|
|
||||||
None => {
|
|
||||||
// handle Kernel[T], KernelInvariant[T]
|
// handle Kernel[T], KernelInvariant[T]
|
||||||
let (annotation, mutable) = match &annotation.node {
|
let (annotation, mutable) = match &annotation.node {
|
||||||
ExprKind::Subscript { value, slice, .. }
|
ast::ExprKind::Subscript { value, slice, .. }
|
||||||
if matches!(
|
if matches!(
|
||||||
&value.node,
|
&value.node,
|
||||||
ast::ExprKind::Name { id, .. } if id == &core_config.kernel_invariant_ann.into()
|
ast::ExprKind::Name { id, .. } if id == &core_config.kernel_invariant_ann.into()
|
||||||
@ -1489,7 +1300,7 @@ impl TopLevelComposer {
|
|||||||
{
|
{
|
||||||
(slice, false)
|
(slice, false)
|
||||||
}
|
}
|
||||||
ExprKind::Subscript { value, slice, .. }
|
ast::ExprKind::Subscript { value, slice, .. }
|
||||||
if matches!(
|
if matches!(
|
||||||
&value.node,
|
&value.node,
|
||||||
ast::ExprKind::Name { id, .. } if core_config.kernel_ann.map_or(false, |c| id == &c.into())
|
ast::ExprKind::Name { id, .. } if core_config.kernel_ann.map_or(false, |c| id == &c.into())
|
||||||
@ -1501,45 +1312,7 @@ impl TopLevelComposer {
|
|||||||
_ => continue, // ignore fields annotated otherwise
|
_ => continue, // ignore fields annotated otherwise
|
||||||
};
|
};
|
||||||
class_fields_def.push((*attr, dummy_field_type, mutable));
|
class_fields_def.push((*attr, dummy_field_type, mutable));
|
||||||
annotation
|
|
||||||
}
|
|
||||||
// Supporting Class Attributes
|
|
||||||
Some(boxed_expr) => {
|
|
||||||
// Class attributes are set as immutable regardless
|
|
||||||
let (annotation, _) = match &annotation.node {
|
|
||||||
ExprKind::Subscript { slice, .. } => (slice, false),
|
|
||||||
_ if core_config.kernel_ann.is_none() => (annotation, false),
|
|
||||||
_ => continue,
|
|
||||||
};
|
|
||||||
|
|
||||||
match &**boxed_expr {
|
|
||||||
ast::Located {location: _, custom: (), node: ExprKind::Constant { value: v, kind: _ }} => {
|
|
||||||
// Restricting the types allowed to be defined as class attributes
|
|
||||||
match v {
|
|
||||||
ast::Constant::Bool(_) | ast::Constant::Str(_) | ast::Constant::Int(_) | ast::Constant::Float(_) => {}
|
|
||||||
_ => {
|
|
||||||
return Err(HashSet::from([
|
|
||||||
format!(
|
|
||||||
"unsupported statement in class definition body (at {})",
|
|
||||||
b.location
|
|
||||||
),
|
|
||||||
]))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
class_attributes_def.push((*attr, dummy_field_type, v.clone()));
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
return Err(HashSet::from([
|
|
||||||
format!(
|
|
||||||
"unsupported statement in class definition body (at {})",
|
|
||||||
b.location
|
|
||||||
),
|
|
||||||
]))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
annotation
|
|
||||||
}
|
|
||||||
};
|
|
||||||
let parsed_annotation = parse_ast_to_type_annotation_kinds(
|
let parsed_annotation = parse_ast_to_type_annotation_kinds(
|
||||||
class_resolver,
|
class_resolver,
|
||||||
temp_def_list,
|
temp_def_list,
|
||||||
@ -1611,14 +1384,7 @@ impl TopLevelComposer {
|
|||||||
type_var_to_concrete_def: &mut HashMap<Type, TypeAnnotation>,
|
type_var_to_concrete_def: &mut HashMap<Type, TypeAnnotation>,
|
||||||
) -> Result<(), HashSet<String>> {
|
) -> Result<(), HashSet<String>> {
|
||||||
let TopLevelDef::Class {
|
let TopLevelDef::Class {
|
||||||
object_id,
|
object_id, ancestors, fields, methods, resolver, type_vars, ..
|
||||||
ancestors,
|
|
||||||
fields,
|
|
||||||
attributes,
|
|
||||||
methods,
|
|
||||||
resolver,
|
|
||||||
type_vars,
|
|
||||||
..
|
|
||||||
} = class_def
|
} = class_def
|
||||||
else {
|
else {
|
||||||
unreachable!("here must be class def ast")
|
unreachable!("here must be class def ast")
|
||||||
@ -1627,11 +1393,10 @@ impl TopLevelComposer {
|
|||||||
_class_id,
|
_class_id,
|
||||||
class_ancestor_def,
|
class_ancestor_def,
|
||||||
class_fields_def,
|
class_fields_def,
|
||||||
class_attribute_def,
|
|
||||||
class_methods_def,
|
class_methods_def,
|
||||||
_class_type_vars_def,
|
_class_type_vars_def,
|
||||||
_class_resolver,
|
_class_resolver,
|
||||||
) = (*object_id, ancestors, fields, attributes, methods, type_vars, resolver);
|
) = (*object_id, ancestors, fields, methods, type_vars, resolver);
|
||||||
|
|
||||||
// since when this function is called, the ancestors of the direct parent
|
// since when this function is called, the ancestors of the direct parent
|
||||||
// are supposed to be already handled, so we only need to deal with the direct parent
|
// are supposed to be already handled, so we only need to deal with the direct parent
|
||||||
@ -1642,7 +1407,7 @@ impl TopLevelComposer {
|
|||||||
|
|
||||||
let base = temp_def_list.get(id.0).unwrap();
|
let base = temp_def_list.get(id.0).unwrap();
|
||||||
let base = base.read();
|
let base = base.read();
|
||||||
let TopLevelDef::Class { methods, fields, attributes, .. } = &*base else {
|
let TopLevelDef::Class { methods, fields, .. } = &*base else {
|
||||||
unreachable!("must be top level class def")
|
unreachable!("must be top level class def")
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -1684,7 +1449,7 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
// use the new_child_methods to replace all the elements in `class_methods_def`
|
// use the new_child_methods to replace all the elements in `class_methods_def`
|
||||||
class_methods_def.clear();
|
class_methods_def.drain(..);
|
||||||
class_methods_def.extend(new_child_methods);
|
class_methods_def.extend(new_child_methods);
|
||||||
|
|
||||||
// handle class fields
|
// handle class fields
|
||||||
@ -1694,9 +1459,7 @@ impl TopLevelComposer {
|
|||||||
let to_be_added = (*anc_field_name, *anc_field_ty, *mutable);
|
let to_be_added = (*anc_field_name, *anc_field_ty, *mutable);
|
||||||
// find if there is a fields with the same name in the child class
|
// find if there is a fields with the same name in the child class
|
||||||
for (class_field_name, ..) in &*class_fields_def {
|
for (class_field_name, ..) in &*class_fields_def {
|
||||||
if class_field_name == anc_field_name
|
if class_field_name == anc_field_name {
|
||||||
|| attributes.iter().any(|f| f.0 == *class_field_name)
|
|
||||||
{
|
|
||||||
return Err(HashSet::from([format!(
|
return Err(HashSet::from([format!(
|
||||||
"field `{class_field_name}` has already declared in the ancestor classes"
|
"field `{class_field_name}` has already declared in the ancestor classes"
|
||||||
)]));
|
)]));
|
||||||
@ -1704,33 +1467,14 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
new_child_fields.push(to_be_added);
|
new_child_fields.push(to_be_added);
|
||||||
}
|
}
|
||||||
|
|
||||||
// handle class attributes
|
|
||||||
let mut new_child_attributes: Vec<(StrRef, Type, ast::Constant)> = Vec::new();
|
|
||||||
for (anc_attr_name, anc_attr_ty, attr_value) in attributes {
|
|
||||||
let to_be_added = (*anc_attr_name, *anc_attr_ty, attr_value.clone());
|
|
||||||
// find if there is a attribute with the same name in the child class
|
|
||||||
for (class_attr_name, ..) in &*class_attribute_def {
|
|
||||||
if class_attr_name == anc_attr_name
|
|
||||||
|| fields.iter().any(|f| f.0 == *class_attr_name)
|
|
||||||
{
|
|
||||||
return Err(HashSet::from([format!(
|
|
||||||
"attribute `{class_attr_name}` has already declared in the ancestor classes"
|
|
||||||
)]));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
new_child_attributes.push(to_be_added);
|
|
||||||
}
|
|
||||||
|
|
||||||
for (class_field_name, class_field_ty, mutable) in &*class_fields_def {
|
for (class_field_name, class_field_ty, mutable) in &*class_fields_def {
|
||||||
if !is_override.contains(class_field_name) {
|
if !is_override.contains(class_field_name) {
|
||||||
new_child_fields.push((*class_field_name, *class_field_ty, *mutable));
|
new_child_fields.push((*class_field_name, *class_field_ty, *mutable));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
class_fields_def.clear();
|
class_fields_def.drain(..);
|
||||||
class_fields_def.extend(new_child_fields);
|
class_fields_def.extend(new_child_fields);
|
||||||
class_attribute_def.clear();
|
|
||||||
class_attribute_def.extend(new_child_attributes);
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1780,6 +1524,7 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let mut errors = HashSet::new();
|
||||||
let mut analyze = |i, def: &Arc<RwLock<TopLevelDef>>, ast: &Option<Stmt>| {
|
let mut analyze = |i, def: &Arc<RwLock<TopLevelDef>>, ast: &Option<Stmt>| {
|
||||||
let class_def = def.read();
|
let class_def = def.read();
|
||||||
if let TopLevelDef::Class {
|
if let TopLevelDef::Class {
|
||||||
@ -1797,7 +1542,6 @@ impl TopLevelComposer {
|
|||||||
let self_type = get_type_from_type_annotation_kinds(
|
let self_type = get_type_from_type_annotation_kinds(
|
||||||
&def_list,
|
&def_list,
|
||||||
unifier,
|
unifier,
|
||||||
primitives_ty,
|
|
||||||
&make_self_type_annotation(type_vars, *object_id),
|
&make_self_type_annotation(type_vars, *object_id),
|
||||||
&mut None,
|
&mut None,
|
||||||
)?;
|
)?;
|
||||||
@ -1814,25 +1558,21 @@ impl TopLevelComposer {
|
|||||||
name: "msg".into(),
|
name: "msg".into(),
|
||||||
ty: string,
|
ty: string,
|
||||||
default_value: Some(SymbolValue::Str(String::new())),
|
default_value: Some(SymbolValue::Str(String::new())),
|
||||||
is_vararg: false,
|
|
||||||
},
|
},
|
||||||
FuncArg {
|
FuncArg {
|
||||||
name: "param0".into(),
|
name: "param0".into(),
|
||||||
ty: int64,
|
ty: int64,
|
||||||
default_value: Some(SymbolValue::I64(0)),
|
default_value: Some(SymbolValue::I64(0)),
|
||||||
is_vararg: false,
|
|
||||||
},
|
},
|
||||||
FuncArg {
|
FuncArg {
|
||||||
name: "param1".into(),
|
name: "param1".into(),
|
||||||
ty: int64,
|
ty: int64,
|
||||||
default_value: Some(SymbolValue::I64(0)),
|
default_value: Some(SymbolValue::I64(0)),
|
||||||
is_vararg: false,
|
|
||||||
},
|
},
|
||||||
FuncArg {
|
FuncArg {
|
||||||
name: "param2".into(),
|
name: "param2".into(),
|
||||||
ty: int64,
|
ty: int64,
|
||||||
default_value: Some(SymbolValue::I64(0)),
|
default_value: Some(SymbolValue::I64(0)),
|
||||||
is_vararg: false,
|
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
ret: self_type,
|
ret: self_type,
|
||||||
@ -1899,12 +1639,7 @@ impl TopLevelComposer {
|
|||||||
if *name != init_str_id {
|
if *name != init_str_id {
|
||||||
unreachable!("must be init function here")
|
unreachable!("must be init function here")
|
||||||
}
|
}
|
||||||
|
let all_inited = Self::get_all_assigned_field(body.as_slice())?;
|
||||||
let all_inited = Self::get_all_assigned_field(
|
|
||||||
object_id.0,
|
|
||||||
definition_ast_list,
|
|
||||||
body.as_slice(),
|
|
||||||
)?;
|
|
||||||
for (f, _, _) in fields {
|
for (f, _, _) in fields {
|
||||||
if !all_inited.contains(f) {
|
if !all_inited.contains(f) {
|
||||||
return Err(HashSet::from([
|
return Err(HashSet::from([
|
||||||
@ -1922,8 +1657,6 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut errors = HashSet::new();
|
|
||||||
for (i, (def, ast)) in definition_ast_list.iter().enumerate().skip(self.builtin_num) {
|
for (i, (def, ast)) in definition_ast_list.iter().enumerate().skip(self.builtin_num) {
|
||||||
if ast.is_none() {
|
if ast.is_none() {
|
||||||
continue;
|
continue;
|
||||||
@ -1961,20 +1694,19 @@ impl TopLevelComposer {
|
|||||||
if ast.is_none() {
|
if ast.is_none() {
|
||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
let mut function_def = def.write();
|
||||||
let (name, simple_name, signature, resolver) = {
|
if let TopLevelDef::Function {
|
||||||
let function_def = def.read();
|
instance_to_stmt,
|
||||||
let TopLevelDef::Function { name, simple_name, signature, resolver, .. } =
|
instance_to_symbol,
|
||||||
&*function_def
|
name,
|
||||||
else {
|
simple_name,
|
||||||
return Ok(());
|
signature,
|
||||||
};
|
resolver,
|
||||||
|
..
|
||||||
(name.clone(), *simple_name, *signature, resolver.clone())
|
} = &mut *function_def
|
||||||
};
|
{
|
||||||
|
let signature_ty_enum = unifier.get_ty(*signature);
|
||||||
let signature_ty_enum = unifier.get_ty(signature);
|
let TypeEnum::TFunc(FunSignature { args, ret, vars }) = signature_ty_enum.as_ref()
|
||||||
let TypeEnum::TFunc(FunSignature { args, ret, vars, .. }) = signature_ty_enum.as_ref()
|
|
||||||
else {
|
else {
|
||||||
unreachable!("must be typeenum::tfunc")
|
unreachable!("must be typeenum::tfunc")
|
||||||
};
|
};
|
||||||
@ -1991,11 +1723,7 @@ impl TopLevelComposer {
|
|||||||
|
|
||||||
let ty_ann = make_self_type_annotation(type_vars, *class_id);
|
let ty_ann = make_self_type_annotation(type_vars, *class_id);
|
||||||
let self_ty = get_type_from_type_annotation_kinds(
|
let self_ty = get_type_from_type_annotation_kinds(
|
||||||
&def_list,
|
&def_list, unifier, &ty_ann, &mut None,
|
||||||
unifier,
|
|
||||||
primitives_ty,
|
|
||||||
&ty_ann,
|
|
||||||
&mut None,
|
|
||||||
)?;
|
)?;
|
||||||
vars.extend(type_vars.iter().map(|ty| {
|
vars.extend(type_vars.iter().map(|ty| {
|
||||||
let TypeEnum::TVar { id, .. } = &*unifier.get_ty(*ty) else {
|
let TypeEnum::TVar { id, .. } = &*unifier.get_ty(*ty) else {
|
||||||
@ -2050,7 +1778,6 @@ impl TopLevelComposer {
|
|||||||
name: a.name,
|
name: a.name,
|
||||||
ty: unifier.subst(a.ty, &subst).unwrap_or(a.ty),
|
ty: unifier.subst(a.ty, &subst).unwrap_or(a.ty),
|
||||||
default_value: a.default_value.clone(),
|
default_value: a.default_value.clone(),
|
||||||
is_vararg: false,
|
|
||||||
})
|
})
|
||||||
.collect_vec()
|
.collect_vec()
|
||||||
};
|
};
|
||||||
@ -2082,11 +1809,11 @@ impl TopLevelComposer {
|
|||||||
})
|
})
|
||||||
};
|
};
|
||||||
let mut identifiers = {
|
let mut identifiers = {
|
||||||
let mut result = HashMap::new();
|
let mut result: HashSet<_> = HashSet::new();
|
||||||
if self_type.is_some() {
|
if self_type.is_some() {
|
||||||
result.insert("self".into(), IdentifierInfo::default());
|
result.insert("self".into());
|
||||||
}
|
}
|
||||||
result.extend(inst_args.iter().map(|x| (x.name, IdentifierInfo::default())));
|
result.extend(inst_args.iter().map(|x| x.name));
|
||||||
result
|
result
|
||||||
};
|
};
|
||||||
let mut calls: HashMap<CodeLocation, CallId> = HashMap::new();
|
let mut calls: HashMap<CodeLocation, CallId> = HashMap::new();
|
||||||
@ -2123,43 +1850,23 @@ impl TopLevelComposer {
|
|||||||
else {
|
else {
|
||||||
unreachable!("must be function def ast")
|
unreachable!("must be function def ast")
|
||||||
};
|
};
|
||||||
|
if !decorator_list.is_empty()
|
||||||
if !decorator_list.is_empty() {
|
&& matches!(&decorator_list[0].node,
|
||||||
if matches!(&decorator_list[0].node, ExprKind::Name { id, .. } if id == &"extern".into())
|
ast::ExprKind::Name{ id, .. } if id == &"extern".into())
|
||||||
{
|
{
|
||||||
let TopLevelDef::Function { instance_to_symbol, .. } = &mut *def.write()
|
|
||||||
else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
instance_to_symbol.insert(String::new(), simple_name.to_string());
|
instance_to_symbol.insert(String::new(), simple_name.to_string());
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
if !decorator_list.is_empty()
|
||||||
if matches!(&decorator_list[0].node, ExprKind::Name { id, .. } if id == &"rpc".into())
|
&& matches!(&decorator_list[0].node,
|
||||||
|
ast::ExprKind::Name{ id, .. } if id == &"rpc".into())
|
||||||
{
|
{
|
||||||
let TopLevelDef::Function { instance_to_symbol, .. } = &mut *def.write()
|
|
||||||
else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
instance_to_symbol.insert(String::new(), simple_name.to_string());
|
instance_to_symbol.insert(String::new(), simple_name.to_string());
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
if let ExprKind::Call { func, .. } = &decorator_list[0].node {
|
let fun_body = body
|
||||||
if matches!(&func.node, ExprKind::Name { id, .. } if id == &"rpc".into()) {
|
.into_iter()
|
||||||
let TopLevelDef::Function { instance_to_symbol, .. } =
|
|
||||||
&mut *def.write()
|
|
||||||
else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
instance_to_symbol.insert(String::new(), simple_name.to_string());
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let fun_body =
|
|
||||||
body.into_iter()
|
|
||||||
.map(|b| inferencer.fold_stmt(b))
|
.map(|b| inferencer.fold_stmt(b))
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
.collect::<Result<Vec<_>, _>>()?;
|
||||||
|
|
||||||
@ -2229,9 +1936,6 @@ impl TopLevelComposer {
|
|||||||
)]));
|
)]));
|
||||||
}
|
}
|
||||||
|
|
||||||
let TopLevelDef::Function { instance_to_stmt, .. } = &mut *def.write() else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
instance_to_stmt.insert(
|
instance_to_stmt.insert(
|
||||||
get_subst_key(
|
get_subst_key(
|
||||||
unifier,
|
unifier,
|
||||||
@ -2247,10 +1951,10 @@ impl TopLevelComposer {
|
|||||||
},
|
},
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
};
|
};
|
||||||
|
|
||||||
for (id, (def, ast)) in self.definition_ast_list.iter().enumerate().skip(self.builtin_num) {
|
for (id, (def, ast)) in self.definition_ast_list.iter().enumerate().skip(self.builtin_num) {
|
||||||
if ast.is_none() {
|
if ast.is_none() {
|
||||||
continue;
|
continue;
|
||||||
@ -2264,59 +1968,4 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Step 6. Analyze and populate the types of global variables.
|
|
||||||
fn analyze_top_level_variables(&mut self) -> Result<(), HashSet<String>> {
|
|
||||||
let def_list = &self.definition_ast_list;
|
|
||||||
let temp_def_list = self.extract_def_list();
|
|
||||||
let unifier = &mut self.unifier;
|
|
||||||
let primitives_store = &self.primitives_ty;
|
|
||||||
|
|
||||||
let mut analyze = |variable_def: &Arc<RwLock<TopLevelDef>>| -> Result<_, HashSet<String>> {
|
|
||||||
let TopLevelDef::Variable { ty: dummy_ty, ty_decl, resolver, loc, .. } =
|
|
||||||
&*variable_def.read()
|
|
||||||
else {
|
|
||||||
// not top level variable def, skip
|
|
||||||
return Ok(());
|
|
||||||
};
|
|
||||||
|
|
||||||
let resolver = &**resolver.as_ref().unwrap();
|
|
||||||
|
|
||||||
if let Some(ty_decl) = ty_decl {
|
|
||||||
let ty_annotation = parse_ast_to_type_annotation_kinds(
|
|
||||||
resolver,
|
|
||||||
&temp_def_list,
|
|
||||||
unifier,
|
|
||||||
primitives_store,
|
|
||||||
ty_decl,
|
|
||||||
HashMap::new(),
|
|
||||||
)?;
|
|
||||||
let ty_from_ty_annotation = get_type_from_type_annotation_kinds(
|
|
||||||
&temp_def_list,
|
|
||||||
unifier,
|
|
||||||
primitives_store,
|
|
||||||
&ty_annotation,
|
|
||||||
&mut None,
|
|
||||||
)?;
|
|
||||||
|
|
||||||
unifier.unify(*dummy_ty, ty_from_ty_annotation).map_err(|e| {
|
|
||||||
HashSet::from([e.at(Some(loc.unwrap())).to_display(unifier).to_string()])
|
|
||||||
})?;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut errors = HashSet::new();
|
|
||||||
for (variable_def, _) in def_list.iter().skip(self.builtin_num) {
|
|
||||||
if let Err(e) = analyze(variable_def) {
|
|
||||||
errors.extend(e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if !errors.is_empty() {
|
|
||||||
return Err(errors);
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -1,22 +1,17 @@
|
|||||||
use std::convert::TryInto;
|
use std::convert::TryInto;
|
||||||
|
|
||||||
|
use crate::symbol_resolver::SymbolValue;
|
||||||
|
use crate::toplevel::numpy::unpack_ndarray_var_tys;
|
||||||
|
use crate::typecheck::typedef::{into_var_map, Mapping, TypeVarId, VarMap};
|
||||||
|
use nac3parser::ast::{Constant, Location};
|
||||||
use strum::IntoEnumIterator;
|
use strum::IntoEnumIterator;
|
||||||
use strum_macros::EnumIter;
|
use strum_macros::EnumIter;
|
||||||
|
|
||||||
use ast::ExprKind;
|
|
||||||
use nac3parser::ast::{Constant, Location};
|
|
||||||
|
|
||||||
use super::*;
|
use super::*;
|
||||||
use crate::{
|
|
||||||
symbol_resolver::SymbolValue,
|
|
||||||
toplevel::numpy::unpack_ndarray_var_tys,
|
|
||||||
typecheck::typedef::{into_var_map, iter_type_vars, Mapping, TypeVarId, VarMap},
|
|
||||||
};
|
|
||||||
|
|
||||||
/// All primitive types and functions in nac3core.
|
/// All primitive types and functions in nac3core.
|
||||||
#[derive(Clone, Copy, Debug, EnumIter, PartialEq, Eq)]
|
#[derive(Clone, Copy, Debug, EnumIter, PartialEq, Eq)]
|
||||||
pub enum PrimDef {
|
pub enum PrimDef {
|
||||||
// Classes
|
|
||||||
Int32,
|
Int32,
|
||||||
Int64,
|
Int64,
|
||||||
Float,
|
Float,
|
||||||
@ -28,25 +23,17 @@ pub enum PrimDef {
|
|||||||
UInt32,
|
UInt32,
|
||||||
UInt64,
|
UInt64,
|
||||||
Option,
|
Option,
|
||||||
List,
|
OptionIsSome,
|
||||||
|
OptionIsNone,
|
||||||
|
OptionUnwrap,
|
||||||
NDArray,
|
NDArray,
|
||||||
|
NDArrayCopy,
|
||||||
// Option methods
|
NDArrayFill,
|
||||||
FunOptionIsSome,
|
FunInt32,
|
||||||
FunOptionIsNone,
|
FunInt64,
|
||||||
FunOptionUnwrap,
|
FunUInt32,
|
||||||
|
FunUInt64,
|
||||||
// Option-related functions
|
FunFloat,
|
||||||
FunSome,
|
|
||||||
|
|
||||||
// NDArray methods
|
|
||||||
FunNDArrayCopy,
|
|
||||||
FunNDArrayFill,
|
|
||||||
|
|
||||||
// Range methods
|
|
||||||
FunRangeInit,
|
|
||||||
|
|
||||||
// NumPy factory functions
|
|
||||||
FunNpNDArray,
|
FunNpNDArray,
|
||||||
FunNpEmpty,
|
FunNpEmpty,
|
||||||
FunNpZeros,
|
FunNpZeros,
|
||||||
@ -55,17 +42,26 @@ pub enum PrimDef {
|
|||||||
FunNpArray,
|
FunNpArray,
|
||||||
FunNpEye,
|
FunNpEye,
|
||||||
FunNpIdentity,
|
FunNpIdentity,
|
||||||
|
FunRound,
|
||||||
// Miscellaneous NumPy & SciPy functions
|
FunRound64,
|
||||||
FunNpRound,
|
FunNpRound,
|
||||||
|
FunRange,
|
||||||
|
FunStr,
|
||||||
|
FunBool,
|
||||||
|
FunFloor,
|
||||||
|
FunFloor64,
|
||||||
FunNpFloor,
|
FunNpFloor,
|
||||||
|
FunCeil,
|
||||||
|
FunCeil64,
|
||||||
FunNpCeil,
|
FunNpCeil,
|
||||||
|
FunLen,
|
||||||
|
FunMin,
|
||||||
FunNpMin,
|
FunNpMin,
|
||||||
FunNpMinimum,
|
FunNpMinimum,
|
||||||
FunNpArgmin,
|
FunMax,
|
||||||
FunNpMax,
|
FunNpMax,
|
||||||
FunNpMaximum,
|
FunNpMaximum,
|
||||||
FunNpArgmax,
|
FunAbs,
|
||||||
FunNpIsNan,
|
FunNpIsNan,
|
||||||
FunNpIsInf,
|
FunNpIsInf,
|
||||||
FunNpSin,
|
FunNpSin,
|
||||||
@ -103,46 +99,15 @@ pub enum PrimDef {
|
|||||||
FunNpLdExp,
|
FunNpLdExp,
|
||||||
FunNpHypot,
|
FunNpHypot,
|
||||||
FunNpNextAfter,
|
FunNpNextAfter,
|
||||||
FunNpTranspose,
|
FunSome,
|
||||||
FunNpReshape,
|
FunNpAny,
|
||||||
|
FunNpAll,
|
||||||
// Linalg functions
|
|
||||||
FunNpDot,
|
|
||||||
FunNpLinalgCholesky,
|
|
||||||
FunNpLinalgQr,
|
|
||||||
FunNpLinalgSvd,
|
|
||||||
FunNpLinalgInv,
|
|
||||||
FunNpLinalgPinv,
|
|
||||||
FunNpLinalgMatrixPower,
|
|
||||||
FunNpLinalgDet,
|
|
||||||
FunSpLinalgLu,
|
|
||||||
FunSpLinalgSchur,
|
|
||||||
FunSpLinalgHessenberg,
|
|
||||||
|
|
||||||
// Miscellaneous Python & NAC3 functions
|
|
||||||
FunInt32,
|
|
||||||
FunInt64,
|
|
||||||
FunUInt32,
|
|
||||||
FunUInt64,
|
|
||||||
FunFloat,
|
|
||||||
FunRound,
|
|
||||||
FunRound64,
|
|
||||||
FunStr,
|
|
||||||
FunBool,
|
|
||||||
FunFloor,
|
|
||||||
FunFloor64,
|
|
||||||
FunCeil,
|
|
||||||
FunCeil64,
|
|
||||||
FunLen,
|
|
||||||
FunMin,
|
|
||||||
FunMax,
|
|
||||||
FunAbs,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Associated details of a [`PrimDef`]
|
/// Associated details of a [`PrimDef`]
|
||||||
pub enum PrimDefDetails {
|
pub enum PrimDefDetails {
|
||||||
PrimFunction { name: &'static str, simple_name: &'static str },
|
PrimFunction { name: &'static str, simple_name: &'static str },
|
||||||
PrimClass { name: &'static str, get_ty_fn: fn(&PrimitiveStore) -> Type },
|
PrimClass { name: &'static str },
|
||||||
}
|
}
|
||||||
|
|
||||||
impl PrimDef {
|
impl PrimDef {
|
||||||
@ -184,17 +149,15 @@ impl PrimDef {
|
|||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn name(&self) -> &'static str {
|
pub fn name(&self) -> &'static str {
|
||||||
match self.details() {
|
match self.details() {
|
||||||
PrimDefDetails::PrimFunction { name, .. } | PrimDefDetails::PrimClass { name, .. } => {
|
PrimDefDetails::PrimFunction { name, .. } | PrimDefDetails::PrimClass { name } => name,
|
||||||
name
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the associated details of this [`PrimDef`]
|
/// Get the associated details of this [`PrimDef`]
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn details(self) -> PrimDefDetails {
|
pub fn details(self) -> PrimDefDetails {
|
||||||
fn class(name: &'static str, get_ty_fn: fn(&PrimitiveStore) -> Type) -> PrimDefDetails {
|
fn class(name: &'static str) -> PrimDefDetails {
|
||||||
PrimDefDetails::PrimClass { name, get_ty_fn }
|
PrimDefDetails::PrimClass { name }
|
||||||
}
|
}
|
||||||
|
|
||||||
fn fun(name: &'static str, simple_name: Option<&'static str>) -> PrimDefDetails {
|
fn fun(name: &'static str, simple_name: Option<&'static str>) -> PrimDefDetails {
|
||||||
@ -202,37 +165,28 @@ impl PrimDef {
|
|||||||
}
|
}
|
||||||
|
|
||||||
match self {
|
match self {
|
||||||
// Classes
|
PrimDef::Int32 => class("int32"),
|
||||||
PrimDef::Int32 => class("int32", |primitives| primitives.int32),
|
PrimDef::Int64 => class("int64"),
|
||||||
PrimDef::Int64 => class("int64", |primitives| primitives.int64),
|
PrimDef::Float => class("float"),
|
||||||
PrimDef::Float => class("float", |primitives| primitives.float),
|
PrimDef::Bool => class("bool"),
|
||||||
PrimDef::Bool => class("bool", |primitives| primitives.bool),
|
PrimDef::None => class("none"),
|
||||||
PrimDef::None => class("none", |primitives| primitives.none),
|
PrimDef::Range => class("range"),
|
||||||
PrimDef::Range => class("range", |primitives| primitives.range),
|
PrimDef::Str => class("str"),
|
||||||
PrimDef::Str => class("str", |primitives| primitives.str),
|
PrimDef::Exception => class("Exception"),
|
||||||
PrimDef::Exception => class("Exception", |primitives| primitives.exception),
|
PrimDef::UInt32 => class("uint32"),
|
||||||
PrimDef::UInt32 => class("uint32", |primitives| primitives.uint32),
|
PrimDef::UInt64 => class("uint64"),
|
||||||
PrimDef::UInt64 => class("uint64", |primitives| primitives.uint64),
|
PrimDef::Option => class("Option"),
|
||||||
PrimDef::Option => class("Option", |primitives| primitives.option),
|
PrimDef::OptionIsSome => fun("Option.is_some", Some("is_some")),
|
||||||
PrimDef::List => class("list", |primitives| primitives.list),
|
PrimDef::OptionIsNone => fun("Option.is_none", Some("is_none")),
|
||||||
PrimDef::NDArray => class("ndarray", |primitives| primitives.ndarray),
|
PrimDef::OptionUnwrap => fun("Option.unwrap", Some("unwrap")),
|
||||||
|
PrimDef::NDArray => class("ndarray"),
|
||||||
// Option methods
|
PrimDef::NDArrayCopy => fun("ndarray.copy", Some("copy")),
|
||||||
PrimDef::FunOptionIsSome => fun("Option.is_some", Some("is_some")),
|
PrimDef::NDArrayFill => fun("ndarray.fill", Some("fill")),
|
||||||
PrimDef::FunOptionIsNone => fun("Option.is_none", Some("is_none")),
|
PrimDef::FunInt32 => fun("int32", None),
|
||||||
PrimDef::FunOptionUnwrap => fun("Option.unwrap", Some("unwrap")),
|
PrimDef::FunInt64 => fun("int64", None),
|
||||||
|
PrimDef::FunUInt32 => fun("uint32", None),
|
||||||
// Option-related functions
|
PrimDef::FunUInt64 => fun("uint64", None),
|
||||||
PrimDef::FunSome => fun("Some", None),
|
PrimDef::FunFloat => fun("float", None),
|
||||||
|
|
||||||
// NDArray methods
|
|
||||||
PrimDef::FunNDArrayCopy => fun("ndarray.copy", Some("copy")),
|
|
||||||
PrimDef::FunNDArrayFill => fun("ndarray.fill", Some("fill")),
|
|
||||||
|
|
||||||
// Range methods
|
|
||||||
PrimDef::FunRangeInit => fun("range.__init__", Some("__init__")),
|
|
||||||
|
|
||||||
// NumPy factory functions
|
|
||||||
PrimDef::FunNpNDArray => fun("np_ndarray", None),
|
PrimDef::FunNpNDArray => fun("np_ndarray", None),
|
||||||
PrimDef::FunNpEmpty => fun("np_empty", None),
|
PrimDef::FunNpEmpty => fun("np_empty", None),
|
||||||
PrimDef::FunNpZeros => fun("np_zeros", None),
|
PrimDef::FunNpZeros => fun("np_zeros", None),
|
||||||
@ -241,17 +195,26 @@ impl PrimDef {
|
|||||||
PrimDef::FunNpArray => fun("np_array", None),
|
PrimDef::FunNpArray => fun("np_array", None),
|
||||||
PrimDef::FunNpEye => fun("np_eye", None),
|
PrimDef::FunNpEye => fun("np_eye", None),
|
||||||
PrimDef::FunNpIdentity => fun("np_identity", None),
|
PrimDef::FunNpIdentity => fun("np_identity", None),
|
||||||
|
PrimDef::FunRound => fun("round", None),
|
||||||
// Miscellaneous NumPy & SciPy functions
|
PrimDef::FunRound64 => fun("round64", None),
|
||||||
PrimDef::FunNpRound => fun("np_round", None),
|
PrimDef::FunNpRound => fun("np_round", None),
|
||||||
|
PrimDef::FunRange => fun("range", None),
|
||||||
|
PrimDef::FunStr => fun("str", None),
|
||||||
|
PrimDef::FunBool => fun("bool", None),
|
||||||
|
PrimDef::FunFloor => fun("floor", None),
|
||||||
|
PrimDef::FunFloor64 => fun("floor64", None),
|
||||||
PrimDef::FunNpFloor => fun("np_floor", None),
|
PrimDef::FunNpFloor => fun("np_floor", None),
|
||||||
|
PrimDef::FunCeil => fun("ceil", None),
|
||||||
|
PrimDef::FunCeil64 => fun("ceil64", None),
|
||||||
PrimDef::FunNpCeil => fun("np_ceil", None),
|
PrimDef::FunNpCeil => fun("np_ceil", None),
|
||||||
|
PrimDef::FunLen => fun("len", None),
|
||||||
|
PrimDef::FunMin => fun("min", None),
|
||||||
PrimDef::FunNpMin => fun("np_min", None),
|
PrimDef::FunNpMin => fun("np_min", None),
|
||||||
PrimDef::FunNpMinimum => fun("np_minimum", None),
|
PrimDef::FunNpMinimum => fun("np_minimum", None),
|
||||||
PrimDef::FunNpArgmin => fun("np_argmin", None),
|
PrimDef::FunMax => fun("max", None),
|
||||||
PrimDef::FunNpMax => fun("np_max", None),
|
PrimDef::FunNpMax => fun("np_max", None),
|
||||||
PrimDef::FunNpMaximum => fun("np_maximum", None),
|
PrimDef::FunNpMaximum => fun("np_maximum", None),
|
||||||
PrimDef::FunNpArgmax => fun("np_argmax", None),
|
PrimDef::FunAbs => fun("abs", None),
|
||||||
PrimDef::FunNpIsNan => fun("np_isnan", None),
|
PrimDef::FunNpIsNan => fun("np_isnan", None),
|
||||||
PrimDef::FunNpIsInf => fun("np_isinf", None),
|
PrimDef::FunNpIsInf => fun("np_isinf", None),
|
||||||
PrimDef::FunNpSin => fun("np_sin", None),
|
PrimDef::FunNpSin => fun("np_sin", None),
|
||||||
@ -289,40 +252,9 @@ impl PrimDef {
|
|||||||
PrimDef::FunNpLdExp => fun("np_ldexp", None),
|
PrimDef::FunNpLdExp => fun("np_ldexp", None),
|
||||||
PrimDef::FunNpHypot => fun("np_hypot", None),
|
PrimDef::FunNpHypot => fun("np_hypot", None),
|
||||||
PrimDef::FunNpNextAfter => fun("np_nextafter", None),
|
PrimDef::FunNpNextAfter => fun("np_nextafter", None),
|
||||||
PrimDef::FunNpTranspose => fun("np_transpose", None),
|
PrimDef::FunSome => fun("Some", None),
|
||||||
PrimDef::FunNpReshape => fun("np_reshape", None),
|
PrimDef::FunNpAny => fun("np_any", None),
|
||||||
|
PrimDef::FunNpAll => fun("np_all", None),
|
||||||
// Linalg functions
|
|
||||||
PrimDef::FunNpDot => fun("np_dot", None),
|
|
||||||
PrimDef::FunNpLinalgCholesky => fun("np_linalg_cholesky", None),
|
|
||||||
PrimDef::FunNpLinalgQr => fun("np_linalg_qr", None),
|
|
||||||
PrimDef::FunNpLinalgSvd => fun("np_linalg_svd", None),
|
|
||||||
PrimDef::FunNpLinalgInv => fun("np_linalg_inv", None),
|
|
||||||
PrimDef::FunNpLinalgPinv => fun("np_linalg_pinv", None),
|
|
||||||
PrimDef::FunNpLinalgMatrixPower => fun("np_linalg_matrix_power", None),
|
|
||||||
PrimDef::FunNpLinalgDet => fun("np_linalg_det", None),
|
|
||||||
PrimDef::FunSpLinalgLu => fun("sp_linalg_lu", None),
|
|
||||||
PrimDef::FunSpLinalgSchur => fun("sp_linalg_schur", None),
|
|
||||||
PrimDef::FunSpLinalgHessenberg => fun("sp_linalg_hessenberg", None),
|
|
||||||
|
|
||||||
// Miscellaneous Python & NAC3 functions
|
|
||||||
PrimDef::FunInt32 => fun("int32", None),
|
|
||||||
PrimDef::FunInt64 => fun("int64", None),
|
|
||||||
PrimDef::FunUInt32 => fun("uint32", None),
|
|
||||||
PrimDef::FunUInt64 => fun("uint64", None),
|
|
||||||
PrimDef::FunFloat => fun("float", None),
|
|
||||||
PrimDef::FunRound => fun("round", None),
|
|
||||||
PrimDef::FunRound64 => fun("round64", None),
|
|
||||||
PrimDef::FunStr => fun("str", None),
|
|
||||||
PrimDef::FunBool => fun("bool", None),
|
|
||||||
PrimDef::FunFloor => fun("floor", None),
|
|
||||||
PrimDef::FunFloor64 => fun("floor64", None),
|
|
||||||
PrimDef::FunCeil => fun("ceil", None),
|
|
||||||
PrimDef::FunCeil64 => fun("ceil64", None),
|
|
||||||
PrimDef::FunLen => fun("len", None),
|
|
||||||
PrimDef::FunMin => fun("min", None),
|
|
||||||
PrimDef::FunMax => fun("max", None),
|
|
||||||
PrimDef::FunAbs => fun("abs", None),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -391,9 +323,6 @@ impl TopLevelDef {
|
|||||||
r
|
r
|
||||||
}
|
}
|
||||||
),
|
),
|
||||||
TopLevelDef::Variable { name, ty, .. } => {
|
|
||||||
format!("Variable {{ name: {name:?}, ty: {:?} }}", unifier.stringify(*ty),)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -429,13 +358,7 @@ impl TopLevelComposer {
|
|||||||
});
|
});
|
||||||
let range = unifier.add_ty(TypeEnum::TObj {
|
let range = unifier.add_ty(TypeEnum::TObj {
|
||||||
obj_id: PrimDef::Range.id(),
|
obj_id: PrimDef::Range.id(),
|
||||||
fields: [
|
fields: HashMap::new(),
|
||||||
("start".into(), (int32, true)),
|
|
||||||
("stop".into(), (int32, true)),
|
|
||||||
("step".into(), (int32, true)),
|
|
||||||
]
|
|
||||||
.into_iter()
|
|
||||||
.collect(),
|
|
||||||
params: VarMap::new(),
|
params: VarMap::new(),
|
||||||
});
|
});
|
||||||
let str = unifier.add_ty(TypeEnum::TObj {
|
let str = unifier.add_ty(TypeEnum::TObj {
|
||||||
@ -476,9 +399,9 @@ impl TopLevelComposer {
|
|||||||
let option = unifier.add_ty(TypeEnum::TObj {
|
let option = unifier.add_ty(TypeEnum::TObj {
|
||||||
obj_id: PrimDef::Option.id(),
|
obj_id: PrimDef::Option.id(),
|
||||||
fields: vec![
|
fields: vec![
|
||||||
(PrimDef::FunOptionIsSome.simple_name().into(), (is_some_type_fun_ty, true)),
|
(PrimDef::OptionIsSome.simple_name().into(), (is_some_type_fun_ty, true)),
|
||||||
(PrimDef::FunOptionIsNone.simple_name().into(), (is_some_type_fun_ty, true)),
|
(PrimDef::OptionIsNone.simple_name().into(), (is_some_type_fun_ty, true)),
|
||||||
(PrimDef::FunOptionUnwrap.simple_name().into(), (unwrap_fun_ty, true)),
|
(PrimDef::OptionUnwrap.simple_name().into(), (unwrap_fun_ty, true)),
|
||||||
]
|
]
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.collect::<HashMap<_, _>>(),
|
.collect::<HashMap<_, _>>(),
|
||||||
@ -491,13 +414,6 @@ impl TopLevelComposer {
|
|||||||
_ => unreachable!(),
|
_ => unreachable!(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let list_elem_tvar = unifier.get_fresh_var(Some("list_elem".into()), None);
|
|
||||||
let list = unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: PrimDef::List.id(),
|
|
||||||
fields: Mapping::new(),
|
|
||||||
params: into_var_map([list_elem_tvar]),
|
|
||||||
});
|
|
||||||
|
|
||||||
let ndarray_dtype_tvar = unifier.get_fresh_var(Some("ndarray_dtype".into()), None);
|
let ndarray_dtype_tvar = unifier.get_fresh_var(Some("ndarray_dtype".into()), None);
|
||||||
let ndarray_ndims_tvar =
|
let ndarray_ndims_tvar =
|
||||||
unifier.get_fresh_const_generic_var(size_t_ty, Some("ndarray_ndims".into()), None);
|
unifier.get_fresh_const_generic_var(size_t_ty, Some("ndarray_ndims".into()), None);
|
||||||
@ -512,7 +428,6 @@ impl TopLevelComposer {
|
|||||||
name: "value".into(),
|
name: "value".into(),
|
||||||
ty: ndarray_dtype_tvar.ty,
|
ty: ndarray_dtype_tvar.ty,
|
||||||
default_value: None,
|
default_value: None,
|
||||||
is_vararg: false,
|
|
||||||
}],
|
}],
|
||||||
ret: none,
|
ret: none,
|
||||||
vars: into_var_map([ndarray_dtype_tvar, ndarray_ndims_tvar]),
|
vars: into_var_map([ndarray_dtype_tvar, ndarray_ndims_tvar]),
|
||||||
@ -520,8 +435,8 @@ impl TopLevelComposer {
|
|||||||
let ndarray = unifier.add_ty(TypeEnum::TObj {
|
let ndarray = unifier.add_ty(TypeEnum::TObj {
|
||||||
obj_id: PrimDef::NDArray.id(),
|
obj_id: PrimDef::NDArray.id(),
|
||||||
fields: Mapping::from([
|
fields: Mapping::from([
|
||||||
(PrimDef::FunNDArrayCopy.simple_name().into(), (ndarray_copy_fun_ty, true)),
|
(PrimDef::NDArrayCopy.simple_name().into(), (ndarray_copy_fun_ty, true)),
|
||||||
(PrimDef::FunNDArrayFill.simple_name().into(), (ndarray_fill_fun_ty, true)),
|
(PrimDef::NDArrayFill.simple_name().into(), (ndarray_fill_fun_ty, true)),
|
||||||
]),
|
]),
|
||||||
params: into_var_map([ndarray_dtype_tvar, ndarray_ndims_tvar]),
|
params: into_var_map([ndarray_dtype_tvar, ndarray_ndims_tvar]),
|
||||||
});
|
});
|
||||||
@ -540,7 +455,6 @@ impl TopLevelComposer {
|
|||||||
str,
|
str,
|
||||||
exception,
|
exception,
|
||||||
option,
|
option,
|
||||||
list,
|
|
||||||
ndarray,
|
ndarray,
|
||||||
size_t,
|
size_t,
|
||||||
};
|
};
|
||||||
@ -564,7 +478,6 @@ impl TopLevelComposer {
|
|||||||
object_id: obj_id,
|
object_id: obj_id,
|
||||||
type_vars: Vec::default(),
|
type_vars: Vec::default(),
|
||||||
fields: Vec::default(),
|
fields: Vec::default(),
|
||||||
attributes: Vec::default(),
|
|
||||||
methods: Vec::default(),
|
methods: Vec::default(),
|
||||||
ancestors: Vec::default(),
|
ancestors: Vec::default(),
|
||||||
constructor,
|
constructor,
|
||||||
@ -595,18 +508,6 @@ impl TopLevelComposer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[must_use]
|
|
||||||
pub fn make_top_level_variable_def(
|
|
||||||
name: String,
|
|
||||||
simple_name: StrRef,
|
|
||||||
ty: Type,
|
|
||||||
ty_decl: Option<Expr>,
|
|
||||||
resolver: Option<Arc<dyn SymbolResolver + Send + Sync>>,
|
|
||||||
loc: Option<Location>,
|
|
||||||
) -> TopLevelDef {
|
|
||||||
TopLevelDef::Variable { name, simple_name, ty, ty_decl, resolver, loc }
|
|
||||||
}
|
|
||||||
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn make_class_method_name(mut class_name: String, method_name: &str) -> String {
|
pub fn make_class_method_name(mut class_name: String, method_name: &str) -> String {
|
||||||
class_name.push('.');
|
class_name.push('.');
|
||||||
@ -752,16 +653,7 @@ impl TopLevelComposer {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// This function returns the fields that have been initialized in the `__init__` function of a class
|
pub fn get_all_assigned_field(stmts: &[Stmt<()>]) -> Result<HashSet<StrRef>, HashSet<String>> {
|
||||||
/// The function takes as input:
|
|
||||||
/// * `class_id`: The `object_id` of the class whose function is being evaluated (check `TopLevelDef::Class`)
|
|
||||||
/// * `definition_ast_list`: A list of ast definitions and statements defined in `TopLevelComposer`
|
|
||||||
/// * `stmts`: The body of function being parsed. Each statment is analyzed to check varaible initialization statements
|
|
||||||
pub fn get_all_assigned_field(
|
|
||||||
class_id: usize,
|
|
||||||
definition_ast_list: &Vec<DefAst>,
|
|
||||||
stmts: &[Stmt<()>],
|
|
||||||
) -> Result<HashSet<StrRef>, HashSet<String>> {
|
|
||||||
let mut result = HashSet::new();
|
let mut result = HashSet::new();
|
||||||
for s in stmts {
|
for s in stmts {
|
||||||
match &s.node {
|
match &s.node {
|
||||||
@ -797,138 +689,30 @@ impl TopLevelComposer {
|
|||||||
// TODO: do not check for For and While?
|
// TODO: do not check for For and While?
|
||||||
ast::StmtKind::For { body, orelse, .. }
|
ast::StmtKind::For { body, orelse, .. }
|
||||||
| ast::StmtKind::While { body, orelse, .. } => {
|
| ast::StmtKind::While { body, orelse, .. } => {
|
||||||
result.extend(Self::get_all_assigned_field(
|
result.extend(Self::get_all_assigned_field(body.as_slice())?);
|
||||||
class_id,
|
result.extend(Self::get_all_assigned_field(orelse.as_slice())?);
|
||||||
definition_ast_list,
|
|
||||||
body.as_slice(),
|
|
||||||
)?);
|
|
||||||
result.extend(Self::get_all_assigned_field(
|
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
orelse.as_slice(),
|
|
||||||
)?);
|
|
||||||
}
|
}
|
||||||
ast::StmtKind::If { body, orelse, .. } => {
|
ast::StmtKind::If { body, orelse, .. } => {
|
||||||
let inited_for_sure = Self::get_all_assigned_field(
|
let inited_for_sure = Self::get_all_assigned_field(body.as_slice())?
|
||||||
class_id,
|
.intersection(&Self::get_all_assigned_field(orelse.as_slice())?)
|
||||||
definition_ast_list,
|
|
||||||
body.as_slice(),
|
|
||||||
)?
|
|
||||||
.intersection(&Self::get_all_assigned_field(
|
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
orelse.as_slice(),
|
|
||||||
)?)
|
|
||||||
.copied()
|
.copied()
|
||||||
.collect::<HashSet<_>>();
|
.collect::<HashSet<_>>();
|
||||||
result.extend(inited_for_sure);
|
result.extend(inited_for_sure);
|
||||||
}
|
}
|
||||||
ast::StmtKind::Try { body, orelse, finalbody, .. } => {
|
ast::StmtKind::Try { body, orelse, finalbody, .. } => {
|
||||||
let inited_for_sure = Self::get_all_assigned_field(
|
let inited_for_sure = Self::get_all_assigned_field(body.as_slice())?
|
||||||
class_id,
|
.intersection(&Self::get_all_assigned_field(orelse.as_slice())?)
|
||||||
definition_ast_list,
|
|
||||||
body.as_slice(),
|
|
||||||
)?
|
|
||||||
.intersection(&Self::get_all_assigned_field(
|
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
orelse.as_slice(),
|
|
||||||
)?)
|
|
||||||
.copied()
|
.copied()
|
||||||
.collect::<HashSet<_>>();
|
.collect::<HashSet<_>>();
|
||||||
result.extend(inited_for_sure);
|
result.extend(inited_for_sure);
|
||||||
result.extend(Self::get_all_assigned_field(
|
result.extend(Self::get_all_assigned_field(finalbody.as_slice())?);
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
finalbody.as_slice(),
|
|
||||||
)?);
|
|
||||||
}
|
}
|
||||||
ast::StmtKind::With { body, .. } => {
|
ast::StmtKind::With { body, .. } => {
|
||||||
result.extend(Self::get_all_assigned_field(
|
result.extend(Self::get_all_assigned_field(body.as_slice())?);
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
body.as_slice(),
|
|
||||||
)?);
|
|
||||||
}
|
|
||||||
// Variables Initialized in function calls
|
|
||||||
ast::StmtKind::Expr { value, .. } => {
|
|
||||||
let ExprKind::Call { func, .. } = &value.node else {
|
|
||||||
continue;
|
|
||||||
};
|
|
||||||
let ExprKind::Attribute { value, attr, .. } = &func.node else {
|
|
||||||
continue;
|
|
||||||
};
|
|
||||||
let ExprKind::Name { id, .. } = &value.node else {
|
|
||||||
continue;
|
|
||||||
};
|
|
||||||
// Need to consider the two cases:
|
|
||||||
// Case 1) Call to class function i.e. id = `self`
|
|
||||||
// Case 2) Call to class ancestor function i.e. id = ancestor_name
|
|
||||||
// We leave checking whether function in case 2 belonged to class ancestor or not to type checker
|
|
||||||
//
|
|
||||||
// According to current handling of `self`, function definition are fixed and do not change regardless
|
|
||||||
// of which object is passed as `self` i.e. virtual polymorphism is not supported
|
|
||||||
// Therefore, we change class id for case 2 to reflect behavior of our compiler
|
|
||||||
|
|
||||||
let class_name = if *id == "self".into() {
|
|
||||||
let ast::StmtKind::ClassDef { name, .. } =
|
|
||||||
&definition_ast_list[class_id].1.as_ref().unwrap().node
|
|
||||||
else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
name
|
|
||||||
} else {
|
|
||||||
id
|
|
||||||
};
|
|
||||||
|
|
||||||
let parent_method = definition_ast_list.iter().find_map(|def| {
|
|
||||||
let (
|
|
||||||
class_def,
|
|
||||||
Some(ast::Located {
|
|
||||||
node: ast::StmtKind::ClassDef { name, body, .. },
|
|
||||||
..
|
|
||||||
}),
|
|
||||||
) = &def
|
|
||||||
else {
|
|
||||||
return None;
|
|
||||||
};
|
|
||||||
let TopLevelDef::Class { object_id: class_id, .. } = &*class_def.read()
|
|
||||||
else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
if name == class_name {
|
|
||||||
body.iter().find_map(|m| {
|
|
||||||
let ast::StmtKind::FunctionDef { name, body, .. } = &m.node else {
|
|
||||||
return None;
|
|
||||||
};
|
|
||||||
if *name == *attr {
|
|
||||||
return Some((body.clone(), class_id.0));
|
|
||||||
}
|
|
||||||
None
|
|
||||||
})
|
|
||||||
} else {
|
|
||||||
None
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
// If method body is none then method does not exist
|
|
||||||
if let Some((method_body, class_id)) = parent_method {
|
|
||||||
result.extend(Self::get_all_assigned_field(
|
|
||||||
class_id,
|
|
||||||
definition_ast_list,
|
|
||||||
method_body.as_slice(),
|
|
||||||
)?);
|
|
||||||
} else {
|
|
||||||
return Err(HashSet::from([format!(
|
|
||||||
"{}.{} not found in class {class_name} at {}",
|
|
||||||
*id, *attr, value.location
|
|
||||||
)]));
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
ast::StmtKind::Pass { .. }
|
ast::StmtKind::Pass { .. }
|
||||||
| ast::StmtKind::Assert { .. }
|
| ast::StmtKind::Assert { .. }
|
||||||
| ast::StmtKind::AnnAssign { .. } => {}
|
| ast::StmtKind::Expr { .. } => {}
|
||||||
|
|
||||||
_ => {
|
_ => {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
@ -1107,9 +891,7 @@ pub fn arraylike_flatten_element_type(unifier: &mut Unifier, ty: Type) -> Type {
|
|||||||
unpack_ndarray_var_tys(unifier, ty).0
|
unpack_ndarray_var_tys(unifier, ty).0
|
||||||
}
|
}
|
||||||
|
|
||||||
TypeEnum::TObj { obj_id, params, .. } if *obj_id == PrimDef::List.id() => {
|
TypeEnum::TList { ty } => arraylike_flatten_element_type(unifier, *ty),
|
||||||
arraylike_flatten_element_type(unifier, iter_type_vars(params).next().unwrap().ty)
|
|
||||||
}
|
|
||||||
_ => ty,
|
_ => ty,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1130,9 +912,7 @@ pub fn arraylike_get_ndims(unifier: &mut Unifier, ty: Type) -> u64 {
|
|||||||
u64::try_from(values[0].clone()).unwrap()
|
u64::try_from(values[0].clone()).unwrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
TypeEnum::TObj { obj_id, params, .. } if *obj_id == PrimDef::List.id() => {
|
TypeEnum::TList { ty } => arraylike_get_ndims(unifier, *ty) + 1,
|
||||||
arraylike_get_ndims(unifier, iter_type_vars(params).next().unwrap().ty) + 1
|
|
||||||
}
|
|
||||||
_ => 0,
|
_ => 0,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -6,23 +6,23 @@ use std::{
|
|||||||
sync::Arc,
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
|
||||||
use inkwell::values::BasicValueEnum;
|
use super::codegen::CodeGenContext;
|
||||||
use itertools::Itertools;
|
use super::typecheck::type_inferencer::PrimitiveStore;
|
||||||
use parking_lot::RwLock;
|
use super::typecheck::typedef::{
|
||||||
|
FunSignature, FuncArg, SharedUnifier, Type, TypeEnum, Unifier, VarMap,
|
||||||
use nac3parser::ast::{self, Expr, Location, Stmt, StrRef};
|
};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{CodeGenContext, CodeGenerator},
|
codegen::CodeGenerator,
|
||||||
symbol_resolver::{SymbolResolver, ValueEnum},
|
symbol_resolver::{SymbolResolver, ValueEnum},
|
||||||
typecheck::{
|
typecheck::{
|
||||||
type_inferencer::{CodeLocation, PrimitiveStore},
|
type_inferencer::CodeLocation,
|
||||||
typedef::{
|
typedef::{CallId, TypeVarId},
|
||||||
CallId, FunSignature, FuncArg, SharedUnifier, Type, TypeEnum, TypeVarId, Unifier,
|
|
||||||
VarMap,
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use inkwell::values::BasicValueEnum;
|
||||||
|
use itertools::Itertools;
|
||||||
|
use nac3parser::ast::{self, Location, Stmt, StrRef};
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
|
||||||
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Copy, Hash, Debug)]
|
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Copy, Hash, Debug)]
|
||||||
pub struct DefinitionId(pub usize);
|
pub struct DefinitionId(pub usize);
|
||||||
@ -103,10 +103,6 @@ pub enum TopLevelDef {
|
|||||||
///
|
///
|
||||||
/// Name and type is mutable.
|
/// Name and type is mutable.
|
||||||
fields: Vec<(StrRef, Type, bool)>,
|
fields: Vec<(StrRef, Type, bool)>,
|
||||||
/// Class Attributes.
|
|
||||||
///
|
|
||||||
/// Name, type, value.
|
|
||||||
attributes: Vec<(StrRef, Type, ast::Constant)>,
|
|
||||||
/// Class methods, pointing to the corresponding function definition.
|
/// Class methods, pointing to the corresponding function definition.
|
||||||
methods: Vec<(StrRef, Type, DefinitionId)>,
|
methods: Vec<(StrRef, Type, DefinitionId)>,
|
||||||
/// Ancestor classes, including itself.
|
/// Ancestor classes, including itself.
|
||||||
@ -148,25 +144,6 @@ pub enum TopLevelDef {
|
|||||||
/// Definition location.
|
/// Definition location.
|
||||||
loc: Option<Location>,
|
loc: Option<Location>,
|
||||||
},
|
},
|
||||||
Variable {
|
|
||||||
/// Qualified name of the global variable, should be unique globally.
|
|
||||||
name: String,
|
|
||||||
|
|
||||||
/// Simple name, the same as in method/function definition.
|
|
||||||
simple_name: StrRef,
|
|
||||||
|
|
||||||
/// Type of the global variable.
|
|
||||||
ty: Type,
|
|
||||||
|
|
||||||
/// The declared type of the global variable, or [`None`] if no type annotation is provided.
|
|
||||||
ty_decl: Option<Expr>,
|
|
||||||
|
|
||||||
/// Symbol resolver of the module defined the class.
|
|
||||||
resolver: Option<Arc<dyn SymbolResolver + Send + Sync>>,
|
|
||||||
|
|
||||||
/// Definition location.
|
|
||||||
loc: Option<Location>,
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct TopLevelContext {
|
pub struct TopLevelContext {
|
||||||
|
@ -1,5 +1,3 @@
|
|||||||
use itertools::Itertools;
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
toplevel::helper::PrimDef,
|
toplevel::helper::PrimDef,
|
||||||
typecheck::{
|
typecheck::{
|
||||||
@ -7,6 +5,7 @@ use crate::{
|
|||||||
typedef::{Type, TypeEnum, TypeVarId, Unifier, VarMap},
|
typedef::{Type, TypeEnum, TypeVarId, Unifier, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use itertools::Itertools;
|
||||||
|
|
||||||
/// Creates a `ndarray` [`Type`] with the given type arguments.
|
/// Creates a `ndarray` [`Type`] with the given type arguments.
|
||||||
///
|
///
|
||||||
|
@ -5,7 +5,7 @@ expression: res_vec
|
|||||||
[
|
[
|
||||||
"Class {\nname: \"Generic_A\",\nancestors: [\"Generic_A[V]\", \"B\"],\nfields: [\"aa\", \"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"foo\", \"fn[[b:T], none]\"), (\"fun\", \"fn[[a:int32], V]\")],\ntype_vars: [\"V\"]\n}\n",
|
"Class {\nname: \"Generic_A\",\nancestors: [\"Generic_A[V]\", \"B\"],\nfields: [\"aa\", \"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"foo\", \"fn[[b:T], none]\"), (\"fun\", \"fn[[a:int32], V]\")],\ntype_vars: [\"V\"]\n}\n",
|
||||||
"Function {\nname: \"Generic_A.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"Generic_A.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"Generic_A.fun\",\nsig: \"fn[[a:int32], V]\",\nvar_id: [TypeVarId(241)]\n}\n",
|
"Function {\nname: \"Generic_A.fun\",\nsig: \"fn[[a:int32], V]\",\nvar_id: [TypeVarId(239)]\n}\n",
|
||||||
"Class {\nname: \"B\",\nancestors: [\"B\"],\nfields: [\"aa\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"foo\", \"fn[[b:T], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"B\",\nancestors: [\"B\"],\nfields: [\"aa\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"foo\", \"fn[[b:T], none]\")],\ntype_vars: []\n}\n",
|
||||||
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"B.foo\",\nsig: \"fn[[b:T], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.foo\",\nsig: \"fn[[b:T], none]\",\nvar_id: []\n}\n",
|
||||||
|
@ -7,7 +7,7 @@ expression: res_vec
|
|||||||
"Function {\nname: \"A.__init__\",\nsig: \"fn[[t:T], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.__init__\",\nsig: \"fn[[t:T], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"A.foo\",\nsig: \"fn[[c:C], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.foo\",\nsig: \"fn[[c:C], none]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"B\",\nancestors: [\"B[typevar230]\", \"A[float]\"],\nfields: [\"a\", \"b\", \"c\", \"d\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\"), (\"foo\", \"fn[[c:C], none]\")],\ntype_vars: [\"typevar230\"]\n}\n",
|
"Class {\nname: \"B\",\nancestors: [\"B[typevar228]\", \"A[float]\"],\nfields: [\"a\", \"b\", \"c\", \"d\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\"), (\"foo\", \"fn[[c:C], none]\")],\ntype_vars: [\"typevar228\"]\n}\n",
|
||||||
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"B.fun\",\nsig: \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.fun\",\nsig: \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"C\",\nancestors: [\"C\", \"B[bool]\", \"A[float]\"],\nfields: [\"a\", \"b\", \"c\", \"d\", \"e\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\"), (\"foo\", \"fn[[c:C], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"C\",\nancestors: [\"C\", \"B[bool]\", \"A[float]\"],\nfields: [\"a\", \"b\", \"c\", \"d\", \"e\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:int32, b:T], list[virtual[B[bool]]]]\"), (\"foo\", \"fn[[c:C], none]\")],\ntype_vars: []\n}\n",
|
||||||
|
@ -5,8 +5,8 @@ expression: res_vec
|
|||||||
[
|
[
|
||||||
"Function {\nname: \"foo\",\nsig: \"fn[[a:list[int32], b:tuple[T, float]], A[B, bool]]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"foo\",\nsig: \"fn[[a:list[int32], b:tuple[T, float]], A[B, bool]]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"A\",\nancestors: [\"A[T, V]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[v:V], none]\"), (\"fun\", \"fn[[a:T], V]\")],\ntype_vars: [\"T\", \"V\"]\n}\n",
|
"Class {\nname: \"A\",\nancestors: [\"A[T, V]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[v:V], none]\"), (\"fun\", \"fn[[a:T], V]\")],\ntype_vars: [\"T\", \"V\"]\n}\n",
|
||||||
"Function {\nname: \"A.__init__\",\nsig: \"fn[[v:V], none]\",\nvar_id: [TypeVarId(243)]\n}\n",
|
"Function {\nname: \"A.__init__\",\nsig: \"fn[[v:V], none]\",\nvar_id: [TypeVarId(241)]\n}\n",
|
||||||
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:T], V]\",\nvar_id: [TypeVarId(248)]\n}\n",
|
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:T], V]\",\nvar_id: [TypeVarId(246)]\n}\n",
|
||||||
"Function {\nname: \"gfun\",\nsig: \"fn[[a:A[list[float], int32]], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"gfun\",\nsig: \"fn[[a:A[list[float], int32]], none]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"B\",\nancestors: [\"B\"],\nfields: [],\nmethods: [(\"__init__\", \"fn[[], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"B\",\nancestors: [\"B\"],\nfields: [],\nmethods: [(\"__init__\", \"fn[[], none]\")],\ntype_vars: []\n}\n",
|
||||||
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
|
@ -3,7 +3,7 @@ source: nac3core/src/toplevel/test.rs
|
|||||||
expression: res_vec
|
expression: res_vec
|
||||||
---
|
---
|
||||||
[
|
[
|
||||||
"Class {\nname: \"A\",\nancestors: [\"A[typevar229, typevar230]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[a:A[float, bool], b:B], none]\"), (\"fun\", \"fn[[a:A[float, bool]], A[bool, int32]]\")],\ntype_vars: [\"typevar229\", \"typevar230\"]\n}\n",
|
"Class {\nname: \"A\",\nancestors: [\"A[typevar227, typevar228]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[a:A[float, bool], b:B], none]\"), (\"fun\", \"fn[[a:A[float, bool]], A[bool, int32]]\")],\ntype_vars: [\"typevar227\", \"typevar228\"]\n}\n",
|
||||||
"Function {\nname: \"A.__init__\",\nsig: \"fn[[a:A[float, bool], b:B], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.__init__\",\nsig: \"fn[[a:A[float, bool], b:B], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:A[float, bool]], A[bool, int32]]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.fun\",\nsig: \"fn[[a:A[float, bool]], A[bool, int32]]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"B\",\nancestors: [\"B\", \"A[int64, bool]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:A[float, bool]], A[bool, int32]]\"), (\"foo\", \"fn[[b:B], B]\"), (\"bar\", \"fn[[a:A[list[B], int32]], tuple[A[virtual[A[B, int32]], bool], B]]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"B\",\nancestors: [\"B\", \"A[int64, bool]\"],\nfields: [\"a\", \"b\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[a:A[float, bool]], A[bool, int32]]\"), (\"foo\", \"fn[[b:B], B]\"), (\"bar\", \"fn[[a:A[list[B], int32]], tuple[A[virtual[A[B, int32]], bool], B]]\")],\ntype_vars: []\n}\n",
|
||||||
|
@ -6,12 +6,12 @@ expression: res_vec
|
|||||||
"Class {\nname: \"A\",\nancestors: [\"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"A\",\nancestors: [\"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
||||||
"Function {\nname: \"A.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"A.fun\",\nsig: \"fn[[b:B], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"A.fun\",\nsig: \"fn[[b:B], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"A.foo\",\nsig: \"fn[[a:T, b:V], none]\",\nvar_id: [TypeVarId(249)]\n}\n",
|
"Function {\nname: \"A.foo\",\nsig: \"fn[[a:T, b:V], none]\",\nvar_id: [TypeVarId(247)]\n}\n",
|
||||||
"Class {\nname: \"B\",\nancestors: [\"B\", \"C\", \"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"B\",\nancestors: [\"B\", \"C\", \"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
||||||
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"B.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Class {\nname: \"C\",\nancestors: [\"C\", \"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
"Class {\nname: \"C\",\nancestors: [\"C\", \"A\"],\nfields: [\"a\"],\nmethods: [(\"__init__\", \"fn[[], none]\"), (\"fun\", \"fn[[b:B], none]\"), (\"foo\", \"fn[[a:T, b:V], none]\")],\ntype_vars: []\n}\n",
|
||||||
"Function {\nname: \"C.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"C.__init__\",\nsig: \"fn[[], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"C.fun\",\nsig: \"fn[[b:B], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"C.fun\",\nsig: \"fn[[b:B], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"foo\",\nsig: \"fn[[a:A], none]\",\nvar_id: []\n}\n",
|
"Function {\nname: \"foo\",\nsig: \"fn[[a:A], none]\",\nvar_id: []\n}\n",
|
||||||
"Function {\nname: \"ff\",\nsig: \"fn[[a:T], V]\",\nvar_id: [TypeVarId(257)]\n}\n",
|
"Function {\nname: \"ff\",\nsig: \"fn[[a:T], V]\",\nvar_id: [TypeVarId(255)]\n}\n",
|
||||||
]
|
]
|
||||||
|
@ -1,24 +1,20 @@
|
|||||||
use std::{collections::HashMap, sync::Arc};
|
|
||||||
|
|
||||||
use indoc::indoc;
|
|
||||||
use parking_lot::Mutex;
|
|
||||||
use test_case::test_case;
|
|
||||||
|
|
||||||
use nac3parser::{
|
|
||||||
ast::{fold::Fold, FileName},
|
|
||||||
parser::parse_program,
|
|
||||||
};
|
|
||||||
|
|
||||||
use super::*;
|
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::CodeGenContext,
|
codegen::CodeGenContext,
|
||||||
symbol_resolver::{SymbolResolver, ValueEnum},
|
symbol_resolver::{SymbolResolver, ValueEnum},
|
||||||
toplevel::{helper::PrimDef, DefinitionId},
|
toplevel::DefinitionId,
|
||||||
typecheck::{
|
typecheck::{
|
||||||
type_inferencer::PrimitiveStore,
|
type_inferencer::PrimitiveStore,
|
||||||
typedef::{into_var_map, Type, Unifier},
|
typedef::{Type, Unifier},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use indoc::indoc;
|
||||||
|
use nac3parser::ast::FileName;
|
||||||
|
use nac3parser::{ast::fold::Fold, parser::parse_program};
|
||||||
|
use parking_lot::Mutex;
|
||||||
|
use std::{collections::HashMap, sync::Arc};
|
||||||
|
use test_case::test_case;
|
||||||
|
|
||||||
|
use super::*;
|
||||||
|
|
||||||
struct ResolverInternal {
|
struct ResolverInternal {
|
||||||
id_to_type: Mutex<HashMap<StrRef, Type>>,
|
id_to_type: Mutex<HashMap<StrRef, Type>>,
|
||||||
@ -65,7 +61,6 @@ impl SymbolResolver for Resolver {
|
|||||||
&self,
|
&self,
|
||||||
_: StrRef,
|
_: StrRef,
|
||||||
_: &mut CodeGenContext<'ctx, '_>,
|
_: &mut CodeGenContext<'ctx, '_>,
|
||||||
_: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>> {
|
) -> Option<ValueEnum<'ctx>> {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
@ -121,8 +116,7 @@ impl SymbolResolver for Resolver {
|
|||||||
"register"
|
"register"
|
||||||
)]
|
)]
|
||||||
fn test_simple_register(source: Vec<&str>) {
|
fn test_simple_register(source: Vec<&str>) {
|
||||||
let mut composer =
|
let mut composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 64).0;
|
||||||
TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 64).0;
|
|
||||||
|
|
||||||
for s in source {
|
for s in source {
|
||||||
let ast = parse_program(s, FileName::default()).unwrap();
|
let ast = parse_program(s, FileName::default()).unwrap();
|
||||||
@ -142,8 +136,7 @@ fn test_simple_register(source: Vec<&str>) {
|
|||||||
"register"
|
"register"
|
||||||
)]
|
)]
|
||||||
fn test_simple_register_without_constructor(source: &str) {
|
fn test_simple_register_without_constructor(source: &str) {
|
||||||
let mut composer =
|
let mut composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 64).0;
|
||||||
TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 64).0;
|
|
||||||
let ast = parse_program(source, FileName::default()).unwrap();
|
let ast = parse_program(source, FileName::default()).unwrap();
|
||||||
let ast = ast[0].clone();
|
let ast = ast[0].clone();
|
||||||
composer.register_top_level(ast, None, "", true).unwrap();
|
composer.register_top_level(ast, None, "", true).unwrap();
|
||||||
@ -177,8 +170,7 @@ fn test_simple_register_without_constructor(source: &str) {
|
|||||||
"function compose"
|
"function compose"
|
||||||
)]
|
)]
|
||||||
fn test_simple_function_analyze(source: &[&str], tys: &[&str], names: &[&str]) {
|
fn test_simple_function_analyze(source: &[&str], tys: &[&str], names: &[&str]) {
|
||||||
let mut composer =
|
let mut composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 64).0;
|
||||||
TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 64).0;
|
|
||||||
|
|
||||||
let internal_resolver = Arc::new(ResolverInternal {
|
let internal_resolver = Arc::new(ResolverInternal {
|
||||||
id_to_def: Mutex::default(),
|
id_to_def: Mutex::default(),
|
||||||
@ -526,8 +518,7 @@ fn test_simple_function_analyze(source: &[&str], tys: &[&str], names: &[&str]) {
|
|||||||
)]
|
)]
|
||||||
fn test_analyze(source: &[&str], res: &[&str]) {
|
fn test_analyze(source: &[&str], res: &[&str]) {
|
||||||
let print = false;
|
let print = false;
|
||||||
let mut composer =
|
let mut composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 64).0;
|
||||||
TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 64).0;
|
|
||||||
|
|
||||||
let internal_resolver = make_internal_resolver_with_tvar(
|
let internal_resolver = make_internal_resolver_with_tvar(
|
||||||
vec![
|
vec![
|
||||||
@ -704,8 +695,7 @@ fn test_analyze(source: &[&str], res: &[&str]) {
|
|||||||
)]
|
)]
|
||||||
fn test_inference(source: Vec<&str>, res: &[&str]) {
|
fn test_inference(source: Vec<&str>, res: &[&str]) {
|
||||||
let print = true;
|
let print = true;
|
||||||
let mut composer =
|
let mut composer = TopLevelComposer::new(Vec::new(), ComposerConfig::default(), 64).0;
|
||||||
TopLevelComposer::new(Vec::new(), Vec::new(), ComposerConfig::default(), 64).0;
|
|
||||||
|
|
||||||
let internal_resolver = make_internal_resolver_with_tvar(
|
let internal_resolver = make_internal_resolver_with_tvar(
|
||||||
vec![
|
vec![
|
||||||
@ -785,15 +775,8 @@ fn make_internal_resolver_with_tvar(
|
|||||||
unifier: &mut Unifier,
|
unifier: &mut Unifier,
|
||||||
print: bool,
|
print: bool,
|
||||||
) -> Arc<ResolverInternal> {
|
) -> Arc<ResolverInternal> {
|
||||||
let list_elem_tvar = unifier.get_fresh_var(Some("list_elem".into()), None);
|
|
||||||
let list = unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: PrimDef::List.id(),
|
|
||||||
fields: HashMap::new(),
|
|
||||||
params: into_var_map([list_elem_tvar]),
|
|
||||||
});
|
|
||||||
|
|
||||||
let res: Arc<ResolverInternal> = ResolverInternal {
|
let res: Arc<ResolverInternal> = ResolverInternal {
|
||||||
id_to_def: Mutex::new(HashMap::from([("list".into(), PrimDef::List.id())])),
|
id_to_def: Mutex::default(),
|
||||||
id_to_type: tvars
|
id_to_type: tvars
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.map(|(name, range)| {
|
.map(|(name, range)| {
|
||||||
@ -807,7 +790,7 @@ fn make_internal_resolver_with_tvar(
|
|||||||
})
|
})
|
||||||
.collect::<HashMap<_, _>>()
|
.collect::<HashMap<_, _>>()
|
||||||
.into(),
|
.into(),
|
||||||
class_names: Mutex::new(HashMap::from([("list".into(), list)])),
|
class_names: Mutex::default(),
|
||||||
}
|
}
|
||||||
.into();
|
.into();
|
||||||
if print {
|
if print {
|
||||||
|
@ -1,13 +1,8 @@
|
|||||||
use strum::IntoEnumIterator;
|
|
||||||
|
|
||||||
use nac3parser::ast::Constant;
|
|
||||||
|
|
||||||
use super::*;
|
use super::*;
|
||||||
use crate::{
|
use crate::symbol_resolver::SymbolValue;
|
||||||
symbol_resolver::SymbolValue,
|
use crate::toplevel::helper::PrimDef;
|
||||||
toplevel::helper::{PrimDef, PrimDefDetails},
|
use crate::typecheck::typedef::VarMap;
|
||||||
typecheck::typedef::VarMap,
|
use nac3parser::ast::Constant;
|
||||||
};
|
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
pub enum TypeAnnotation {
|
pub enum TypeAnnotation {
|
||||||
@ -23,6 +18,7 @@ pub enum TypeAnnotation {
|
|||||||
TypeVar(Type),
|
TypeVar(Type),
|
||||||
/// A `Literal` allowing a subset of literals.
|
/// A `Literal` allowing a subset of literals.
|
||||||
Literal(Vec<Constant>),
|
Literal(Vec<Constant>),
|
||||||
|
List(Box<TypeAnnotation>),
|
||||||
Tuple(Vec<TypeAnnotation>),
|
Tuple(Vec<TypeAnnotation>),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -55,6 +51,7 @@ impl TypeAnnotation {
|
|||||||
format!("Literal({})", values.iter().map(|v| format!("{v:?}")).join(", "))
|
format!("Literal({})", values.iter().map(|v| format!("{v:?}")).join(", "))
|
||||||
}
|
}
|
||||||
Virtual(ty) => format!("virtual[{}]", ty.stringify(unifier)),
|
Virtual(ty) => format!("virtual[{}]", ty.stringify(unifier)),
|
||||||
|
List(ty) => format!("list[{}]", ty.stringify(unifier)),
|
||||||
Tuple(types) => {
|
Tuple(types) => {
|
||||||
format!(
|
format!(
|
||||||
"tuple[{}]",
|
"tuple[{}]",
|
||||||
@ -148,7 +145,9 @@ pub fn parse_ast_to_type_annotation_kinds<T, S: std::hash::BuildHasher + Clone>(
|
|||||||
slice: &ast::Expr<T>,
|
slice: &ast::Expr<T>,
|
||||||
unifier: &mut Unifier,
|
unifier: &mut Unifier,
|
||||||
mut locked: HashMap<DefinitionId, Vec<Type>, S>| {
|
mut locked: HashMap<DefinitionId, Vec<Type>, S>| {
|
||||||
if ["virtual".into(), "Generic".into(), "tuple".into(), "Option".into()].contains(id) {
|
if ["virtual".into(), "Generic".into(), "list".into(), "tuple".into(), "Option".into()]
|
||||||
|
.contains(id)
|
||||||
|
{
|
||||||
return Err(HashSet::from([format!(
|
return Err(HashSet::from([format!(
|
||||||
"keywords cannot be class name (at {})",
|
"keywords cannot be class name (at {})",
|
||||||
expr.location
|
expr.location
|
||||||
@ -237,6 +236,23 @@ pub fn parse_ast_to_type_annotation_kinds<T, S: std::hash::BuildHasher + Clone>(
|
|||||||
Ok(TypeAnnotation::Virtual(def.into()))
|
Ok(TypeAnnotation::Virtual(def.into()))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// list
|
||||||
|
ast::ExprKind::Subscript { value, slice, .. }
|
||||||
|
if {
|
||||||
|
matches!(&value.node, ast::ExprKind::Name { id, .. } if id == &"list".into())
|
||||||
|
} =>
|
||||||
|
{
|
||||||
|
let def_ann = parse_ast_to_type_annotation_kinds(
|
||||||
|
resolver,
|
||||||
|
top_level_defs,
|
||||||
|
unifier,
|
||||||
|
primitives,
|
||||||
|
slice.as_ref(),
|
||||||
|
locked,
|
||||||
|
)?;
|
||||||
|
Ok(TypeAnnotation::List(def_ann.into()))
|
||||||
|
}
|
||||||
|
|
||||||
// option
|
// option
|
||||||
ast::ExprKind::Subscript { value, slice, .. }
|
ast::ExprKind::Subscript { value, slice, .. }
|
||||||
if {
|
if {
|
||||||
@ -362,7 +378,6 @@ pub fn parse_ast_to_type_annotation_kinds<T, S: std::hash::BuildHasher + Clone>(
|
|||||||
pub fn get_type_from_type_annotation_kinds(
|
pub fn get_type_from_type_annotation_kinds(
|
||||||
top_level_defs: &[Arc<RwLock<TopLevelDef>>],
|
top_level_defs: &[Arc<RwLock<TopLevelDef>>],
|
||||||
unifier: &mut Unifier,
|
unifier: &mut Unifier,
|
||||||
primitives: &PrimitiveStore,
|
|
||||||
ann: &TypeAnnotation,
|
ann: &TypeAnnotation,
|
||||||
subst_list: &mut Option<Vec<Type>>,
|
subst_list: &mut Option<Vec<Type>>,
|
||||||
) -> Result<Type, HashSet<String>> {
|
) -> Result<Type, HashSet<String>> {
|
||||||
@ -385,43 +400,10 @@ pub fn get_type_from_type_annotation_kinds(
|
|||||||
let param_ty = params
|
let param_ty = params
|
||||||
.iter()
|
.iter()
|
||||||
.map(|x| {
|
.map(|x| {
|
||||||
get_type_from_type_annotation_kinds(
|
get_type_from_type_annotation_kinds(top_level_defs, unifier, x, subst_list)
|
||||||
top_level_defs,
|
|
||||||
unifier,
|
|
||||||
primitives,
|
|
||||||
x,
|
|
||||||
subst_list,
|
|
||||||
)
|
|
||||||
})
|
})
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
.collect::<Result<Vec<_>, _>>()?;
|
||||||
|
|
||||||
let ty = if let Some(prim_def) = PrimDef::iter().find(|prim| prim.id() == *obj_id) {
|
|
||||||
// Primitive TopLevelDefs do not contain all fields that are present in their Type
|
|
||||||
// counterparts, so directly perform subst on the Type instead.
|
|
||||||
|
|
||||||
let PrimDefDetails::PrimClass { get_ty_fn, .. } = prim_def.details() else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
let base_ty = get_ty_fn(primitives);
|
|
||||||
let params =
|
|
||||||
if let TypeEnum::TObj { params, .. } = &*unifier.get_ty_immutable(base_ty) {
|
|
||||||
params.clone()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
unifier
|
|
||||||
.subst(
|
|
||||||
get_ty_fn(primitives),
|
|
||||||
¶ms
|
|
||||||
.iter()
|
|
||||||
.zip(param_ty)
|
|
||||||
.map(|(obj_tv, param)| (*obj_tv.0, param))
|
|
||||||
.collect(),
|
|
||||||
)
|
|
||||||
.unwrap_or(base_ty)
|
|
||||||
} else {
|
|
||||||
let subst = {
|
let subst = {
|
||||||
// check for compatible range
|
// check for compatible range
|
||||||
// TODO: if allow type var to be applied(now this disallowed in the parse_to_type_annotation), need more check
|
// TODO: if allow type var to be applied(now this disallowed in the parse_to_type_annotation), need more check
|
||||||
@ -463,15 +445,12 @@ pub fn get_type_from_type_annotation_kinds(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
TypeEnum::TVar {
|
TypeEnum::TVar { id, range, name, loc, is_const_generic: true, .. } => {
|
||||||
id, range, name, loc, is_const_generic: true, ..
|
|
||||||
} => {
|
|
||||||
let ty = range[0];
|
let ty = range[0];
|
||||||
let ok: bool = {
|
let ok: bool = {
|
||||||
// create a temp type var and unify to check compatibility
|
// create a temp type var and unify to check compatibility
|
||||||
p == *tvar || {
|
p == *tvar || {
|
||||||
let temp =
|
let temp = unifier.get_fresh_const_generic_var(ty, *name, *loc);
|
||||||
unifier.get_fresh_const_generic_var(ty, *name, *loc);
|
|
||||||
unifier.unify(temp.ty, p).is_ok()
|
unifier.unify(temp.ty, p).is_ok()
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -491,7 +470,6 @@ pub fn get_type_from_type_annotation_kinds(
|
|||||||
}
|
}
|
||||||
result
|
result
|
||||||
};
|
};
|
||||||
// Class Attributes keep a copy with Class Definition and are not added to objects
|
|
||||||
let mut tobj_fields = methods
|
let mut tobj_fields = methods
|
||||||
.iter()
|
.iter()
|
||||||
.map(|(name, ty, _)| {
|
.map(|(name, ty, _)| {
|
||||||
@ -510,16 +488,11 @@ pub fn get_type_from_type_annotation_kinds(
|
|||||||
fields: tobj_fields,
|
fields: tobj_fields,
|
||||||
params: subst,
|
params: subst,
|
||||||
});
|
});
|
||||||
|
|
||||||
if need_subst {
|
if need_subst {
|
||||||
if let Some(wl) = subst_list.as_mut() {
|
if let Some(wl) = subst_list.as_mut() {
|
||||||
wl.push(ty);
|
wl.push(ty);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ty
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok(ty)
|
Ok(ty)
|
||||||
}
|
}
|
||||||
TypeAnnotation::Primitive(ty) | TypeAnnotation::TypeVar(ty) => Ok(*ty),
|
TypeAnnotation::Primitive(ty) | TypeAnnotation::TypeVar(ty) => Ok(*ty),
|
||||||
@ -537,26 +510,28 @@ pub fn get_type_from_type_annotation_kinds(
|
|||||||
let ty = get_type_from_type_annotation_kinds(
|
let ty = get_type_from_type_annotation_kinds(
|
||||||
top_level_defs,
|
top_level_defs,
|
||||||
unifier,
|
unifier,
|
||||||
primitives,
|
|
||||||
ty.as_ref(),
|
ty.as_ref(),
|
||||||
subst_list,
|
subst_list,
|
||||||
)?;
|
)?;
|
||||||
Ok(unifier.add_ty(TypeEnum::TVirtual { ty }))
|
Ok(unifier.add_ty(TypeEnum::TVirtual { ty }))
|
||||||
}
|
}
|
||||||
|
TypeAnnotation::List(ty) => {
|
||||||
|
let ty = get_type_from_type_annotation_kinds(
|
||||||
|
top_level_defs,
|
||||||
|
unifier,
|
||||||
|
ty.as_ref(),
|
||||||
|
subst_list,
|
||||||
|
)?;
|
||||||
|
Ok(unifier.add_ty(TypeEnum::TList { ty }))
|
||||||
|
}
|
||||||
TypeAnnotation::Tuple(tys) => {
|
TypeAnnotation::Tuple(tys) => {
|
||||||
let tys = tys
|
let tys = tys
|
||||||
.iter()
|
.iter()
|
||||||
.map(|x| {
|
.map(|x| {
|
||||||
get_type_from_type_annotation_kinds(
|
get_type_from_type_annotation_kinds(top_level_defs, unifier, x, subst_list)
|
||||||
top_level_defs,
|
|
||||||
unifier,
|
|
||||||
primitives,
|
|
||||||
x,
|
|
||||||
subst_list,
|
|
||||||
)
|
|
||||||
})
|
})
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
.collect::<Result<Vec<_>, _>>()?;
|
||||||
Ok(unifier.add_ty(TypeEnum::TTuple { ty: tys, is_vararg_ctx: false }))
|
Ok(unifier.add_ty(TypeEnum::TTuple { ty: tys }))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -589,7 +564,7 @@ pub fn get_type_var_contained_in_type_annotation(ann: &TypeAnnotation) -> Vec<Ty
|
|||||||
let mut result: Vec<TypeAnnotation> = Vec::new();
|
let mut result: Vec<TypeAnnotation> = Vec::new();
|
||||||
match ann {
|
match ann {
|
||||||
TypeAnnotation::TypeVar(..) => result.push(ann.clone()),
|
TypeAnnotation::TypeVar(..) => result.push(ann.clone()),
|
||||||
TypeAnnotation::Virtual(ann) => {
|
TypeAnnotation::Virtual(ann) | TypeAnnotation::List(ann) => {
|
||||||
result.extend(get_type_var_contained_in_type_annotation(ann.as_ref()));
|
result.extend(get_type_var_contained_in_type_annotation(ann.as_ref()));
|
||||||
}
|
}
|
||||||
TypeAnnotation::CustomClass { params, .. } => {
|
TypeAnnotation::CustomClass { params, .. } => {
|
||||||
@ -630,7 +605,8 @@ pub fn check_overload_type_annotation_compatible(
|
|||||||
|
|
||||||
a == b
|
a == b
|
||||||
}
|
}
|
||||||
(TypeAnnotation::Virtual(a), TypeAnnotation::Virtual(b)) => {
|
(TypeAnnotation::Virtual(a), TypeAnnotation::Virtual(b))
|
||||||
|
| (TypeAnnotation::List(a), TypeAnnotation::List(b)) => {
|
||||||
check_overload_type_annotation_compatible(a.as_ref(), b.as_ref(), unifier)
|
check_overload_type_annotation_compatible(a.as_ref(), b.as_ref(), unifier)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1,19 +1,13 @@
|
|||||||
use std::{
|
use crate::typecheck::typedef::TypeEnum;
|
||||||
collections::{HashMap, HashSet},
|
|
||||||
iter::once,
|
|
||||||
};
|
|
||||||
|
|
||||||
|
use super::type_inferencer::Inferencer;
|
||||||
|
use super::typedef::Type;
|
||||||
use nac3parser::ast::{
|
use nac3parser::ast::{
|
||||||
self, Constant, Expr, ExprKind,
|
self, Constant, Expr, ExprKind,
|
||||||
Operator::{LShift, RShift},
|
Operator::{LShift, RShift},
|
||||||
Stmt, StmtKind, StrRef,
|
Stmt, StmtKind, StrRef,
|
||||||
};
|
};
|
||||||
|
use std::{collections::HashSet, iter::once};
|
||||||
use super::{
|
|
||||||
type_inferencer::{DeclarationSource, IdentifierInfo, Inferencer},
|
|
||||||
typedef::{Type, TypeEnum},
|
|
||||||
};
|
|
||||||
use crate::toplevel::helper::PrimDef;
|
|
||||||
|
|
||||||
impl<'a> Inferencer<'a> {
|
impl<'a> Inferencer<'a> {
|
||||||
fn should_have_value(&mut self, expr: &Expr<Option<Type>>) -> Result<(), HashSet<String>> {
|
fn should_have_value(&mut self, expr: &Expr<Option<Type>>) -> Result<(), HashSet<String>> {
|
||||||
@ -27,45 +21,26 @@ impl<'a> Inferencer<'a> {
|
|||||||
fn check_pattern(
|
fn check_pattern(
|
||||||
&mut self,
|
&mut self,
|
||||||
pattern: &Expr<Option<Type>>,
|
pattern: &Expr<Option<Type>>,
|
||||||
defined_identifiers: &mut HashMap<StrRef, IdentifierInfo>,
|
defined_identifiers: &mut HashSet<StrRef>,
|
||||||
) -> Result<(), HashSet<String>> {
|
) -> Result<(), HashSet<String>> {
|
||||||
match &pattern.node {
|
match &pattern.node {
|
||||||
ExprKind::Name { id, .. } if id == &"none".into() => {
|
ExprKind::Name { id, .. } if id == &"none".into() => {
|
||||||
Err(HashSet::from([format!("cannot assign to a `none` (at {})", pattern.location)]))
|
Err(HashSet::from([format!("cannot assign to a `none` (at {})", pattern.location)]))
|
||||||
}
|
}
|
||||||
ExprKind::Name { id, .. } => {
|
ExprKind::Name { id, .. } => {
|
||||||
// If `id` refers to a declared symbol, reject this assignment if it is used in the
|
if !defined_identifiers.contains(id) {
|
||||||
// context of an (implicit) global variable
|
defined_identifiers.insert(*id);
|
||||||
if let Some(id_info) = defined_identifiers.get(id) {
|
|
||||||
if matches!(
|
|
||||||
id_info.source,
|
|
||||||
DeclarationSource::Global { is_explicit: Some(false) }
|
|
||||||
) {
|
|
||||||
return Err(HashSet::from([format!(
|
|
||||||
"cannot access local variable '{id}' before it is declared (at {})",
|
|
||||||
pattern.location
|
|
||||||
)]));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if !defined_identifiers.contains_key(id) {
|
|
||||||
defined_identifiers.insert(*id, IdentifierInfo::default());
|
|
||||||
}
|
}
|
||||||
self.should_have_value(pattern)?;
|
self.should_have_value(pattern)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
ExprKind::List { elts, .. } | ExprKind::Tuple { elts, .. } => {
|
ExprKind::Tuple { elts, .. } => {
|
||||||
for elt in elts {
|
for elt in elts {
|
||||||
self.check_pattern(elt, defined_identifiers)?;
|
self.check_pattern(elt, defined_identifiers)?;
|
||||||
self.should_have_value(elt)?;
|
self.should_have_value(elt)?;
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
ExprKind::Starred { value, .. } => {
|
|
||||||
self.check_pattern(value, defined_identifiers)?;
|
|
||||||
self.should_have_value(value)?;
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
ExprKind::Subscript { value, slice, .. } => {
|
ExprKind::Subscript { value, slice, .. } => {
|
||||||
self.check_expr(value, defined_identifiers)?;
|
self.check_expr(value, defined_identifiers)?;
|
||||||
self.should_have_value(value)?;
|
self.should_have_value(value)?;
|
||||||
@ -89,12 +64,11 @@ impl<'a> Inferencer<'a> {
|
|||||||
fn check_expr(
|
fn check_expr(
|
||||||
&mut self,
|
&mut self,
|
||||||
expr: &Expr<Option<Type>>,
|
expr: &Expr<Option<Type>>,
|
||||||
defined_identifiers: &mut HashMap<StrRef, IdentifierInfo>,
|
defined_identifiers: &mut HashSet<StrRef>,
|
||||||
) -> Result<(), HashSet<String>> {
|
) -> Result<(), HashSet<String>> {
|
||||||
// there are some cases where the custom field is None
|
// there are some cases where the custom field is None
|
||||||
if let Some(ty) = &expr.custom {
|
if let Some(ty) = &expr.custom {
|
||||||
if !matches!(&expr.node, ExprKind::Constant { value: Constant::Ellipsis, .. })
|
if !matches!(&expr.node, ExprKind::Constant { value: Constant::Ellipsis, .. })
|
||||||
&& !ty.obj_id(self.unifier).is_some_and(|id| id == PrimDef::List.id())
|
|
||||||
&& !self.unifier.is_concrete(*ty, &self.function_data.bound_variables)
|
&& !self.unifier.is_concrete(*ty, &self.function_data.bound_variables)
|
||||||
{
|
{
|
||||||
return Err(HashSet::from([format!(
|
return Err(HashSet::from([format!(
|
||||||
@ -110,7 +84,7 @@ impl<'a> Inferencer<'a> {
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
self.should_have_value(expr)?;
|
self.should_have_value(expr)?;
|
||||||
if !defined_identifiers.contains_key(id) {
|
if !defined_identifiers.contains(id) {
|
||||||
match self.function_data.resolver.get_symbol_type(
|
match self.function_data.resolver.get_symbol_type(
|
||||||
self.unifier,
|
self.unifier,
|
||||||
&self.top_level.definitions.read(),
|
&self.top_level.definitions.read(),
|
||||||
@ -118,22 +92,7 @@ impl<'a> Inferencer<'a> {
|
|||||||
*id,
|
*id,
|
||||||
) {
|
) {
|
||||||
Ok(_) => {
|
Ok(_) => {
|
||||||
let is_global = self.is_id_global(*id);
|
self.defined_identifiers.insert(*id);
|
||||||
|
|
||||||
defined_identifiers.insert(
|
|
||||||
*id,
|
|
||||||
IdentifierInfo {
|
|
||||||
source: match is_global {
|
|
||||||
Some(true) => {
|
|
||||||
DeclarationSource::Global { is_explicit: Some(false) }
|
|
||||||
}
|
|
||||||
Some(false) => {
|
|
||||||
DeclarationSource::Global { is_explicit: None }
|
|
||||||
}
|
|
||||||
None => DeclarationSource::Local,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
return Err(HashSet::from([format!(
|
return Err(HashSet::from([format!(
|
||||||
@ -206,7 +165,9 @@ impl<'a> Inferencer<'a> {
|
|||||||
let mut defined_identifiers = defined_identifiers.clone();
|
let mut defined_identifiers = defined_identifiers.clone();
|
||||||
for arg in &args.args {
|
for arg in &args.args {
|
||||||
// TODO: should we check the types here?
|
// TODO: should we check the types here?
|
||||||
defined_identifiers.entry(arg.node.arg).or_default();
|
if !defined_identifiers.contains(&arg.node.arg) {
|
||||||
|
defined_identifiers.insert(arg.node.arg);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
self.check_expr(body, &mut defined_identifiers)?;
|
self.check_expr(body, &mut defined_identifiers)?;
|
||||||
}
|
}
|
||||||
@ -245,9 +206,6 @@ impl<'a> Inferencer<'a> {
|
|||||||
/// This is a workaround preventing the caller from using a variable `alloca`-ed in the body, which
|
/// This is a workaround preventing the caller from using a variable `alloca`-ed in the body, which
|
||||||
/// is freed when the function returns.
|
/// is freed when the function returns.
|
||||||
fn check_return_value_ty(&mut self, ret_ty: Type) -> bool {
|
fn check_return_value_ty(&mut self, ret_ty: Type) -> bool {
|
||||||
if cfg!(feature = "no-escape-analysis") {
|
|
||||||
true
|
|
||||||
} else {
|
|
||||||
match &*self.unifier.get_ty_immutable(ret_ty) {
|
match &*self.unifier.get_ty_immutable(ret_ty) {
|
||||||
TypeEnum::TObj { .. } => [
|
TypeEnum::TObj { .. } => [
|
||||||
self.primitives.int32,
|
self.primitives.int32,
|
||||||
@ -259,17 +217,16 @@ impl<'a> Inferencer<'a> {
|
|||||||
]
|
]
|
||||||
.iter()
|
.iter()
|
||||||
.any(|allowed_ty| self.unifier.unioned(ret_ty, *allowed_ty)),
|
.any(|allowed_ty| self.unifier.unioned(ret_ty, *allowed_ty)),
|
||||||
TypeEnum::TTuple { ty, .. } => ty.iter().all(|t| self.check_return_value_ty(*t)),
|
TypeEnum::TTuple { ty } => ty.iter().all(|t| self.check_return_value_ty(*t)),
|
||||||
_ => false,
|
_ => false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
// check statements for proper identifier def-use and return on all paths
|
// check statements for proper identifier def-use and return on all paths
|
||||||
fn check_stmt(
|
fn check_stmt(
|
||||||
&mut self,
|
&mut self,
|
||||||
stmt: &Stmt<Option<Type>>,
|
stmt: &Stmt<Option<Type>>,
|
||||||
defined_identifiers: &mut HashMap<StrRef, IdentifierInfo>,
|
defined_identifiers: &mut HashSet<StrRef>,
|
||||||
) -> Result<bool, HashSet<String>> {
|
) -> Result<bool, HashSet<String>> {
|
||||||
match &stmt.node {
|
match &stmt.node {
|
||||||
StmtKind::For { target, iter, body, orelse, .. } => {
|
StmtKind::For { target, iter, body, orelse, .. } => {
|
||||||
@ -295,11 +252,9 @@ impl<'a> Inferencer<'a> {
|
|||||||
let body_returned = self.check_block(body, &mut body_identifiers)?;
|
let body_returned = self.check_block(body, &mut body_identifiers)?;
|
||||||
let orelse_returned = self.check_block(orelse, &mut orelse_identifiers)?;
|
let orelse_returned = self.check_block(orelse, &mut orelse_identifiers)?;
|
||||||
|
|
||||||
for ident in body_identifiers.keys() {
|
for ident in &body_identifiers {
|
||||||
if !defined_identifiers.contains_key(ident)
|
if !defined_identifiers.contains(ident) && orelse_identifiers.contains(ident) {
|
||||||
&& orelse_identifiers.contains_key(ident)
|
defined_identifiers.insert(*ident);
|
||||||
{
|
|
||||||
defined_identifiers.insert(*ident, IdentifierInfo::default());
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Ok(body_returned && orelse_returned)
|
Ok(body_returned && orelse_returned)
|
||||||
@ -330,7 +285,7 @@ impl<'a> Inferencer<'a> {
|
|||||||
let mut defined_identifiers = defined_identifiers.clone();
|
let mut defined_identifiers = defined_identifiers.clone();
|
||||||
let ast::ExcepthandlerKind::ExceptHandler { name, body, .. } = &handler.node;
|
let ast::ExcepthandlerKind::ExceptHandler { name, body, .. } = &handler.node;
|
||||||
if let Some(name) = name {
|
if let Some(name) = name {
|
||||||
defined_identifiers.insert(*name, IdentifierInfo::default());
|
defined_identifiers.insert(*name);
|
||||||
}
|
}
|
||||||
self.check_block(body, &mut defined_identifiers)?;
|
self.check_block(body, &mut defined_identifiers)?;
|
||||||
}
|
}
|
||||||
@ -394,44 +349,6 @@ impl<'a> Inferencer<'a> {
|
|||||||
}
|
}
|
||||||
Ok(true)
|
Ok(true)
|
||||||
}
|
}
|
||||||
StmtKind::Global { names, .. } => {
|
|
||||||
for id in names {
|
|
||||||
if let Some(id_info) = defined_identifiers.get(id) {
|
|
||||||
if id_info.source == DeclarationSource::Local {
|
|
||||||
return Err(HashSet::from([format!(
|
|
||||||
"name '{id}' is referenced prior to global declaration at {}",
|
|
||||||
stmt.location,
|
|
||||||
)]));
|
|
||||||
}
|
|
||||||
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
match self.function_data.resolver.get_symbol_type(
|
|
||||||
self.unifier,
|
|
||||||
&self.top_level.definitions.read(),
|
|
||||||
self.primitives,
|
|
||||||
*id,
|
|
||||||
) {
|
|
||||||
Ok(_) => {
|
|
||||||
defined_identifiers.insert(
|
|
||||||
*id,
|
|
||||||
IdentifierInfo {
|
|
||||||
source: DeclarationSource::Global { is_explicit: Some(true) },
|
|
||||||
},
|
|
||||||
);
|
|
||||||
}
|
|
||||||
Err(e) => {
|
|
||||||
return Err(HashSet::from([format!(
|
|
||||||
"type error at identifier `{}` ({}) at {}",
|
|
||||||
id, e, stmt.location
|
|
||||||
)]))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(false)
|
|
||||||
}
|
|
||||||
// break, raise, etc.
|
// break, raise, etc.
|
||||||
_ => Ok(false),
|
_ => Ok(false),
|
||||||
}
|
}
|
||||||
@ -440,7 +357,7 @@ impl<'a> Inferencer<'a> {
|
|||||||
pub fn check_block(
|
pub fn check_block(
|
||||||
&mut self,
|
&mut self,
|
||||||
block: &[Stmt<Option<Type>>],
|
block: &[Stmt<Option<Type>>],
|
||||||
defined_identifiers: &mut HashMap<StrRef, IdentifierInfo>,
|
defined_identifiers: &mut HashSet<StrRef>,
|
||||||
) -> Result<bool, HashSet<String>> {
|
) -> Result<bool, HashSet<String>> {
|
||||||
let mut ret = false;
|
let mut ret = false;
|
||||||
for stmt in block {
|
for stmt in block {
|
||||||
|
@ -1,151 +1,76 @@
|
|||||||
use std::{cmp::max, collections::HashMap, rc::Rc};
|
use crate::symbol_resolver::SymbolValue;
|
||||||
|
use crate::toplevel::helper::PrimDef;
|
||||||
use itertools::{iproduct, Itertools};
|
use crate::toplevel::numpy::{make_ndarray_ty, unpack_ndarray_var_tys};
|
||||||
use strum::IntoEnumIterator;
|
use crate::typecheck::{
|
||||||
|
|
||||||
use nac3parser::ast::{Cmpop, Operator, StrRef, Unaryop};
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
symbol_resolver::SymbolValue,
|
|
||||||
toplevel::{
|
|
||||||
helper::PrimDef,
|
|
||||||
numpy::{make_ndarray_ty, unpack_ndarray_var_tys},
|
|
||||||
},
|
|
||||||
typecheck::{
|
|
||||||
type_inferencer::*,
|
type_inferencer::*,
|
||||||
typedef::{FunSignature, FuncArg, Type, TypeEnum, Unifier, VarMap},
|
typedef::{FunSignature, FuncArg, Type, TypeEnum, Unifier, VarMap},
|
||||||
},
|
|
||||||
};
|
};
|
||||||
|
use itertools::Itertools;
|
||||||
|
use nac3parser::ast::StrRef;
|
||||||
|
use nac3parser::ast::{Cmpop, Operator, Unaryop};
|
||||||
|
use std::cmp::max;
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::rc::Rc;
|
||||||
|
use strum::IntoEnumIterator;
|
||||||
|
|
||||||
/// The variant of a binary operator.
|
|
||||||
#[derive(Debug, Clone, Copy, PartialEq, Eq)]
|
|
||||||
pub enum BinopVariant {
|
|
||||||
/// The normal variant.
|
|
||||||
/// For addition, it would be `+`.
|
|
||||||
Normal,
|
|
||||||
/// The "Augmented Assigning Operator" variant.
|
|
||||||
/// For addition, it would be `+=`.
|
|
||||||
AugAssign,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A binary operator with its variant.
|
|
||||||
#[derive(Debug, Clone, Copy)]
|
|
||||||
pub struct Binop {
|
|
||||||
/// The base [`Operator`] of this binary operator.
|
|
||||||
pub base: Operator,
|
|
||||||
/// The variant of this binary operator.
|
|
||||||
pub variant: BinopVariant,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Binop {
|
|
||||||
/// Make a [`Binop`] of the normal variant from an [`Operator`].
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn normal(base: Operator) -> Self {
|
pub fn binop_name(op: Operator) -> &'static str {
|
||||||
Binop { base, variant: BinopVariant::Normal }
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Make a [`Binop`] of the aug assign variant from an [`Operator`].
|
|
||||||
#[must_use]
|
|
||||||
pub fn aug_assign(base: Operator) -> Self {
|
|
||||||
Binop { base, variant: BinopVariant::AugAssign }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Details about an operator (unary, binary, etc...) in Python
|
|
||||||
#[derive(Debug, Clone, Copy)]
|
|
||||||
pub struct OpInfo {
|
|
||||||
/// The method name of the binary operator.
|
|
||||||
/// For addition, this would be `__add__`, and `__iadd__` if
|
|
||||||
/// it is the augmented assigning variant.
|
|
||||||
pub method_name: &'static str,
|
|
||||||
/// The symbol of the binary operator.
|
|
||||||
/// For addition, this would be `+`, and `+=` if
|
|
||||||
/// it is the augmented assigning variant.
|
|
||||||
pub symbol: &'static str,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Helper macro to conveniently build an [`OpInfo`].
|
|
||||||
///
|
|
||||||
/// Example usage: `make_info("add", "+")` generates `OpInfo { name: "__add__", symbol: "+" }`
|
|
||||||
macro_rules! make_op_info {
|
|
||||||
($name:expr, $symbol:expr) => {
|
|
||||||
OpInfo { method_name: concat!("__", $name, "__"), symbol: $symbol }
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
pub trait HasOpInfo {
|
|
||||||
fn op_info(&self) -> OpInfo;
|
|
||||||
}
|
|
||||||
|
|
||||||
fn try_get_cmpop_info(op: Cmpop) -> Option<OpInfo> {
|
|
||||||
match op {
|
match op {
|
||||||
Cmpop::Lt => Some(make_op_info!("lt", "<")),
|
Operator::Add => "__add__",
|
||||||
Cmpop::LtE => Some(make_op_info!("le", "<=")),
|
Operator::Sub => "__sub__",
|
||||||
Cmpop::Gt => Some(make_op_info!("gt", ">")),
|
Operator::Div => "__truediv__",
|
||||||
Cmpop::GtE => Some(make_op_info!("ge", ">=")),
|
Operator::Mod => "__mod__",
|
||||||
Cmpop::Eq => Some(make_op_info!("eq", "==")),
|
Operator::Mult => "__mul__",
|
||||||
Cmpop::NotEq => Some(make_op_info!("ne", "!=")),
|
Operator::Pow => "__pow__",
|
||||||
_ => None,
|
Operator::BitOr => "__or__",
|
||||||
|
Operator::BitXor => "__xor__",
|
||||||
|
Operator::BitAnd => "__and__",
|
||||||
|
Operator::LShift => "__lshift__",
|
||||||
|
Operator::RShift => "__rshift__",
|
||||||
|
Operator::FloorDiv => "__floordiv__",
|
||||||
|
Operator::MatMult => "__matmul__",
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl OpInfo {
|
|
||||||
#[must_use]
|
#[must_use]
|
||||||
pub fn supports_cmpop(op: Cmpop) -> bool {
|
pub fn binop_assign_name(op: Operator) -> &'static str {
|
||||||
try_get_cmpop_info(op).is_some()
|
match op {
|
||||||
|
Operator::Add => "__iadd__",
|
||||||
|
Operator::Sub => "__isub__",
|
||||||
|
Operator::Div => "__itruediv__",
|
||||||
|
Operator::Mod => "__imod__",
|
||||||
|
Operator::Mult => "__imul__",
|
||||||
|
Operator::Pow => "__ipow__",
|
||||||
|
Operator::BitOr => "__ior__",
|
||||||
|
Operator::BitXor => "__ixor__",
|
||||||
|
Operator::BitAnd => "__iand__",
|
||||||
|
Operator::LShift => "__ilshift__",
|
||||||
|
Operator::RShift => "__irshift__",
|
||||||
|
Operator::FloorDiv => "__ifloordiv__",
|
||||||
|
Operator::MatMult => "__imatmul__",
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl HasOpInfo for Cmpop {
|
#[must_use]
|
||||||
fn op_info(&self) -> OpInfo {
|
pub fn unaryop_name(op: Unaryop) -> &'static str {
|
||||||
try_get_cmpop_info(*self).expect("{self:?} is not supported")
|
match op {
|
||||||
|
Unaryop::UAdd => "__pos__",
|
||||||
|
Unaryop::USub => "__neg__",
|
||||||
|
Unaryop::Not => "__not__",
|
||||||
|
Unaryop::Invert => "__inv__",
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl HasOpInfo for Binop {
|
#[must_use]
|
||||||
fn op_info(&self) -> OpInfo {
|
pub fn comparison_name(op: Cmpop) -> Option<&'static str> {
|
||||||
// Helper macro to generate both the normal variant [`OpInfo`] and the
|
match op {
|
||||||
// augmented assigning variant [`OpInfo`] for a binary operator conveniently.
|
Cmpop::Lt => Some("__lt__"),
|
||||||
macro_rules! info {
|
Cmpop::LtE => Some("__le__"),
|
||||||
($name:literal, $symbol:literal) => {
|
Cmpop::Gt => Some("__gt__"),
|
||||||
(
|
Cmpop::GtE => Some("__ge__"),
|
||||||
make_op_info!($name, $symbol),
|
Cmpop::Eq => Some("__eq__"),
|
||||||
make_op_info!(concat!("i", $name), concat!($symbol, "=")),
|
Cmpop::NotEq => Some("__ne__"),
|
||||||
)
|
_ => None,
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
let (normal_variant, aug_assign_variant) = match self.base {
|
|
||||||
Operator::Add => info!("add", "+"),
|
|
||||||
Operator::Sub => info!("sub", "-"),
|
|
||||||
Operator::Div => info!("truediv", "/"),
|
|
||||||
Operator::Mod => info!("mod", "%"),
|
|
||||||
Operator::Mult => info!("mul", "*"),
|
|
||||||
Operator::Pow => info!("pow", "**"),
|
|
||||||
Operator::BitOr => info!("or", "|"),
|
|
||||||
Operator::BitXor => info!("xor", "^"),
|
|
||||||
Operator::BitAnd => info!("and", "&"),
|
|
||||||
Operator::LShift => info!("lshift", "<<"),
|
|
||||||
Operator::RShift => info!("rshift", ">>"),
|
|
||||||
Operator::FloorDiv => info!("floordiv", "//"),
|
|
||||||
Operator::MatMult => info!("matmul", "@"),
|
|
||||||
};
|
|
||||||
|
|
||||||
match self.variant {
|
|
||||||
BinopVariant::Normal => normal_variant,
|
|
||||||
BinopVariant::AugAssign => aug_assign_variant,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl HasOpInfo for Unaryop {
|
|
||||||
fn op_info(&self) -> OpInfo {
|
|
||||||
match self {
|
|
||||||
Unaryop::UAdd => make_op_info!("pos", "+"),
|
|
||||||
Unaryop::USub => make_op_info!("neg", "-"),
|
|
||||||
Unaryop::Not => make_op_info!("not", "not"), // i.e., `not False`, so the symbol is just `not`.
|
|
||||||
Unaryop::Invert => make_op_info!("inv", "~"),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -190,9 +115,23 @@ pub fn impl_binop(
|
|||||||
|
|
||||||
let ret_ty = ret_ty.unwrap_or_else(|| unifier.get_fresh_var(None, None).ty);
|
let ret_ty = ret_ty.unwrap_or_else(|| unifier.get_fresh_var(None, None).ty);
|
||||||
|
|
||||||
for (base_op, variant) in iproduct!(ops, [BinopVariant::Normal, BinopVariant::AugAssign]) {
|
for op in ops {
|
||||||
let op = Binop { base: *base_op, variant };
|
fields.insert(binop_name(*op).into(), {
|
||||||
fields.insert(op.op_info().method_name.into(), {
|
(
|
||||||
|
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
|
ret: ret_ty,
|
||||||
|
vars: function_vars.clone(),
|
||||||
|
args: vec![FuncArg {
|
||||||
|
ty: other_ty,
|
||||||
|
default_value: None,
|
||||||
|
name: "other".into(),
|
||||||
|
}],
|
||||||
|
})),
|
||||||
|
false,
|
||||||
|
)
|
||||||
|
});
|
||||||
|
|
||||||
|
fields.insert(binop_assign_name(*op).into(), {
|
||||||
(
|
(
|
||||||
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
ret: ret_ty,
|
ret: ret_ty,
|
||||||
@ -201,7 +140,6 @@ pub fn impl_binop(
|
|||||||
ty: other_ty,
|
ty: other_ty,
|
||||||
default_value: None,
|
default_value: None,
|
||||||
name: "other".into(),
|
name: "other".into(),
|
||||||
is_vararg: false,
|
|
||||||
}],
|
}],
|
||||||
})),
|
})),
|
||||||
false,
|
false,
|
||||||
@ -217,7 +155,7 @@ pub fn impl_unaryop(unifier: &mut Unifier, ty: Type, ret_ty: Option<Type>, ops:
|
|||||||
|
|
||||||
for op in ops {
|
for op in ops {
|
||||||
fields.insert(
|
fields.insert(
|
||||||
op.op_info().method_name.into(),
|
unaryop_name(*op).into(),
|
||||||
(
|
(
|
||||||
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
ret: ret_ty,
|
ret: ret_ty,
|
||||||
@ -257,7 +195,7 @@ pub fn impl_cmpop(
|
|||||||
|
|
||||||
for op in ops {
|
for op in ops {
|
||||||
fields.insert(
|
fields.insert(
|
||||||
op.op_info().method_name.into(),
|
comparison_name(*op).unwrap().into(),
|
||||||
(
|
(
|
||||||
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
ret: ret_ty,
|
ret: ret_ty,
|
||||||
@ -266,7 +204,6 @@ pub fn impl_cmpop(
|
|||||||
ty: other_ty,
|
ty: other_ty,
|
||||||
default_value: None,
|
default_value: None,
|
||||||
name: "other".into(),
|
name: "other".into(),
|
||||||
is_vararg: false,
|
|
||||||
}],
|
}],
|
||||||
})),
|
})),
|
||||||
false,
|
false,
|
||||||
@ -492,29 +429,12 @@ pub fn typeof_binop(
|
|||||||
lhs: Type,
|
lhs: Type,
|
||||||
rhs: Type,
|
rhs: Type,
|
||||||
) -> Result<Option<Type>, String> {
|
) -> Result<Option<Type>, String> {
|
||||||
let op = Binop { base: op, variant: BinopVariant::Normal };
|
|
||||||
|
|
||||||
let is_left_list = lhs.obj_id(unifier).is_some_and(|id| id == PrimDef::List.id());
|
|
||||||
let is_right_list = rhs.obj_id(unifier).is_some_and(|id| id == PrimDef::List.id());
|
|
||||||
let is_left_ndarray = lhs.obj_id(unifier).is_some_and(|id| id == PrimDef::NDArray.id());
|
let is_left_ndarray = lhs.obj_id(unifier).is_some_and(|id| id == PrimDef::NDArray.id());
|
||||||
let is_right_ndarray = rhs.obj_id(unifier).is_some_and(|id| id == PrimDef::NDArray.id());
|
let is_right_ndarray = rhs.obj_id(unifier).is_some_and(|id| id == PrimDef::NDArray.id());
|
||||||
|
|
||||||
Ok(Some(match op.base {
|
Ok(Some(match op {
|
||||||
Operator::Add | Operator::Sub | Operator::Mult | Operator::Mod | Operator::FloorDiv => {
|
Operator::Add | Operator::Sub | Operator::Mult | Operator::Mod | Operator::FloorDiv => {
|
||||||
if is_left_list || is_right_list {
|
if is_left_ndarray || is_right_ndarray {
|
||||||
if ![Operator::Add, Operator::Mult].contains(&op.base) {
|
|
||||||
return Err(format!(
|
|
||||||
"Binary operator {} not supported for list",
|
|
||||||
op.op_info().symbol
|
|
||||||
));
|
|
||||||
}
|
|
||||||
|
|
||||||
if is_left_list {
|
|
||||||
lhs
|
|
||||||
} else {
|
|
||||||
rhs
|
|
||||||
}
|
|
||||||
} else if is_left_ndarray || is_right_ndarray {
|
|
||||||
typeof_ndarray_broadcast(unifier, primitives, lhs, rhs)?
|
typeof_ndarray_broadcast(unifier, primitives, lhs, rhs)?
|
||||||
} else if unifier.unioned(lhs, rhs) {
|
} else if unifier.unioned(lhs, rhs) {
|
||||||
lhs
|
lhs
|
||||||
@ -524,23 +444,6 @@ pub fn typeof_binop(
|
|||||||
}
|
}
|
||||||
|
|
||||||
Operator::MatMult => {
|
Operator::MatMult => {
|
||||||
// NOTE: NumPy matmul's LHS and RHS must both be ndarrays. Scalars are not allowed.
|
|
||||||
match (&*unifier.get_ty(lhs), &*unifier.get_ty(rhs)) {
|
|
||||||
(
|
|
||||||
TypeEnum::TObj { obj_id: lhs_obj_id, .. },
|
|
||||||
TypeEnum::TObj { obj_id: rhs_obj_id, .. },
|
|
||||||
) if *lhs_obj_id == primitives.ndarray.obj_id(unifier).unwrap()
|
|
||||||
&& *rhs_obj_id == primitives.ndarray.obj_id(unifier).unwrap() =>
|
|
||||||
{
|
|
||||||
// LHS and RHS have valid types
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
let lhs_str = unifier.stringify(lhs);
|
|
||||||
let rhs_str = unifier.stringify(rhs);
|
|
||||||
return Err(format!("ndarray.__matmul__ only accepts ndarray operands, but left operand has type {lhs_str}, and right operand has type {rhs_str}"));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let (_, lhs_ndims) = unpack_ndarray_var_tys(unifier, lhs);
|
let (_, lhs_ndims) = unpack_ndarray_var_tys(unifier, lhs);
|
||||||
let lhs_ndims = match &*unifier.get_ty_immutable(lhs_ndims) {
|
let lhs_ndims = match &*unifier.get_ty_immutable(lhs_ndims) {
|
||||||
TypeEnum::TLiteral { values, .. } => {
|
TypeEnum::TLiteral { values, .. } => {
|
||||||
@ -701,8 +604,6 @@ pub fn set_primitives_magic_methods(store: &PrimitiveStore, unifier: &mut Unifie
|
|||||||
bool: bool_t,
|
bool: bool_t,
|
||||||
uint32: uint32_t,
|
uint32: uint32_t,
|
||||||
uint64: uint64_t,
|
uint64: uint64_t,
|
||||||
str: str_t,
|
|
||||||
list: list_t,
|
|
||||||
ndarray: ndarray_t,
|
ndarray: ndarray_t,
|
||||||
..
|
..
|
||||||
} = *store;
|
} = *store;
|
||||||
@ -747,14 +648,6 @@ pub fn set_primitives_magic_methods(store: &PrimitiveStore, unifier: &mut Unifie
|
|||||||
impl_sign(unifier, store, bool_t, Some(int32_t));
|
impl_sign(unifier, store, bool_t, Some(int32_t));
|
||||||
impl_eq(unifier, store, bool_t, &[bool_t, ndarray_bool_t], None);
|
impl_eq(unifier, store, bool_t, &[bool_t, ndarray_bool_t], None);
|
||||||
|
|
||||||
/* str ========= */
|
|
||||||
impl_cmpop(unifier, store, str_t, &[str_t], &[Cmpop::Eq, Cmpop::NotEq], Some(bool_t));
|
|
||||||
|
|
||||||
/* list ======== */
|
|
||||||
impl_binop(unifier, store, list_t, &[list_t], Some(list_t), &[Operator::Add]);
|
|
||||||
impl_binop(unifier, store, list_t, &[int32_t, int64_t], Some(list_t), &[Operator::Mult]);
|
|
||||||
impl_cmpop(unifier, store, list_t, &[list_t], &[Cmpop::Eq, Cmpop::NotEq], Some(bool_t));
|
|
||||||
|
|
||||||
/* ndarray ===== */
|
/* ndarray ===== */
|
||||||
let ndarray_usized_ndims_tvar =
|
let ndarray_usized_ndims_tvar =
|
||||||
unifier.get_fresh_const_generic_var(size_t, Some("ndarray_ndims".into()), None);
|
unifier.get_fresh_const_generic_var(size_t, Some("ndarray_ndims".into()), None);
|
||||||
|
@ -1,46 +1,24 @@
|
|||||||
use std::{collections::HashMap, fmt::Display};
|
use std::collections::HashMap;
|
||||||
|
use std::fmt::Display;
|
||||||
|
|
||||||
use itertools::Itertools;
|
use crate::typecheck::typedef::TypeEnum;
|
||||||
|
|
||||||
use nac3parser::ast::{Cmpop, Location, StrRef};
|
use super::typedef::{RecordKey, Type, Unifier};
|
||||||
|
use nac3parser::ast::{Location, StrRef};
|
||||||
use super::{
|
|
||||||
magic_methods::Binop,
|
|
||||||
typedef::{RecordKey, Type, Unifier},
|
|
||||||
};
|
|
||||||
use crate::typecheck::{magic_methods::HasOpInfo, typedef::TypeEnum};
|
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
pub enum TypeErrorKind {
|
pub enum TypeErrorKind {
|
||||||
GotMultipleValues {
|
|
||||||
name: StrRef,
|
|
||||||
},
|
|
||||||
TooManyArguments {
|
TooManyArguments {
|
||||||
expected_min_count: usize,
|
expected: usize,
|
||||||
expected_max_count: usize,
|
got: usize,
|
||||||
got_count: usize,
|
|
||||||
},
|
|
||||||
MissingArgs {
|
|
||||||
missing_arg_names: Vec<StrRef>,
|
|
||||||
},
|
},
|
||||||
|
MissingArgs(String),
|
||||||
UnknownArgName(StrRef),
|
UnknownArgName(StrRef),
|
||||||
IncorrectArgType {
|
IncorrectArgType {
|
||||||
name: StrRef,
|
name: StrRef,
|
||||||
expected: Type,
|
expected: Type,
|
||||||
got: Type,
|
got: Type,
|
||||||
},
|
},
|
||||||
UnsupportedBinaryOpTypes {
|
|
||||||
operator: Binop,
|
|
||||||
lhs_type: Type,
|
|
||||||
rhs_type: Type,
|
|
||||||
expected_rhs_type: Type,
|
|
||||||
},
|
|
||||||
UnsupportedComparsionOpTypes {
|
|
||||||
operator: Cmpop,
|
|
||||||
lhs_type: Type,
|
|
||||||
rhs_type: Type,
|
|
||||||
expected_rhs_type: Type,
|
|
||||||
},
|
|
||||||
FieldUnificationError {
|
FieldUnificationError {
|
||||||
field: RecordKey,
|
field: RecordKey,
|
||||||
types: (Type, Type),
|
types: (Type, Type),
|
||||||
@ -56,7 +34,6 @@ pub enum TypeErrorKind {
|
|||||||
},
|
},
|
||||||
RequiresTypeAnn,
|
RequiresTypeAnn,
|
||||||
PolymorphicFunctionPointer,
|
PolymorphicFunctionPointer,
|
||||||
NoSuchAttribute(RecordKey, Type),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
@ -100,49 +77,19 @@ impl<'a> Display for DisplayTypeError<'a> {
|
|||||||
use TypeErrorKind::*;
|
use TypeErrorKind::*;
|
||||||
let mut notes = Some(HashMap::new());
|
let mut notes = Some(HashMap::new());
|
||||||
match &self.err.kind {
|
match &self.err.kind {
|
||||||
GotMultipleValues { name } => {
|
TooManyArguments { expected, got } => {
|
||||||
write!(f, "For multiple values for parameter {name}")
|
write!(f, "Too many arguments. Expected {expected} but got {got}")
|
||||||
}
|
}
|
||||||
TooManyArguments { expected_min_count, expected_max_count, got_count } => {
|
MissingArgs(args) => {
|
||||||
debug_assert!(expected_min_count <= expected_max_count);
|
|
||||||
if expected_min_count == expected_max_count {
|
|
||||||
let expected_count = expected_min_count; // or expected_max_count
|
|
||||||
write!(f, "Too many arguments. Expected {expected_count} but got {got_count}")
|
|
||||||
} else {
|
|
||||||
write!(f, "Too many arguments. Expected {expected_min_count} to {expected_max_count} arguments but got {got_count}")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
MissingArgs { missing_arg_names } => {
|
|
||||||
let args = missing_arg_names.iter().join(", ");
|
|
||||||
write!(f, "Missing arguments: {args}")
|
write!(f, "Missing arguments: {args}")
|
||||||
}
|
}
|
||||||
UnsupportedBinaryOpTypes { operator, lhs_type, rhs_type, expected_rhs_type } => {
|
|
||||||
let op_symbol = operator.op_info().symbol;
|
|
||||||
|
|
||||||
let lhs_type_str = self.unifier.stringify_with_notes(*lhs_type, &mut notes);
|
|
||||||
let rhs_type_str = self.unifier.stringify_with_notes(*rhs_type, &mut notes);
|
|
||||||
let expected_rhs_type_str =
|
|
||||||
self.unifier.stringify_with_notes(*expected_rhs_type, &mut notes);
|
|
||||||
|
|
||||||
write!(f, "Unsupported operand type(s) for {op_symbol}: '{lhs_type_str}' and '{rhs_type_str}' (right operand should have type {expected_rhs_type_str})")
|
|
||||||
}
|
|
||||||
UnsupportedComparsionOpTypes { operator, lhs_type, rhs_type, expected_rhs_type } => {
|
|
||||||
let op_symbol = operator.op_info().symbol;
|
|
||||||
|
|
||||||
let lhs_type_str = self.unifier.stringify_with_notes(*lhs_type, &mut notes);
|
|
||||||
let rhs_type_str = self.unifier.stringify_with_notes(*rhs_type, &mut notes);
|
|
||||||
let expected_rhs_type_str =
|
|
||||||
self.unifier.stringify_with_notes(*expected_rhs_type, &mut notes);
|
|
||||||
|
|
||||||
write!(f, "'{op_symbol}' not supported between instances of '{lhs_type_str}' and '{rhs_type_str}' (right operand should have type {expected_rhs_type_str})")
|
|
||||||
}
|
|
||||||
UnknownArgName(name) => {
|
UnknownArgName(name) => {
|
||||||
write!(f, "Unknown argument name: {name}")
|
write!(f, "Unknown argument name: {name}")
|
||||||
}
|
}
|
||||||
IncorrectArgType { name, expected, got } => {
|
IncorrectArgType { name, expected, got } => {
|
||||||
let expected = self.unifier.stringify_with_notes(*expected, &mut notes);
|
let expected = self.unifier.stringify_with_notes(*expected, &mut notes);
|
||||||
let got = self.unifier.stringify_with_notes(*got, &mut notes);
|
let got = self.unifier.stringify_with_notes(*got, &mut notes);
|
||||||
write!(f, "Incorrect argument type for parameter {name}. Expected {expected}, but got {got}")
|
write!(f, "Incorrect argument type for {name}. Expected {expected}, but got {got}")
|
||||||
}
|
}
|
||||||
FieldUnificationError { field, types, loc } => {
|
FieldUnificationError { field, types, loc } => {
|
||||||
let lhs = self.unifier.stringify_with_notes(types.0, &mut notes);
|
let lhs = self.unifier.stringify_with_notes(types.0, &mut notes);
|
||||||
@ -183,10 +130,9 @@ impl<'a> Display for DisplayTypeError<'a> {
|
|||||||
}
|
}
|
||||||
result
|
result
|
||||||
}
|
}
|
||||||
(
|
(TypeEnum::TTuple { ty: ty1 }, TypeEnum::TTuple { ty: ty2 })
|
||||||
TypeEnum::TTuple { ty: ty1, is_vararg_ctx: is_vararg1 },
|
if ty1.len() != ty2.len() =>
|
||||||
TypeEnum::TTuple { ty: ty2, is_vararg_ctx: is_vararg2 },
|
{
|
||||||
) if !is_vararg1 && !is_vararg2 && ty1.len() != ty2.len() => {
|
|
||||||
let t1 = self.unifier.stringify_with_notes(*t1, &mut notes);
|
let t1 = self.unifier.stringify_with_notes(*t1, &mut notes);
|
||||||
let t2 = self.unifier.stringify_with_notes(*t2, &mut notes);
|
let t2 = self.unifier.stringify_with_notes(*t2, &mut notes);
|
||||||
write!(f, "Tuple length mismatch: got {t1} and {t2}")
|
write!(f, "Tuple length mismatch: got {t1} and {t2}")
|
||||||
@ -210,10 +156,6 @@ impl<'a> Display for DisplayTypeError<'a> {
|
|||||||
let t = self.unifier.stringify_with_notes(*t, &mut notes);
|
let t = self.unifier.stringify_with_notes(*t, &mut notes);
|
||||||
write!(f, "`{t}::{name}` field/method does not exist")
|
write!(f, "`{t}::{name}` field/method does not exist")
|
||||||
}
|
}
|
||||||
NoSuchAttribute(name, t) => {
|
|
||||||
let t = self.unifier.stringify_with_notes(*t, &mut notes);
|
|
||||||
write!(f, "`{t}::{name}` is not a class attribute")
|
|
||||||
}
|
|
||||||
TupleIndexOutOfBounds { index, len } => {
|
TupleIndexOutOfBounds { index, len } => {
|
||||||
write!(
|
write!(
|
||||||
f,
|
f,
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -1,19 +1,17 @@
|
|||||||
use std::iter::zip;
|
use super::super::{magic_methods::with_fields, typedef::*};
|
||||||
|
|
||||||
use indexmap::IndexMap;
|
|
||||||
use indoc::indoc;
|
|
||||||
use parking_lot::RwLock;
|
|
||||||
use test_case::test_case;
|
|
||||||
|
|
||||||
use nac3parser::{ast::FileName, parser::parse_program};
|
|
||||||
|
|
||||||
use super::*;
|
use super::*;
|
||||||
use crate::{
|
use crate::{
|
||||||
codegen::{CodeGenContext, CodeGenerator},
|
codegen::CodeGenContext,
|
||||||
symbol_resolver::ValueEnum,
|
symbol_resolver::ValueEnum,
|
||||||
toplevel::{helper::PrimDef, DefinitionId, TopLevelDef},
|
toplevel::{helper::PrimDef, DefinitionId, TopLevelDef},
|
||||||
typecheck::{magic_methods::with_fields, typedef::*},
|
|
||||||
};
|
};
|
||||||
|
use indexmap::IndexMap;
|
||||||
|
use indoc::indoc;
|
||||||
|
use nac3parser::ast::FileName;
|
||||||
|
use nac3parser::parser::parse_program;
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
use std::iter::zip;
|
||||||
|
use test_case::test_case;
|
||||||
|
|
||||||
struct Resolver {
|
struct Resolver {
|
||||||
id_to_type: HashMap<StrRef, Type>,
|
id_to_type: HashMap<StrRef, Type>,
|
||||||
@ -43,7 +41,6 @@ impl SymbolResolver for Resolver {
|
|||||||
&self,
|
&self,
|
||||||
_: StrRef,
|
_: StrRef,
|
||||||
_: &mut CodeGenContext<'ctx, '_>,
|
_: &mut CodeGenContext<'ctx, '_>,
|
||||||
_: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>> {
|
) -> Option<ValueEnum<'ctx>> {
|
||||||
unimplemented!()
|
unimplemented!()
|
||||||
}
|
}
|
||||||
@ -86,12 +83,7 @@ impl TestEnvironment {
|
|||||||
});
|
});
|
||||||
with_fields(&mut unifier, int32, |unifier, fields| {
|
with_fields(&mut unifier, int32, |unifier, fields| {
|
||||||
let add_ty = unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
let add_ty = unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
args: vec![FuncArg {
|
args: vec![FuncArg { name: "other".into(), ty: int32, default_value: None }],
|
||||||
name: "other".into(),
|
|
||||||
ty: int32,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
}],
|
|
||||||
ret: int32,
|
ret: int32,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
}));
|
}));
|
||||||
@ -147,12 +139,6 @@ impl TestEnvironment {
|
|||||||
fields: HashMap::new(),
|
fields: HashMap::new(),
|
||||||
params: VarMap::new(),
|
params: VarMap::new(),
|
||||||
});
|
});
|
||||||
let list_elem_tvar = unifier.get_fresh_var(Some("list_elem".into()), None);
|
|
||||||
let list = unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: PrimDef::List.id(),
|
|
||||||
fields: HashMap::new(),
|
|
||||||
params: into_var_map([list_elem_tvar]),
|
|
||||||
});
|
|
||||||
let ndarray_dtype_tvar = unifier.get_fresh_var(Some("ndarray_dtype".into()), None);
|
let ndarray_dtype_tvar = unifier.get_fresh_var(Some("ndarray_dtype".into()), None);
|
||||||
let ndarray_ndims_tvar =
|
let ndarray_ndims_tvar =
|
||||||
unifier.get_fresh_const_generic_var(uint64, Some("ndarray_ndims".into()), None);
|
unifier.get_fresh_const_generic_var(uint64, Some("ndarray_ndims".into()), None);
|
||||||
@ -173,7 +159,6 @@ impl TestEnvironment {
|
|||||||
uint32,
|
uint32,
|
||||||
uint64,
|
uint64,
|
||||||
option,
|
option,
|
||||||
list,
|
|
||||||
ndarray,
|
ndarray,
|
||||||
size_t: 64,
|
size_t: 64,
|
||||||
};
|
};
|
||||||
@ -232,12 +217,7 @@ impl TestEnvironment {
|
|||||||
});
|
});
|
||||||
with_fields(&mut unifier, int32, |unifier, fields| {
|
with_fields(&mut unifier, int32, |unifier, fields| {
|
||||||
let add_ty = unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
let add_ty = unifier.add_ty(TypeEnum::TFunc(FunSignature {
|
||||||
args: vec![FuncArg {
|
args: vec![FuncArg { name: "other".into(), ty: int32, default_value: None }],
|
||||||
name: "other".into(),
|
|
||||||
ty: int32,
|
|
||||||
default_value: None,
|
|
||||||
is_vararg: false,
|
|
||||||
}],
|
|
||||||
ret: int32,
|
ret: int32,
|
||||||
vars: VarMap::new(),
|
vars: VarMap::new(),
|
||||||
}));
|
}));
|
||||||
@ -293,33 +273,13 @@ impl TestEnvironment {
|
|||||||
fields: HashMap::new(),
|
fields: HashMap::new(),
|
||||||
params: VarMap::new(),
|
params: VarMap::new(),
|
||||||
});
|
});
|
||||||
let list_elem_tvar = unifier.get_fresh_var(Some("list_elem".into()), None);
|
|
||||||
let list = unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: PrimDef::List.id(),
|
|
||||||
fields: HashMap::new(),
|
|
||||||
params: into_var_map([list_elem_tvar]),
|
|
||||||
});
|
|
||||||
let ndarray = unifier.add_ty(TypeEnum::TObj {
|
let ndarray = unifier.add_ty(TypeEnum::TObj {
|
||||||
obj_id: PrimDef::NDArray.id(),
|
obj_id: PrimDef::NDArray.id(),
|
||||||
fields: HashMap::new(),
|
fields: HashMap::new(),
|
||||||
params: VarMap::new(),
|
params: VarMap::new(),
|
||||||
});
|
});
|
||||||
identifier_mapping.insert("None".into(), none);
|
identifier_mapping.insert("None".into(), none);
|
||||||
for (i, name) in [
|
for (i, name) in ["int32", "int64", "float", "bool", "none", "range", "str", "Exception"]
|
||||||
"int32",
|
|
||||||
"int64",
|
|
||||||
"float",
|
|
||||||
"bool",
|
|
||||||
"none",
|
|
||||||
"range",
|
|
||||||
"str",
|
|
||||||
"Exception",
|
|
||||||
"uint32",
|
|
||||||
"uint64",
|
|
||||||
"Option",
|
|
||||||
"list",
|
|
||||||
"ndarray",
|
|
||||||
]
|
|
||||||
.iter()
|
.iter()
|
||||||
.enumerate()
|
.enumerate()
|
||||||
{
|
{
|
||||||
@ -329,7 +289,6 @@ impl TestEnvironment {
|
|||||||
object_id: DefinitionId(i),
|
object_id: DefinitionId(i),
|
||||||
type_vars: Vec::default(),
|
type_vars: Vec::default(),
|
||||||
fields: Vec::default(),
|
fields: Vec::default(),
|
||||||
attributes: Vec::default(),
|
|
||||||
methods: Vec::default(),
|
methods: Vec::default(),
|
||||||
ancestors: Vec::default(),
|
ancestors: Vec::default(),
|
||||||
resolver: None,
|
resolver: None,
|
||||||
@ -339,7 +298,7 @@ impl TestEnvironment {
|
|||||||
.into(),
|
.into(),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
let defs = 12;
|
let defs = 7;
|
||||||
|
|
||||||
let primitives = PrimitiveStore {
|
let primitives = PrimitiveStore {
|
||||||
int32,
|
int32,
|
||||||
@ -353,7 +312,6 @@ impl TestEnvironment {
|
|||||||
uint32,
|
uint32,
|
||||||
uint64,
|
uint64,
|
||||||
option,
|
option,
|
||||||
list,
|
|
||||||
ndarray,
|
ndarray,
|
||||||
size_t: 64,
|
size_t: 64,
|
||||||
};
|
};
|
||||||
@ -373,7 +331,6 @@ impl TestEnvironment {
|
|||||||
object_id: DefinitionId(defs + 1),
|
object_id: DefinitionId(defs + 1),
|
||||||
type_vars: vec![tvar.ty],
|
type_vars: vec![tvar.ty],
|
||||||
fields: [("a".into(), tvar.ty, true)].into(),
|
fields: [("a".into(), tvar.ty, true)].into(),
|
||||||
attributes: Vec::default(),
|
|
||||||
methods: Vec::default(),
|
methods: Vec::default(),
|
||||||
ancestors: Vec::default(),
|
ancestors: Vec::default(),
|
||||||
resolver: None,
|
resolver: None,
|
||||||
@ -408,7 +365,6 @@ impl TestEnvironment {
|
|||||||
object_id: DefinitionId(defs + 2),
|
object_id: DefinitionId(defs + 2),
|
||||||
type_vars: Vec::default(),
|
type_vars: Vec::default(),
|
||||||
fields: [("a".into(), int32, true), ("b".into(), fun, true)].into(),
|
fields: [("a".into(), int32, true), ("b".into(), fun, true)].into(),
|
||||||
attributes: Vec::default(),
|
|
||||||
methods: Vec::default(),
|
methods: Vec::default(),
|
||||||
ancestors: Vec::default(),
|
ancestors: Vec::default(),
|
||||||
resolver: None,
|
resolver: None,
|
||||||
@ -437,7 +393,6 @@ impl TestEnvironment {
|
|||||||
object_id: DefinitionId(defs + 3),
|
object_id: DefinitionId(defs + 3),
|
||||||
type_vars: Vec::default(),
|
type_vars: Vec::default(),
|
||||||
fields: [("a".into(), bool, true), ("b".into(), fun, false)].into(),
|
fields: [("a".into(), bool, true), ("b".into(), fun, false)].into(),
|
||||||
attributes: Vec::default(),
|
|
||||||
methods: Vec::default(),
|
methods: Vec::default(),
|
||||||
ancestors: Vec::default(),
|
ancestors: Vec::default(),
|
||||||
resolver: None,
|
resolver: None,
|
||||||
@ -465,11 +420,6 @@ impl TestEnvironment {
|
|||||||
"range".into(),
|
"range".into(),
|
||||||
"str".into(),
|
"str".into(),
|
||||||
"exception".into(),
|
"exception".into(),
|
||||||
"uint32".into(),
|
|
||||||
"uint64".into(),
|
|
||||||
"option".into(),
|
|
||||||
"list".into(),
|
|
||||||
"ndarray".into(),
|
|
||||||
"Foo".into(),
|
"Foo".into(),
|
||||||
"Bar".into(),
|
"Bar".into(),
|
||||||
"Bar2".into(),
|
"Bar2".into(),
|
||||||
@ -520,7 +470,7 @@ impl TestEnvironment {
|
|||||||
primitives: &mut self.primitives,
|
primitives: &mut self.primitives,
|
||||||
virtual_checks: &mut self.virtual_checks,
|
virtual_checks: &mut self.virtual_checks,
|
||||||
calls: &mut self.calls,
|
calls: &mut self.calls,
|
||||||
defined_identifiers: HashMap::default(),
|
defined_identifiers: HashSet::default(),
|
||||||
in_handler: false,
|
in_handler: false,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -596,9 +546,8 @@ fn test_basic(source: &str, mapping: &HashMap<&str, &str>, virtuals: &[(&str, &s
|
|||||||
println!("source:\n{source}");
|
println!("source:\n{source}");
|
||||||
let mut env = TestEnvironment::new();
|
let mut env = TestEnvironment::new();
|
||||||
let id_to_name = std::mem::take(&mut env.id_to_name);
|
let id_to_name = std::mem::take(&mut env.id_to_name);
|
||||||
let mut defined_identifiers: HashMap<_, _> =
|
let mut defined_identifiers: HashSet<_> = env.identifier_mapping.keys().copied().collect();
|
||||||
env.identifier_mapping.keys().copied().map(|id| (id, IdentifierInfo::default())).collect();
|
defined_identifiers.insert("virtual".into());
|
||||||
defined_identifiers.insert("virtual".into(), IdentifierInfo::default());
|
|
||||||
let mut inferencer = env.get_inferencer();
|
let mut inferencer = env.get_inferencer();
|
||||||
inferencer.defined_identifiers.clone_from(&defined_identifiers);
|
inferencer.defined_identifiers.clone_from(&defined_identifiers);
|
||||||
let statements = parse_program(source, FileName::default()).unwrap();
|
let statements = parse_program(source, FileName::default()).unwrap();
|
||||||
@ -743,9 +692,8 @@ fn test_primitive_magic_methods(source: &str, mapping: &HashMap<&str, &str>) {
|
|||||||
println!("source:\n{source}");
|
println!("source:\n{source}");
|
||||||
let mut env = TestEnvironment::basic_test_env();
|
let mut env = TestEnvironment::basic_test_env();
|
||||||
let id_to_name = std::mem::take(&mut env.id_to_name);
|
let id_to_name = std::mem::take(&mut env.id_to_name);
|
||||||
let mut defined_identifiers: HashMap<_, _> =
|
let mut defined_identifiers: HashSet<_> = env.identifier_mapping.keys().copied().collect();
|
||||||
env.identifier_mapping.keys().copied().map(|id| (id, IdentifierInfo::default())).collect();
|
defined_identifiers.insert("virtual".into());
|
||||||
defined_identifiers.insert("virtual".into(), IdentifierInfo::default());
|
|
||||||
let mut inferencer = env.get_inferencer();
|
let mut inferencer = env.get_inferencer();
|
||||||
inferencer.defined_identifiers.clone_from(&defined_identifiers);
|
inferencer.defined_identifiers.clone_from(&defined_identifiers);
|
||||||
let statements = parse_program(source, FileName::default()).unwrap();
|
let statements = parse_program(source, FileName::default()).unwrap();
|
||||||
|
@ -1,28 +1,20 @@
|
|||||||
use std::{
|
|
||||||
borrow::Cow,
|
|
||||||
cell::RefCell,
|
|
||||||
collections::{HashMap, HashSet},
|
|
||||||
fmt::{self, Display},
|
|
||||||
iter::{repeat, zip},
|
|
||||||
rc::Rc,
|
|
||||||
sync::{Arc, Mutex},
|
|
||||||
};
|
|
||||||
|
|
||||||
use indexmap::IndexMap;
|
use indexmap::IndexMap;
|
||||||
use itertools::{repeat_n, Itertools};
|
use itertools::Itertools;
|
||||||
|
use std::cell::RefCell;
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::fmt::{self, Display};
|
||||||
|
use std::iter::zip;
|
||||||
|
use std::rc::Rc;
|
||||||
|
use std::sync::{Arc, Mutex};
|
||||||
|
use std::{borrow::Cow, collections::HashSet};
|
||||||
|
|
||||||
use nac3parser::ast::{Cmpop, Location, StrRef, Unaryop};
|
use nac3parser::ast::{Location, StrRef};
|
||||||
|
|
||||||
use super::{
|
use super::type_error::{TypeError, TypeErrorKind};
|
||||||
magic_methods::{Binop, HasOpInfo},
|
use super::unification_table::{UnificationKey, UnificationTable};
|
||||||
type_error::{TypeError, TypeErrorKind},
|
use crate::symbol_resolver::SymbolValue;
|
||||||
unification_table::{UnificationKey, UnificationTable},
|
use crate::toplevel::{DefinitionId, TopLevelContext, TopLevelDef};
|
||||||
};
|
use crate::typecheck::type_inferencer::PrimitiveStore;
|
||||||
use crate::{
|
|
||||||
symbol_resolver::SymbolValue,
|
|
||||||
toplevel::{helper::PrimDef, DefinitionId, TopLevelContext, TopLevelDef},
|
|
||||||
typecheck::{magic_methods::OpInfo, type_inferencer::PrimitiveStore},
|
|
||||||
};
|
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod test;
|
mod test;
|
||||||
@ -81,28 +73,6 @@ pub fn iter_type_vars(var_map: &VarMap) -> impl Iterator<Item = TypeVar> + '_ {
|
|||||||
var_map.iter().map(|(&id, &ty)| TypeVar { id, ty })
|
var_map.iter().map(|(&id, &ty)| TypeVar { id, ty })
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
|
||||||
pub enum OperatorInfo {
|
|
||||||
/// The call was written as an unary operation, e.g., `~a` or `not a`.
|
|
||||||
IsUnaryOp {
|
|
||||||
/// The [`Type`] of the `self` object
|
|
||||||
self_type: Type,
|
|
||||||
operator: Unaryop,
|
|
||||||
},
|
|
||||||
/// The call was written as a binary operation, e.g., `a + b` or `a += b`.
|
|
||||||
IsBinaryOp {
|
|
||||||
/// The [`Type`] of the `self` object
|
|
||||||
self_type: Type,
|
|
||||||
operator: Binop,
|
|
||||||
},
|
|
||||||
/// The call was written as a binary comparison operation, e.g., `a < b`.
|
|
||||||
IsComparisonOp {
|
|
||||||
/// The [`Type`] of the `self` object
|
|
||||||
self_type: Type,
|
|
||||||
operator: Cmpop,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct Call {
|
pub struct Call {
|
||||||
pub posargs: Vec<Type>,
|
pub posargs: Vec<Type>,
|
||||||
@ -110,9 +80,6 @@ pub struct Call {
|
|||||||
pub ret: Type,
|
pub ret: Type,
|
||||||
pub fun: RefCell<Option<Type>>,
|
pub fun: RefCell<Option<Type>>,
|
||||||
pub loc: Option<Location>,
|
pub loc: Option<Location>,
|
||||||
|
|
||||||
/// Details about the associated Python user operator expression of this call, if any.
|
|
||||||
pub operator_info: Option<OperatorInfo>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
@ -120,14 +87,6 @@ pub struct FuncArg {
|
|||||||
pub name: StrRef,
|
pub name: StrRef,
|
||||||
pub ty: Type,
|
pub ty: Type,
|
||||||
pub default_value: Option<SymbolValue>,
|
pub default_value: Option<SymbolValue>,
|
||||||
pub is_vararg: bool,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl FuncArg {
|
|
||||||
#[must_use]
|
|
||||||
pub fn is_required(&self) -> bool {
|
|
||||||
self.default_value.is_none()
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Debug, Clone)]
|
||||||
@ -239,12 +198,12 @@ pub enum TypeEnum {
|
|||||||
TTuple {
|
TTuple {
|
||||||
/// The types of elements present in this tuple.
|
/// The types of elements present in this tuple.
|
||||||
ty: Vec<Type>,
|
ty: Vec<Type>,
|
||||||
|
},
|
||||||
|
|
||||||
/// Whether this tuple is used in a vararg context.
|
/// A list type.
|
||||||
///
|
TList {
|
||||||
/// If `true`, `ty` must only contain one type, and the tuple is assumed to contain any
|
/// The type of elements present in this list.
|
||||||
/// number of `ty`-typed values.
|
ty: Type,
|
||||||
is_vararg_ctx: bool,
|
|
||||||
},
|
},
|
||||||
|
|
||||||
/// An object type.
|
/// An object type.
|
||||||
@ -280,6 +239,7 @@ impl TypeEnum {
|
|||||||
TypeEnum::TVar { .. } => "TVar",
|
TypeEnum::TVar { .. } => "TVar",
|
||||||
TypeEnum::TLiteral { .. } => "TConstant",
|
TypeEnum::TLiteral { .. } => "TConstant",
|
||||||
TypeEnum::TTuple { .. } => "TTuple",
|
TypeEnum::TTuple { .. } => "TTuple",
|
||||||
|
TypeEnum::TList { .. } => "TList",
|
||||||
TypeEnum::TObj { .. } => "TObj",
|
TypeEnum::TObj { .. } => "TObj",
|
||||||
TypeEnum::TVirtual { .. } => "TVirtual",
|
TypeEnum::TVirtual { .. } => "TVirtual",
|
||||||
TypeEnum::TCall { .. } => "TCall",
|
TypeEnum::TCall { .. } => "TCall",
|
||||||
@ -515,31 +475,13 @@ impl Unifier {
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
TypeEnum::TObj { obj_id, params, .. } if *obj_id == PrimDef::List.id() => {
|
TypeEnum::TList { ty } => self
|
||||||
let tv = iter_type_vars(params).nth(0).unwrap();
|
.get_instantiations(*ty)
|
||||||
|
.map(|ty| ty.iter().map(|&ty| self.add_ty(TypeEnum::TList { ty })).collect_vec()),
|
||||||
let tv_id = if let TypeEnum::TVar { id, .. } =
|
|
||||||
self.unification_table.probe_value(tv.ty).as_ref()
|
|
||||||
{
|
|
||||||
*id
|
|
||||||
} else {
|
|
||||||
tv.id
|
|
||||||
};
|
|
||||||
|
|
||||||
self.get_instantiations(tv.ty).map(|ty_insts| {
|
|
||||||
ty_insts
|
|
||||||
.iter()
|
|
||||||
.map(|&ty_inst| {
|
|
||||||
self.subst(ty, &into_var_map([TypeVar { id: tv_id, ty: ty_inst }]))
|
|
||||||
.unwrap_or(ty)
|
|
||||||
})
|
|
||||||
.collect()
|
|
||||||
})
|
|
||||||
}
|
|
||||||
TypeEnum::TVirtual { ty } => self.get_instantiations(*ty).map(|ty| {
|
TypeEnum::TVirtual { ty } => self.get_instantiations(*ty).map(|ty| {
|
||||||
ty.iter().map(|&ty| self.add_ty(TypeEnum::TVirtual { ty })).collect_vec()
|
ty.iter().map(|&ty| self.add_ty(TypeEnum::TVirtual { ty })).collect_vec()
|
||||||
}),
|
}),
|
||||||
TypeEnum::TTuple { ty, is_vararg_ctx } => {
|
TypeEnum::TTuple { ty } => {
|
||||||
let tuples = ty
|
let tuples = ty
|
||||||
.iter()
|
.iter()
|
||||||
.map(|ty| self.get_instantiations(*ty).unwrap_or_else(|| vec![*ty]))
|
.map(|ty| self.get_instantiations(*ty).unwrap_or_else(|| vec![*ty]))
|
||||||
@ -549,12 +491,7 @@ impl Unifier {
|
|||||||
None
|
None
|
||||||
} else {
|
} else {
|
||||||
Some(
|
Some(
|
||||||
tuples
|
tuples.into_iter().map(|ty| self.add_ty(TypeEnum::TTuple { ty })).collect(),
|
||||||
.into_iter()
|
|
||||||
.map(|ty| {
|
|
||||||
self.add_ty(TypeEnum::TTuple { ty, is_vararg_ctx: *is_vararg_ctx })
|
|
||||||
})
|
|
||||||
.collect(),
|
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -597,8 +534,10 @@ impl Unifier {
|
|||||||
|
|
||||||
TVar { .. } => allowed_typevars.iter().any(|b| self.unification_table.unioned(a, *b)),
|
TVar { .. } => allowed_typevars.iter().any(|b| self.unification_table.unioned(a, *b)),
|
||||||
TCall { .. } => false,
|
TCall { .. } => false,
|
||||||
TVirtual { ty } => self.is_concrete(*ty, allowed_typevars),
|
TList { ty }
|
||||||
TTuple { ty, .. } => ty.iter().all(|ty| self.is_concrete(*ty, allowed_typevars)),
|
| TVirtual { ty } => self.is_concrete(*ty, allowed_typevars),
|
||||||
|
|
||||||
|
TTuple { ty } => ty.iter().all(|ty| self.is_concrete(*ty, allowed_typevars)),
|
||||||
TObj { params: vars, .. } => {
|
TObj { params: vars, .. } => {
|
||||||
vars.values().all(|ty| self.is_concrete(*ty, allowed_typevars))
|
vars.values().all(|ty| self.is_concrete(*ty, allowed_typevars))
|
||||||
}
|
}
|
||||||
@ -623,232 +562,61 @@ impl Unifier {
|
|||||||
call: &Call,
|
call: &Call,
|
||||||
b: Type,
|
b: Type,
|
||||||
signature: &FunSignature,
|
signature: &FunSignature,
|
||||||
|
required: &[StrRef],
|
||||||
) -> Result<(), TypeError> {
|
) -> Result<(), TypeError> {
|
||||||
/*
|
|
||||||
NOTE: scenarios to consider:
|
|
||||||
|
|
||||||
```python
|
|
||||||
def func1(x: int32, y: int32, z: int32 = 5): pass
|
|
||||||
|
|
||||||
# Normal scenarios
|
|
||||||
func1(23, 45) # OK, z has default
|
|
||||||
func1(23, 45, 67) # OK, z's default is overwritten
|
|
||||||
func1(x = 23, y = 45) # OK, user is using kwargs to set positional args
|
|
||||||
func1(y = 45, x = 23) # OK, kwargs order doesn't matter
|
|
||||||
|
|
||||||
# Error scenarios
|
|
||||||
func1() # ERROR: Missing arguments: x, y
|
|
||||||
func1(23) # ERROR: Missing arguments: y
|
|
||||||
func1(z = 23) # ERROR: Missing arguments: x, y
|
|
||||||
func1(x = 23) # ERROR: Missing arguments: y
|
|
||||||
func1(23, 45, x = 5) # ERROR: Got multiple values for x
|
|
||||||
func1(23, 45, x = 5, y = 6) # ERROR: Got multiple values for x (y too but Python does not report it)
|
|
||||||
func1(23, 45, 67, z = 89) # ERROR: Got multiple values for z
|
|
||||||
func1(23, 45, 67, 89) # ERROR: Function only takes from 2 to 3 positional arguments but 4 were given.
|
|
||||||
func1(23, 45, 67, w = 3) # ERROR: Got an unexpected keyword argument 'w'
|
|
||||||
|
|
||||||
# Error scenarios that do not need to be handled here.
|
|
||||||
func1(23, 45, z = 67, z = 89) # ERROR: Keyword argument repeated: z, the parser panics on this.
|
|
||||||
```
|
|
||||||
*/
|
|
||||||
|
|
||||||
struct ParamInfo<'a> {
|
|
||||||
/// Has this parameter been supplied with an argument already?
|
|
||||||
has_been_supplied: bool,
|
|
||||||
/// The corresponding [`FuncArg`] instance of this parameter (for fast table lookups)
|
|
||||||
param: &'a FuncArg,
|
|
||||||
}
|
|
||||||
|
|
||||||
let snapshot = self.unification_table.get_snapshot();
|
let snapshot = self.unification_table.get_snapshot();
|
||||||
if self.snapshot.is_none() {
|
if self.snapshot.is_none() {
|
||||||
self.snapshot = Some(snapshot);
|
self.snapshot = Some(snapshot);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get details about the function signature/parameters.
|
let Call { posargs, kwargs, ret, fun, loc } = call;
|
||||||
let num_params = signature.args.len();
|
let instantiated = self.instantiate_fun(b, signature);
|
||||||
let is_vararg = signature.args.iter().any(|arg| arg.is_vararg);
|
let r = self.get_ty(instantiated);
|
||||||
|
let r = r.as_ref();
|
||||||
// Force the type vars in `b` and `signature' to be up-to-date.
|
let TypeEnum::TFunc(signature) = r else { unreachable!() };
|
||||||
let b = self.instantiate_fun(b, signature);
|
// we check to make sure that all required arguments (those without default
|
||||||
let TypeEnum::TFunc(signature) = &*self.get_ty(b) else { unreachable!() };
|
// arguments) are provided, and do not provide the same argument twice.
|
||||||
|
let mut required = required.to_vec();
|
||||||
// Get details about the input arguments
|
let mut all_names: Vec<_> = signature.args.iter().map(|v| (v.name, v.ty)).rev().collect();
|
||||||
let Call { posargs, kwargs, ret, fun, loc, operator_info } = call;
|
for (i, t) in posargs.iter().enumerate() {
|
||||||
let num_args = posargs.len() + kwargs.len();
|
if signature.args.len() <= i {
|
||||||
|
|
||||||
// Now we check the arguments against the parameters,
|
|
||||||
// and depending on what `call_info` is, we might change how `unify_call()` behaves
|
|
||||||
// to improve user error messages when type checking fails.
|
|
||||||
match operator_info {
|
|
||||||
Some(OperatorInfo::IsBinaryOp { self_type, operator }) => {
|
|
||||||
// The call is written in the form of (say) `a + b`.
|
|
||||||
// Technically, it is `a.__add__(b)`, and they have the following constraints:
|
|
||||||
assert_eq!(posargs.len(), 1);
|
|
||||||
assert_eq!(kwargs.len(), 0);
|
|
||||||
assert_eq!(num_params, 1);
|
|
||||||
|
|
||||||
let other_type = posargs[0]; // the second operand
|
|
||||||
let expected_other_type = signature.args[0].ty;
|
|
||||||
|
|
||||||
let ok = self.unify_impl(expected_other_type, other_type, false).is_ok();
|
|
||||||
if !ok {
|
|
||||||
self.restore_snapshot();
|
|
||||||
return Err(TypeError::new(
|
|
||||||
TypeErrorKind::UnsupportedBinaryOpTypes {
|
|
||||||
operator: *operator,
|
|
||||||
lhs_type: *self_type,
|
|
||||||
rhs_type: other_type,
|
|
||||||
expected_rhs_type: expected_other_type,
|
|
||||||
},
|
|
||||||
*loc,
|
|
||||||
));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Some(OperatorInfo::IsComparisonOp { self_type, operator })
|
|
||||||
if OpInfo::supports_cmpop(*operator) // Otherwise that comparison operator is not supported.
|
|
||||||
=>
|
|
||||||
{
|
|
||||||
// The call is written in the form of (say) `a <= b`.
|
|
||||||
// Technically, it is `a.__le__(b)`, and they have the following constraints:
|
|
||||||
assert_eq!(posargs.len(), 1);
|
|
||||||
assert_eq!(kwargs.len(), 0);
|
|
||||||
assert_eq!(num_params, 1);
|
|
||||||
|
|
||||||
let other_type = posargs[0]; // the second operand
|
|
||||||
let expected_other_type = signature.args[0].ty;
|
|
||||||
|
|
||||||
let ok = self.unify_impl(expected_other_type, other_type, false).is_ok();
|
|
||||||
if !ok {
|
|
||||||
self.restore_snapshot();
|
|
||||||
return Err(TypeError::new(
|
|
||||||
TypeErrorKind::UnsupportedComparsionOpTypes {
|
|
||||||
operator: *operator,
|
|
||||||
lhs_type: *self_type,
|
|
||||||
rhs_type: other_type,
|
|
||||||
expected_rhs_type: expected_other_type,
|
|
||||||
},
|
|
||||||
*loc,
|
|
||||||
));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
// Handle [`CallInfo::IsNormalFunctionCall`] and other uninteresting variants
|
|
||||||
// of [`CallInfo`] (e.g, `CallInfo::IsUnaryOp` and unsupported comparison operators)
|
|
||||||
|
|
||||||
// Helper lambdas
|
|
||||||
let mut type_check_arg = |param_name, expected_arg_ty, arg_ty| {
|
|
||||||
let ok = self.unify_impl(expected_arg_ty, arg_ty, false).is_ok();
|
|
||||||
if ok {
|
|
||||||
Ok(())
|
|
||||||
} else {
|
|
||||||
// Typecheck failed, throw an error.
|
|
||||||
self.restore_snapshot();
|
|
||||||
Err(TypeError::new(
|
|
||||||
TypeErrorKind::IncorrectArgType {
|
|
||||||
name: param_name,
|
|
||||||
expected: expected_arg_ty,
|
|
||||||
got: arg_ty,
|
|
||||||
},
|
|
||||||
*loc,
|
|
||||||
))
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
// Check for "too many arguments"
|
|
||||||
if !is_vararg && num_params < posargs.len() {
|
|
||||||
let expected_min_count =
|
|
||||||
signature.args.iter().filter(|param| param.is_required()).count();
|
|
||||||
let expected_max_count = num_params;
|
|
||||||
|
|
||||||
self.restore_snapshot();
|
self.restore_snapshot();
|
||||||
return Err(TypeError::new(
|
return Err(TypeError::new(
|
||||||
TypeErrorKind::TooManyArguments {
|
TypeErrorKind::TooManyArguments {
|
||||||
expected_min_count,
|
expected: signature.args.len(),
|
||||||
expected_max_count,
|
got: posargs.len() + kwargs.len(),
|
||||||
got_count: num_args,
|
|
||||||
},
|
},
|
||||||
*loc,
|
*loc,
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
|
required.pop();
|
||||||
// NOTE: order of `param_info_by_name` is leveraged, so use an IndexMap
|
let (name, expected) = all_names.pop().unwrap();
|
||||||
let mut param_info_by_name: IndexMap<StrRef, ParamInfo> = signature
|
self.unify_impl(expected, *t, false).map_err(|_| {
|
||||||
.args
|
self.restore_snapshot();
|
||||||
.iter()
|
TypeError::new(TypeErrorKind::IncorrectArgType { name, expected, got: *t }, *loc)
|
||||||
.map(|arg| (arg.name, ParamInfo { has_been_supplied: false, param: arg }))
|
})?;
|
||||||
.collect();
|
|
||||||
|
|
||||||
// Now consume all positional arguments and typecheck them.
|
|
||||||
for (&arg_ty, param) in zip(posargs, signature.args.iter()) {
|
|
||||||
// We will also use this opportunity to mark the corresponding `param_info` as having been supplied.
|
|
||||||
let param_info = param_info_by_name.get_mut(¶m.name).unwrap();
|
|
||||||
param_info.has_been_supplied = true;
|
|
||||||
|
|
||||||
// Typecheck
|
|
||||||
type_check_arg(param.name, param.ty, arg_ty)?;
|
|
||||||
}
|
}
|
||||||
|
for (k, t) in kwargs {
|
||||||
if is_vararg {
|
if let Some(i) = required.iter().position(|v| v == k) {
|
||||||
debug_assert!(!signature.args.is_empty());
|
required.remove(i);
|
||||||
|
|
||||||
let vararg_args = posargs.iter().skip(signature.args.len());
|
|
||||||
let vararg_param = signature.args.last().unwrap();
|
|
||||||
|
|
||||||
for (&arg_ty, param) in zip(vararg_args, repeat(vararg_param)) {
|
|
||||||
// `param_info` for this argument would've already been marked as supplied
|
|
||||||
// during non-vararg posarg typecheck
|
|
||||||
type_check_arg(param.name, param.ty, arg_ty)?;
|
|
||||||
}
|
}
|
||||||
|
let i = all_names.iter().position(|v| &v.0 == k).ok_or_else(|| {
|
||||||
|
self.restore_snapshot();
|
||||||
|
TypeError::new(TypeErrorKind::UnknownArgName(*k), *loc)
|
||||||
|
})?;
|
||||||
|
let (name, expected) = all_names.remove(i);
|
||||||
|
self.unify_impl(expected, *t, false).map_err(|_| {
|
||||||
|
self.restore_snapshot();
|
||||||
|
TypeError::new(TypeErrorKind::IncorrectArgType { name, expected, got: *t }, *loc)
|
||||||
|
})?;
|
||||||
}
|
}
|
||||||
|
if !required.is_empty() {
|
||||||
// Now consume all keyword arguments and typecheck them.
|
|
||||||
for (¶m_name, &arg_ty) in kwargs {
|
|
||||||
// We will also use this opportunity to check if this keyword argument is "legal".
|
|
||||||
|
|
||||||
let Some(param_info) = param_info_by_name.get_mut(¶m_name) else {
|
|
||||||
self.restore_snapshot();
|
self.restore_snapshot();
|
||||||
return Err(TypeError::new(
|
return Err(TypeError::new(
|
||||||
TypeErrorKind::UnknownArgName(param_name),
|
TypeErrorKind::MissingArgs(required.iter().join(", ")),
|
||||||
*loc,
|
|
||||||
));
|
|
||||||
};
|
|
||||||
|
|
||||||
if param_info.has_been_supplied {
|
|
||||||
// NOTE: Duplicate keyword argument (i.e., `hello(1, 2, 3, arg = 4, arg = 5)`)
|
|
||||||
// is IMPOSSIBLE as the parser would have already failed.
|
|
||||||
// We only have to care about "got multiple values for XYZ"
|
|
||||||
|
|
||||||
self.restore_snapshot();
|
|
||||||
return Err(TypeError::new(
|
|
||||||
TypeErrorKind::GotMultipleValues { name: param_name },
|
|
||||||
*loc,
|
*loc,
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
|
|
||||||
param_info.has_been_supplied = true;
|
|
||||||
|
|
||||||
// Typecheck
|
|
||||||
type_check_arg(param_name, param_info.param.ty, arg_ty)?;
|
|
||||||
}
|
|
||||||
|
|
||||||
// After checking posargs and kwargs, check if there are any
|
|
||||||
// unsupplied required parameters, and throw an error if they exist.
|
|
||||||
let missing_arg_names = param_info_by_name
|
|
||||||
.values()
|
|
||||||
.filter(|param_info| {
|
|
||||||
param_info.param.is_required() && !param_info.has_been_supplied
|
|
||||||
})
|
|
||||||
.map(|param_info| param_info.param.name)
|
|
||||||
.collect_vec();
|
|
||||||
if !missing_arg_names.is_empty() {
|
|
||||||
self.restore_snapshot();
|
|
||||||
return Err(TypeError::new(
|
|
||||||
TypeErrorKind::MissingArgs { missing_arg_names },
|
|
||||||
*loc,
|
|
||||||
));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Finally, check the Call's return type
|
|
||||||
self.unify_impl(*ret, signature.ret, false).map_err(|mut err| {
|
self.unify_impl(*ret, signature.ret, false).map_err(|mut err| {
|
||||||
self.restore_snapshot();
|
self.restore_snapshot();
|
||||||
if err.loc.is_none() {
|
if err.loc.is_none() {
|
||||||
@ -856,10 +624,7 @@ impl Unifier {
|
|||||||
}
|
}
|
||||||
err
|
err
|
||||||
})?;
|
})?;
|
||||||
}
|
*fun.borrow_mut() = Some(instantiated);
|
||||||
}
|
|
||||||
|
|
||||||
*fun.borrow_mut() = Some(b);
|
|
||||||
|
|
||||||
self.discard_snapshot(snapshot);
|
self.discard_snapshot(snapshot);
|
||||||
Ok(())
|
Ok(())
|
||||||
@ -990,10 +755,7 @@ impl Unifier {
|
|||||||
self.unify_impl(x, b, false)?;
|
self.unify_impl(x, b, false)?;
|
||||||
self.set_a_to_b(a, x);
|
self.set_a_to_b(a, x);
|
||||||
}
|
}
|
||||||
(
|
(TVar { fields: Some(fields), range, is_const_generic: false, .. }, TTuple { ty }) => {
|
||||||
TVar { fields: Some(fields), range, is_const_generic: false, .. },
|
|
||||||
TTuple { ty, .. },
|
|
||||||
) => {
|
|
||||||
let len = i32::try_from(ty.len()).unwrap();
|
let len = i32::try_from(ty.len()).unwrap();
|
||||||
for (k, v) in fields {
|
for (k, v) in fields {
|
||||||
match *k {
|
match *k {
|
||||||
@ -1014,18 +776,8 @@ impl Unifier {
|
|||||||
self.unify_impl(v.ty, ty[ind as usize], false)
|
self.unify_impl(v.ty, ty[ind as usize], false)
|
||||||
.map_err(|e| e.at(v.loc))?;
|
.map_err(|e| e.at(v.loc))?;
|
||||||
}
|
}
|
||||||
RecordKey::Str(s) => {
|
RecordKey::Str(_) => {
|
||||||
let tuple_fns = [
|
return Err(TypeError::new(TypeErrorKind::NoSuchField(*k, b), v.loc))
|
||||||
Cmpop::Eq.op_info().method_name,
|
|
||||||
Cmpop::NotEq.op_info().method_name,
|
|
||||||
];
|
|
||||||
|
|
||||||
if !tuple_fns.into_iter().any(|op| s.to_string() == op) {
|
|
||||||
return Err(TypeError::new(
|
|
||||||
TypeErrorKind::NoSuchField(*k, b),
|
|
||||||
v.loc,
|
|
||||||
));
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1033,6 +785,22 @@ impl Unifier {
|
|||||||
self.unify_impl(x, b, false)?;
|
self.unify_impl(x, b, false)?;
|
||||||
self.set_a_to_b(a, x);
|
self.set_a_to_b(a, x);
|
||||||
}
|
}
|
||||||
|
(TVar { fields: Some(fields), range, is_const_generic: false, .. }, TList { ty }) => {
|
||||||
|
for (k, v) in fields {
|
||||||
|
match *k {
|
||||||
|
RecordKey::Int(_) => {
|
||||||
|
self.unify_impl(v.ty, *ty, false).map_err(|e| e.at(v.loc))?;
|
||||||
|
}
|
||||||
|
RecordKey::Str(_) => {
|
||||||
|
return Err(TypeError::new(TypeErrorKind::NoSuchField(*k, b), v.loc))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let x = self.check_var_compatibility(b, range)?.unwrap_or(b);
|
||||||
|
self.unify_impl(x, b, false)?;
|
||||||
|
self.set_a_to_b(a, x);
|
||||||
|
}
|
||||||
|
|
||||||
(
|
(
|
||||||
TVar { id: id1, range: ty1, is_const_generic: true, .. },
|
TVar { id: id1, range: ty1, is_const_generic: true, .. },
|
||||||
TVar { id: id2, range: ty2, .. },
|
TVar { id: id2, range: ty2, .. },
|
||||||
@ -1100,50 +868,24 @@ impl Unifier {
|
|||||||
self.set_a_to_b(a, b);
|
self.set_a_to_b(a, b);
|
||||||
}
|
}
|
||||||
|
|
||||||
(
|
(TTuple { ty: ty1 }, TTuple { ty: ty2 }) => {
|
||||||
TTuple { ty: ty1, is_vararg_ctx: is_vararg1 },
|
|
||||||
TTuple { ty: ty2, is_vararg_ctx: is_vararg2 },
|
|
||||||
) => {
|
|
||||||
// Rules for Tuples:
|
|
||||||
// - ty1: is_vararg && ty2: is_vararg -> ty1[0] == ty2[0]
|
|
||||||
// - ty1: is_vararg && ty2: !is_vararg -> type error (not enough info to infer the correct number of arguments)
|
|
||||||
// - ty1: !is_vararg && ty2: is_vararg -> ty1[..] == ty2[0]
|
|
||||||
// - ty1: !is_vararg && ty2: !is_vararg -> ty1.len() == ty2.len() && ty1[i] == ty2[i]
|
|
||||||
|
|
||||||
debug_assert!(!is_vararg1 || ty1.len() == 1);
|
|
||||||
debug_assert!(!is_vararg2 || ty2.len() == 1);
|
|
||||||
|
|
||||||
match (*is_vararg1, *is_vararg2) {
|
|
||||||
(true, true) => {
|
|
||||||
if self.unify_impl(ty1[0], ty2[0], false).is_err() {
|
|
||||||
return Self::incompatible_types(a, b);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
(true, false) => return Self::incompatible_types(a, b),
|
|
||||||
|
|
||||||
(false, true) => {
|
|
||||||
for y in ty2 {
|
|
||||||
if self.unify_impl(ty1[0], *y, false).is_err() {
|
|
||||||
return Self::incompatible_types(a, b);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
(false, false) => {
|
|
||||||
if ty1.len() != ty2.len() {
|
if ty1.len() != ty2.len() {
|
||||||
return Self::incompatible_types(a, b);
|
return Err(TypeError::new(TypeErrorKind::IncompatibleTypes(a, b), None));
|
||||||
}
|
}
|
||||||
|
|
||||||
for (x, y) in ty1.iter().zip(ty2.iter()) {
|
for (x, y) in ty1.iter().zip(ty2.iter()) {
|
||||||
if self.unify_impl(*x, *y, false).is_err() {
|
if self.unify_impl(*x, *y, false).is_err() {
|
||||||
return Self::incompatible_types(a, b);
|
return Err(TypeError::new(TypeErrorKind::IncompatibleTypes(a, b), None));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
self.set_a_to_b(a, b);
|
self.set_a_to_b(a, b);
|
||||||
}
|
}
|
||||||
(TVar { fields: Some(map), range, .. }, TObj { obj_id, fields, params }) => {
|
(TList { ty: ty1 }, TList { ty: ty2 }) => {
|
||||||
|
if self.unify_impl(*ty1, *ty2, false).is_err() {
|
||||||
|
return Err(TypeError::new(TypeErrorKind::IncompatibleTypes(a, b), None));
|
||||||
|
}
|
||||||
|
self.set_a_to_b(a, b);
|
||||||
|
}
|
||||||
|
(TVar { fields: Some(map), range, .. }, TObj { fields, .. }) => {
|
||||||
for (k, field) in map {
|
for (k, field) in map {
|
||||||
match *k {
|
match *k {
|
||||||
RecordKey::Str(s) => {
|
RecordKey::Str(s) => {
|
||||||
@ -1162,18 +904,10 @@ impl Unifier {
|
|||||||
self.unify_impl(field.ty, ty, false).map_err(|v| v.at(field.loc))?;
|
self.unify_impl(field.ty, ty, false).map_err(|v| v.at(field.loc))?;
|
||||||
}
|
}
|
||||||
RecordKey::Int(_) => {
|
RecordKey::Int(_) => {
|
||||||
// Allow expressions such as list[0]
|
|
||||||
if *obj_id == PrimDef::List.id() {
|
|
||||||
let ty = iter_type_vars(params).nth(0).unwrap().ty;
|
|
||||||
|
|
||||||
self.unify_impl(field.ty, ty, false)
|
|
||||||
.map_err(|e| e.at(field.loc))?;
|
|
||||||
} else {
|
|
||||||
return Err(TypeError::new(
|
return Err(TypeError::new(
|
||||||
TypeErrorKind::NoSuchField(*k, b),
|
TypeErrorKind::NoSuchField(*k, b),
|
||||||
field.loc,
|
field.loc,
|
||||||
));
|
))
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1256,10 +990,17 @@ impl Unifier {
|
|||||||
self.unification_table.set_value(b, Rc::new(TCall(calls)));
|
self.unification_table.set_value(b, Rc::new(TCall(calls)));
|
||||||
}
|
}
|
||||||
(TCall(calls), TFunc(signature)) => {
|
(TCall(calls), TFunc(signature)) => {
|
||||||
|
let required: Vec<StrRef> = signature
|
||||||
|
.args
|
||||||
|
.iter()
|
||||||
|
.filter(|v| v.default_value.is_none())
|
||||||
|
.map(|v| v.name)
|
||||||
|
.rev()
|
||||||
|
.collect();
|
||||||
// we unify every calls to the function signature.
|
// we unify every calls to the function signature.
|
||||||
for c in calls {
|
for c in calls {
|
||||||
let call = self.calls[c.0].clone();
|
let call = self.calls[c.0].clone();
|
||||||
self.unify_call(&call, b, signature)?;
|
self.unify_call(&call, b, signature, &required)?;
|
||||||
}
|
}
|
||||||
self.set_a_to_b(a, b);
|
self.set_a_to_b(a, b);
|
||||||
}
|
}
|
||||||
@ -1383,22 +1124,13 @@ impl Unifier {
|
|||||||
TypeEnum::TLiteral { values, .. } => {
|
TypeEnum::TLiteral { values, .. } => {
|
||||||
format!("const({})", values.iter().map(|v| format!("{v:?}")).join(", "))
|
format!("const({})", values.iter().map(|v| format!("{v:?}")).join(", "))
|
||||||
}
|
}
|
||||||
TypeEnum::TTuple { ty, is_vararg_ctx } => {
|
TypeEnum::TTuple { ty } => {
|
||||||
if *is_vararg_ctx {
|
let mut fields =
|
||||||
debug_assert_eq!(ty.len(), 1);
|
ty.iter().map(|v| self.internal_stringify(*v, obj_to_name, var_to_name, notes));
|
||||||
let field = self.internal_stringify(
|
|
||||||
*ty.iter().next().unwrap(),
|
|
||||||
obj_to_name,
|
|
||||||
var_to_name,
|
|
||||||
notes,
|
|
||||||
);
|
|
||||||
format!("tuple[*{field}]")
|
|
||||||
} else {
|
|
||||||
let mut fields = ty
|
|
||||||
.iter()
|
|
||||||
.map(|v| self.internal_stringify(*v, obj_to_name, var_to_name, notes));
|
|
||||||
format!("tuple[{}]", fields.join(", "))
|
format!("tuple[{}]", fields.join(", "))
|
||||||
}
|
}
|
||||||
|
TypeEnum::TList { ty } => {
|
||||||
|
format!("list[{}]", self.internal_stringify(*ty, obj_to_name, var_to_name, notes))
|
||||||
}
|
}
|
||||||
TypeEnum::TVirtual { ty } => {
|
TypeEnum::TVirtual { ty } => {
|
||||||
format!(
|
format!(
|
||||||
@ -1423,21 +1155,17 @@ impl Unifier {
|
|||||||
.args
|
.args
|
||||||
.iter()
|
.iter()
|
||||||
.map(|arg| {
|
.map(|arg| {
|
||||||
let vararg_prefix = if arg.is_vararg { "*" } else { "" };
|
|
||||||
|
|
||||||
if let Some(dv) = &arg.default_value {
|
if let Some(dv) = &arg.default_value {
|
||||||
format!(
|
format!(
|
||||||
"{}:{}{}={}",
|
"{}:{}={}",
|
||||||
arg.name,
|
arg.name,
|
||||||
vararg_prefix,
|
|
||||||
self.internal_stringify(arg.ty, obj_to_name, var_to_name, notes),
|
self.internal_stringify(arg.ty, obj_to_name, var_to_name, notes),
|
||||||
dv
|
dv
|
||||||
)
|
)
|
||||||
} else {
|
} else {
|
||||||
format!(
|
format!(
|
||||||
"{}:{}{}",
|
"{}:{}",
|
||||||
arg.name,
|
arg.name,
|
||||||
vararg_prefix,
|
|
||||||
self.internal_stringify(arg.ty, obj_to_name, var_to_name, notes)
|
self.internal_stringify(arg.ty, obj_to_name, var_to_name, notes)
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
@ -1523,7 +1251,7 @@ impl Unifier {
|
|||||||
match &*ty {
|
match &*ty {
|
||||||
TypeEnum::TRigidVar { .. } | TypeEnum::TLiteral { .. } => None,
|
TypeEnum::TRigidVar { .. } | TypeEnum::TLiteral { .. } => None,
|
||||||
TypeEnum::TVar { id, .. } => mapping.get(id).copied(),
|
TypeEnum::TVar { id, .. } => mapping.get(id).copied(),
|
||||||
TypeEnum::TTuple { ty, is_vararg_ctx } => {
|
TypeEnum::TTuple { ty } => {
|
||||||
let mut new_ty = Cow::from(ty);
|
let mut new_ty = Cow::from(ty);
|
||||||
for (i, t) in ty.iter().enumerate() {
|
for (i, t) in ty.iter().enumerate() {
|
||||||
if let Some(t1) = self.subst_impl(*t, mapping, cache) {
|
if let Some(t1) = self.subst_impl(*t, mapping, cache) {
|
||||||
@ -1531,14 +1259,14 @@ impl Unifier {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if matches!(new_ty, Cow::Owned(_)) {
|
if matches!(new_ty, Cow::Owned(_)) {
|
||||||
Some(self.add_ty(TypeEnum::TTuple {
|
Some(self.add_ty(TypeEnum::TTuple { ty: new_ty.into_owned() }))
|
||||||
ty: new_ty.into_owned(),
|
|
||||||
is_vararg_ctx: *is_vararg_ctx,
|
|
||||||
}))
|
|
||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
TypeEnum::TList { ty } => {
|
||||||
|
self.subst_impl(*ty, mapping, cache).map(|t| self.add_ty(TypeEnum::TList { ty: t }))
|
||||||
|
}
|
||||||
TypeEnum::TVirtual { ty } => self
|
TypeEnum::TVirtual { ty } => self
|
||||||
.subst_impl(*ty, mapping, cache)
|
.subst_impl(*ty, mapping, cache)
|
||||||
.map(|t| self.add_ty(TypeEnum::TVirtual { ty: t })),
|
.map(|t| self.add_ty(TypeEnum::TVirtual { ty: t })),
|
||||||
@ -1549,7 +1277,6 @@ impl Unifier {
|
|||||||
// This is also used to prevent infinite substitution...
|
// This is also used to prevent infinite substitution...
|
||||||
let need_subst = params.values().any(|v| {
|
let need_subst = params.values().any(|v| {
|
||||||
let ty = self.unification_table.probe_value(*v);
|
let ty = self.unification_table.probe_value(*v);
|
||||||
// TODO(Derppening): #444
|
|
||||||
if let TypeEnum::TVar { id, .. } = ty.as_ref() {
|
if let TypeEnum::TVar { id, .. } = ty.as_ref() {
|
||||||
mapping.contains_key(id)
|
mapping.contains_key(id)
|
||||||
} else {
|
} else {
|
||||||
@ -1694,55 +1421,20 @@ impl Unifier {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
(TVar { range, .. }, _) => self.check_var_compatibility(b, range).or(Err(())),
|
(TVar { range, .. }, _) => self.check_var_compatibility(b, range).or(Err(())),
|
||||||
(
|
(TTuple { ty: ty1 }, TTuple { ty: ty2 }) if ty1.len() == ty2.len() => {
|
||||||
TTuple { ty: ty1, is_vararg_ctx: is_vararg1 },
|
let ty: Vec<_> = zip(ty1.iter(), ty2.iter())
|
||||||
TTuple { ty: ty2, is_vararg_ctx: is_vararg2 },
|
.map(|(a, b)| self.get_intersection(*a, *b))
|
||||||
) => {
|
.try_collect()?;
|
||||||
if *is_vararg1 && *is_vararg2 {
|
if ty.iter().any(Option::is_some) {
|
||||||
let isect_ty = self.get_intersection(ty1[0], ty2[0])?;
|
Ok(Some(self.add_ty(TTuple {
|
||||||
Ok(isect_ty.map(|ty| self.add_ty(TTuple { ty: vec![ty], is_vararg_ctx: true })))
|
|
||||||
} else {
|
|
||||||
let zip_iter: Box<dyn Iterator<Item = (&Type, &Type)>> =
|
|
||||||
match (*is_vararg1, *is_vararg2) {
|
|
||||||
(true, _) => Box::new(repeat_n(&ty1[0], ty2.len()).zip(ty2.iter())),
|
|
||||||
(_, false) => Box::new(ty1.iter().zip(repeat_n(&ty2[0], ty1.len()))),
|
|
||||||
_ => {
|
|
||||||
if ty1.len() != ty2.len() {
|
|
||||||
return Err(());
|
|
||||||
}
|
|
||||||
|
|
||||||
Box::new(ty1.iter().zip(ty2.iter()))
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let ty: Vec<_> =
|
|
||||||
zip_iter.map(|(a, b)| self.get_intersection(*a, *b)).try_collect()?;
|
|
||||||
Ok(if ty.iter().any(Option::is_some) {
|
|
||||||
Some(self.add_ty(TTuple {
|
|
||||||
ty: zip(ty, ty1.iter()).map(|(a, b)| a.unwrap_or(*b)).collect(),
|
ty: zip(ty, ty1.iter()).map(|(a, b)| a.unwrap_or(*b)).collect(),
|
||||||
is_vararg_ctx: false,
|
})))
|
||||||
}))
|
|
||||||
} else {
|
} else {
|
||||||
None
|
Ok(None)
|
||||||
})
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// TODO(Derppening): #444
|
(TList { ty: ty1 }, TList { ty: ty2 }) => {
|
||||||
(
|
Ok(self.get_intersection(*ty1, *ty2)?.map(|ty| self.add_ty(TList { ty })))
|
||||||
TObj { obj_id: id1, fields, params: params1 },
|
|
||||||
TObj { obj_id: id2, params: params2, .. },
|
|
||||||
) if *id1 == PrimDef::List.id() && *id2 == PrimDef::List.id() => {
|
|
||||||
let tv_id = iter_type_vars(params1).nth(0).unwrap().id;
|
|
||||||
let ty1 = iter_type_vars(params1).nth(0).unwrap().ty;
|
|
||||||
let ty2 = iter_type_vars(params2).nth(0).unwrap().ty;
|
|
||||||
|
|
||||||
Ok(self.get_intersection(ty1, ty2)?.map(|ty| {
|
|
||||||
self.add_ty(TObj {
|
|
||||||
obj_id: *id1,
|
|
||||||
fields: fields.clone(),
|
|
||||||
params: into_var_map([TypeVar { id: tv_id, ty }]),
|
|
||||||
})
|
|
||||||
}))
|
|
||||||
}
|
}
|
||||||
(TVirtual { ty: ty1 }, TVirtual { ty: ty2 }) => {
|
(TVirtual { ty: ty1 }, TVirtual { ty: ty2 }) => {
|
||||||
Ok(self.get_intersection(*ty1, *ty2)?.map(|ty| self.add_ty(TVirtual { ty })))
|
Ok(self.get_intersection(*ty1, *ty2)?.map(|ty| self.add_ty(TVirtual { ty })))
|
||||||
|
@ -1,12 +1,10 @@
|
|||||||
use std::collections::HashMap;
|
use super::super::magic_methods::with_fields;
|
||||||
|
use super::*;
|
||||||
use indoc::indoc;
|
use indoc::indoc;
|
||||||
use itertools::Itertools;
|
use itertools::Itertools;
|
||||||
|
use std::collections::HashMap;
|
||||||
use test_case::test_case;
|
use test_case::test_case;
|
||||||
|
|
||||||
use super::*;
|
|
||||||
use crate::typecheck::magic_methods::with_fields;
|
|
||||||
|
|
||||||
impl Unifier {
|
impl Unifier {
|
||||||
/// Check whether two types are equal.
|
/// Check whether two types are equal.
|
||||||
fn eq(&mut self, a: Type, b: Type) -> bool {
|
fn eq(&mut self, a: Type, b: Type) -> bool {
|
||||||
@ -30,14 +28,14 @@ impl Unifier {
|
|||||||
TypeEnum::TVar { fields: Some(map1), .. },
|
TypeEnum::TVar { fields: Some(map1), .. },
|
||||||
TypeEnum::TVar { fields: Some(map2), .. },
|
TypeEnum::TVar { fields: Some(map2), .. },
|
||||||
) => self.map_eq2(map1, map2),
|
) => self.map_eq2(map1, map2),
|
||||||
(
|
(TypeEnum::TTuple { ty: ty1 }, TypeEnum::TTuple { ty: ty2 }) => {
|
||||||
TypeEnum::TTuple { ty: ty1, is_vararg_ctx: false },
|
|
||||||
TypeEnum::TTuple { ty: ty2, is_vararg_ctx: false },
|
|
||||||
) => {
|
|
||||||
ty1.len() == ty2.len()
|
ty1.len() == ty2.len()
|
||||||
&& ty1.iter().zip(ty2.iter()).all(|(t1, t2)| self.eq(*t1, *t2))
|
&& ty1.iter().zip(ty2.iter()).all(|(t1, t2)| self.eq(*t1, *t2))
|
||||||
}
|
}
|
||||||
(TypeEnum::TVirtual { ty: ty1 }, TypeEnum::TVirtual { ty: ty2 }) => self.eq(*ty1, *ty2),
|
(TypeEnum::TList { ty: ty1 }, TypeEnum::TList { ty: ty2 })
|
||||||
|
| (TypeEnum::TVirtual { ty: ty1 }, TypeEnum::TVirtual { ty: ty2 }) => {
|
||||||
|
self.eq(*ty1, *ty2)
|
||||||
|
}
|
||||||
(
|
(
|
||||||
TypeEnum::TObj { obj_id: id1, params: params1, .. },
|
TypeEnum::TObj { obj_id: id1, params: params1, .. },
|
||||||
TypeEnum::TObj { obj_id: id2, params: params2, .. },
|
TypeEnum::TObj { obj_id: id2, params: params2, .. },
|
||||||
@ -121,15 +119,6 @@ impl TestEnvironment {
|
|||||||
params: into_var_map([tvar]),
|
params: into_var_map([tvar]),
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
let tvar = unifier.get_dummy_var();
|
|
||||||
type_mapping.insert(
|
|
||||||
"list".into(),
|
|
||||||
unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: PrimDef::List.id(),
|
|
||||||
fields: HashMap::new(),
|
|
||||||
params: into_var_map([tvar]),
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
|
|
||||||
TestEnvironment { unifier, type_mapping }
|
TestEnvironment { unifier, type_mapping }
|
||||||
}
|
}
|
||||||
@ -144,36 +133,6 @@ impl TestEnvironment {
|
|||||||
// for testing only, so we can just panic when the input is malformed
|
// for testing only, so we can just panic when the input is malformed
|
||||||
let end = typ.find(|c| ['[', ',', ']', '='].contains(&c)).unwrap_or(typ.len());
|
let end = typ.find(|c| ['[', ',', ']', '='].contains(&c)).unwrap_or(typ.len());
|
||||||
match &typ[..end] {
|
match &typ[..end] {
|
||||||
"list" => {
|
|
||||||
let mut s = &typ[end..];
|
|
||||||
assert_eq!(&s[0..1], "[");
|
|
||||||
let mut ty = Vec::new();
|
|
||||||
while &s[0..1] != "]" {
|
|
||||||
let result = self.internal_parse(&s[1..], mapping);
|
|
||||||
ty.push(result.0);
|
|
||||||
s = result.1;
|
|
||||||
}
|
|
||||||
|
|
||||||
assert_eq!(ty.len(), 1);
|
|
||||||
|
|
||||||
let list_elem_tvar = if let TypeEnum::TObj { params, .. } =
|
|
||||||
&*self.unifier.get_ty_immutable(self.type_mapping["list"])
|
|
||||||
{
|
|
||||||
iter_type_vars(params).next().unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
(
|
|
||||||
self.unifier
|
|
||||||
.subst(
|
|
||||||
self.type_mapping["list"],
|
|
||||||
&into_var_map([TypeVar { id: list_elem_tvar.id, ty: ty[0] }]),
|
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
&s[1..],
|
|
||||||
)
|
|
||||||
}
|
|
||||||
"tuple" => {
|
"tuple" => {
|
||||||
let mut s = &typ[end..];
|
let mut s = &typ[end..];
|
||||||
assert_eq!(&s[0..1], "[");
|
assert_eq!(&s[0..1], "[");
|
||||||
@ -183,7 +142,13 @@ impl TestEnvironment {
|
|||||||
ty.push(result.0);
|
ty.push(result.0);
|
||||||
s = result.1;
|
s = result.1;
|
||||||
}
|
}
|
||||||
(self.unifier.add_ty(TypeEnum::TTuple { ty, is_vararg_ctx: false }), &s[1..])
|
(self.unifier.add_ty(TypeEnum::TTuple { ty }), &s[1..])
|
||||||
|
}
|
||||||
|
"list" => {
|
||||||
|
assert_eq!(&typ[end..=end], "[");
|
||||||
|
let (ty, s) = self.internal_parse(&typ[end + 1..], mapping);
|
||||||
|
assert_eq!(&s[0..1], "]");
|
||||||
|
(self.unifier.add_ty(TypeEnum::TList { ty }), &s[1..])
|
||||||
}
|
}
|
||||||
"Record" => {
|
"Record" => {
|
||||||
let mut s = &typ[end..];
|
let mut s = &typ[end..];
|
||||||
@ -309,7 +274,7 @@ fn test_unify(
|
|||||||
("v1", "tuple[int]"),
|
("v1", "tuple[int]"),
|
||||||
("v2", "list[int]"),
|
("v2", "list[int]"),
|
||||||
],
|
],
|
||||||
(("v1", "v2"), "Incompatible types: 11[0] and tuple[0]")
|
(("v1", "v2"), "Incompatible types: list[0] and tuple[0]")
|
||||||
; "type mismatch"
|
; "type mismatch"
|
||||||
)]
|
)]
|
||||||
#[test_case(2,
|
#[test_case(2,
|
||||||
@ -333,7 +298,7 @@ fn test_unify(
|
|||||||
("v1", "Record[a=float,b=int]"),
|
("v1", "Record[a=float,b=int]"),
|
||||||
("v2", "Foo[v3]"),
|
("v2", "Foo[v3]"),
|
||||||
],
|
],
|
||||||
(("v1", "v2"), "`3[typevar5]::b` field/method does not exist")
|
(("v1", "v2"), "`3[typevar4]::b` field/method does not exist")
|
||||||
; "record obj merge"
|
; "record obj merge"
|
||||||
)]
|
)]
|
||||||
/// Test cases for invalid unifications.
|
/// Test cases for invalid unifications.
|
||||||
@ -423,14 +388,6 @@ fn test_typevar_range() {
|
|||||||
let int_list = env.parse("list[int]", &HashMap::new());
|
let int_list = env.parse("list[int]", &HashMap::new());
|
||||||
let float_list = env.parse("list[float]", &HashMap::new());
|
let float_list = env.parse("list[float]", &HashMap::new());
|
||||||
|
|
||||||
let list_elem_tvar = if let TypeEnum::TObj { params, .. } =
|
|
||||||
&*env.unifier.get_ty_immutable(env.type_mapping["list"])
|
|
||||||
{
|
|
||||||
iter_type_vars(params).next().unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
// unification between v and int
|
// unification between v and int
|
||||||
// where v in (int, bool)
|
// where v in (int, bool)
|
||||||
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
||||||
@ -441,7 +398,7 @@ fn test_typevar_range() {
|
|||||||
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
env.unify(int_list, v),
|
env.unify(int_list, v),
|
||||||
Err("Expected any one of these types: 0, 2, but got 11[0]".to_string())
|
Err("Expected any one of these types: 0, 2, but got list[0]".to_string())
|
||||||
);
|
);
|
||||||
|
|
||||||
// unification between v and float
|
// unification between v and float
|
||||||
@ -453,11 +410,7 @@ fn test_typevar_range() {
|
|||||||
);
|
);
|
||||||
|
|
||||||
let v1 = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
let v1 = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
||||||
let v1_list = env.unifier.add_ty(TypeEnum::TObj {
|
let v1_list = env.unifier.add_ty(TypeEnum::TList { ty: v1 });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: v1 }]),
|
|
||||||
});
|
|
||||||
let v = env.unifier.get_fresh_var_with_range(&[int, v1_list], None, None).ty;
|
let v = env.unifier.get_fresh_var_with_range(&[int, v1_list], None, None).ty;
|
||||||
// unification between v and int
|
// unification between v and int
|
||||||
// where v in (int, list[v1]), v1 in (int, bool)
|
// where v in (int, list[v1]), v1 in (int, bool)
|
||||||
@ -471,10 +424,9 @@ fn test_typevar_range() {
|
|||||||
let v = env.unifier.get_fresh_var_with_range(&[int, v1_list], None, None).ty;
|
let v = env.unifier.get_fresh_var_with_range(&[int, v1_list], None, None).ty;
|
||||||
// unification between v and list[float]
|
// unification between v and list[float]
|
||||||
// where v in (int, list[v1]), v1 in (int, bool)
|
// where v in (int, list[v1]), v1 in (int, bool)
|
||||||
println!("float_list: {}, v: {}", env.unifier.stringify(float_list), env.unifier.stringify(v));
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
env.unify(float_list, v),
|
env.unify(float_list, v),
|
||||||
Err("Expected any one of these types: 0, 11[typevar6], but got 11[1]\n\nNotes:\n typevar6 ∈ {0, 2}".to_string())
|
Err("Expected any one of these types: 0, list[typevar5], but got list[1]\n\nNotes:\n typevar5 ∈ {0, 2}".to_string())
|
||||||
);
|
);
|
||||||
|
|
||||||
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
||||||
@ -489,66 +441,34 @@ fn test_typevar_range() {
|
|||||||
|
|
||||||
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
||||||
let b = env.unifier.get_fresh_var_with_range(&[boolean, float], None, None).ty;
|
let b = env.unifier.get_fresh_var_with_range(&[boolean, float], None, None).ty;
|
||||||
let a_list = env.unifier.add_ty(TypeEnum::TObj {
|
let a_list = env.unifier.add_ty(TypeEnum::TList { ty: a });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: a }]),
|
|
||||||
});
|
|
||||||
let a_list = env.unifier.get_fresh_var_with_range(&[a_list], None, None).ty;
|
let a_list = env.unifier.get_fresh_var_with_range(&[a_list], None, None).ty;
|
||||||
let b_list = env.unifier.add_ty(TypeEnum::TObj {
|
let b_list = env.unifier.add_ty(TypeEnum::TList { ty: b });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: b }]),
|
|
||||||
});
|
|
||||||
let b_list = env.unifier.get_fresh_var_with_range(&[b_list], None, None).ty;
|
let b_list = env.unifier.get_fresh_var_with_range(&[b_list], None, None).ty;
|
||||||
env.unifier.unify(a_list, b_list).unwrap();
|
env.unifier.unify(a_list, b_list).unwrap();
|
||||||
let float_list = env.unifier.add_ty(TypeEnum::TObj {
|
let float_list = env.unifier.add_ty(TypeEnum::TList { ty: float });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: float }]),
|
|
||||||
});
|
|
||||||
env.unifier.unify(a_list, float_list).unwrap();
|
env.unifier.unify(a_list, float_list).unwrap();
|
||||||
// previous unifications should not affect a and b
|
// previous unifications should not affect a and b
|
||||||
env.unifier.unify(a, int).unwrap();
|
env.unifier.unify(a, int).unwrap();
|
||||||
|
|
||||||
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
||||||
let b = env.unifier.get_fresh_var_with_range(&[boolean, float], None, None).ty;
|
let b = env.unifier.get_fresh_var_with_range(&[boolean, float], None, None).ty;
|
||||||
let a_list = env.unifier.add_ty(TypeEnum::TObj {
|
let a_list = env.unifier.add_ty(TypeEnum::TList { ty: a });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
let b_list = env.unifier.add_ty(TypeEnum::TList { ty: b });
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: a }]),
|
|
||||||
});
|
|
||||||
let b_list = env.unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: b }]),
|
|
||||||
});
|
|
||||||
env.unifier.unify(a_list, b_list).unwrap();
|
env.unifier.unify(a_list, b_list).unwrap();
|
||||||
let int_list = env.unifier.add_ty(TypeEnum::TObj {
|
let int_list = env.unifier.add_ty(TypeEnum::TList { ty: int });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: int }]),
|
|
||||||
});
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
env.unify(a_list, int_list),
|
env.unify(a_list, int_list),
|
||||||
Err("Incompatible types: 11[typevar23] and 11[0]\
|
Err("Incompatible types: list[typevar22] and list[0]\
|
||||||
\n\nNotes:\n typevar23 ∈ {1}"
|
\n\nNotes:\n typevar22 ∈ {1}"
|
||||||
.into())
|
.into())
|
||||||
);
|
);
|
||||||
|
|
||||||
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
let a = env.unifier.get_fresh_var_with_range(&[int, float], None, None).ty;
|
||||||
let b = env.unifier.get_dummy_var().ty;
|
let b = env.unifier.get_dummy_var().ty;
|
||||||
let a_list = env.unifier.add_ty(TypeEnum::TObj {
|
let a_list = env.unifier.add_ty(TypeEnum::TList { ty: a });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: a }]),
|
|
||||||
});
|
|
||||||
let a_list = env.unifier.get_fresh_var_with_range(&[a_list], None, None).ty;
|
let a_list = env.unifier.get_fresh_var_with_range(&[a_list], None, None).ty;
|
||||||
let b_list = env.unifier.add_ty(TypeEnum::TObj {
|
let b_list = env.unifier.add_ty(TypeEnum::TList { ty: b });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: b }]),
|
|
||||||
});
|
|
||||||
env.unifier.unify(a_list, b_list).unwrap();
|
env.unifier.unify(a_list, b_list).unwrap();
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
env.unify(b, boolean),
|
env.unify(b, boolean),
|
||||||
@ -562,25 +482,16 @@ fn test_rigid_var() {
|
|||||||
let a = env.unifier.get_fresh_rigid_var(None, None).ty;
|
let a = env.unifier.get_fresh_rigid_var(None, None).ty;
|
||||||
let b = env.unifier.get_fresh_rigid_var(None, None).ty;
|
let b = env.unifier.get_fresh_rigid_var(None, None).ty;
|
||||||
let x = env.unifier.get_dummy_var().ty;
|
let x = env.unifier.get_dummy_var().ty;
|
||||||
let list_elem_tvar = env.unifier.get_fresh_var(Some("list_elem".into()), None);
|
let list_a = env.unifier.add_ty(TypeEnum::TList { ty: a });
|
||||||
let list_a = env.unifier.add_ty(TypeEnum::TObj {
|
let list_x = env.unifier.add_ty(TypeEnum::TList { ty: x });
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: a }]),
|
|
||||||
});
|
|
||||||
let list_x = env.unifier.add_ty(TypeEnum::TObj {
|
|
||||||
obj_id: env.type_mapping["list"].obj_id(&env.unifier).unwrap(),
|
|
||||||
fields: Mapping::default(),
|
|
||||||
params: into_var_map([TypeVar { id: list_elem_tvar.id, ty: x }]),
|
|
||||||
});
|
|
||||||
let int = env.parse("int", &HashMap::new());
|
let int = env.parse("int", &HashMap::new());
|
||||||
let list_int = env.parse("list[int]", &HashMap::new());
|
let list_int = env.parse("list[int]", &HashMap::new());
|
||||||
|
|
||||||
assert_eq!(env.unify(a, b), Err("Incompatible types: typevar4 and typevar3".to_string()));
|
assert_eq!(env.unify(a, b), Err("Incompatible types: typevar3 and typevar2".to_string()));
|
||||||
env.unifier.unify(list_a, list_x).unwrap();
|
env.unifier.unify(list_a, list_x).unwrap();
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
env.unify(list_x, list_int),
|
env.unify(list_x, list_int),
|
||||||
Err("Incompatible types: 11[typevar3] and 11[0]".to_string())
|
Err("Incompatible types: list[typevar2] and list[0]".to_string())
|
||||||
);
|
);
|
||||||
|
|
||||||
env.unifier.replace_rigid_var(a, int);
|
env.unifier.replace_rigid_var(a, int);
|
||||||
@ -595,25 +506,14 @@ fn test_instantiation() {
|
|||||||
let float = env.parse("float", &HashMap::new());
|
let float = env.parse("float", &HashMap::new());
|
||||||
let list_int = env.parse("list[int]", &HashMap::new());
|
let list_int = env.parse("list[int]", &HashMap::new());
|
||||||
|
|
||||||
let list_elem_tvar = if let TypeEnum::TObj { params, .. } =
|
let obj_map: HashMap<_, _> = [(0usize, "int"), (1, "float"), (2, "bool")].into();
|
||||||
&*env.unifier.get_ty_immutable(env.type_mapping["list"])
|
|
||||||
{
|
|
||||||
iter_type_vars(params).next().unwrap()
|
|
||||||
} else {
|
|
||||||
unreachable!()
|
|
||||||
};
|
|
||||||
|
|
||||||
let obj_map: HashMap<_, _> = [(0usize, "int"), (1, "float"), (2, "bool"), (11, "list")].into();
|
|
||||||
|
|
||||||
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
let v = env.unifier.get_fresh_var_with_range(&[int, boolean], None, None).ty;
|
||||||
let list_v = env
|
let list_v = env.unifier.add_ty(TypeEnum::TList { ty: v });
|
||||||
.unifier
|
|
||||||
.subst(env.type_mapping["list"], &into_var_map([TypeVar { id: list_elem_tvar.id, ty: v }]))
|
|
||||||
.unwrap();
|
|
||||||
let v1 = env.unifier.get_fresh_var_with_range(&[list_v, int], None, None).ty;
|
let v1 = env.unifier.get_fresh_var_with_range(&[list_v, int], None, None).ty;
|
||||||
let v2 = env.unifier.get_fresh_var_with_range(&[list_int, float], None, None).ty;
|
let v2 = env.unifier.get_fresh_var_with_range(&[list_int, float], None, None).ty;
|
||||||
let t = env.unifier.get_dummy_var().ty;
|
let t = env.unifier.get_dummy_var().ty;
|
||||||
let tuple = env.unifier.add_ty(TypeEnum::TTuple { ty: vec![v, v1, v2], is_vararg_ctx: false });
|
let tuple = env.unifier.add_ty(TypeEnum::TTuple { ty: vec![v, v1, v2] });
|
||||||
let v3 = env.unifier.get_fresh_var_with_range(&[tuple, t], None, None).ty;
|
let v3 = env.unifier.get_fresh_var_with_range(&[tuple, t], None, None).ty;
|
||||||
// t = TypeVar('t')
|
// t = TypeVar('t')
|
||||||
// v = TypeVar('v', int, bool)
|
// v = TypeVar('v', int, bool)
|
||||||
@ -636,7 +536,7 @@ fn test_instantiation() {
|
|||||||
tuple[int, list[bool], list[int]]
|
tuple[int, list[bool], list[int]]
|
||||||
tuple[int, list[int], float]
|
tuple[int, list[int], float]
|
||||||
tuple[int, list[int], list[int]]
|
tuple[int, list[int], list[int]]
|
||||||
v6"
|
v5"
|
||||||
}
|
}
|
||||||
.split('\n')
|
.split('\n')
|
||||||
.collect_vec();
|
.collect_vec();
|
||||||
|
@ -238,7 +238,7 @@ impl<'a> EH_Frame<'a> {
|
|||||||
/// From the [specification](https://refspecs.linuxfoundation.org/LSB_5.0.0/LSB-Core-generic/LSB-Core-generic/ehframechpt.html):
|
/// From the [specification](https://refspecs.linuxfoundation.org/LSB_5.0.0/LSB-Core-generic/LSB-Core-generic/ehframechpt.html):
|
||||||
///
|
///
|
||||||
/// > Each CFI record contains a Common Information Entry (CIE) record followed by 1 or more Frame
|
/// > Each CFI record contains a Common Information Entry (CIE) record followed by 1 or more Frame
|
||||||
/// > Description Entry (FDE) records.
|
/// Description Entry (FDE) records.
|
||||||
pub struct CFI_Record<'a> {
|
pub struct CFI_Record<'a> {
|
||||||
// It refers to the augmentation data that corresponds to 'R' in the augmentation string
|
// It refers to the augmentation data that corresponds to 'R' in the augmentation string
|
||||||
fde_pointer_encoding: u8,
|
fde_pointer_encoding: u8,
|
||||||
|
@ -21,12 +21,13 @@
|
|||||||
clippy::wildcard_imports
|
clippy::wildcard_imports
|
||||||
)]
|
)]
|
||||||
|
|
||||||
use std::{collections::HashMap, mem, ptr, slice, str};
|
|
||||||
|
|
||||||
use byteorder::{ByteOrder, LittleEndian};
|
|
||||||
|
|
||||||
use dwarf::*;
|
use dwarf::*;
|
||||||
use elf::*;
|
use elf::*;
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::{mem, ptr, slice, str};
|
||||||
|
|
||||||
|
extern crate byteorder;
|
||||||
|
use byteorder::{ByteOrder, LittleEndian};
|
||||||
|
|
||||||
mod dwarf;
|
mod dwarf;
|
||||||
mod elf;
|
mod elf;
|
||||||
|
@ -8,15 +8,15 @@ license = "MIT"
|
|||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[build-dependencies]
|
[build-dependencies]
|
||||||
lalrpop = "0.22"
|
lalrpop = "0.20"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
nac3ast = { path = "../nac3ast" }
|
nac3ast = { path = "../nac3ast" }
|
||||||
lalrpop-util = "0.22"
|
lalrpop-util = "0.20"
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
unic-emoji-char = "0.9"
|
unic-emoji-char = "0.9"
|
||||||
unic-ucd-ident = "0.9"
|
unic-ucd-ident = "0.9"
|
||||||
unicode_names2 = "1.3"
|
unicode_names2 = "1.2"
|
||||||
phf = { version = "0.11", features = ["macros"] }
|
phf = { version = "0.11", features = ["macros"] }
|
||||||
ahash = "0.8"
|
ahash = "0.8"
|
||||||
|
|
||||||
|
@ -1,10 +1,8 @@
|
|||||||
use crate::{
|
use crate::ast::Ident;
|
||||||
ast::{Ident, Location},
|
use crate::ast::Location;
|
||||||
error::*,
|
use crate::error::*;
|
||||||
token::Tok,
|
use crate::token::Tok;
|
||||||
};
|
|
||||||
use lalrpop_util::ParseError;
|
use lalrpop_util::ParseError;
|
||||||
|
|
||||||
use nac3ast::*;
|
use nac3ast::*;
|
||||||
|
|
||||||
pub fn make_config_comment(
|
pub fn make_config_comment(
|
||||||
|
@ -1,11 +1,12 @@
|
|||||||
//! Define internal parse error types
|
//! Define internal parse error types
|
||||||
//! The goal is to provide a matching and a safe error API, maksing errors from LALR
|
//! The goal is to provide a matching and a safe error API, maksing errors from LALR
|
||||||
use std::error::Error;
|
|
||||||
use std::fmt;
|
|
||||||
|
|
||||||
use lalrpop_util::ParseError as LalrpopError;
|
use lalrpop_util::ParseError as LalrpopError;
|
||||||
|
|
||||||
use crate::{ast::Location, token::Tok};
|
use crate::ast::Location;
|
||||||
|
use crate::token::Tok;
|
||||||
|
|
||||||
|
use std::error::Error;
|
||||||
|
use std::fmt;
|
||||||
|
|
||||||
/// Represents an error during lexical scanning.
|
/// Represents an error during lexical scanning.
|
||||||
#[derive(Debug, PartialEq)]
|
#[derive(Debug, PartialEq)]
|
||||||
|
@ -1,11 +1,12 @@
|
|||||||
use std::{iter, mem, str};
|
use std::iter;
|
||||||
|
use std::mem;
|
||||||
|
use std::str;
|
||||||
|
|
||||||
|
use crate::ast::{Constant, ConversionFlag, Expr, ExprKind, Location};
|
||||||
|
use crate::error::{FStringError, FStringErrorType, ParseError};
|
||||||
|
use crate::parser::parse_expression;
|
||||||
|
|
||||||
use self::FStringErrorType::*;
|
use self::FStringErrorType::*;
|
||||||
use crate::{
|
|
||||||
ast::{Constant, ConversionFlag, Expr, ExprKind, Location},
|
|
||||||
error::{FStringError, FStringErrorType, ParseError},
|
|
||||||
parser::parse_expression,
|
|
||||||
};
|
|
||||||
|
|
||||||
struct FStringParser<'a> {
|
struct FStringParser<'a> {
|
||||||
chars: iter::Peekable<str::Chars<'a>>,
|
chars: iter::Peekable<str::Chars<'a>>,
|
||||||
|
@ -1,11 +1,8 @@
|
|||||||
|
use ahash::RandomState;
|
||||||
use std::collections::HashSet;
|
use std::collections::HashSet;
|
||||||
|
|
||||||
use ahash::RandomState;
|
use crate::ast;
|
||||||
|
use crate::error::{LexicalError, LexicalErrorType};
|
||||||
use crate::{
|
|
||||||
ast,
|
|
||||||
error::{LexicalError, LexicalErrorType},
|
|
||||||
};
|
|
||||||
|
|
||||||
pub struct ArgumentList {
|
pub struct ArgumentList {
|
||||||
pub args: Vec<ast::Expr>,
|
pub args: Vec<ast::Expr>,
|
||||||
|
@ -1,16 +1,16 @@
|
|||||||
//! This module takes care of lexing python source text.
|
//! This module takes care of lexing python source text.
|
||||||
//!
|
//!
|
||||||
//! This means source code is translated into separate tokens.
|
//! This means source code is translated into separate tokens.
|
||||||
use std::{char, cmp::Ordering, num::IntErrorKind, str::FromStr};
|
|
||||||
|
|
||||||
use unic_emoji_char::is_emoji_presentation;
|
|
||||||
use unic_ucd_ident::{is_xid_continue, is_xid_start};
|
|
||||||
|
|
||||||
pub use super::token::Tok;
|
pub use super::token::Tok;
|
||||||
use crate::{
|
use crate::ast::{FileName, Location};
|
||||||
ast::{FileName, Location},
|
use crate::error::{LexicalError, LexicalErrorType};
|
||||||
error::{LexicalError, LexicalErrorType},
|
use std::char;
|
||||||
};
|
use std::cmp::Ordering;
|
||||||
|
use std::num::IntErrorKind;
|
||||||
|
use std::str::FromStr;
|
||||||
|
use unic_emoji_char::is_emoji_presentation;
|
||||||
|
use unic_ucd_ident::{is_xid_continue, is_xid_start};
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Debug, Default)]
|
#[derive(Clone, Copy, PartialEq, Debug, Default)]
|
||||||
struct IndentationLevel {
|
struct IndentationLevel {
|
||||||
|
@ -5,16 +5,14 @@
|
|||||||
//! parse a whole program, a single statement, or a single
|
//! parse a whole program, a single statement, or a single
|
||||||
//! expression.
|
//! expression.
|
||||||
|
|
||||||
|
use nac3ast::Location;
|
||||||
use std::iter;
|
use std::iter;
|
||||||
|
|
||||||
use nac3ast::Location;
|
use crate::ast::{self, FileName};
|
||||||
|
use crate::error::ParseError;
|
||||||
|
use crate::lexer;
|
||||||
pub use crate::mode::Mode;
|
pub use crate::mode::Mode;
|
||||||
use crate::{
|
use crate::python;
|
||||||
ast::{self, FileName},
|
|
||||||
error::ParseError,
|
|
||||||
lexer, python,
|
|
||||||
};
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Parse python code.
|
* Parse python code.
|
||||||
|
@ -1,8 +1,7 @@
|
|||||||
//! Different token definitions.
|
//! Different token definitions.
|
||||||
//! Loosely based on token.h from CPython source:
|
//! Loosely based on token.h from CPython source:
|
||||||
use std::fmt::{self, Write};
|
|
||||||
|
|
||||||
use crate::ast;
|
use crate::ast;
|
||||||
|
use std::fmt::{self, Write};
|
||||||
|
|
||||||
/// Python source code can be tokenized in a sequence of these tokens.
|
/// Python source code can be tokenized in a sequence of these tokens.
|
||||||
#[derive(Clone, Debug, PartialEq)]
|
#[derive(Clone, Debug, PartialEq)]
|
||||||
|
@ -4,13 +4,16 @@ version = "0.1.0"
|
|||||||
authors = ["M-Labs"]
|
authors = ["M-Labs"]
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
|
|
||||||
[features]
|
|
||||||
no-escape-analysis = ["nac3core/no-escape-analysis"]
|
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
parking_lot = "0.12"
|
parking_lot = "0.12"
|
||||||
|
nac3parser = { path = "../nac3parser" }
|
||||||
nac3core = { path = "../nac3core" }
|
nac3core = { path = "../nac3core" }
|
||||||
|
|
||||||
[dependencies.clap]
|
[dependencies.clap]
|
||||||
version = "4.5"
|
version = "4.5"
|
||||||
features = ["derive"]
|
features = ["derive"]
|
||||||
|
|
||||||
|
[dependencies.inkwell]
|
||||||
|
version = "0.4"
|
||||||
|
default-features = false
|
||||||
|
features = ["llvm14-0", "target-x86", "target-arm", "target-riscv", "no-libffi-linking"]
|
||||||
|
@ -3,66 +3,23 @@
|
|||||||
set -e
|
set -e
|
||||||
|
|
||||||
if [ -z "$1" ]; then
|
if [ -z "$1" ]; then
|
||||||
echo "No argument supplied"
|
echo "Requires at least one argument"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
declare -a nac3args
|
declare -a nac3args
|
||||||
while [ $# -gt 1 ]; do
|
|
||||||
case "$1" in
|
|
||||||
--help)
|
|
||||||
echo "Usage: check_demo.sh [--debug] [-i686] -- [NAC3ARGS...] demo"
|
|
||||||
exit
|
|
||||||
;;
|
|
||||||
--debug)
|
|
||||||
debug=1
|
|
||||||
;;
|
|
||||||
-i686)
|
|
||||||
i686=1
|
|
||||||
;;
|
|
||||||
--)
|
|
||||||
shift
|
|
||||||
break
|
|
||||||
;;
|
|
||||||
*)
|
|
||||||
echo "Unrecognized argument \"$1\""
|
|
||||||
exit 1
|
|
||||||
;;
|
|
||||||
esac
|
|
||||||
shift
|
|
||||||
done
|
|
||||||
|
|
||||||
while [ $# -gt 1 ]; do
|
while [ $# -gt 1 ]; do
|
||||||
nac3args+=("$1")
|
nac3args+=("$1")
|
||||||
shift
|
shift
|
||||||
done
|
done
|
||||||
demo="$1"
|
demo="$1"
|
||||||
|
|
||||||
|
echo -n "Checking $demo... "
|
||||||
echo "### Checking $demo..."
|
|
||||||
|
|
||||||
echo ">>>>>> Running $demo with the Python interpreter"
|
|
||||||
./interpret_demo.py "$demo" > interpreted.log
|
./interpret_demo.py "$demo" > interpreted.log
|
||||||
|
./run_demo.sh --out run.log "${nac3args[@]}" "$demo"
|
||||||
|
./run_demo.sh --lli --out run_lli.log "${nac3args[@]}" "$demo"
|
||||||
|
diff -Nau interpreted.log run.log
|
||||||
|
diff -Nau interpreted.log run_lli.log
|
||||||
|
echo "ok"
|
||||||
|
|
||||||
if [ -n "$i686" ]; then
|
rm -f interpreted.log run.log run_lli.log
|
||||||
echo "...... Trying NAC3's 32-bit code generator output"
|
|
||||||
if [ -n "$debug" ]; then
|
|
||||||
./run_demo.sh --debug -i686 --out run_32.log -- "${nac3args[@]}" "$demo"
|
|
||||||
else
|
|
||||||
./run_demo.sh -i686 --out run_32.log -- "${nac3args[@]}" "$demo"
|
|
||||||
fi
|
|
||||||
diff -Nau interpreted.log run_32.log
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "...... Trying NAC3's 64-bit code generator output"
|
|
||||||
if [ -n "$debug" ]; then
|
|
||||||
./run_demo.sh --debug --out run_64.log -- "${nac3args[@]}" "$demo"
|
|
||||||
else
|
|
||||||
./run_demo.sh --out run_64.log -- "${nac3args[@]}" "$demo"
|
|
||||||
fi
|
|
||||||
diff -Nau interpreted.log run_64.log
|
|
||||||
|
|
||||||
echo "...... OK"
|
|
||||||
|
|
||||||
rm -f interpreted.log \
|
|
||||||
run_32.log run_64.log
|
|
@ -2,11 +2,6 @@
|
|||||||
|
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
if [ "$1" == "--help" ]; then
|
|
||||||
echo "Usage: check_demos.sh [CHECKARGS...] [--] [NAC3ARGS...]"
|
|
||||||
exit
|
|
||||||
fi
|
|
||||||
|
|
||||||
count=0
|
count=0
|
||||||
for demo in src/*.py; do
|
for demo in src/*.py; do
|
||||||
./check_demo.sh "$@" "$demo"
|
./check_demo.sh "$@" "$demo"
|
||||||
|
@ -6,6 +6,8 @@
|
|||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
#include <string.h>
|
#include <string.h>
|
||||||
|
|
||||||
|
#define usize size_t
|
||||||
|
|
||||||
double dbl_nan(void) {
|
double dbl_nan(void) {
|
||||||
return NAN;
|
return NAN;
|
||||||
}
|
}
|
||||||
@ -42,15 +44,6 @@ void output_float64(double x) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void output_range(int32_t range[3]) {
|
|
||||||
printf("range(");
|
|
||||||
printf("%d, %d", range[0], range[1]);
|
|
||||||
if (range[2] != 1) {
|
|
||||||
printf(", %d", range[2]);
|
|
||||||
}
|
|
||||||
puts(")");
|
|
||||||
}
|
|
||||||
|
|
||||||
void output_asciiart(int32_t x) {
|
void output_asciiart(int32_t x) {
|
||||||
static const char *chars = " .,-:;i+hHM$*#@ ";
|
static const char *chars = " .,-:;i+hHM$*#@ ";
|
||||||
if (x < 0) {
|
if (x < 0) {
|
||||||
@ -62,14 +55,14 @@ void output_asciiart(int32_t x) {
|
|||||||
|
|
||||||
struct cslice {
|
struct cslice {
|
||||||
void *data;
|
void *data;
|
||||||
size_t len;
|
usize len;
|
||||||
};
|
};
|
||||||
|
|
||||||
void output_int32_list(struct cslice *slice) {
|
void output_int32_list(struct cslice *slice) {
|
||||||
const int32_t *data = (int32_t *) slice->data;
|
const int32_t *data = (int32_t *) slice->data;
|
||||||
|
|
||||||
putchar('[');
|
putchar('[');
|
||||||
for (size_t i = 0; i < slice->len; ++i) {
|
for (usize i = 0; i < slice->len; ++i) {
|
||||||
if (i == slice->len - 1) {
|
if (i == slice->len - 1) {
|
||||||
printf("%d", data[i]);
|
printf("%d", data[i]);
|
||||||
} else {
|
} else {
|
||||||
@ -83,13 +76,9 @@ void output_int32_list(struct cslice* slice) {
|
|||||||
void output_str(struct cslice *slice) {
|
void output_str(struct cslice *slice) {
|
||||||
const char *data = (const char *) slice->data;
|
const char *data = (const char *) slice->data;
|
||||||
|
|
||||||
for (size_t i = 0; i < slice->len; ++i) {
|
for (usize i = 0; i < slice->len; ++i) {
|
||||||
putchar(data[i]);
|
putchar(data[i]);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
void output_strln(struct cslice* slice) {
|
|
||||||
output_str(slice);
|
|
||||||
putchar('\n');
|
putchar('\n');
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -105,26 +94,8 @@ uint32_t __nac3_personality(uint32_t state, uint32_t exception_object, uint32_t
|
|||||||
__builtin_unreachable();
|
__builtin_unreachable();
|
||||||
}
|
}
|
||||||
|
|
||||||
// See `struct Exception<'a>` in
|
uint32_t __nac3_raise(uint32_t state, uint32_t exception_object, uint32_t context) {
|
||||||
// https://github.com/m-labs/artiq/blob/master/artiq/firmware/libeh/eh_artiq.rs
|
printf("__nac3_raise(state: %u, exception_object: %u, context: %u)\n", state, exception_object, context);
|
||||||
struct Exception {
|
|
||||||
uint32_t id;
|
|
||||||
struct cslice file;
|
|
||||||
uint32_t line;
|
|
||||||
uint32_t column;
|
|
||||||
struct cslice function;
|
|
||||||
struct cslice message;
|
|
||||||
int64_t param[3];
|
|
||||||
};
|
|
||||||
|
|
||||||
uint32_t __nac3_raise(struct Exception* e) {
|
|
||||||
printf("__nac3_raise called. Exception details:\n");
|
|
||||||
printf(" ID: %" PRIu32 "\n", e->id);
|
|
||||||
printf(" Location: %*s:%" PRIu32 ":%" PRIu32 "\n", (int)e->file.len, (const char*)e->file.data, e->line,
|
|
||||||
e->column);
|
|
||||||
printf(" Function: %*s\n", (int)e->function.len, (const char*)e->function.data);
|
|
||||||
printf(" Message: \"%*s\"\n", (int)e->message.len, (const char*)e->message.data);
|
|
||||||
printf(" Params: {0}=%" PRId64 ", {1}=%" PRId64 ", {2}=%" PRId64 "\n", e->param[0], e->param[1], e->param[2]);
|
|
||||||
exit(101);
|
exit(101);
|
||||||
__builtin_unreachable();
|
__builtin_unreachable();
|
||||||
}
|
}
|
||||||
|
@ -6,7 +6,6 @@ import importlib.machinery
|
|||||||
import math
|
import math
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import numpy.typing as npt
|
import numpy.typing as npt
|
||||||
import scipy as sp
|
|
||||||
import pathlib
|
import pathlib
|
||||||
|
|
||||||
from numpy import int32, int64, uint32, uint64
|
from numpy import int32, int64, uint32, uint64
|
||||||
@ -108,9 +107,6 @@ def patch(module):
|
|||||||
def output_float(x):
|
def output_float(x):
|
||||||
print("%f" % x)
|
print("%f" % x)
|
||||||
|
|
||||||
def output_strln(x):
|
|
||||||
print(x, end='')
|
|
||||||
|
|
||||||
def dbg_stack_address(_):
|
def dbg_stack_address(_):
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
@ -124,8 +120,6 @@ def patch(module):
|
|||||||
return output_asciiart
|
return output_asciiart
|
||||||
elif name == "output_float64":
|
elif name == "output_float64":
|
||||||
return output_float
|
return output_float
|
||||||
elif name == "output_str":
|
|
||||||
return output_strln
|
|
||||||
elif name in {
|
elif name in {
|
||||||
"output_bool",
|
"output_bool",
|
||||||
"output_int32",
|
"output_int32",
|
||||||
@ -133,8 +127,7 @@ def patch(module):
|
|||||||
"output_int32_list",
|
"output_int32_list",
|
||||||
"output_uint32",
|
"output_uint32",
|
||||||
"output_uint64",
|
"output_uint64",
|
||||||
"output_strln",
|
"output_str",
|
||||||
"output_range",
|
|
||||||
}:
|
}:
|
||||||
return print
|
return print
|
||||||
elif name == "dbg_stack_address":
|
elif name == "dbg_stack_address":
|
||||||
@ -168,7 +161,7 @@ def patch(module):
|
|||||||
module.ceil64 = _ceil
|
module.ceil64 = _ceil
|
||||||
module.np_ceil = np.ceil
|
module.np_ceil = np.ceil
|
||||||
|
|
||||||
# NumPy NDArray factory functions
|
# NumPy ndarray functions
|
||||||
module.ndarray = NDArray
|
module.ndarray = NDArray
|
||||||
module.np_ndarray = np.ndarray
|
module.np_ndarray = np.ndarray
|
||||||
module.np_empty = np.empty
|
module.np_empty = np.empty
|
||||||
@ -184,10 +177,8 @@ def patch(module):
|
|||||||
module.np_isinf = np.isinf
|
module.np_isinf = np.isinf
|
||||||
module.np_min = np.min
|
module.np_min = np.min
|
||||||
module.np_minimum = np.minimum
|
module.np_minimum = np.minimum
|
||||||
module.np_argmin = np.argmin
|
|
||||||
module.np_max = np.max
|
module.np_max = np.max
|
||||||
module.np_maximum = np.maximum
|
module.np_maximum = np.maximum
|
||||||
module.np_argmax = np.argmax
|
|
||||||
module.np_sin = np.sin
|
module.np_sin = np.sin
|
||||||
module.np_cos = np.cos
|
module.np_cos = np.cos
|
||||||
module.np_exp = np.exp
|
module.np_exp = np.exp
|
||||||
@ -218,10 +209,8 @@ def patch(module):
|
|||||||
module.np_ldexp = np.ldexp
|
module.np_ldexp = np.ldexp
|
||||||
module.np_hypot = np.hypot
|
module.np_hypot = np.hypot
|
||||||
module.np_nextafter = np.nextafter
|
module.np_nextafter = np.nextafter
|
||||||
module.np_transpose = np.transpose
|
|
||||||
module.np_reshape = np.reshape
|
|
||||||
|
|
||||||
# SciPy Math functions
|
# SciPy Math Functions
|
||||||
module.sp_spec_erf = special.erf
|
module.sp_spec_erf = special.erf
|
||||||
module.sp_spec_erfc = special.erfc
|
module.sp_spec_erfc = special.erfc
|
||||||
module.sp_spec_gamma = special.gamma
|
module.sp_spec_gamma = special.gamma
|
||||||
@ -229,19 +218,16 @@ def patch(module):
|
|||||||
module.sp_spec_j0 = special.j0
|
module.sp_spec_j0 = special.j0
|
||||||
module.sp_spec_j1 = special.j1
|
module.sp_spec_j1 = special.j1
|
||||||
|
|
||||||
# Linalg functions
|
# NumPy NDArray Functions
|
||||||
module.np_dot = np.dot
|
module.np_ndarray = np.ndarray
|
||||||
module.np_linalg_cholesky = np.linalg.cholesky
|
module.np_empty = np.empty
|
||||||
module.np_linalg_qr = np.linalg.qr
|
module.np_zeros = np.zeros
|
||||||
module.np_linalg_svd = np.linalg.svd
|
module.np_ones = np.ones
|
||||||
module.np_linalg_inv = np.linalg.inv
|
module.np_full = np.full
|
||||||
module.np_linalg_pinv = np.linalg.pinv
|
module.np_eye = np.eye
|
||||||
module.np_linalg_matrix_power = np.linalg.matrix_power
|
module.np_identity = np.identity
|
||||||
module.np_linalg_det = np.linalg.det
|
module.np_any = np.any
|
||||||
|
module.np_all = np.all
|
||||||
module.sp_linalg_lu = lambda x: sp.linalg.lu(x, True)
|
|
||||||
module.sp_linalg_schur = sp.linalg.schur
|
|
||||||
module.sp_linalg_hessenberg = lambda x: sp.linalg.hessenberg(x, True)
|
|
||||||
|
|
||||||
def file_import(filename, prefix="file_import_"):
|
def file_import(filename, prefix="file_import_"):
|
||||||
filename = pathlib.Path(filename)
|
filename = pathlib.Path(filename)
|
||||||
|
114
nac3standalone/demo/linalg/Cargo.lock
generated
114
nac3standalone/demo/linalg/Cargo.lock
generated
@ -1,114 +0,0 @@
|
|||||||
# This file is automatically @generated by Cargo.
|
|
||||||
# It is not intended for manual editing.
|
|
||||||
version = 3
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "approx"
|
|
||||||
version = "0.5.1"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "cab112f0a86d568ea0e627cc1d6be74a1e9cd55214684db5561995f6dad897c6"
|
|
||||||
dependencies = [
|
|
||||||
"num-traits",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "autocfg"
|
|
||||||
version = "1.3.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "0c4b4d0bd25bd0b74681c0ad21497610ce1b7c91b1022cd21c80c6fbdd9476b0"
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "cslice"
|
|
||||||
version = "0.3.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "0f8cb7306107e4b10e64994de6d3274bd08996a7c1322a27b86482392f96be0a"
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "libm"
|
|
||||||
version = "0.2.8"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "4ec2a862134d2a7d32d7983ddcdd1c4923530833c9f2ea1a44fc5fa473989058"
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "linalg"
|
|
||||||
version = "0.1.0"
|
|
||||||
dependencies = [
|
|
||||||
"cslice",
|
|
||||||
"nalgebra",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "nalgebra"
|
|
||||||
version = "0.32.6"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "7b5c17de023a86f59ed79891b2e5d5a94c705dbe904a5b5c9c952ea6221b03e4"
|
|
||||||
dependencies = [
|
|
||||||
"approx",
|
|
||||||
"num-complex",
|
|
||||||
"num-rational",
|
|
||||||
"num-traits",
|
|
||||||
"simba",
|
|
||||||
"typenum",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "num-complex"
|
|
||||||
version = "0.4.6"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "73f88a1307638156682bada9d7604135552957b7818057dcef22705b4d509495"
|
|
||||||
dependencies = [
|
|
||||||
"num-traits",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "num-integer"
|
|
||||||
version = "0.1.46"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "7969661fd2958a5cb096e56c8e1ad0444ac2bbcd0061bd28660485a44879858f"
|
|
||||||
dependencies = [
|
|
||||||
"num-traits",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "num-rational"
|
|
||||||
version = "0.4.2"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "f83d14da390562dca69fc84082e73e548e1ad308d24accdedd2720017cb37824"
|
|
||||||
dependencies = [
|
|
||||||
"num-integer",
|
|
||||||
"num-traits",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "num-traits"
|
|
||||||
version = "0.2.19"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "071dfc062690e90b734c0b2273ce72ad0ffa95f0c74596bc250dcfd960262841"
|
|
||||||
dependencies = [
|
|
||||||
"autocfg",
|
|
||||||
"libm",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "paste"
|
|
||||||
version = "1.0.15"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "57c0d7b74b563b49d38dae00a0c37d4d6de9b432382b2892f0574ddcae73fd0a"
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "simba"
|
|
||||||
version = "0.8.1"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "061507c94fc6ab4ba1c9a0305018408e312e17c041eb63bef8aa726fa33aceae"
|
|
||||||
dependencies = [
|
|
||||||
"approx",
|
|
||||||
"num-complex",
|
|
||||||
"num-traits",
|
|
||||||
"paste",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "typenum"
|
|
||||||
version = "1.17.0"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "42ff0bf0c66b8238c6f3b578df37d0b7848e55df8577b3f74f92a69acceeb825"
|
|
@ -1,13 +0,0 @@
|
|||||||
[package]
|
|
||||||
name = "linalg"
|
|
||||||
version = "0.1.0"
|
|
||||||
edition = "2021"
|
|
||||||
|
|
||||||
[lib]
|
|
||||||
crate-type = ["staticlib"]
|
|
||||||
|
|
||||||
[dependencies]
|
|
||||||
nalgebra = {version = "0.32.6", default-features = false, features = ["libm", "alloc"]}
|
|
||||||
cslice = "0.3.0"
|
|
||||||
|
|
||||||
[workspace]
|
|
@ -1,406 +0,0 @@
|
|||||||
// Uses `nalgebra` crate to invoke `np_linalg` and `sp_linalg` functions
|
|
||||||
// When converting between `nalgebra::Matrix` and `NDArray` following considerations are necessary
|
|
||||||
//
|
|
||||||
// * Both `nalgebra::Matrix` and `NDArray` require their content to be stored in row-major order
|
|
||||||
// * `NDArray` data pointer can be directly read and converted to `nalgebra::Matrix` (row and column number must be known)
|
|
||||||
// * `nalgebra::Matrix::as_slice` returns the content of matrix in column-major order and initial data needs to be transposed before storing it in `NDArray` data pointer
|
|
||||||
|
|
||||||
use core::slice;
|
|
||||||
use nalgebra::DMatrix;
|
|
||||||
|
|
||||||
fn report_error(
|
|
||||||
error_name: &str,
|
|
||||||
fn_name: &str,
|
|
||||||
file_name: &str,
|
|
||||||
line_num: u32,
|
|
||||||
col_num: u32,
|
|
||||||
err_msg: &str,
|
|
||||||
) -> ! {
|
|
||||||
panic!(
|
|
||||||
"Exception {} from {} in {}:{}:{}, message: {}",
|
|
||||||
error_name, fn_name, file_name, line_num, col_num, err_msg
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
pub struct InputMatrix {
|
|
||||||
pub ndims: usize,
|
|
||||||
pub dims: *const usize,
|
|
||||||
pub data: *mut f64,
|
|
||||||
}
|
|
||||||
impl InputMatrix {
|
|
||||||
fn get_dims(&mut self) -> Vec<usize> {
|
|
||||||
let dims = unsafe { slice::from_raw_parts(self.dims, self.ndims) };
|
|
||||||
dims.to_vec()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_cholesky(mat1: *mut InputMatrix, out: *mut InputMatrix) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out = out.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_cholesky", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
if dim1[0] != dim1[1] {
|
|
||||||
let err_msg =
|
|
||||||
format!("last 2 dimensions of the array must be square: {0} != {1}", dim1[0], dim1[1]);
|
|
||||||
report_error("LinAlgError", "np_linalg_cholesky", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let outdim = out.get_dims();
|
|
||||||
let out_slice = unsafe { slice::from_raw_parts_mut(out.data, outdim[0] * outdim[1]) };
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
|
|
||||||
let matrix1 = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let result = matrix1.cholesky();
|
|
||||||
match result {
|
|
||||||
Some(res) => {
|
|
||||||
out_slice.copy_from_slice(res.unpack().transpose().as_slice());
|
|
||||||
}
|
|
||||||
None => {
|
|
||||||
report_error(
|
|
||||||
"LinAlgError",
|
|
||||||
"np_linalg_cholesky",
|
|
||||||
file!(),
|
|
||||||
line!(),
|
|
||||||
column!(),
|
|
||||||
"Matrix is not positive definite",
|
|
||||||
);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_qr(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
out_q: *mut InputMatrix,
|
|
||||||
out_r: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out_q = out_q.as_mut().unwrap();
|
|
||||||
let out_r = out_r.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_cholesky", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let outq_dim = (*out_q).get_dims();
|
|
||||||
let outr_dim = (*out_r).get_dims();
|
|
||||||
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
let out_q_slice = unsafe { slice::from_raw_parts_mut(out_q.data, outq_dim[0] * outq_dim[1]) };
|
|
||||||
let out_r_slice = unsafe { slice::from_raw_parts_mut(out_r.data, outr_dim[0] * outr_dim[1]) };
|
|
||||||
|
|
||||||
// Refer to https://github.com/dimforge/nalgebra/issues/735
|
|
||||||
let matrix1 = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
|
|
||||||
let res = matrix1.qr();
|
|
||||||
let (q, r) = res.unpack();
|
|
||||||
|
|
||||||
// Uses different algo need to match numpy
|
|
||||||
out_q_slice.copy_from_slice(q.transpose().as_slice());
|
|
||||||
out_r_slice.copy_from_slice(r.transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_svd(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
outu: *mut InputMatrix,
|
|
||||||
outs: *mut InputMatrix,
|
|
||||||
outvh: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let outu = outu.as_mut().unwrap();
|
|
||||||
let outs = outs.as_mut().unwrap();
|
|
||||||
let outvh = outvh.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_svd", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let outu_dim = (*outu).get_dims();
|
|
||||||
let outs_dim = (*outs).get_dims();
|
|
||||||
let outvh_dim = (*outvh).get_dims();
|
|
||||||
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
let out_u_slice = unsafe { slice::from_raw_parts_mut(outu.data, outu_dim[0] * outu_dim[1]) };
|
|
||||||
let out_s_slice = unsafe { slice::from_raw_parts_mut(outs.data, outs_dim[0]) };
|
|
||||||
let out_vh_slice =
|
|
||||||
unsafe { slice::from_raw_parts_mut(outvh.data, outvh_dim[0] * outvh_dim[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let result = matrix.svd(true, true);
|
|
||||||
out_u_slice.copy_from_slice(result.u.unwrap().transpose().as_slice());
|
|
||||||
out_s_slice.copy_from_slice(result.singular_values.as_slice());
|
|
||||||
out_vh_slice.copy_from_slice(result.v_t.unwrap().transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_inv(mat1: *mut InputMatrix, out: *mut InputMatrix) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out = out.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_inv", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
|
|
||||||
if dim1[0] != dim1[1] {
|
|
||||||
let err_msg =
|
|
||||||
format!("last 2 dimensions of the array must be square: {0} != {1}", dim1[0], dim1[1]);
|
|
||||||
report_error("LinAlgError", "np_linalg_inv", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let outdim = out.get_dims();
|
|
||||||
let out_slice = unsafe { slice::from_raw_parts_mut(out.data, outdim[0] * outdim[1]) };
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
if !matrix.is_invertible() {
|
|
||||||
report_error(
|
|
||||||
"LinAlgError",
|
|
||||||
"np_linalg_inv",
|
|
||||||
file!(),
|
|
||||||
line!(),
|
|
||||||
column!(),
|
|
||||||
"no inverse for Singular Matrix",
|
|
||||||
);
|
|
||||||
}
|
|
||||||
let inv = matrix.try_inverse().unwrap();
|
|
||||||
out_slice.copy_from_slice(inv.transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_pinv(mat1: *mut InputMatrix, out: *mut InputMatrix) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out = out.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_pinv", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let outdim = out.get_dims();
|
|
||||||
let out_slice = unsafe { slice::from_raw_parts_mut(out.data, outdim[0] * outdim[1]) };
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let svd = matrix.svd(true, true);
|
|
||||||
let inv = svd.pseudo_inverse(1e-15);
|
|
||||||
|
|
||||||
match inv {
|
|
||||||
Ok(m) => {
|
|
||||||
out_slice.copy_from_slice(m.transpose().as_slice());
|
|
||||||
}
|
|
||||||
Err(err_msg) => {
|
|
||||||
report_error("LinAlgError", "np_linalg_pinv", file!(), line!(), column!(), err_msg);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_matrix_power(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
mat2: *mut InputMatrix,
|
|
||||||
out: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let mat2 = mat2.as_mut().unwrap();
|
|
||||||
let out = out.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_matrix_power", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let power = unsafe { slice::from_raw_parts_mut(mat2.data, 1) };
|
|
||||||
let power = power[0];
|
|
||||||
let outdim = out.get_dims();
|
|
||||||
let out_slice = unsafe { slice::from_raw_parts_mut(out.data, outdim[0] * outdim[1]) };
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
|
|
||||||
let abs_pow = power.abs();
|
|
||||||
let matrix1 = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let mut result = matrix1.pow(abs_pow as u32);
|
|
||||||
|
|
||||||
if power < 0.0 {
|
|
||||||
if !result.is_invertible() {
|
|
||||||
report_error(
|
|
||||||
"LinAlgError",
|
|
||||||
"np_linalg_inv",
|
|
||||||
file!(),
|
|
||||||
line!(),
|
|
||||||
column!(),
|
|
||||||
"no inverse for Singular Matrix",
|
|
||||||
);
|
|
||||||
}
|
|
||||||
result = result.try_inverse().unwrap();
|
|
||||||
}
|
|
||||||
out_slice.copy_from_slice(result.transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn np_linalg_det(mat1: *mut InputMatrix, out: *mut InputMatrix) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out = out.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "np_linalg_det", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let out_slice = unsafe { slice::from_raw_parts_mut(out.data, 1) };
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
if !matrix.is_square() {
|
|
||||||
let err_msg =
|
|
||||||
format!("last 2 dimensions of the array must be square: {0} != {1}", dim1[0], dim1[1]);
|
|
||||||
report_error("LinAlgError", "np_linalg_inv", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
out_slice[0] = matrix.determinant();
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn sp_linalg_lu(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
out_l: *mut InputMatrix,
|
|
||||||
out_u: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out_l = out_l.as_mut().unwrap();
|
|
||||||
let out_u = out_u.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "sp_linalg_lu", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
let outl_dim = (*out_l).get_dims();
|
|
||||||
let outu_dim = (*out_u).get_dims();
|
|
||||||
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
let out_l_slice = unsafe { slice::from_raw_parts_mut(out_l.data, outl_dim[0] * outl_dim[1]) };
|
|
||||||
let out_u_slice = unsafe { slice::from_raw_parts_mut(out_u.data, outu_dim[0] * outu_dim[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let (_, l, u) = matrix.lu().unpack();
|
|
||||||
|
|
||||||
out_l_slice.copy_from_slice(l.transpose().as_slice());
|
|
||||||
out_u_slice.copy_from_slice(u.transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn sp_linalg_schur(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
out_t: *mut InputMatrix,
|
|
||||||
out_z: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out_t = out_t.as_mut().unwrap();
|
|
||||||
let out_z = out_z.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "sp_linalg_schur", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
|
|
||||||
if dim1[0] != dim1[1] {
|
|
||||||
let err_msg =
|
|
||||||
format!("last 2 dimensions of the array must be square: {0} != {1}", dim1[0], dim1[1]);
|
|
||||||
report_error("LinAlgError", "np_linalg_schur", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let out_t_dim = (*out_t).get_dims();
|
|
||||||
let out_z_dim = (*out_z).get_dims();
|
|
||||||
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
let out_t_slice = unsafe { slice::from_raw_parts_mut(out_t.data, out_t_dim[0] * out_t_dim[1]) };
|
|
||||||
let out_z_slice = unsafe { slice::from_raw_parts_mut(out_z.data, out_z_dim[0] * out_z_dim[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let (z, t) = matrix.schur().unpack();
|
|
||||||
|
|
||||||
out_t_slice.copy_from_slice(t.transpose().as_slice());
|
|
||||||
out_z_slice.copy_from_slice(z.transpose().as_slice());
|
|
||||||
}
|
|
||||||
|
|
||||||
/// # Safety
|
|
||||||
///
|
|
||||||
/// `mat1` should point to a valid 2DArray of `f64` floats in row-major order
|
|
||||||
#[no_mangle]
|
|
||||||
pub unsafe extern "C" fn sp_linalg_hessenberg(
|
|
||||||
mat1: *mut InputMatrix,
|
|
||||||
out_h: *mut InputMatrix,
|
|
||||||
out_q: *mut InputMatrix,
|
|
||||||
) {
|
|
||||||
let mat1 = mat1.as_mut().unwrap();
|
|
||||||
let out_h = out_h.as_mut().unwrap();
|
|
||||||
let out_q = out_q.as_mut().unwrap();
|
|
||||||
|
|
||||||
if mat1.ndims != 2 {
|
|
||||||
let err_msg = format!("expected 2D Vector Input, but received {}D input", mat1.ndims);
|
|
||||||
report_error("ValueError", "sp_linalg_hessenberg", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let dim1 = (*mat1).get_dims();
|
|
||||||
|
|
||||||
if dim1[0] != dim1[1] {
|
|
||||||
let err_msg =
|
|
||||||
format!("last 2 dimensions of the array must be square: {} != {}", dim1[0], dim1[1]);
|
|
||||||
report_error("LinAlgError", "sp_linalg_hessenberg", file!(), line!(), column!(), &err_msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
let out_h_dim = (*out_h).get_dims();
|
|
||||||
let out_q_dim = (*out_q).get_dims();
|
|
||||||
|
|
||||||
let data_slice1 = unsafe { slice::from_raw_parts_mut(mat1.data, dim1[0] * dim1[1]) };
|
|
||||||
let out_h_slice = unsafe { slice::from_raw_parts_mut(out_h.data, out_h_dim[0] * out_h_dim[1]) };
|
|
||||||
let out_q_slice = unsafe { slice::from_raw_parts_mut(out_q.data, out_q_dim[0] * out_q_dim[1]) };
|
|
||||||
|
|
||||||
let matrix = DMatrix::from_row_slice(dim1[0], dim1[1], data_slice1);
|
|
||||||
let (q, h) = matrix.hessenberg().unpack();
|
|
||||||
|
|
||||||
out_h_slice.copy_from_slice(h.transpose().as_slice());
|
|
||||||
out_q_slice.copy_from_slice(q.transpose().as_slice());
|
|
||||||
}
|
|
@ -2,9 +2,6 @@
|
|||||||
|
|
||||||
set -e
|
set -e
|
||||||
|
|
||||||
: "${DEMO_LINALG_STUB:=linalg/target/release/liblinalg.a}"
|
|
||||||
: "${DEMO_LINALG_STUB32:=linalg/target/i686-unknown-linux-gnu/release/liblinalg.a}"
|
|
||||||
|
|
||||||
if [ -z "$1" ]; then
|
if [ -z "$1" ]; then
|
||||||
echo "No argument supplied"
|
echo "No argument supplied"
|
||||||
exit 1
|
exit 1
|
||||||
@ -14,26 +11,25 @@ declare -a nac3args
|
|||||||
while [ $# -ge 1 ]; do
|
while [ $# -ge 1 ]; do
|
||||||
case "$1" in
|
case "$1" in
|
||||||
--help)
|
--help)
|
||||||
echo "Usage: run_demo.sh [--help] [--out OUTFILE] [--debug] [-i686] -- [NAC3ARGS...] demo"
|
echo "Usage: run_demo.sh [--help] [--out OUTFILE] [--lli] [--debug] -- [NAC3ARGS...]"
|
||||||
exit
|
exit
|
||||||
;;
|
;;
|
||||||
--out)
|
--out)
|
||||||
shift
|
shift
|
||||||
outfile="$1"
|
outfile="$1"
|
||||||
;;
|
;;
|
||||||
|
--lli)
|
||||||
|
use_lli=1
|
||||||
|
;;
|
||||||
--debug)
|
--debug)
|
||||||
debug=1
|
debug=1
|
||||||
;;
|
;;
|
||||||
-i686)
|
|
||||||
i686=1
|
|
||||||
;;
|
|
||||||
--)
|
--)
|
||||||
shift
|
shift
|
||||||
break
|
break
|
||||||
;;
|
;;
|
||||||
*)
|
*)
|
||||||
echo "Unrecognized argument \"$1\""
|
break
|
||||||
exit 1
|
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
shift
|
shift
|
||||||
@ -54,19 +50,29 @@ else
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
rm -f ./*.o ./*.bc demo
|
rm -f ./*.o ./*.bc demo
|
||||||
|
if [ -z "$use_lli" ]; then
|
||||||
if [ -z "$i686" ]; then
|
|
||||||
$nac3standalone "${nac3args[@]}"
|
$nac3standalone "${nac3args[@]}"
|
||||||
|
|
||||||
clang -c -std=gnu11 -Wall -Wextra -O3 -o demo.o demo.c
|
clang -c -std=gnu11 -Wall -Wextra -O3 -o demo.o demo.c
|
||||||
clang -o demo module.o demo.o $DEMO_LINALG_STUB -lm -Wl,--no-warn-search-mismatch
|
clang -lm -o demo module.o demo.o
|
||||||
else
|
|
||||||
$nac3standalone --triple i686-unknown-linux-gnu --target-features +sse2 "${nac3args[@]}"
|
|
||||||
clang -m32 -c -std=gnu11 -Wall -Wextra -O3 -msse2 -o demo.o demo.c
|
|
||||||
clang -m32 -o demo module.o demo.o $DEMO_LINALG_STUB32 -lm -Wl,--no-warn-search-mismatch
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ -z "$outfile" ]; then
|
if [ -z "$outfile" ]; then
|
||||||
./demo
|
./demo
|
||||||
else
|
else
|
||||||
./demo > "$outfile"
|
./demo > "$outfile"
|
||||||
fi
|
fi
|
||||||
|
else
|
||||||
|
$nac3standalone --emit-llvm "${nac3args[@]}"
|
||||||
|
|
||||||
|
clang -c -std=gnu11 -Wall -Wextra -O3 -emit-llvm -o demo.bc demo.c
|
||||||
|
|
||||||
|
shopt -s nullglob
|
||||||
|
llvm-link -o nac3out.bc module*.bc main.bc
|
||||||
|
shopt -u nullglob
|
||||||
|
|
||||||
|
if [ -z "$outfile" ]; then
|
||||||
|
lli --extra-module demo.bc --extra-module irrt.bc nac3out.bc
|
||||||
|
else
|
||||||
|
lli --extra-module demo.bc --extra-module irrt.bc nac3out.bc > "$outfile"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
@ -1,76 +0,0 @@
|
|||||||
@extern
|
|
||||||
def output_int32(x: int32):
|
|
||||||
...
|
|
||||||
|
|
||||||
@extern
|
|
||||||
def output_bool(x: bool):
|
|
||||||
...
|
|
||||||
|
|
||||||
def example1():
|
|
||||||
x, *ys, z = (1, 2, 3, 4, 5)
|
|
||||||
output_int32(x)
|
|
||||||
output_int32(len(ys))
|
|
||||||
output_int32(ys[0])
|
|
||||||
output_int32(ys[1])
|
|
||||||
output_int32(ys[2])
|
|
||||||
output_int32(z)
|
|
||||||
|
|
||||||
def example2():
|
|
||||||
x, y, *zs = (1, 2, 3, 4, 5)
|
|
||||||
output_int32(x)
|
|
||||||
output_int32(y)
|
|
||||||
output_int32(len(zs))
|
|
||||||
output_int32(zs[0])
|
|
||||||
output_int32(zs[1])
|
|
||||||
output_int32(zs[2])
|
|
||||||
|
|
||||||
def example3():
|
|
||||||
*xs, y, z = (1, 2, 3, 4, 5)
|
|
||||||
output_int32(len(xs))
|
|
||||||
output_int32(xs[0])
|
|
||||||
output_int32(xs[1])
|
|
||||||
output_int32(xs[2])
|
|
||||||
output_int32(y)
|
|
||||||
output_int32(z)
|
|
||||||
|
|
||||||
def example4():
|
|
||||||
*xs, y, z = (4, 5)
|
|
||||||
output_int32(len(xs))
|
|
||||||
output_int32(y)
|
|
||||||
output_int32(z)
|
|
||||||
|
|
||||||
def example5():
|
|
||||||
# Example from: https://docs.python.org/3/reference/simple_stmts.html#assignment-statements
|
|
||||||
x = [0, 1]
|
|
||||||
i = 0
|
|
||||||
i, x[i] = 1, 2 # i is updated, then x[i] is updated
|
|
||||||
output_int32(i)
|
|
||||||
output_int32(x[0])
|
|
||||||
output_int32(x[1])
|
|
||||||
|
|
||||||
class A:
|
|
||||||
value: int32
|
|
||||||
def __init__(self):
|
|
||||||
self.value = 1000
|
|
||||||
|
|
||||||
def example6():
|
|
||||||
ws = [88, 7, 8]
|
|
||||||
a = A()
|
|
||||||
x, [y, *ys, a.value], ws[0], (ws[0],) = 1, (2, False, 4, 5), 99, (6,)
|
|
||||||
output_int32(x)
|
|
||||||
output_int32(y)
|
|
||||||
output_bool(ys[0])
|
|
||||||
output_int32(ys[1])
|
|
||||||
output_int32(a.value)
|
|
||||||
output_int32(ws[0])
|
|
||||||
output_int32(ws[1])
|
|
||||||
output_int32(ws[2])
|
|
||||||
|
|
||||||
def run() -> int32:
|
|
||||||
example1()
|
|
||||||
example2()
|
|
||||||
example3()
|
|
||||||
example4()
|
|
||||||
example5()
|
|
||||||
example6()
|
|
||||||
return 0
|
|
@ -7,7 +7,7 @@ def output_int64(x: int64):
|
|||||||
...
|
...
|
||||||
|
|
||||||
@extern
|
@extern
|
||||||
def output_strln(x: str):
|
def output_str(x: str):
|
||||||
...
|
...
|
||||||
|
|
||||||
|
|
||||||
@ -33,7 +33,7 @@ class A:
|
|||||||
|
|
||||||
class Initless:
|
class Initless:
|
||||||
def foo(self):
|
def foo(self):
|
||||||
output_strln("hello")
|
output_str("hello")
|
||||||
|
|
||||||
def run() -> int32:
|
def run() -> int32:
|
||||||
a = A(10)
|
a = A(10)
|
||||||
|
@ -22,10 +22,6 @@ def output_uint64(x: uint64):
|
|||||||
def output_float64(x: float):
|
def output_float64(x: float):
|
||||||
...
|
...
|
||||||
|
|
||||||
@extern
|
|
||||||
def output_range(x: range):
|
|
||||||
...
|
|
||||||
|
|
||||||
@extern
|
@extern
|
||||||
def output_int32_list(x: list[int32]):
|
def output_int32_list(x: list[int32]):
|
||||||
...
|
...
|
||||||
@ -38,10 +34,6 @@ def output_asciiart(x: int32):
|
|||||||
def output_str(x: str):
|
def output_str(x: str):
|
||||||
...
|
...
|
||||||
|
|
||||||
@extern
|
|
||||||
def output_strln(x: str):
|
|
||||||
...
|
|
||||||
|
|
||||||
def test_output_bool():
|
def test_output_bool():
|
||||||
output_bool(True)
|
output_bool(True)
|
||||||
output_bool(False)
|
output_bool(False)
|
||||||
@ -67,15 +59,6 @@ def test_output_float64():
|
|||||||
output_float64(16.25)
|
output_float64(16.25)
|
||||||
output_float64(-16.25)
|
output_float64(-16.25)
|
||||||
|
|
||||||
def test_output_range():
|
|
||||||
r = range(1, 100, 5)
|
|
||||||
output_int32(r.start)
|
|
||||||
output_int32(r.stop)
|
|
||||||
output_int32(r.step)
|
|
||||||
output_range(range(10))
|
|
||||||
output_range(range(1, 10))
|
|
||||||
output_range(range(1, 10, 2))
|
|
||||||
|
|
||||||
def test_output_asciiart():
|
def test_output_asciiart():
|
||||||
for i in range(17):
|
for i in range(17):
|
||||||
output_asciiart(i)
|
output_asciiart(i)
|
||||||
@ -85,8 +68,7 @@ def test_output_int32_list():
|
|||||||
output_int32_list([0, 1, 3, 5, 10])
|
output_int32_list([0, 1, 3, 5, 10])
|
||||||
|
|
||||||
def test_output_str_family():
|
def test_output_str_family():
|
||||||
output_str("hello")
|
output_str("hello world")
|
||||||
output_strln(" world")
|
|
||||||
|
|
||||||
def run() -> int32:
|
def run() -> int32:
|
||||||
test_output_bool()
|
test_output_bool()
|
||||||
@ -95,7 +77,6 @@ def run() -> int32:
|
|||||||
test_output_uint32()
|
test_output_uint32()
|
||||||
test_output_uint64()
|
test_output_uint64()
|
||||||
test_output_float64()
|
test_output_float64()
|
||||||
test_output_range()
|
|
||||||
test_output_asciiart()
|
test_output_asciiart()
|
||||||
test_output_int32_list()
|
test_output_int32_list()
|
||||||
test_output_str_family()
|
test_output_str_family()
|
||||||
|
@ -1,31 +0,0 @@
|
|||||||
@extern
|
|
||||||
def output_int32(x: int32):
|
|
||||||
...
|
|
||||||
|
|
||||||
@extern
|
|
||||||
def output_int64(x: int64):
|
|
||||||
...
|
|
||||||
|
|
||||||
X: int32 = 0
|
|
||||||
Y = int64(1)
|
|
||||||
|
|
||||||
def f():
|
|
||||||
global X, Y
|
|
||||||
X = 1
|
|
||||||
Y = int64(2)
|
|
||||||
|
|
||||||
def run() -> int32:
|
|
||||||
global X, Y
|
|
||||||
|
|
||||||
output_int32(X)
|
|
||||||
output_int64(Y)
|
|
||||||
f()
|
|
||||||
output_int32(X)
|
|
||||||
output_int64(Y)
|
|
||||||
|
|
||||||
X = 0
|
|
||||||
Y = int64(0)
|
|
||||||
output_int32(X)
|
|
||||||
output_int64(Y)
|
|
||||||
|
|
||||||
return 0
|
|
@ -10,58 +10,23 @@ class A:
|
|||||||
def __init__(self, a: int32):
|
def __init__(self, a: int32):
|
||||||
self.a = a
|
self.a = a
|
||||||
|
|
||||||
def output_all_fields(self):
|
def f1(self):
|
||||||
output_int32(self.a)
|
self.f2()
|
||||||
|
|
||||||
def set_a(self, a: int32):
|
def f2(self):
|
||||||
self.a = a
|
output_int32(self.a)
|
||||||
|
|
||||||
class B(A):
|
class B(A):
|
||||||
b: int32
|
b: int32
|
||||||
|
|
||||||
def __init__(self, b: int32):
|
def __init__(self, b: int32):
|
||||||
A.__init__(self, b + 1)
|
self.a = b + 1
|
||||||
self.set_b(b)
|
|
||||||
|
|
||||||
def output_parent_fields(self):
|
|
||||||
A.output_all_fields(self)
|
|
||||||
|
|
||||||
def output_all_fields(self):
|
|
||||||
A.output_all_fields(self)
|
|
||||||
output_int32(self.b)
|
|
||||||
|
|
||||||
def set_b(self, b: int32):
|
|
||||||
self.b = b
|
self.b = b
|
||||||
|
|
||||||
class C(B):
|
|
||||||
c: int32
|
|
||||||
|
|
||||||
def __init__(self, c: int32):
|
|
||||||
B.__init__(self, c + 1)
|
|
||||||
self.c = c
|
|
||||||
|
|
||||||
def output_parent_fields(self):
|
|
||||||
B.output_all_fields(self)
|
|
||||||
|
|
||||||
def output_all_fields(self):
|
|
||||||
B.output_all_fields(self)
|
|
||||||
output_int32(self.c)
|
|
||||||
|
|
||||||
def set_c(self, c: int32):
|
|
||||||
self.c = c
|
|
||||||
|
|
||||||
def run() -> int32:
|
def run() -> int32:
|
||||||
ccc = C(10)
|
aaa = A(5)
|
||||||
ccc.output_all_fields()
|
bbb = B(2)
|
||||||
ccc.set_a(1)
|
aaa.f1()
|
||||||
ccc.set_b(2)
|
bbb.f1()
|
||||||
ccc.set_c(3)
|
|
||||||
ccc.output_all_fields()
|
|
||||||
|
|
||||||
bbb = B(10)
|
|
||||||
bbb.set_a(9)
|
|
||||||
bbb.set_b(8)
|
|
||||||
bbb.output_all_fields()
|
|
||||||
ccc.output_all_fields()
|
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
@ -1,7 +1,3 @@
|
|||||||
@extern
|
|
||||||
def output_bool(x: bool):
|
|
||||||
...
|
|
||||||
|
|
||||||
@extern
|
@extern
|
||||||
def output_int32_list(x: list[int32]):
|
def output_int32_list(x: list[int32]):
|
||||||
...
|
...
|
||||||
@ -34,32 +30,6 @@ def run() -> int32:
|
|||||||
|
|
||||||
get_list_slice()
|
get_list_slice()
|
||||||
list_slice_assignment()
|
list_slice_assignment()
|
||||||
|
|
||||||
output_int32_list([1, 2, 3] + [4, 5, 6])
|
|
||||||
output_int32_list([1, 2, 3] * 3)
|
|
||||||
output_bool([] == [])
|
|
||||||
output_bool([0] == [])
|
|
||||||
output_bool([0] == [0])
|
|
||||||
output_bool([0, 1] == [0])
|
|
||||||
output_bool([0, 1] == [0, 1])
|
|
||||||
output_bool([] != [])
|
|
||||||
output_bool([0] != [])
|
|
||||||
output_bool([0] != [0])
|
|
||||||
output_bool([0] != [0, 1])
|
|
||||||
output_bool([0, 1] != [0, 1])
|
|
||||||
output_bool([] == [] == [])
|
|
||||||
output_bool([0] == [0] == [0])
|
|
||||||
output_bool([0, 1] == [0] == [0, 1])
|
|
||||||
output_bool([0, 1] == [0, 1] == [0])
|
|
||||||
output_bool([0] == [0, 1] == [0, 1])
|
|
||||||
output_bool([0, 1] == [0, 1] == [0, 1])
|
|
||||||
output_bool([] != [] != [])
|
|
||||||
output_bool([0] != [0] != [0])
|
|
||||||
output_bool([0, 1] != [0] != [0, 1])
|
|
||||||
output_bool([0, 1] != [0, 1] != [0])
|
|
||||||
output_bool([0] != [0, 1] != [0, 1])
|
|
||||||
output_bool([0, 1] != [0, 1] != [0, 1])
|
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
def get_list_slice():
|
def get_list_slice():
|
||||||
|
@ -23,12 +23,11 @@ def run() -> int32:
|
|||||||
output_int32(x)
|
output_int32(x)
|
||||||
output_str(" * ")
|
output_str(" * ")
|
||||||
output_float64(n / x)
|
output_float64(n / x)
|
||||||
output_str("\n")
|
|
||||||
except: # Assume this is intended to catch x == 0
|
except: # Assume this is intended to catch x == 0
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
# loop fell through without finding a factor
|
# loop fell through without finding a factor
|
||||||
output_int32(n)
|
output_int32(n)
|
||||||
output_str(" is a prime number\n")
|
output_str(" is a prime number")
|
||||||
|
|
||||||
return 0
|
return 0
|
@ -37,7 +37,7 @@ def test_round64():
|
|||||||
output_int64(round64(x))
|
output_int64(round64(x))
|
||||||
|
|
||||||
def test_np_round():
|
def test_np_round():
|
||||||
for x in [-1.5, -0.5, 0.5, 1.5, dbl_inf(), -dbl_inf(), dbl_nan(), 0.0, -0.0, 1.6, 1.4, -1.4, -1.6]:
|
for x in [-1.5, -0.5, 0.5, 1.5, dbl_inf(), -dbl_inf(), dbl_nan()]:
|
||||||
output_float64(np_round(x))
|
output_float64(np_round(x))
|
||||||
|
|
||||||
def test_np_isnan():
|
def test_np_isnan():
|
||||||
|
@ -71,65 +71,28 @@ def output_ndarray_float_2(n: ndarray[float, Literal[2]]):
|
|||||||
def consume_ndarray_1(n: ndarray[float, Literal[1]]):
|
def consume_ndarray_1(n: ndarray[float, Literal[1]]):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def consume_ndarray_2(n: ndarray[float, Literal[2]]):
|
|
||||||
pass
|
|
||||||
|
|
||||||
def test_ndarray_ctor():
|
def test_ndarray_ctor():
|
||||||
n: ndarray[float, Literal[1]] = np_ndarray([1])
|
n: ndarray[float, Literal[1]] = np_ndarray([1])
|
||||||
consume_ndarray_1(n)
|
consume_ndarray_1(n)
|
||||||
|
|
||||||
def test_ndarray_empty():
|
def test_ndarray_empty():
|
||||||
n1: ndarray[float, 1] = np_empty([1])
|
n: ndarray[float, 1] = np_empty([1])
|
||||||
consume_ndarray_1(n1)
|
consume_ndarray_1(n)
|
||||||
|
|
||||||
n2: ndarray[float, 1] = np_empty(10)
|
|
||||||
consume_ndarray_1(n2)
|
|
||||||
|
|
||||||
n3: ndarray[float, 1] = np_empty((2,))
|
|
||||||
consume_ndarray_1(n3)
|
|
||||||
|
|
||||||
n4: ndarray[float, 2] = np_empty((4, 4))
|
|
||||||
consume_ndarray_2(n4)
|
|
||||||
|
|
||||||
dim4 = (5, 2)
|
|
||||||
n5: ndarray[float, 2] = np_empty(dim4)
|
|
||||||
consume_ndarray_2(n5)
|
|
||||||
|
|
||||||
def test_ndarray_zeros():
|
def test_ndarray_zeros():
|
||||||
n1: ndarray[float, 1] = np_zeros([1])
|
n: ndarray[float, 1] = np_zeros([1])
|
||||||
output_ndarray_float_1(n1)
|
output_ndarray_float_1(n)
|
||||||
|
|
||||||
k = 3 + int32(n1[0]) # to test variable shape inputs
|
|
||||||
n2: ndarray[float, 1] = np_zeros(k * k)
|
|
||||||
output_ndarray_float_1(n2)
|
|
||||||
|
|
||||||
n3: ndarray[float, 1] = np_zeros((k * 2,))
|
|
||||||
output_ndarray_float_1(n3)
|
|
||||||
|
|
||||||
dim4 = (3, 2 * k)
|
|
||||||
n4: ndarray[float, 2] = np_zeros(dim4)
|
|
||||||
output_ndarray_float_2(n4)
|
|
||||||
|
|
||||||
def test_ndarray_ones():
|
def test_ndarray_ones():
|
||||||
n: ndarray[float, 1] = np_ones([1])
|
n: ndarray[float, 1] = np_ones([1])
|
||||||
output_ndarray_float_1(n)
|
output_ndarray_float_1(n)
|
||||||
|
|
||||||
dim = (1,)
|
|
||||||
n_tup: ndarray[float, 1] = np_ones(dim)
|
|
||||||
output_ndarray_float_1(n_tup)
|
|
||||||
|
|
||||||
def test_ndarray_full():
|
def test_ndarray_full():
|
||||||
n_float: ndarray[float, 1] = np_full([1], 2.0)
|
n_float: ndarray[float, 1] = np_full([1], 2.0)
|
||||||
output_ndarray_float_1(n_float)
|
output_ndarray_float_1(n_float)
|
||||||
n_i32: ndarray[int32, 1] = np_full([1], 2)
|
n_i32: ndarray[int32, 1] = np_full([1], 2)
|
||||||
output_ndarray_int32_1(n_i32)
|
output_ndarray_int32_1(n_i32)
|
||||||
|
|
||||||
dim = (1,)
|
|
||||||
n_float_tup: ndarray[float, 1] = np_full(dim, 2.0)
|
|
||||||
output_ndarray_float_1(n_float_tup)
|
|
||||||
n_i32_tup: ndarray[int32, 1] = np_full(dim, 2)
|
|
||||||
output_ndarray_int32_1(n_i32_tup)
|
|
||||||
|
|
||||||
def test_ndarray_eye():
|
def test_ndarray_eye():
|
||||||
n: ndarray[float, 2] = np_eye(2)
|
n: ndarray[float, 2] = np_eye(2)
|
||||||
output_ndarray_float_2(n)
|
output_ndarray_float_2(n)
|
||||||
@ -877,13 +840,6 @@ def test_ndarray_minimum_broadcast_rhs_scalar():
|
|||||||
output_ndarray_float_2(min_x_zeros)
|
output_ndarray_float_2(min_x_zeros)
|
||||||
output_ndarray_float_2(min_x_ones)
|
output_ndarray_float_2(min_x_ones)
|
||||||
|
|
||||||
def test_ndarray_argmin():
|
|
||||||
x = np_array([[1., 2.], [3., 4.]])
|
|
||||||
y = np_argmin(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_int64(y)
|
|
||||||
|
|
||||||
def test_ndarray_max():
|
def test_ndarray_max():
|
||||||
x = np_identity(2)
|
x = np_identity(2)
|
||||||
y = np_max(x)
|
y = np_max(x)
|
||||||
@ -927,13 +883,6 @@ def test_ndarray_maximum_broadcast_rhs_scalar():
|
|||||||
output_ndarray_float_2(max_x_zeros)
|
output_ndarray_float_2(max_x_zeros)
|
||||||
output_ndarray_float_2(max_x_ones)
|
output_ndarray_float_2(max_x_ones)
|
||||||
|
|
||||||
def test_ndarray_argmax():
|
|
||||||
x = np_array([[1., 2.], [3., 4.]])
|
|
||||||
y = np_argmax(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_int64(y)
|
|
||||||
|
|
||||||
def test_ndarray_abs():
|
def test_ndarray_abs():
|
||||||
x = np_identity(2)
|
x = np_identity(2)
|
||||||
y = abs(x)
|
y = abs(x)
|
||||||
@ -1439,141 +1388,47 @@ def test_ndarray_nextafter_broadcast_rhs_scalar():
|
|||||||
output_ndarray_float_2(nextafter_x_zeros)
|
output_ndarray_float_2(nextafter_x_zeros)
|
||||||
output_ndarray_float_2(nextafter_x_ones)
|
output_ndarray_float_2(nextafter_x_ones)
|
||||||
|
|
||||||
def test_ndarray_transpose():
|
def test_ndarray_any():
|
||||||
x: ndarray[float, 2] = np_array([[1., 2., 3.], [4., 5., 6.]])
|
x1 = np_identity(5)
|
||||||
y = np_transpose(x)
|
y1 = np_any(x1)
|
||||||
z = np_transpose(y)
|
output_ndarray_float_2(x1)
|
||||||
|
output_bool(y1)
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
x2 = np_identity(1)
|
||||||
output_ndarray_float_2(y)
|
y2 = np_any(x2)
|
||||||
|
output_ndarray_float_2(x2)
|
||||||
|
output_bool(y2)
|
||||||
|
|
||||||
def test_ndarray_reshape():
|
x3 = np_array([[1.0, 2.0], [3.0, 4.0]])
|
||||||
w: ndarray[float, 1] = np_array([1., 2., 3., 4., 5., 6., 7., 8., 9., 10.])
|
y3 = np_any(x3)
|
||||||
x = np_reshape(w, (1, 2, 1, -1))
|
output_ndarray_float_2(x3)
|
||||||
y = np_reshape(x, [2, -1])
|
output_bool(y3)
|
||||||
z = np_reshape(y, 10)
|
|
||||||
|
|
||||||
x1: ndarray[int32, 1] = np_array([1, 2, 3, 4])
|
x4 = np_zeros([3, 5])
|
||||||
x2: ndarray[int32, 2] = np_reshape(x1, (2, 2))
|
y4 = np_any(x4)
|
||||||
|
output_ndarray_float_2(x4)
|
||||||
|
output_bool(y4)
|
||||||
|
|
||||||
output_ndarray_float_1(w)
|
def test_ndarray_all():
|
||||||
output_ndarray_float_2(y)
|
x1 = np_identity(5)
|
||||||
output_ndarray_float_1(z)
|
y1 = np_all(x1)
|
||||||
|
output_ndarray_float_2(x1)
|
||||||
|
output_bool(y1)
|
||||||
|
|
||||||
def test_ndarray_dot():
|
x2 = np_identity(1)
|
||||||
x1: ndarray[float, 1] = np_array([5.0, 1.0, 4.0, 2.0])
|
y2 = np_all(x2)
|
||||||
y1: ndarray[float, 1] = np_array([5.0, 1.0, 6.0, 6.0])
|
output_ndarray_float_2(x2)
|
||||||
z1 = np_dot(x1, y1)
|
output_bool(y2)
|
||||||
|
|
||||||
x2: ndarray[int32, 1] = np_array([5, 1, 4, 2])
|
x3 = np_array([[1.0, 2.0], [3.0, 4.0]])
|
||||||
y2: ndarray[int32, 1] = np_array([5, 1, 6, 6])
|
y3 = np_all(x3)
|
||||||
z2 = np_dot(x2, y2)
|
output_ndarray_float_2(x3)
|
||||||
|
output_bool(y3)
|
||||||
x3: ndarray[bool, 1] = np_array([True, True, True, True])
|
|
||||||
y3: ndarray[bool, 1] = np_array([True, True, True, True])
|
|
||||||
z3 = np_dot(x3, y3)
|
|
||||||
|
|
||||||
z4 = np_dot(2, 3)
|
|
||||||
z5 = np_dot(2., 3.)
|
|
||||||
z6 = np_dot(True, False)
|
|
||||||
|
|
||||||
output_float64(z1)
|
|
||||||
output_int32(z2)
|
|
||||||
output_bool(z3)
|
|
||||||
output_int32(z4)
|
|
||||||
output_float64(z5)
|
|
||||||
output_bool(z6)
|
|
||||||
|
|
||||||
def test_ndarray_cholesky():
|
|
||||||
x: ndarray[float, 2] = np_array([[5.0, 1.0], [1.0, 4.0]])
|
|
||||||
y = np_linalg_cholesky(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_ndarray_float_2(y)
|
|
||||||
|
|
||||||
def test_ndarray_qr():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
y, z = np_linalg_qr(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
|
|
||||||
# QR Factorization is not unique and gives different results in numpy and nalgebra
|
|
||||||
# Reverting the decomposition to compare the initial arrays
|
|
||||||
a = y @ z
|
|
||||||
output_ndarray_float_2(a)
|
|
||||||
|
|
||||||
def test_ndarray_linalg_inv():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
y = np_linalg_inv(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_ndarray_float_2(y)
|
|
||||||
|
|
||||||
def test_ndarray_pinv():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5]])
|
|
||||||
y = np_linalg_pinv(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_ndarray_float_2(y)
|
|
||||||
|
|
||||||
def test_ndarray_matrix_power():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
y = np_linalg_matrix_power(x, -9)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_ndarray_float_2(y)
|
|
||||||
|
|
||||||
def test_ndarray_det():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
y = np_linalg_det(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_float64(y)
|
|
||||||
|
|
||||||
def test_ndarray_schur():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
t, z = sp_linalg_schur(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
|
|
||||||
# Schur Factorization is not unique and gives different results in scipy and nalgebra
|
|
||||||
# Reverting the decomposition to compare the initial arrays
|
|
||||||
a = (z @ t) @ np_linalg_inv(z)
|
|
||||||
output_ndarray_float_2(a)
|
|
||||||
|
|
||||||
def test_ndarray_hessenberg():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 5.0, 8.5]])
|
|
||||||
h, q = sp_linalg_hessenberg(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
|
|
||||||
# Hessenberg Factorization is not unique and gives different results in scipy and nalgebra
|
|
||||||
# Reverting the decomposition to compare the initial arrays
|
|
||||||
a = (q @ h) @ np_linalg_inv(q)
|
|
||||||
output_ndarray_float_2(a)
|
|
||||||
|
|
||||||
|
|
||||||
def test_ndarray_lu():
|
|
||||||
x: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5]])
|
|
||||||
l, u = sp_linalg_lu(x)
|
|
||||||
|
|
||||||
output_ndarray_float_2(x)
|
|
||||||
output_ndarray_float_2(l)
|
|
||||||
output_ndarray_float_2(u)
|
|
||||||
|
|
||||||
|
|
||||||
def test_ndarray_svd():
|
|
||||||
w: ndarray[float, 2] = np_array([[-5.0, -1.0, 2.0], [-1.0, 4.0, 7.5], [-1.0, 8.0, -8.5]])
|
|
||||||
x, y, z = np_linalg_svd(w)
|
|
||||||
|
|
||||||
output_ndarray_float_2(w)
|
|
||||||
|
|
||||||
# SVD Factorization is not unique and gives different results in numpy and nalgebra
|
|
||||||
# Reverting the decomposition to compare the initial arrays
|
|
||||||
a = x @ z
|
|
||||||
output_ndarray_float_2(a)
|
|
||||||
output_ndarray_float_1(y)
|
|
||||||
|
|
||||||
|
x4 = np_zeros([3, 5])
|
||||||
|
y4 = np_all(x4)
|
||||||
|
output_ndarray_float_2(x4)
|
||||||
|
output_bool(y4)
|
||||||
|
|
||||||
def run() -> int32:
|
def run() -> int32:
|
||||||
test_ndarray_ctor()
|
test_ndarray_ctor()
|
||||||
@ -1679,19 +1534,16 @@ def run() -> int32:
|
|||||||
|
|
||||||
test_ndarray_round()
|
test_ndarray_round()
|
||||||
test_ndarray_floor()
|
test_ndarray_floor()
|
||||||
test_ndarray_ceil()
|
|
||||||
test_ndarray_min()
|
test_ndarray_min()
|
||||||
test_ndarray_minimum()
|
test_ndarray_minimum()
|
||||||
test_ndarray_minimum_broadcast()
|
test_ndarray_minimum_broadcast()
|
||||||
test_ndarray_minimum_broadcast_lhs_scalar()
|
test_ndarray_minimum_broadcast_lhs_scalar()
|
||||||
test_ndarray_minimum_broadcast_rhs_scalar()
|
test_ndarray_minimum_broadcast_rhs_scalar()
|
||||||
test_ndarray_argmin()
|
|
||||||
test_ndarray_max()
|
test_ndarray_max()
|
||||||
test_ndarray_maximum()
|
test_ndarray_maximum()
|
||||||
test_ndarray_maximum_broadcast()
|
test_ndarray_maximum_broadcast()
|
||||||
test_ndarray_maximum_broadcast_lhs_scalar()
|
test_ndarray_maximum_broadcast_lhs_scalar()
|
||||||
test_ndarray_maximum_broadcast_rhs_scalar()
|
test_ndarray_maximum_broadcast_rhs_scalar()
|
||||||
test_ndarray_argmax()
|
|
||||||
test_ndarray_abs()
|
test_ndarray_abs()
|
||||||
test_ndarray_isnan()
|
test_ndarray_isnan()
|
||||||
test_ndarray_isinf()
|
test_ndarray_isinf()
|
||||||
@ -1754,18 +1606,8 @@ def run() -> int32:
|
|||||||
test_ndarray_nextafter_broadcast()
|
test_ndarray_nextafter_broadcast()
|
||||||
test_ndarray_nextafter_broadcast_lhs_scalar()
|
test_ndarray_nextafter_broadcast_lhs_scalar()
|
||||||
test_ndarray_nextafter_broadcast_rhs_scalar()
|
test_ndarray_nextafter_broadcast_rhs_scalar()
|
||||||
test_ndarray_transpose()
|
|
||||||
test_ndarray_reshape()
|
|
||||||
|
|
||||||
test_ndarray_dot()
|
test_ndarray_any()
|
||||||
test_ndarray_cholesky()
|
test_ndarray_all()
|
||||||
test_ndarray_qr()
|
|
||||||
test_ndarray_svd()
|
|
||||||
test_ndarray_linalg_inv()
|
|
||||||
test_ndarray_pinv()
|
|
||||||
test_ndarray_matrix_power()
|
|
||||||
test_ndarray_det()
|
|
||||||
test_ndarray_lu()
|
|
||||||
test_ndarray_schur()
|
|
||||||
test_ndarray_hessenberg()
|
|
||||||
return 0
|
return 0
|
||||||
|
@ -1,30 +0,0 @@
|
|||||||
@extern
|
|
||||||
def output_bool(x: bool):
|
|
||||||
...
|
|
||||||
|
|
||||||
|
|
||||||
def str_eq():
|
|
||||||
output_bool("" == "")
|
|
||||||
output_bool("a" == "")
|
|
||||||
output_bool("a" == "b")
|
|
||||||
output_bool("b" == "a")
|
|
||||||
output_bool("a" == "a")
|
|
||||||
output_bool("test string" == "test string")
|
|
||||||
output_bool("test string1" == "test string2")
|
|
||||||
|
|
||||||
|
|
||||||
def str_ne():
|
|
||||||
output_bool("" != "")
|
|
||||||
output_bool("a" != "")
|
|
||||||
output_bool("a" != "b")
|
|
||||||
output_bool("b" != "a")
|
|
||||||
output_bool("a" != "a")
|
|
||||||
output_bool("test string" != "test string")
|
|
||||||
output_bool("test string1" != "test string2")
|
|
||||||
|
|
||||||
|
|
||||||
def run() -> int32:
|
|
||||||
str_eq()
|
|
||||||
str_ne()
|
|
||||||
|
|
||||||
return 0
|
|
@ -1,7 +1,3 @@
|
|||||||
@extern
|
|
||||||
def output_bool(b: bool):
|
|
||||||
...
|
|
||||||
|
|
||||||
@extern
|
@extern
|
||||||
def output_int32_list(x: list[int32]):
|
def output_int32_list(x: list[int32]):
|
||||||
...
|
...
|
||||||
@ -17,41 +13,6 @@ class A:
|
|||||||
self.a = a
|
self.a = a
|
||||||
self.b = b
|
self.b = b
|
||||||
|
|
||||||
|
|
||||||
def test_tuple_eq():
|
|
||||||
# 0-len
|
|
||||||
output_bool(() == ())
|
|
||||||
# 1-len
|
|
||||||
output_bool((1,) == ())
|
|
||||||
output_bool(() == (1,))
|
|
||||||
output_bool((1,) == (1,))
|
|
||||||
output_bool((1,) == (2,))
|
|
||||||
# # 2-len
|
|
||||||
output_bool((1, 2) == ())
|
|
||||||
output_bool(() == (1, 2))
|
|
||||||
output_bool((1,) == (1, 2))
|
|
||||||
output_bool((1, 2) == (1,))
|
|
||||||
output_bool((2, 2) == (1, 2))
|
|
||||||
output_bool((1, 2) == (2, 2))
|
|
||||||
|
|
||||||
|
|
||||||
def test_tuple_ne():
|
|
||||||
# 0-len
|
|
||||||
output_bool(() != ())
|
|
||||||
# 1-len
|
|
||||||
output_bool((1,) != ())
|
|
||||||
output_bool(() != (1,))
|
|
||||||
output_bool((1,) != (1,))
|
|
||||||
output_bool((1,) != (2,))
|
|
||||||
# 2-len
|
|
||||||
output_bool((1, 2) != ())
|
|
||||||
output_bool(() != (1, 2))
|
|
||||||
output_bool((1,) != (1, 2))
|
|
||||||
output_bool((1, 2) != (1,))
|
|
||||||
output_bool((2, 2) != (1, 2))
|
|
||||||
output_bool((1, 2) != (2, 2))
|
|
||||||
|
|
||||||
|
|
||||||
def run() -> int32:
|
def run() -> int32:
|
||||||
data = [0, 1, 2, 3]
|
data = [0, 1, 2, 3]
|
||||||
|
|
||||||
@ -65,14 +26,4 @@ def run() -> int32:
|
|||||||
output_int32(tl[0][1])
|
output_int32(tl[0][1])
|
||||||
output_int32(tl[1])
|
output_int32(tl[1])
|
||||||
|
|
||||||
output_int32(len(()))
|
|
||||||
output_int32(len((1,)))
|
|
||||||
output_int32(len((1, 2)))
|
|
||||||
output_int32(len((1, 2, 3)))
|
|
||||||
output_int32(len((1, 2, 3, 4)))
|
|
||||||
output_int32(len((1, 2, 3, 4, 5)))
|
|
||||||
|
|
||||||
test_tuple_eq()
|
|
||||||
test_tuple_ne()
|
|
||||||
|
|
||||||
return 0
|
return 0
|
@ -1,11 +0,0 @@
|
|||||||
def f(*args: int32):
|
|
||||||
pass
|
|
||||||
|
|
||||||
|
|
||||||
def run() -> int32:
|
|
||||||
f()
|
|
||||||
f(1)
|
|
||||||
f(1, 2)
|
|
||||||
f(1, 2, 3)
|
|
||||||
|
|
||||||
return 0
|
|
@ -1,14 +1,5 @@
|
|||||||
use std::{
|
|
||||||
collections::{HashMap, HashSet},
|
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
use parking_lot::{Mutex, RwLock};
|
|
||||||
|
|
||||||
use nac3core::{
|
use nac3core::{
|
||||||
codegen::{CodeGenContext, CodeGenerator},
|
codegen::CodeGenContext,
|
||||||
inkwell::{module::Linkage, values::BasicValue},
|
|
||||||
nac3parser::ast::{self, StrRef},
|
|
||||||
symbol_resolver::{SymbolResolver, SymbolValue, ValueEnum},
|
symbol_resolver::{SymbolResolver, SymbolValue, ValueEnum},
|
||||||
toplevel::{DefinitionId, TopLevelDef},
|
toplevel::{DefinitionId, TopLevelDef},
|
||||||
typecheck::{
|
typecheck::{
|
||||||
@ -16,10 +7,15 @@ use nac3core::{
|
|||||||
typedef::{Type, Unifier},
|
typedef::{Type, Unifier},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use nac3parser::ast::{self, StrRef};
|
||||||
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
use std::collections::HashSet;
|
||||||
|
use std::{collections::HashMap, sync::Arc};
|
||||||
|
|
||||||
pub struct ResolverInternal {
|
pub struct ResolverInternal {
|
||||||
pub id_to_type: Mutex<HashMap<StrRef, Type>>,
|
pub id_to_type: Mutex<HashMap<StrRef, Type>>,
|
||||||
pub id_to_def: Mutex<HashMap<StrRef, DefinitionId>>,
|
pub id_to_def: Mutex<HashMap<StrRef, DefinitionId>>,
|
||||||
|
pub class_names: Mutex<HashMap<StrRef, Type>>,
|
||||||
pub module_globals: Mutex<HashMap<StrRef, SymbolValue>>,
|
pub module_globals: Mutex<HashMap<StrRef, SymbolValue>>,
|
||||||
pub str_store: Mutex<HashMap<String, i32>>,
|
pub str_store: Mutex<HashMap<String, i32>>,
|
||||||
}
|
}
|
||||||
@ -50,51 +46,20 @@ impl SymbolResolver for Resolver {
|
|||||||
|
|
||||||
fn get_symbol_type(
|
fn get_symbol_type(
|
||||||
&self,
|
&self,
|
||||||
unifier: &mut Unifier,
|
_: &mut Unifier,
|
||||||
_: &[Arc<RwLock<TopLevelDef>>],
|
_: &[Arc<RwLock<TopLevelDef>>],
|
||||||
primitives: &PrimitiveStore,
|
_: &PrimitiveStore,
|
||||||
str: StrRef,
|
str: StrRef,
|
||||||
) -> Result<Type, String> {
|
) -> Result<Type, String> {
|
||||||
self.0
|
self.0.id_to_type.lock().get(&str).copied().ok_or(format!("cannot get type of {str}"))
|
||||||
.id_to_type
|
|
||||||
.lock()
|
|
||||||
.get(&str)
|
|
||||||
.copied()
|
|
||||||
.or_else(|| {
|
|
||||||
self.0
|
|
||||||
.module_globals
|
|
||||||
.lock()
|
|
||||||
.get(&str)
|
|
||||||
.cloned()
|
|
||||||
.map(|v| v.get_type(primitives, unifier))
|
|
||||||
})
|
|
||||||
.ok_or(format!("cannot get type of {str}"))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_symbol_value<'ctx>(
|
fn get_symbol_value<'ctx>(
|
||||||
&self,
|
&self,
|
||||||
str: StrRef,
|
_: StrRef,
|
||||||
ctx: &mut CodeGenContext<'ctx, '_>,
|
_: &mut CodeGenContext<'ctx, '_>,
|
||||||
generator: &mut dyn CodeGenerator,
|
|
||||||
) -> Option<ValueEnum<'ctx>> {
|
) -> Option<ValueEnum<'ctx>> {
|
||||||
self.0.module_globals.lock().get(&str).cloned().map(|v| {
|
unimplemented!()
|
||||||
ctx.module
|
|
||||||
.get_global(&str.to_string())
|
|
||||||
.unwrap_or_else(|| {
|
|
||||||
let ty = v.get_type(&ctx.primitives, &mut ctx.unifier);
|
|
||||||
|
|
||||||
let init_val = ctx.gen_symbol_val(generator, &v, ty);
|
|
||||||
let llvm_ty = init_val.get_type();
|
|
||||||
|
|
||||||
let global = ctx.module.add_global(llvm_ty, None, &str.to_string());
|
|
||||||
global.set_linkage(Linkage::LinkOnceAny);
|
|
||||||
global.set_initializer(&init_val);
|
|
||||||
|
|
||||||
global
|
|
||||||
})
|
|
||||||
.as_basic_value_enum()
|
|
||||||
.into()
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_identifier_def(&self, id: StrRef) -> Result<DefinitionId, HashSet<String>> {
|
fn get_identifier_def(&self, id: StrRef) -> Result<DefinitionId, HashSet<String>> {
|
||||||
|
@ -8,30 +8,21 @@
|
|||||||
#![warn(clippy::pedantic)]
|
#![warn(clippy::pedantic)]
|
||||||
#![allow(clippy::too_many_lines, clippy::wildcard_imports)]
|
#![allow(clippy::too_many_lines, clippy::wildcard_imports)]
|
||||||
|
|
||||||
use std::{
|
|
||||||
collections::{HashMap, HashSet},
|
|
||||||
fs,
|
|
||||||
num::NonZeroUsize,
|
|
||||||
path::Path,
|
|
||||||
sync::Arc,
|
|
||||||
};
|
|
||||||
|
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
|
use inkwell::{
|
||||||
|
memory_buffer::MemoryBuffer, passes::PassBuilderOptions, support::is_multithreaded, targets::*,
|
||||||
|
OptimizationLevel,
|
||||||
|
};
|
||||||
use parking_lot::{Mutex, RwLock};
|
use parking_lot::{Mutex, RwLock};
|
||||||
|
use std::collections::HashSet;
|
||||||
|
use std::num::NonZeroUsize;
|
||||||
|
use std::{collections::HashMap, fs, path::Path, sync::Arc};
|
||||||
|
|
||||||
use nac3core::{
|
use nac3core::{
|
||||||
codegen::{
|
codegen::{
|
||||||
concrete_type::ConcreteTypeStore, irrt::load_irrt, CodeGenLLVMOptions,
|
concrete_type::ConcreteTypeStore, irrt::load_irrt, CodeGenLLVMOptions,
|
||||||
CodeGenTargetMachineOptions, CodeGenTask, DefaultCodeGenerator, WithCall, WorkerRegistry,
|
CodeGenTargetMachineOptions, CodeGenTask, DefaultCodeGenerator, WithCall, WorkerRegistry,
|
||||||
},
|
},
|
||||||
inkwell::{
|
|
||||||
memory_buffer::MemoryBuffer, module::Linkage, passes::PassBuilderOptions,
|
|
||||||
support::is_multithreaded, targets::*, OptimizationLevel,
|
|
||||||
},
|
|
||||||
nac3parser::{
|
|
||||||
ast::{Constant, Expr, ExprKind, StmtKind, StrRef},
|
|
||||||
parser,
|
|
||||||
},
|
|
||||||
symbol_resolver::SymbolResolver,
|
symbol_resolver::SymbolResolver,
|
||||||
toplevel::{
|
toplevel::{
|
||||||
composer::{ComposerConfig, TopLevelComposer},
|
composer::{ComposerConfig, TopLevelComposer},
|
||||||
@ -44,6 +35,10 @@ use nac3core::{
|
|||||||
typedef::{FunSignature, Type, Unifier, VarMap},
|
typedef::{FunSignature, Type, Unifier, VarMap},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
use nac3parser::{
|
||||||
|
ast::{Constant, Expr, ExprKind, StmtKind, StrRef},
|
||||||
|
parser,
|
||||||
|
};
|
||||||
|
|
||||||
mod basic_symbol_resolver;
|
mod basic_symbol_resolver;
|
||||||
use basic_symbol_resolver::*;
|
use basic_symbol_resolver::*;
|
||||||
@ -118,9 +113,7 @@ fn handle_typevar_definition(
|
|||||||
x,
|
x,
|
||||||
HashMap::new(),
|
HashMap::new(),
|
||||||
)?;
|
)?;
|
||||||
get_type_from_type_annotation_kinds(
|
get_type_from_type_annotation_kinds(def_list, unifier, &ty, &mut None)
|
||||||
def_list, unifier, primitives, &ty, &mut None,
|
|
||||||
)
|
|
||||||
})
|
})
|
||||||
.collect::<Result<Vec<_>, _>>()?;
|
.collect::<Result<Vec<_>, _>>()?;
|
||||||
let loc = func.location;
|
let loc = func.location;
|
||||||
@ -159,7 +152,7 @@ fn handle_typevar_definition(
|
|||||||
HashMap::new(),
|
HashMap::new(),
|
||||||
)?;
|
)?;
|
||||||
let constraint =
|
let constraint =
|
||||||
get_type_from_type_annotation_kinds(def_list, unifier, primitives, &ty, &mut None)?;
|
get_type_from_type_annotation_kinds(def_list, unifier, &ty, &mut None)?;
|
||||||
let loc = func.location;
|
let loc = func.location;
|
||||||
|
|
||||||
Ok(unifier.get_fresh_const_generic_var(constraint, Some(generic_name), Some(loc)).ty)
|
Ok(unifier.get_fresh_const_generic_var(constraint, Some(generic_name), Some(loc)).ty)
|
||||||
@ -174,49 +167,46 @@ fn handle_typevar_definition(
|
|||||||
fn handle_assignment_pattern(
|
fn handle_assignment_pattern(
|
||||||
targets: &[Expr],
|
targets: &[Expr],
|
||||||
value: &Expr,
|
value: &Expr,
|
||||||
resolver: Arc<dyn SymbolResolver + Send + Sync>,
|
resolver: &(dyn SymbolResolver + Send + Sync),
|
||||||
internal_resolver: &ResolverInternal,
|
internal_resolver: &ResolverInternal,
|
||||||
composer: &mut TopLevelComposer,
|
def_list: &[Arc<RwLock<TopLevelDef>>],
|
||||||
|
unifier: &mut Unifier,
|
||||||
|
primitives: &PrimitiveStore,
|
||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
if targets.len() == 1 {
|
if targets.len() == 1 {
|
||||||
let target = &targets[0];
|
match &targets[0].node {
|
||||||
|
|
||||||
match &target.node {
|
|
||||||
ExprKind::Name { id, .. } => {
|
ExprKind::Name { id, .. } => {
|
||||||
let def_list = composer.extract_def_list();
|
|
||||||
let unifier = &mut composer.unifier;
|
|
||||||
let primitives = &composer.primitives_ty;
|
|
||||||
|
|
||||||
if let Ok(var) =
|
if let Ok(var) =
|
||||||
handle_typevar_definition(value, &*resolver, &def_list, unifier, primitives)
|
handle_typevar_definition(value, resolver, def_list, unifier, primitives)
|
||||||
{
|
{
|
||||||
internal_resolver.add_id_type(*id, var);
|
internal_resolver.add_id_type(*id, var);
|
||||||
Ok(())
|
Ok(())
|
||||||
} else if let Ok(val) = parse_parameter_default_value(value, &*resolver) {
|
} else if let Ok(val) = parse_parameter_default_value(value, resolver) {
|
||||||
internal_resolver.add_module_global(*id, val);
|
internal_resolver.add_module_global(*id, val);
|
||||||
let (name, def_id, _) = composer
|
|
||||||
.register_top_level_var(
|
|
||||||
*id,
|
|
||||||
None,
|
|
||||||
Some(resolver.clone()),
|
|
||||||
"__main__",
|
|
||||||
target.location,
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
internal_resolver.add_id_def(name, def_id);
|
|
||||||
Ok(())
|
Ok(())
|
||||||
} else {
|
} else {
|
||||||
Err(format!("fails to evaluate this expression `{:?}` as a constant or generic parameter at {}",
|
Err(format!("fails to evaluate this expression `{:?}` as a constant or generic parameter at {}",
|
||||||
target.node,
|
targets[0].node,
|
||||||
target.location,
|
targets[0].location,
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ExprKind::List { elts, .. } | ExprKind::Tuple { elts, .. } => {
|
ExprKind::List { elts, .. } | ExprKind::Tuple { elts, .. } => {
|
||||||
handle_assignment_pattern(elts, value, resolver, internal_resolver, composer)?;
|
handle_assignment_pattern(
|
||||||
|
elts,
|
||||||
|
value,
|
||||||
|
resolver,
|
||||||
|
internal_resolver,
|
||||||
|
def_list,
|
||||||
|
unifier,
|
||||||
|
primitives,
|
||||||
|
)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
_ => Err(format!("assignment to {target:?} is not supported at {}", target.location)),
|
_ => Err(format!(
|
||||||
|
"assignment to {:?} is not supported at {}",
|
||||||
|
targets[0], targets[0].location
|
||||||
|
)),
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
match &value.node {
|
match &value.node {
|
||||||
@ -226,9 +216,11 @@ fn handle_assignment_pattern(
|
|||||||
handle_assignment_pattern(
|
handle_assignment_pattern(
|
||||||
std::slice::from_ref(tar),
|
std::slice::from_ref(tar),
|
||||||
val,
|
val,
|
||||||
resolver.clone(),
|
resolver,
|
||||||
internal_resolver,
|
internal_resolver,
|
||||||
composer,
|
def_list,
|
||||||
|
unifier,
|
||||||
|
primitives,
|
||||||
)?;
|
)?;
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
@ -246,40 +238,9 @@ fn handle_assignment_pattern(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn handle_global_var(
|
|
||||||
target: &Expr,
|
|
||||||
value: Option<&Expr>,
|
|
||||||
resolver: &Arc<dyn SymbolResolver + Send + Sync>,
|
|
||||||
internal_resolver: &ResolverInternal,
|
|
||||||
composer: &mut TopLevelComposer,
|
|
||||||
) -> Result<(), String> {
|
|
||||||
let ExprKind::Name { id, .. } = target.node else {
|
|
||||||
return Err(format!(
|
|
||||||
"global variable declaration must be an identifier (at {})",
|
|
||||||
target.location,
|
|
||||||
));
|
|
||||||
};
|
|
||||||
|
|
||||||
let Some(value) = value else {
|
|
||||||
return Err(format!("global variable `{id}` must be initialized in its definition"));
|
|
||||||
};
|
|
||||||
|
|
||||||
if let Ok(val) = parse_parameter_default_value(value, &**resolver) {
|
|
||||||
internal_resolver.add_module_global(id, val);
|
|
||||||
let (name, def_id, _) = composer
|
|
||||||
.register_top_level_var(id, None, Some(resolver.clone()), "__main__", target.location)
|
|
||||||
.unwrap();
|
|
||||||
internal_resolver.add_id_def(name, def_id);
|
|
||||||
Ok(())
|
|
||||||
} else {
|
|
||||||
Err(format!(
|
|
||||||
"failed to evaluate this expression `{:?}` as a constant at {}",
|
|
||||||
target.node, target.location,
|
|
||||||
))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
|
const SIZE_T: u32 = usize::BITS;
|
||||||
|
|
||||||
let cli = CommandLineArgs::parse();
|
let cli = CommandLineArgs::parse();
|
||||||
let CommandLineArgs { file_name, threads, opt_level, emit_llvm, triple, mcpu, target_features } =
|
let CommandLineArgs { file_name, threads, opt_level, emit_llvm, triple, mcpu, target_features } =
|
||||||
cli;
|
cli;
|
||||||
@ -312,36 +273,22 @@ fn main() {
|
|||||||
_ => OptimizationLevel::Aggressive,
|
_ => OptimizationLevel::Aggressive,
|
||||||
};
|
};
|
||||||
|
|
||||||
let target_machine_options = CodeGenTargetMachineOptions {
|
|
||||||
triple,
|
|
||||||
cpu: mcpu,
|
|
||||||
features: target_features,
|
|
||||||
reloc_mode: RelocMode::PIC,
|
|
||||||
..host_target_machine
|
|
||||||
};
|
|
||||||
let target_machine = target_machine_options
|
|
||||||
.create_target_machine(opt_level)
|
|
||||||
.expect("couldn't create target machine");
|
|
||||||
|
|
||||||
let context = nac3core::inkwell::context::Context::create();
|
|
||||||
|
|
||||||
let size_t =
|
|
||||||
context.ptr_sized_int_type(&target_machine.get_target_data(), None).get_bit_width();
|
|
||||||
|
|
||||||
let program = match fs::read_to_string(file_name.clone()) {
|
let program = match fs::read_to_string(file_name.clone()) {
|
||||||
Ok(program) => program,
|
Ok(program) => program,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
panic!("Cannot open input file: {err}");
|
println!("Cannot open input file: {err}");
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
let primitive: PrimitiveStore = TopLevelComposer::make_primitives(size_t).0;
|
let primitive: PrimitiveStore = TopLevelComposer::make_primitives(SIZE_T).0;
|
||||||
let (mut composer, builtins_def, builtins_ty) =
|
let (mut composer, builtins_def, builtins_ty) =
|
||||||
TopLevelComposer::new(vec![], vec![], ComposerConfig::default(), size_t);
|
TopLevelComposer::new(vec![], ComposerConfig::default(), SIZE_T);
|
||||||
|
|
||||||
let internal_resolver: Arc<ResolverInternal> = ResolverInternal {
|
let internal_resolver: Arc<ResolverInternal> = ResolverInternal {
|
||||||
id_to_type: builtins_ty.into(),
|
id_to_type: builtins_ty.into(),
|
||||||
id_to_def: builtins_def.into(),
|
id_to_def: builtins_def.into(),
|
||||||
|
class_names: Mutex::default(),
|
||||||
module_globals: Mutex::default(),
|
module_globals: Mutex::default(),
|
||||||
str_store: Mutex::default(),
|
str_store: Mutex::default(),
|
||||||
}
|
}
|
||||||
@ -349,41 +296,27 @@ fn main() {
|
|||||||
let resolver =
|
let resolver =
|
||||||
Arc::new(Resolver(internal_resolver.clone())) as Arc<dyn SymbolResolver + Send + Sync>;
|
Arc::new(Resolver(internal_resolver.clone())) as Arc<dyn SymbolResolver + Send + Sync>;
|
||||||
|
|
||||||
// Process IRRT
|
|
||||||
let irrt = load_irrt(&context, resolver.as_ref());
|
|
||||||
if emit_llvm {
|
|
||||||
irrt.write_bitcode_to_path(Path::new("irrt.bc"));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Process the Python script
|
|
||||||
let parser_result = parser::parse_program(&program, file_name.into()).unwrap();
|
let parser_result = parser::parse_program(&program, file_name.into()).unwrap();
|
||||||
|
|
||||||
for stmt in parser_result {
|
for stmt in parser_result {
|
||||||
match &stmt.node {
|
match &stmt.node {
|
||||||
StmtKind::Assign { targets, value, .. } => {
|
StmtKind::Assign { targets, value, .. } => {
|
||||||
|
let def_list = composer.extract_def_list();
|
||||||
|
let unifier = &mut composer.unifier;
|
||||||
|
let primitives = &composer.primitives_ty;
|
||||||
if let Err(err) = handle_assignment_pattern(
|
if let Err(err) = handle_assignment_pattern(
|
||||||
targets,
|
targets,
|
||||||
value,
|
value,
|
||||||
resolver.clone(),
|
resolver.as_ref(),
|
||||||
internal_resolver.as_ref(),
|
internal_resolver.as_ref(),
|
||||||
&mut composer,
|
&def_list,
|
||||||
|
unifier,
|
||||||
|
primitives,
|
||||||
) {
|
) {
|
||||||
panic!("{err}");
|
eprintln!("{err}");
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
StmtKind::AnnAssign { target, value, .. } => {
|
|
||||||
if let Err(err) = handle_global_var(
|
|
||||||
target,
|
|
||||||
value.as_ref().map(Box::as_ref),
|
|
||||||
&resolver,
|
|
||||||
internal_resolver.as_ref(),
|
|
||||||
&mut composer,
|
|
||||||
) {
|
|
||||||
panic!("{err}");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// allow (and ignore) "from __future__ import annotations"
|
// allow (and ignore) "from __future__ import annotations"
|
||||||
StmtKind::ImportFrom { module, names, .. }
|
StmtKind::ImportFrom { module, names, .. }
|
||||||
if module == &Some("__future__".into())
|
if module == &Some("__future__".into())
|
||||||
@ -407,19 +340,7 @@ fn main() {
|
|||||||
let signature = store.from_signature(&mut composer.unifier, &primitive, &signature, &mut cache);
|
let signature = store.from_signature(&mut composer.unifier, &primitive, &signature, &mut cache);
|
||||||
let signature = store.add_cty(signature);
|
let signature = store.add_cty(signature);
|
||||||
|
|
||||||
if let Err(errors) = composer.start_analysis(true) {
|
composer.start_analysis(true).unwrap();
|
||||||
let error_count = errors.len();
|
|
||||||
eprintln!("{error_count} error(s) occurred during top level analysis.");
|
|
||||||
|
|
||||||
for (error_i, error) in errors.iter().enumerate() {
|
|
||||||
let error_num = error_i + 1;
|
|
||||||
eprintln!("=========== ERROR {error_num}/{error_count} ============");
|
|
||||||
eprintln!("{error}");
|
|
||||||
}
|
|
||||||
eprintln!("==================================");
|
|
||||||
|
|
||||||
panic!("top level analysis failed");
|
|
||||||
}
|
|
||||||
|
|
||||||
let top_level = Arc::new(composer.make_top_level_context());
|
let top_level = Arc::new(composer.make_top_level_context());
|
||||||
|
|
||||||
@ -438,7 +359,16 @@ fn main() {
|
|||||||
instance_to_stmt[""].clone()
|
instance_to_stmt[""].clone()
|
||||||
};
|
};
|
||||||
|
|
||||||
let llvm_options = CodeGenLLVMOptions { opt_level, target: target_machine_options };
|
let llvm_options = CodeGenLLVMOptions {
|
||||||
|
opt_level,
|
||||||
|
target: CodeGenTargetMachineOptions {
|
||||||
|
triple,
|
||||||
|
cpu: mcpu,
|
||||||
|
features: target_features,
|
||||||
|
reloc_mode: RelocMode::PIC,
|
||||||
|
..host_target_machine
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
let task = CodeGenTask {
|
let task = CodeGenTask {
|
||||||
subst: Vec::default(),
|
subst: Vec::default(),
|
||||||
@ -461,14 +391,14 @@ fn main() {
|
|||||||
membuffer.lock().push(buffer);
|
membuffer.lock().push(buffer);
|
||||||
})));
|
})));
|
||||||
let threads = (0..threads)
|
let threads = (0..threads)
|
||||||
.map(|i| Box::new(DefaultCodeGenerator::new(format!("module{i}"), size_t)))
|
.map(|i| Box::new(DefaultCodeGenerator::new(format!("module{i}"), SIZE_T)))
|
||||||
.collect();
|
.collect();
|
||||||
let (registry, handles) = WorkerRegistry::create_workers(threads, top_level, &llvm_options, &f);
|
let (registry, handles) = WorkerRegistry::create_workers(threads, top_level, &llvm_options, &f);
|
||||||
registry.add_task(task);
|
registry.add_task(task);
|
||||||
registry.wait_tasks_complete(handles);
|
registry.wait_tasks_complete(handles);
|
||||||
|
|
||||||
// Link all modules together into `main`
|
|
||||||
let buffers = membuffers.lock();
|
let buffers = membuffers.lock();
|
||||||
|
let context = inkwell::context::Context::create();
|
||||||
let main = context
|
let main = context
|
||||||
.create_module_from_ir(MemoryBuffer::create_from_memory_range(&buffers[0], "main"))
|
.create_module_from_ir(MemoryBuffer::create_from_memory_range(&buffers[0], "main"))
|
||||||
.unwrap();
|
.unwrap();
|
||||||
@ -488,18 +418,25 @@ fn main() {
|
|||||||
main.link_in_module(other).unwrap();
|
main.link_in_module(other).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let irrt = load_irrt(&context);
|
||||||
|
if emit_llvm {
|
||||||
|
irrt.write_bitcode_to_path(Path::new("irrt.bc"));
|
||||||
|
}
|
||||||
main.link_in_module(irrt).unwrap();
|
main.link_in_module(irrt).unwrap();
|
||||||
|
|
||||||
// Private all functions except "run"
|
|
||||||
let mut function_iter = main.get_first_function();
|
let mut function_iter = main.get_first_function();
|
||||||
while let Some(func) = function_iter {
|
while let Some(func) = function_iter {
|
||||||
if func.count_basic_blocks() > 0 && func.get_name().to_str().unwrap() != "run" {
|
if func.count_basic_blocks() > 0 && func.get_name().to_str().unwrap() != "run" {
|
||||||
func.set_linkage(Linkage::Private);
|
func.set_linkage(inkwell::module::Linkage::Private);
|
||||||
}
|
}
|
||||||
function_iter = func.get_next_function();
|
function_iter = func.get_next_function();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Optimize `main`
|
let target_machine = llvm_options
|
||||||
|
.target
|
||||||
|
.create_target_machine(llvm_options.opt_level)
|
||||||
|
.expect("couldn't create target machine");
|
||||||
|
|
||||||
let pass_options = PassBuilderOptions::create();
|
let pass_options = PassBuilderOptions::create();
|
||||||
pass_options.set_merge_functions(true);
|
pass_options.set_merge_functions(true);
|
||||||
let passes = format!("default<O{}>", opt_level as u32);
|
let passes = format!("default<O{}>", opt_level as u32);
|
||||||
@ -508,7 +445,6 @@ fn main() {
|
|||||||
panic!("Failed to run optimization for module `main`: {}", err.to_string());
|
panic!("Failed to run optimization for module `main`: {}", err.to_string());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Write output
|
|
||||||
target_machine
|
target_machine
|
||||||
.write_to_file(&main, FileType::Object, Path::new("module.o"))
|
.write_to_file(&main, FileType::Object, Path::new("module.o"))
|
||||||
.expect("couldn't write module to file");
|
.expect("couldn't write module to file");
|
||||||
|
@ -21,6 +21,6 @@ build() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
package() {
|
package() {
|
||||||
mkdir -p $pkgdir/clang64/lib/python3.12/site-packages
|
mkdir -p $pkgdir/clang64/lib/python3.11/site-packages
|
||||||
cp ${srcdir}/nac3artiq.pyd $pkgdir/clang64/lib/python3.12/site-packages
|
cp ${srcdir}/nac3artiq.pyd $pkgdir/clang64/lib/python3.11/site-packages
|
||||||
}
|
}
|
||||||
|
@ -21,10 +21,10 @@ let
|
|||||||
text =
|
text =
|
||||||
''
|
''
|
||||||
implementation=CPython
|
implementation=CPython
|
||||||
version=3.12
|
version=3.11
|
||||||
shared=true
|
shared=true
|
||||||
abi3=false
|
abi3=false
|
||||||
lib_name=python3.12
|
lib_name=python3.11
|
||||||
lib_dir=${msys2-env}/clang64/lib
|
lib_dir=${msys2-env}/clang64/lib
|
||||||
pointer_width=64
|
pointer_width=64
|
||||||
build_flags=WITH_THREAD
|
build_flags=WITH_THREAD
|
||||||
|
@ -6,11 +6,11 @@ cd $(dirname $0)
|
|||||||
MSYS2DIR=`pwd`/msys2
|
MSYS2DIR=`pwd`/msys2
|
||||||
mkdir -p $MSYS2DIR/var/lib/pacman $MSYS2DIR/msys/etc
|
mkdir -p $MSYS2DIR/var/lib/pacman $MSYS2DIR/msys/etc
|
||||||
|
|
||||||
curl -L https://repo.msys2.org/msys/x86_64/pacman-mirrors-20240523-1-any.pkg.tar.zst | tar xvf - -C $MSYS2DIR --zstd
|
curl -L https://mirror.msys2.org/msys/x86_64/pacman-mirrors-20220205-1-any.pkg.tar.zst | tar xvf - -C $MSYS2DIR --zstd
|
||||||
curl -L https://raw.githubusercontent.com/msys2/MSYS2-packages/master/pacman/pacman.conf | sed "s|SigLevel = Required|SigLevel = Never|g" | sed "s|/etc/pacman.d|$MSYS2DIR/etc/pacman.d|g" > $MSYS2DIR/etc/pacman.conf
|
curl -L https://raw.githubusercontent.com/msys2/MSYS2-packages/master/pacman/pacman.conf | sed "s|SigLevel = Required|SigLevel = Never|g" | sed "s|/etc/pacman.d|$MSYS2DIR/etc/pacman.d|g" > $MSYS2DIR/etc/pacman.conf
|
||||||
|
|
||||||
fakeroot pacman --root $MSYS2DIR --config $MSYS2DIR/etc/pacman.conf -Syy
|
fakeroot pacman --root $MSYS2DIR --config $MSYS2DIR/etc/pacman.conf -Syy
|
||||||
pacman --root $MSYS2DIR --config $MSYS2DIR/etc/pacman.conf --cachedir $MSYS2DIR/msys/cache -Sp mingw-w64-clang-x86_64-rust mingw-w64-clang-x86_64-cmake mingw-w64-clang-x86_64-ninja mingw-w64-clang-x86_64-python-numpy mingw-w64-clang-x86_64-python-setuptools > $MSYS2DIR/packages.txt
|
pacman --root $MSYS2DIR --config $MSYS2DIR/etc/pacman.conf --cachedir $MSYS2DIR/msys/cache -Sp mingw-w64-clang-x86_64-rust mingw-w64-clang-x86_64-cmake mingw-w64-clang-x86_64-ninja mingw-w64-clang-x86_64-python3.11 mingw-w64-clang-x86_64-python-numpy mingw-w64-clang-x86_64-python-setuptools > $MSYS2DIR/packages.txt
|
||||||
|
|
||||||
echo "{ pkgs } : [" > msys2_packages.nix
|
echo "{ pkgs } : [" > msys2_packages.nix
|
||||||
while read package; do
|
while read package; do
|
||||||
|
@ -1,15 +1,15 @@
|
|||||||
{ pkgs } : [
|
{ pkgs } : [
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libunwind-18.1.8-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libunwind-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "0f9m76dx40iy794nfks0360gvjhdg6yngb2lyhwp4xd76rn5081m";
|
sha256 = "0ksz7xz1lbwsmdr9sa1444k0dlfkbd8k11pq7w08ir7r1wjy6fid";
|
||||||
name = "mingw-w64-clang-x86_64-libunwind-18.1.8-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libunwind-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libc++-18.1.8-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libc++-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "17savj9wys9my2ji7vyba7wwqkvzdjwnkb3k4858wxrjbzbfa6lk";
|
sha256 = "0r8skyjqv4cpkqif0niakx4hdpkscil1zf6mzj34pqna0j5gdnq2";
|
||||||
name = "mingw-w64-clang-x86_64-libc++-18.1.8-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libc++-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -31,9 +31,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-xz-5.6.2-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-xz-5.6.1-1-any.pkg.tar.zst";
|
||||||
sha256 = "0phb9hwqksk1rg29yhwlc7si78zav19c2kac0i841pc7mc2n9gzx";
|
sha256 = "14p4xxaxjjy6j1ingji82xhai1mc1gls5ali6z40fbb2ylxkaggs";
|
||||||
name = "mingw-w64-clang-x86_64-xz-5.6.2-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-xz-5.6.1-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -43,93 +43,93 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libxml2-2.12.9-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libxml2-2.12.6-1-any.pkg.tar.zst";
|
||||||
sha256 = "0cjz2vj9yz6k5xj601cp0yk631rrr0z94ciamwqrvclb0yhakf25";
|
sha256 = "177b3rmsknqq6hf0zqwva71s3avh20ca7vzznp2ls2z5qm8vhhlp";
|
||||||
name = "mingw-w64-clang-x86_64-libxml2-2.12.9-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libxml2-2.12.6-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-zstd-1.5.6-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-zstd-1.5.5-1-any.pkg.tar.zst";
|
||||||
sha256 = "02cp5ci8w50k7xn38mpkwnr8sn898v18wcc07y8f9sfla7vcyfix";
|
sha256 = "07739wmwgxf0d6db4p8w302a6jwcm01aafr1s8jvcl5k1h5a1m2m";
|
||||||
name = "mingw-w64-clang-x86_64-zstd-1.5.6-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-zstd-1.5.5-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-llvm-libs-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-llvm-libs-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "0rpbgvvinsqflhd3nhfxk0g0yy8j80zzw5yx6573ak0m78a9fa06";
|
sha256 = "0ibiy01v16naik9pj32ch7a9pkbw4yrn3gyq7p0y6kcc63fkjazy";
|
||||||
name = "mingw-w64-clang-x86_64-llvm-libs-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-llvm-libs-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-llvm-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-llvm-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "185g5h8q3x3rav9lp2njln58ny2idh2067fd02j3nsbik6glshpf";
|
sha256 = "1hcfz6nb6svmmcqzfrdi96az2x7mzj0cispdv2ssbgn7nkf19pi0";
|
||||||
name = "mingw-w64-clang-x86_64-llvm-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-llvm-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-clang-libs-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-clang-libs-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "089hji3yd7wsd03v9mdfgc99l5k1dql8kg7p3hy13vrbgfsabxhc";
|
sha256 = "1k17d18g7rmq2ph4kq1mf84vs8133jzf52nkv6syh39ypjga67wa";
|
||||||
name = "mingw-w64-clang-x86_64-clang-libs-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-clang-libs-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-compiler-rt-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-compiler-rt-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "1dwcxnv1k5ljim5ys4h1c3jlrdpi0054z094ynav7if65i8zjj4a";
|
sha256 = "1w2j0vs888haz9shjr1l8dc4j957sk1p0377zzipkbqnzqwjf1z8";
|
||||||
name = "mingw-w64-clang-x86_64-compiler-rt-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-compiler-rt-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-headers-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-headers-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
sha256 = "0163jzjlvq7inpafy3h48pkwag3ysk6x56xm84yfcz5q52fnfzq5";
|
sha256 = "18csfwlk2h9pr4411crx1b41qjzn5jgbssm3h109nzwbdizkp62h";
|
||||||
name = "mingw-w64-clang-x86_64-headers-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-headers-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-crt-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-crt-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
sha256 = "00cn1mi29mfys7qy4hvgnjd0smqvnkdn3ibnrr6a3wy1h2vaykgq";
|
sha256 = "03l1zkrxgxxssp430xcv2gch1d03rbnbk1c0vgiqxigcs8lljh2g";
|
||||||
name = "mingw-w64-clang-x86_64-crt-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-crt-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-lld-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-lld-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "1vpij5d06m4kjy3qv8bizwlkl21gcv6fv0r2f1j9bclgm6k3144x";
|
sha256 = "1ai4gl7ybpk9n10jmbpf3zzfa893m1krj5qhf44ajln0jabdfnbn";
|
||||||
name = "mingw-w64-clang-x86_64-lld-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-lld-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libwinpthread-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libwinpthread-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
sha256 = "1zkzqqd31xpkv817wja3qssjjx891bsdxw07037hv2sk0qr4ffn9";
|
sha256 = "1svhjzwhvl4ldl439jhgfy47g05y2af1cjqvydgijn1dd4g8y8vq";
|
||||||
name = "mingw-w64-clang-x86_64-libwinpthread-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libwinpthread-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-winpthreads-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-winpthreads-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
sha256 = "02ynia88ad3l03r08nyldmnajwqkyxcjd191lyamkbj4d6zck323";
|
sha256 = "0jxdhkl256vnr13xf1x3fyjrdf764zg70xcs3gki3rg109f0a6xk";
|
||||||
name = "mingw-w64-clang-x86_64-winpthreads-git-12.0.0.r250.gc6bf4bdf6-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-winpthreads-git-11.0.0.r655.gdbfdf8025-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-clang-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-clang-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "1qny934nv4g75k9gb5sf31v24bgafkg6qw7r35xv3in491w6annq";
|
sha256 = "0ahfic7vdfv96k5v7fdkgk1agk28l833xjn2igrmbvqg96ak0w6n";
|
||||||
name = "mingw-w64-clang-x86_64-clang-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-clang-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-c-ares-1.33.1-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-c-ares-1.27.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "14r6jjsvfbapbkv2zqp2yglva4vz4srzkgk7f186ri3kcafjspgq";
|
sha256 = "06y3sgqv6a0gr3dsbzs36jrj8adklssgjqi2ms5clsyq6ay4f91r";
|
||||||
name = "mingw-w64-clang-x86_64-c-ares-1.33.1-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-c-ares-1.27.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-brotli-1.1.0-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-brotli-1.1.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "1q01lz9lcyrjmkhv9rddgjazmk7warlcmwhc4qkq9y6h0yfsb71n";
|
sha256 = "113mha41q53cx0hw13cq1xdf7zbsd58sh8cl1cd7xzg1q69n60w2";
|
||||||
name = "mingw-w64-clang-x86_64-brotli-1.1.0-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-brotli-1.1.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libunistring-1.2-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libunistring-1.1-1-any.pkg.tar.zst";
|
||||||
sha256 = "13nz49li39z1zgfx1q9jg4vrmyrmqb6qdq0nqshidaqc6zr16k3g";
|
sha256 = "16myvbg33q5s7jl30w5qd8n8f1r05335ms8r61234vn52n32l2c4";
|
||||||
name = "mingw-w64-clang-x86_64-libunistring-1.2-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libunistring-1.1-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -151,9 +151,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-p11-kit-0.25.5-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-p11-kit-0.25.3-2-any.pkg.tar.zst";
|
||||||
sha256 = "00yz6cmr1ldlrskv811n345xcia88mj7w4fyx4m9z5848jxgsabd";
|
sha256 = "1jrwkc4lvw5hm5rqmi5gqh7mfkbqfa5gi81zjij0krnl0gaxw3c8";
|
||||||
name = "mingw-w64-clang-x86_64-p11-kit-0.25.5-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-p11-kit-0.25.3-2-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -163,9 +163,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-openssl-3.3.2-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-openssl-3.2.1-1-any.pkg.tar.zst";
|
||||||
sha256 = "1djgpcz447yvhdy1yq5wh8l5d0821izxklx9afyszbw0pbr7f24y";
|
sha256 = "0ix2r4ll09m2z5vz2k94gmwfs0pp3ipvjdimwzx7v6xhcs2l25lz";
|
||||||
name = "mingw-w64-clang-x86_64-openssl-3.3.2-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-openssl-3.2.1-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -175,39 +175,33 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-nghttp2-1.63.0-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-nghttp2-1.60.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "0lfqrlmapsc7ilxjmhr7hxi578vclqlhpqimbvzq0c70c0iwk864";
|
sha256 = "0wxw8266hf4qd2m4zpgb1wvlrnaksmcrs0kh5y9zpf2y5sy8f2bq";
|
||||||
name = "mingw-w64-clang-x86_64-nghttp2-1.63.0-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-nghttp2-1.60.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-nghttp3-1.5.0-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-curl-8.6.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "1ljl9kdasf91bxkqcmbbjchp5g00ahv8jn2zab38899z6j3x43nz";
|
sha256 = "1racc7cyzj22kink9w8m8jv73ji5hfg6r6d1ka9dqmvcbx04r8p0";
|
||||||
name = "mingw-w64-clang-x86_64-nghttp3-1.5.0-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-curl-8.6.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-curl-8.9.1-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-rust-1.76.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "1zr6kgqp9i4qqrfckh3kfmz4x1cwv4xis9sfqsx7xji88priax64";
|
sha256 = "0ny3bvwvn5wmqrxzhdfw34akr0kj0m7rg9lg3w5yibqz2mkqhk11";
|
||||||
name = "mingw-w64-clang-x86_64-curl-8.9.1-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-rust-1.76.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-rust-1.80.1-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-pkgconf-1~2.1.1-1-any.pkg.tar.zst";
|
||||||
sha256 = "1dm4vlrfi9m6xl09zpn0yjr7qcjjr4x738z1rjfwysfnc0awq4x8";
|
sha256 = "00kxqg9ds4q74lxrzjh8z0858smqbi1j9r06s0zjadsql0ln98cq";
|
||||||
name = "mingw-w64-clang-x86_64-rust-1.80.1-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-pkgconf-12.1.1-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-cppdap-1.65-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-expat-2.6.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "0phhwkcqp30dsyj5vr6w99sgm1jfm5rzg0w5x5mv9md4x7lm9lmh";
|
sha256 = "0kj1vzjh3qh7d2g47avlgk7a6j4nc62111hy1m63jwq0alc01k38";
|
||||||
name = "mingw-w64-clang-x86_64-cppdap-1.65-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-expat-2.6.2-1-any.pkg.tar.zst";
|
||||||
})
|
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-expat-2.6.3-1-any.pkg.tar.zst";
|
|
||||||
sha256 = "19xfl1q78q1k8j0lr5aspcf668pmfg01fgib73zq7ff7y5y5fcyi";
|
|
||||||
name = "mingw-w64-clang-x86_64-expat-2.6.3-1-any.pkg.tar.zst";
|
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -229,9 +223,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-lz4-1.10.0-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-lz4-1.9.4-1-any.pkg.tar.zst";
|
||||||
sha256 = "0kznnw9z9zqxkmn8qbypm2rpsfaapbgls1ks3zzpfnfjz9cpw8py";
|
sha256 = "0nn7cy25j53q5ckkx4n4f77w00xdwwf5wjswm374shvvs58nlln0";
|
||||||
name = "mingw-w64-clang-x86_64-lz4-1.10.0-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-lz4-1.9.4-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -247,9 +241,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libarchive-3.7.4-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-libarchive-3.7.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "1ykw6imllgxv6lsgwxx1miqjr4l1iryqkrj286jcbfrb8ghpzhv5";
|
sha256 = "1p84yh6yzkdpmr02vyvgz16x5gycckah25jkdc2py09l7iw96bmw";
|
||||||
name = "mingw-w64-clang-x86_64-libarchive-3.7.4-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-libarchive-3.7.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -259,15 +253,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-ninja-1.12.1-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-ninja-1.11.1-3-any.pkg.tar.zst";
|
||||||
sha256 = "1vj9qaa43v316daz8k4ricmz3f33nhjpj7r0vn979nwmy7hzs7jx";
|
sha256 = "13wjfmyfr952n3ydpldjlwx1nla5xpyvr96ng8pfbyw4z900v5ms";
|
||||||
name = "mingw-w64-clang-x86_64-ninja-1.12.1-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-ninja-1.11.1-3-any.pkg.tar.zst";
|
||||||
})
|
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-pkgconf-1~2.3.0-1-any.pkg.tar.zst";
|
|
||||||
sha256 = "15i7x6akkgs7aa7aa804k93p2iipnvygsy7z8hsafskka3h150fa";
|
|
||||||
name = "mingw-w64-clang-x86_64-pkgconf-12.3.0-1-any.pkg.tar.zst";
|
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -277,9 +265,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-cmake-3.30.3-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-cmake-3.29.0-1-any.pkg.tar.zst";
|
||||||
sha256 = "0fjwf6xxzli6rcsbzr1razldmm538ibkyf5kw132lpaz5wma9bj8";
|
sha256 = "0l79lf6zihn0k8hz93qnjnq259y45yq19235g9c444jc2w093si1";
|
||||||
name = "mingw-w64-clang-x86_64-cmake-3.30.3-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-cmake-3.29.0-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -301,9 +289,9 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-readline-8.2.013-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-readline-8.2.010-1-any.pkg.tar.zst";
|
||||||
sha256 = "0pv1ypqfgm4mimzr0amq9anr1ysqmzrwv6gfk7rrlzhihadknsvr";
|
sha256 = "1s47pd5iz8y3hspsxn4pnp0v3m05ccia40v5nfvx0rmwgvcaz82v";
|
||||||
name = "mingw-w64-clang-x86_64-readline-8.2.013-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-readline-8.2.010-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -313,15 +301,15 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-sqlite3-3.46.1-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-sqlite3-3.45.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "1axplxyjnaz411qzjjqwbj55fbrh4akq3plm2p1sx64jp844xpyq";
|
sha256 = "1icvw3f08cgi94p0177i46v72wgpsxw95p6kd0sm2w3vj0qlqbcw";
|
||||||
name = "mingw-w64-clang-x86_64-sqlite3-3.46.1-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-sqlite3-3.45.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-tk-8.6.13-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-tk-8.6.12-2-any.pkg.tar.zst";
|
||||||
sha256 = "12f6lqx1sglczcnz2ns6sxw9cxwm1klxajqzcrbnfwln1nllz2nd";
|
sha256 = "0pi74q91vl6vw8vvmmwnvrgai3b1aanp0zhca5qsmv8ljh2wdgzx";
|
||||||
name = "mingw-w64-clang-x86_64-tk-8.6.13-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-tk-8.6.12-2-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
@ -331,38 +319,32 @@
|
|||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python3.12-3.12.1-2-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-3.11.8-1-any.pkg.tar.zst";
|
||||||
sha256 = "0wmd39wl9z237w093a7c6hl5pclca9yvwxn0kiw6i2njk3sjv51a";
|
sha256 = "0djpf4k8s25nys6nrm2x2v134lcgzhhbjs37ihkg0b3sxmmc3b0p";
|
||||||
name = "mingw-w64-clang-x86_64-python3.12-3.12.1-2-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-python-3.11.8-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-3.11.9-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-openmp-18.1.2-1-any.pkg.tar.zst";
|
||||||
sha256 = "0ah1idjqxg7jc07a1gz9z766rjjd0f0c6ri4hpcsimsrbj1zjd3c";
|
sha256 = "1v9wm3ja3a7a7yna2bpqky481qf244wc98kfdl7l03k7rkvvydpl";
|
||||||
name = "mingw-w64-clang-x86_64-python-3.11.9-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-openmp-18.1.2-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-llvm-openmp-18.1.8-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-openblas-0.3.26-1-any.pkg.tar.zst";
|
||||||
sha256 = "0cy2v0l4af24j34mzj5q5nlzcqhackfajlfj1rpf6mb3rbz23qw9";
|
sha256 = "0kdr72y5lc9dl9s1bjrw8g21qmv2iwd1xvn1r21170i277wsmqiv";
|
||||||
name = "mingw-w64-clang-x86_64-llvm-openmp-18.1.8-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-openblas-0.3.26-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-openblas-0.3.28-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-numpy-1.26.4-1-any.pkg.tar.zst";
|
||||||
sha256 = "1pskcqc1lg9p8m8rk7bw3mz7mn7vw5fpl7zxa23bhjn02p5b79qq";
|
sha256 = "00h0ap954cjwlsc3p01fjwy7s3nlzs90v0kmnrzxm0rljmvn4jkf";
|
||||||
name = "mingw-w64-clang-x86_64-openblas-0.3.28-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-python-numpy-1.26.4-1-any.pkg.tar.zst";
|
||||||
})
|
})
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
(pkgs.fetchurl {
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-numpy-2.0.1-1-any.pkg.tar.zst";
|
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-setuptools-69.1.1-1-any.pkg.tar.zst";
|
||||||
sha256 = "0ks6q8v58h4wmr2pzsjl2xm4f63g0psvfm0jwlz24mqfxp8gqfcc";
|
sha256 = "1mc56anasj0v92nlg84m3pa7dbqgjakxw0b4ibqlrr9cq0xzsg4b";
|
||||||
name = "mingw-w64-clang-x86_64-python-numpy-2.0.1-1-any.pkg.tar.zst";
|
name = "mingw-w64-clang-x86_64-python-setuptools-69.1.1-1-any.pkg.tar.zst";
|
||||||
})
|
|
||||||
|
|
||||||
(pkgs.fetchurl {
|
|
||||||
url = "https://mirror.msys2.org/mingw/clang64/mingw-w64-clang-x86_64-python-setuptools-74.0.0-1-any.pkg.tar.zst";
|
|
||||||
sha256 = "0xc95z5jzzjf5lw35bs4yn5rlwkrkmrh78yi5rranqpz5nn0wsa0";
|
|
||||||
name = "mingw-w64-clang-x86_64-python-setuptools-74.0.0-1-any.pkg.tar.zst";
|
|
||||||
})
|
})
|
||||||
]
|
]
|
||||||
|
Loading…
Reference in New Issue
Block a user