2025-07-12 13:39.16: New job: test ahrefs/ocannl https://github.com/ahrefs/ocannl.git#refs/heads/master (90bf5fc6cd4eb365478ffa2c291ca2d69a834b7e) (linux-x86_64:alpine-3.21-5.3_opam-2.3) Base: ocaml/opam:alpine-3.21-ocaml-5.3@sha256:04615dc1fa88c0285b28799f65b32f2814c854db498aea42fb9ff8c430a28380 Opam project build To reproduce locally: git clone --recursive "https://github.com/ahrefs/ocannl.git" -b "master" && cd "ocannl" && git reset --hard 90bf5fc6 cat > Dockerfile <<'END-OF-DOCKERFILE' FROM ocaml/opam:alpine-3.21-ocaml-5.3@sha256:04615dc1fa88c0285b28799f65b32f2814c854db498aea42fb9ff8c430a28380 # alpine-3.21-5.3_opam-2.3 USER 1000:1000 ENV CLICOLOR_FORCE="1" ENV OPAMCOLOR="always" WORKDIR /src RUN sudo ln -f /usr/bin/opam-2.3 /usr/bin/opam RUN opam init --reinit -ni RUN uname -rs && opam exec -- ocaml -version && opam --version WORKDIR /src RUN sudo chown opam /src RUN cd ~/opam-repository && (git cat-file -e 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 || git fetch origin master) && git reset -q --hard 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 && git log --no-decorate -n1 --oneline && opam update -u COPY --chown=1000:1000 neural_nets_lib.opam arrayjit.opam ./ RUN opam pin add -yn neural_nets_lib.dev './' && \ opam pin add -yn arrayjit.dev './' RUN echo '(lang dune 3.0)' > './dune-project' ENV DEPS="angstrom.0.16.1 astring.0.8.5 backoff.0.1.1 base.v0.17.3 base-bigarray.base base-bytes.base base-domains.base base-effects.base base-nnp.base base-threads.base base-unix.base bigarray-compat.1.1.0 bigstringaf.0.10.0 camlp-streams.5.0.1 camlzip.1.13 cmdliner.1.3.0 conf-libcurl.2 conf-libffi.2.0.0 conf-pkg-config.4 conf-zlib.1 cppo.1.8.0 csexp.1.5.2 csv.2.4 ctypes.0.23.0 ctypes-foreign.0.23.0 curl.0.10.0 dune.3.19.1 dune-configurator.3.19.1 fieldslib.v0.17.0 fmt.0.10.0 integers.0.7.0 jane-street-headers.v0.17.0 jst-config.v0.17.0 logs.0.9.0 mdx.2.5.0 mtime.2.1.0 multicore-magic.2.3.1 num.1.6 ocaml.5.3.0 ocaml-base-compiler.5.3.0 ocaml-compiler.5.3.0 ocaml-compiler-libs.v0.17.0 ocaml-config.3 ocaml-options-vanilla.1 ocaml-syntax-shims.1.0.0 ocaml-version.4.0.1 ocaml_intrinsics_kernel.v0.17.1 ocamlbuild.0.16.1 ocamlfind.1.9.8 parsexp.v0.17.0 pprint.20230830 ppx_assert.v0.17.0 ppx_base.v0.17.0 ppx_cold.v0.17.0 ppx_compare.v0.17.0 ppx_derivers.1.2.1 ppx_deriving.6.1.0 ppx_enumerate.v0.17.0 ppx_expect.v0.17.3 ppx_fields_conv.v0.17.0 ppx_globalize.v0.17.2 ppx_hash.v0.17.0 ppx_here.v0.17.0 ppx_inline_test.v0.17.1 ppx_minidebug.2.3.0 ppx_optcomp.v0.17.1 ppx_sexp_conv.v0.17.1 ppx_string.v0.17.0 ppx_variants_conv.v0.17.1 ppxlib.0.36.0 ppxlib_jane.v0.17.4 printbox.0.12 printbox-ext-plot.0.12 printbox-html.0.12 printbox-md.0.12 printbox-text.0.12 ptime.1.2.0 re.1.13.2 result.1.5 saturn_lockfree.0.5.0 seq.base sexplib.v0.17.0 sexplib0.v0.17.0 stdio.v0.17.0 stdlib-shims.0.3.0 thread-local-storage.0.2 time_now.v0.17.0 topkg.1.0.8 tyxml.4.6.0 uucp.16.0.0 uutf.1.0.4 variantslib.v0.17.0" ENV CI="true" ENV OCAMLCI="true" RUN opam update --depexts && opam install --cli=2.3 --depext-only -y neural_nets_lib.dev arrayjit.dev $DEPS RUN opam install $DEPS COPY --chown=1000:1000 . /src RUN opam exec -- dune build @install @check @runtest && rm -rf _build END-OF-DOCKERFILE docker build . END-REPRO-BLOCK 2025-07-12 13:39.16: Using cache hint "ahrefs/ocannl-ocaml/opam:alpine-3.21-ocaml-5.3@sha256:04615dc1fa88c0285b28799f65b32f2814c854db498aea42fb9ff8c430a28380-alpine-3.21-5.3_opam-2.3-ee06dfbb3e21290708c2c8a6224283e3" 2025-07-12 13:39.16: Using OBuilder spec: ((from ocaml/opam:alpine-3.21-ocaml-5.3@sha256:04615dc1fa88c0285b28799f65b32f2814c854db498aea42fb9ff8c430a28380) (comment alpine-3.21-5.3_opam-2.3) (user (uid 1000) (gid 1000)) (env CLICOLOR_FORCE 1) (env OPAMCOLOR always) (workdir /src) (run (shell "sudo ln -f /usr/bin/opam-2.3 /usr/bin/opam")) (run (shell "opam init --reinit -ni")) (run (shell "uname -rs && opam exec -- ocaml -version && opam --version")) (workdir /src) (run (shell "sudo chown opam /src")) (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "cd ~/opam-repository && (git cat-file -e 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 || git fetch origin master) && git reset -q --hard 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 && git log --no-decorate -n1 --oneline && opam update -u")) (copy (src neural_nets_lib.opam arrayjit.opam) (dst ./)) (run (network host) (shell "opam pin add -yn neural_nets_lib.dev './' && \ \nopam pin add -yn arrayjit.dev './'")) (run (network host) (shell "echo '(lang dune 3.0)' > './dune-project'")) (env DEPS "angstrom.0.16.1 astring.0.8.5 backoff.0.1.1 base.v0.17.3 base-bigarray.base base-bytes.base base-domains.base base-effects.base base-nnp.base base-threads.base base-unix.base bigarray-compat.1.1.0 bigstringaf.0.10.0 camlp-streams.5.0.1 camlzip.1.13 cmdliner.1.3.0 conf-libcurl.2 conf-libffi.2.0.0 conf-pkg-config.4 conf-zlib.1 cppo.1.8.0 csexp.1.5.2 csv.2.4 ctypes.0.23.0 ctypes-foreign.0.23.0 curl.0.10.0 dune.3.19.1 dune-configurator.3.19.1 fieldslib.v0.17.0 fmt.0.10.0 integers.0.7.0 jane-street-headers.v0.17.0 jst-config.v0.17.0 logs.0.9.0 mdx.2.5.0 mtime.2.1.0 multicore-magic.2.3.1 num.1.6 ocaml.5.3.0 ocaml-base-compiler.5.3.0 ocaml-compiler.5.3.0 ocaml-compiler-libs.v0.17.0 ocaml-config.3 ocaml-options-vanilla.1 ocaml-syntax-shims.1.0.0 ocaml-version.4.0.1 ocaml_intrinsics_kernel.v0.17.1 ocamlbuild.0.16.1 ocamlfind.1.9.8 parsexp.v0.17.0 pprint.20230830 ppx_assert.v0.17.0 ppx_base.v0.17.0 ppx_cold.v0.17.0 ppx_compare.v0.17.0 ppx_derivers.1.2.1 ppx_deriving.6.1.0 ppx_enumerate.v0.17.0 ppx_expect.v0.17.3 ppx_fields_conv.v0.17.0 ppx_globalize.v0.17.2 ppx_hash.v0.17.0 ppx_here.v0.17.0 ppx_inline_test.v0.17.1 ppx_minidebug.2.3.0 ppx_optcomp.v0.17.1 ppx_sexp_conv.v0.17.1 ppx_string.v0.17.0 ppx_variants_conv.v0.17.1 ppxlib.0.36.0 ppxlib_jane.v0.17.4 printbox.0.12 printbox-ext-plot.0.12 printbox-html.0.12 printbox-md.0.12 printbox-text.0.12 ptime.1.2.0 re.1.13.2 result.1.5 saturn_lockfree.0.5.0 seq.base sexplib.v0.17.0 sexplib0.v0.17.0 stdio.v0.17.0 stdlib-shims.0.3.0 thread-local-storage.0.2 time_now.v0.17.0 topkg.1.0.8 tyxml.4.6.0 uucp.16.0.0 uutf.1.0.4 variantslib.v0.17.0") (env CI true) (env OCAMLCI true) (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "opam update --depexts && opam install --cli=2.3 --depext-only -y neural_nets_lib.dev arrayjit.dev $DEPS")) (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "opam install $DEPS")) (copy (src .) (dst /src)) (run (shell "opam exec -- dune build @install @check @runtest && rm -rf _build")) ) 2025-07-12 13:39.16: Waiting for resource in pool OCluster 2025-07-12 13:39.16: Waiting for worker… 2025-07-12 13:39.17: Got resource from pool OCluster Building on odawa.caelum.ci.dev All commits already cached HEAD is now at 90bf5fc6 Also rename Ops.map_prec -> apply_prec (see recent commit on renaming Ndarray.map) (from ocaml/opam:alpine-3.21-ocaml-5.3@sha256:04615dc1fa88c0285b28799f65b32f2814c854db498aea42fb9ff8c430a28380) 2025-07-12 13:39.18 ---> using "b1eb2a8af4058a254870e5cad61297a06793cbcf25fc0fa70ddab0692ceef147" from cache /: (comment alpine-3.21-5.3_opam-2.3) /: (user (uid 1000) (gid 1000)) /: (env CLICOLOR_FORCE 1) /: (env OPAMCOLOR always) /: (workdir /src) /src: (run (shell "sudo ln -f /usr/bin/opam-2.3 /usr/bin/opam")) 2025-07-12 13:39.18 ---> using "72a43fed50ca4b22fad6057f4d18290b4cb956a1be690ef2ac7f62bd3176987a" from cache /src: (run (shell "opam init --reinit -ni")) Configuring from /home/opam/.opamrc and then from built-in defaults. Checking for available remotes: rsync and local, git. - you won't be able to use mercurial repositories unless you install the hg command on your system. - you won't be able to use darcs repositories unless you install the darcs command on your system. This development version of opam requires an update to the layout of /home/opam/.opam from version 2.0 to version 2.2, which can't be reverted. You may want to back it up before going further. Continue? [y/n] y Format upgrade done. <><> Updating repositories ><><><><><><><><><><><><><><><><><><><><><><><><><><> [default] Initialised 2025-07-12 13:39.18 ---> using "dffd9938039301ebc6cca6d597e878cfb8a245c9dfa32f07900ab56da35e37dc" from cache /src: (run (shell "uname -rs && opam exec -- ocaml -version && opam --version")) Linux 6.8.0-60-generic The OCaml toplevel, version 5.3.0 2.3.0 2025-07-12 13:39.18 ---> using "8f9a1993785784c76c1507256d4c3a0742ea404a0d05aaa640ab910b812307e9" from cache /src: (workdir /src) /src: (run (shell "sudo chown opam /src")) 2025-07-12 13:39.18 ---> using "d0c65e3f07a69cbc2fcd3ad4787a51328d54858d7b412a5dcb90232d4889600b" from cache /src: (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "cd ~/opam-repository && (git cat-file -e 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 || git fetch origin master) && git reset -q --hard 64169c3d1d6774ef4ce0fcc471269bb09cd215f5 && git log --no-decorate -n1 --oneline && opam update -u")) From https://github.com/ocaml/opam-repository * branch master -> FETCH_HEAD 6dea2f9390..0094db1438 master -> origin/master 64169c3d1d Merge pull request #28149 from dbuenzli/b0-publish-logs.0.9.0 <><> Updating package repositories ><><><><><><><><><><><><><><><><><><><><><><> [default] synchronised from git+file:///home/opam/opam-repository Everything as up-to-date as possible (run with --verbose to show unavailable upgrades). However, you may "opam upgrade" these packages explicitly, which will ask permission to downgrade or uninstall the conflicting packages. Nothing to do. # To update the current shell environment, run: eval $(opam env) 2025-07-12 13:39.18 ---> using "3f9aba96ac7a1654110ad5283d4982861b08bd9757244130aaa4d9e51eeeba9b" from cache /src: (copy (src neural_nets_lib.opam arrayjit.opam) (dst ./)) 2025-07-12 13:39.18 ---> using "491ef324c749344438559e51fe82f034c2b4e4da04b44daf62a1a595288d145f" from cache /src: (run (network host) (shell "opam pin add -yn neural_nets_lib.dev './' && \ \nopam pin add -yn arrayjit.dev './'")) [neural_nets_lib.dev] synchronised (file:///src) neural_nets_lib is now pinned to file:///src (version dev) [arrayjit.dev] synchronised (file:///src) arrayjit is now pinned to file:///src (version dev) 2025-07-12 13:39.18 ---> using "a7d8da6b00a43533102b7f81275e435588cfa02e7849ec55e2ed6877a1c8830c" from cache /src: (run (network host) (shell "echo '(lang dune 3.0)' > './dune-project'")) 2025-07-12 13:39.18 ---> using "d6c480068350a06691a677f00217bbb777850e27f8e462e10080cd62391bb32f" from cache /src: (env DEPS "angstrom.0.16.1 astring.0.8.5 backoff.0.1.1 base.v0.17.3 base-bigarray.base base-bytes.base base-domains.base base-effects.base base-nnp.base base-threads.base base-unix.base bigarray-compat.1.1.0 bigstringaf.0.10.0 camlp-streams.5.0.1 camlzip.1.13 cmdliner.1.3.0 conf-libcurl.2 conf-libffi.2.0.0 conf-pkg-config.4 conf-zlib.1 cppo.1.8.0 csexp.1.5.2 csv.2.4 ctypes.0.23.0 ctypes-foreign.0.23.0 curl.0.10.0 dune.3.19.1 dune-configurator.3.19.1 fieldslib.v0.17.0 fmt.0.10.0 integers.0.7.0 jane-street-headers.v0.17.0 jst-config.v0.17.0 logs.0.9.0 mdx.2.5.0 mtime.2.1.0 multicore-magic.2.3.1 num.1.6 ocaml.5.3.0 ocaml-base-compiler.5.3.0 ocaml-compiler.5.3.0 ocaml-compiler-libs.v0.17.0 ocaml-config.3 ocaml-options-vanilla.1 ocaml-syntax-shims.1.0.0 ocaml-version.4.0.1 ocaml_intrinsics_kernel.v0.17.1 ocamlbuild.0.16.1 ocamlfind.1.9.8 parsexp.v0.17.0 pprint.20230830 ppx_assert.v0.17.0 ppx_base.v0.17.0 ppx_cold.v0.17.0 ppx_compare.v0.17.0 ppx_derivers.1.2.1 ppx_deriving.6.1.0 ppx_enumerate.v0.17.0 ppx_expect.v0.17.3 ppx_fields_conv.v0.17.0 ppx_globalize.v0.17.2 ppx_hash.v0.17.0 ppx_here.v0.17.0 ppx_inline_test.v0.17.1 ppx_minidebug.2.3.0 ppx_optcomp.v0.17.1 ppx_sexp_conv.v0.17.1 ppx_string.v0.17.0 ppx_variants_conv.v0.17.1 ppxlib.0.36.0 ppxlib_jane.v0.17.4 printbox.0.12 printbox-ext-plot.0.12 printbox-html.0.12 printbox-md.0.12 printbox-text.0.12 ptime.1.2.0 re.1.13.2 result.1.5 saturn_lockfree.0.5.0 seq.base sexplib.v0.17.0 sexplib0.v0.17.0 stdio.v0.17.0 stdlib-shims.0.3.0 thread-local-storage.0.2 time_now.v0.17.0 topkg.1.0.8 tyxml.4.6.0 uucp.16.0.0 uutf.1.0.4 variantslib.v0.17.0") /src: (env CI true) /src: (env OCAMLCI true) /src: (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "opam update --depexts && opam install --cli=2.3 --depext-only -y neural_nets_lib.dev arrayjit.dev $DEPS")) + /usr/bin/sudo "apk" "update" - fetch https://dl-cdn.alpinelinux.org/alpine/v3.21/main/x86_64/APKINDEX.tar.gz - fetch https://dl-cdn.alpinelinux.org/alpine/v3.21/community/x86_64/APKINDEX.tar.gz - fetch https://dl-cdn.alpinelinux.org/alpine/edge/main/x86_64/APKINDEX.tar.gz - fetch https://dl-cdn.alpinelinux.org/alpine/edge/community/x86_64/APKINDEX.tar.gz - fetch https://dl-cdn.alpinelinux.org/alpine/edge/testing/x86_64/APKINDEX.tar.gz - v3.21.3-642-ge1f66e202eb [https://dl-cdn.alpinelinux.org/alpine/v3.21/main] - v3.21.3-634-g4638c219547 [https://dl-cdn.alpinelinux.org/alpine/v3.21/community] - v3.23.0_alpha20250612-2135-g1d76ecdcac4 [https://dl-cdn.alpinelinux.org/alpine/edge/main] - v3.23.0_alpha20250612-2141-g4c66e361d86 [https://dl-cdn.alpinelinux.org/alpine/edge/community] - v3.23.0_alpha20250612-2141-g4c66e361d86 [https://dl-cdn.alpinelinux.org/alpine/edge/testing] - OK: 59057 distinct packages available <><> Synchronising pinned packages ><><><><><><><><><><><><><><><><><><><><><><> [arrayjit.dev] synchronised (file:///src) [neural_nets_lib.dev] synchronised (file:///src) [NOTE] Package ocaml-options-vanilla is already installed (current version is 1). [NOTE] Package ocaml-config is already installed (current version is 3). [NOTE] Package ocaml-compiler is already installed (current version is 5.3.0). [NOTE] Package ocaml-base-compiler is already installed (current version is 5.3.0). [NOTE] Package ocaml is already installed (current version is 5.3.0). [NOTE] Package base-unix is already installed (current version is base). [NOTE] Package base-threads is already installed (current version is base). [NOTE] Package base-nnp is already installed (current version is base). [NOTE] Package base-effects is already installed (current version is base). [NOTE] Package base-domains is already installed (current version is base). [NOTE] Package base-bigarray is already installed (current version is base). The following system packages will first need to be installed: curl-dev libffi-dev zlib-dev <><> Handling external dependencies <><><><><><><><><><><><><><><><><><><><><><> + /usr/bin/sudo "apk" "add" "curl-dev" "libffi-dev" "zlib-dev" - (1/15) Upgrading libcrypto3 (3.3.3-r0 -> 3.3.4-r0) - (2/15) Upgrading libssl3 (3.3.3-r0 -> 3.3.4-r0) - (3/15) Installing brotli (1.1.0-r2) - (4/15) Installing brotli-dev (1.1.0-r2) - (5/15) Installing c-ares-dev (1.34.5-r0) - (6/15) Installing libidn2-dev (2.3.7-r0) - (7/15) Installing libpsl-utils (0.21.5-r3) - (8/15) Installing libpsl-dev (0.21.5-r3) - (9/15) Installing nghttp2-dev (1.64.0-r0) - (10/15) Installing openssl-dev (3.3.4-r0) - (11/15) Installing zlib-dev (1.3.1-r2) - (12/15) Installing zstd-dev (1.5.6-r2) - (13/15) Installing curl-dev (8.12.1-r1) - (14/15) Installing linux-headers (6.6-r1) - (15/15) Installing libffi-dev (3.4.7-r0) - Executing busybox-1.37.0-r12.trigger - Executing ca-certificates-20241121-r1.trigger - OK: 317 MiB in 115 packages 2025-07-12 13:39.18 ---> using "ccc66300745983069d411af699905a3e59bf29147f07943986ec7e967f60f585" from cache /src: (run (cache (opam-archives (target /home/opam/.opam/download-cache))) (network host) (shell "opam install $DEPS")) [NOTE] Package ocaml-options-vanilla is already installed (current version is 1). [NOTE] Package ocaml-config is already installed (current version is 3). [NOTE] Package ocaml-compiler is already installed (current version is 5.3.0). [NOTE] Package ocaml-base-compiler is already installed (current version is 5.3.0). [NOTE] Package ocaml is already installed (current version is 5.3.0). [NOTE] Package base-unix is already installed (current version is base). [NOTE] Package base-threads is already installed (current version is base). [NOTE] Package base-nnp is already installed (current version is base). [NOTE] Package base-effects is already installed (current version is base). [NOTE] Package base-domains is already installed (current version is base). [NOTE] Package base-bigarray is already installed (current version is base). The following actions will be performed: === install 81 packages - install angstrom 0.16.1 - install astring 0.8.5 - install backoff 0.1.1 - install base v0.17.3 - install base-bytes base - install bigarray-compat 1.1.0 - install bigstringaf 0.10.0 - install camlp-streams 5.0.1 - install camlzip 1.13 - install cmdliner 1.3.0 - install conf-libcurl 2 - install conf-libffi 2.0.0 - install conf-pkg-config 4 - install conf-zlib 1 - install cppo 1.8.0 - install csexp 1.5.2 - install csv 2.4 - install ctypes 0.23.0 - install ctypes-foreign 0.23.0 - install curl 0.10.0 - install dune 3.19.1 - install dune-configurator 3.19.1 - install fieldslib v0.17.0 - install fmt 0.10.0 - install integers 0.7.0 - install jane-street-headers v0.17.0 - install jst-config v0.17.0 - install logs 0.9.0 - install mdx 2.5.0 - install mtime 2.1.0 - install multicore-magic 2.3.1 - install num 1.6 - install ocaml-compiler-libs v0.17.0 - install ocaml-syntax-shims 1.0.0 - install ocaml-version 4.0.1 - install ocaml_intrinsics_kernel v0.17.1 - install ocamlbuild 0.16.1 - install ocamlfind 1.9.8 - install parsexp v0.17.0 - install pprint 20230830 - install ppx_assert v0.17.0 - install ppx_base v0.17.0 - install ppx_cold v0.17.0 - install ppx_compare v0.17.0 - install ppx_derivers 1.2.1 - install ppx_deriving 6.1.0 - install ppx_enumerate v0.17.0 - install ppx_expect v0.17.3 - install ppx_fields_conv v0.17.0 - install ppx_globalize v0.17.2 - install ppx_hash v0.17.0 - install ppx_here v0.17.0 - install ppx_inline_test v0.17.1 - install ppx_minidebug 2.3.0 - install ppx_optcomp v0.17.1 - install ppx_sexp_conv v0.17.1 - install ppx_string v0.17.0 - install ppx_variants_conv v0.17.1 - install ppxlib 0.36.0 - install ppxlib_jane v0.17.4 - install printbox 0.12 - install printbox-ext-plot 0.12 - install printbox-html 0.12 - install printbox-md 0.12 - install printbox-text 0.12 - install ptime 1.2.0 - install re 1.13.2 - install result 1.5 - install saturn_lockfree 0.5.0 - install seq base - install sexplib v0.17.0 - install sexplib0 v0.17.0 - install stdio v0.17.0 - install stdlib-shims 0.3.0 - install thread-local-storage 0.2 - install time_now v0.17.0 - install topkg 1.0.8 - install tyxml 4.6.0 - install uucp 16.0.0 - install uutf 1.0.4 - install variantslib v0.17.0 <><> Processing actions <><><><><><><><><><><><><><><><><><><><><><><><><><><><> -> retrieved backoff.0.1.1 (cached) -> retrieved astring.0.8.5 (cached) -> retrieved angstrom.0.16.1 (cached) -> retrieved base.v0.17.3 (cached) -> retrieved bigarray-compat.1.1.0 (cached) -> retrieved bigstringaf.0.10.0 (cached) -> retrieved camlp-streams.5.0.1 (cached) -> retrieved camlzip.1.13 (cached) -> retrieved cmdliner.1.3.0 (cached) -> retrieved cppo.1.8.0 (cached) -> installed conf-libcurl.2 -> installed conf-pkg-config.4 -> retrieved csexp.1.5.2 (cached) -> retrieved csv.2.4 (cached) -> retrieved ctypes.0.23.0, ctypes-foreign.0.23.0 (cached) -> installed conf-libffi.2.0.0 -> installed conf-zlib.1 -> retrieved curl.0.10.0 (cached) -> retrieved fieldslib.v0.17.0 (cached) -> retrieved integers.0.7.0 (cached) -> retrieved fmt.0.10.0 (cached) -> retrieved jane-street-headers.v0.17.0 (cached) -> retrieved jst-config.v0.17.0 (cached) -> retrieved logs.0.9.0 (cached) -> retrieved mtime.2.1.0 (cached) -> retrieved multicore-magic.2.3.1 (cached) -> retrieved mdx.2.5.0 (cached) -> retrieved num.1.6 (cached) -> retrieved ocaml-compiler-libs.v0.17.0 (cached) -> retrieved ocaml-syntax-shims.1.0.0 (cached) -> retrieved ocaml-version.4.0.1 (cached) -> retrieved ocaml_intrinsics_kernel.v0.17.1 (cached) -> retrieved ocamlbuild.0.16.1 (cached) -> retrieved ocamlfind.1.9.8 (cached) -> retrieved parsexp.v0.17.0 (cached) -> retrieved pprint.20230830 (cached) -> retrieved ppx_assert.v0.17.0 (cached) -> retrieved ppx_base.v0.17.0 (cached) -> retrieved ppx_cold.v0.17.0 (cached) -> retrieved ppx_compare.v0.17.0 (cached) -> retrieved ppx_derivers.1.2.1 (cached) -> retrieved ppx_enumerate.v0.17.0 (cached) -> retrieved ppx_deriving.6.1.0 (cached) -> retrieved ppx_expect.v0.17.3 (cached) -> retrieved dune.3.19.1, dune-configurator.3.19.1 (cached) -> retrieved ppx_fields_conv.v0.17.0 (cached) -> retrieved ppx_globalize.v0.17.2 (cached) -> retrieved ppx_hash.v0.17.0 (cached) -> retrieved ppx_here.v0.17.0 (cached) -> retrieved ppx_inline_test.v0.17.1 (cached) -> retrieved ppx_optcomp.v0.17.1 (cached) -> retrieved ppx_sexp_conv.v0.17.1 (cached) -> retrieved ppx_string.v0.17.0 (cached) -> retrieved ppx_variants_conv.v0.17.1 (cached) -> installed cmdliner.1.3.0 -> installed num.1.6 -> installed ocamlbuild.0.16.1 -> retrieved ppx_minidebug.2.3.0 (cached) -> retrieved ppxlib.0.36.0 (cached) -> retrieved ppxlib_jane.v0.17.4 (cached) -> retrieved ptime.1.2.0 (cached) -> retrieved re.1.13.2 (cached) -> retrieved result.1.5 (cached) -> retrieved seq.base (cached) -> installed seq.base -> retrieved sexplib.v0.17.0 (cached) -> retrieved saturn_lockfree.0.5.0 (cached) -> retrieved sexplib0.v0.17.0 (cached) -> retrieved stdio.v0.17.0 (cached) -> retrieved stdlib-shims.0.3.0 (cached) -> retrieved thread-local-storage.0.2 (cached) -> retrieved time_now.v0.17.0 (cached) -> retrieved printbox.0.12, printbox-ext-plot.0.12, printbox-html.0.12, printbox-md.0.12, printbox-text.0.12 (cached) -> retrieved topkg.1.0.8 (cached) -> retrieved tyxml.4.6.0 (cached) -> retrieved uutf.1.0.4 (cached) -> retrieved variantslib.v0.17.0 (cached) -> retrieved uucp.16.0.0 (cached) -> installed ocamlfind.1.9.8 -> installed base-bytes.base -> installed camlzip.1.13 -> installed topkg.1.0.8 -> installed uutf.1.0.4 -> installed mtime.2.1.0 -> installed fmt.0.10.0 -> installed ptime.1.2.0 -> installed astring.0.8.5 -> installed logs.0.9.0 -> installed dune.3.19.1 -> installed jane-street-headers.v0.17.0 -> installed ppx_derivers.1.2.1 -> installed csexp.1.5.2 -> installed backoff.0.1.1 -> installed bigarray-compat.1.1.0 -> installed camlp-streams.5.0.1 -> installed cppo.1.8.0 -> installed csv.2.4 -> installed multicore-magic.2.3.1 -> installed ocaml-compiler-libs.v0.17.0 -> installed ocaml-syntax-shims.1.0.0 -> installed ocaml-version.4.0.1 -> installed ocaml_intrinsics_kernel.v0.17.1 -> installed pprint.20230830 -> installed printbox.0.12 -> installed re.1.13.2 -> installed result.1.5 -> installed sexplib0.v0.17.0 -> installed stdlib-shims.0.3.0 -> installed thread-local-storage.0.2 -> installed saturn_lockfree.0.5.0 -> installed integers.0.7.0 -> installed parsexp.v0.17.0 -> installed dune-configurator.3.19.1 -> installed bigstringaf.0.10.0 -> installed sexplib.v0.17.0 -> installed mdx.2.5.0 -> installed tyxml.4.6.0 -> installed angstrom.0.16.1 -> installed printbox-html.0.12 -> installed curl.0.10.0 -> installed uucp.16.0.0 -> installed printbox-text.0.12 -> installed printbox-md.0.12 -> installed ctypes.0.23.0 -> installed printbox-ext-plot.0.12 -> installed base.v0.17.3 -> installed fieldslib.v0.17.0 -> installed variantslib.v0.17.0 -> installed stdio.v0.17.0 -> installed ctypes-foreign.0.23.0 -> installed ppxlib.0.36.0 -> installed ppx_optcomp.v0.17.1 -> installed ppxlib_jane.v0.17.4 -> installed ppx_cold.v0.17.0 -> installed ppx_here.v0.17.0 -> installed ppx_variants_conv.v0.17.1 -> installed ppx_fields_conv.v0.17.0 -> installed ppx_enumerate.v0.17.0 -> installed ppx_globalize.v0.17.2 -> installed ppx_compare.v0.17.0 -> installed ppx_deriving.6.1.0 -> installed ppx_sexp_conv.v0.17.1 -> installed ppx_hash.v0.17.0 -> installed ppx_assert.v0.17.0 -> installed ppx_minidebug.2.3.0 -> installed ppx_base.v0.17.0 -> installed jst-config.v0.17.0 -> installed ppx_string.v0.17.0 -> installed time_now.v0.17.0 -> installed ppx_inline_test.v0.17.1 -> installed ppx_expect.v0.17.3 Done. # To update the current shell environment, run: eval $(opam env) 2025-07-12 13:39.18 ---> using "e87220be021c1f237769871505436dbde675ba40bfe94f1a0edef90ab08915c3" from cache /src: (copy (src .) (dst /src)) 2025-07-12 13:39.18 ---> saved as "435558e2d414bbc90f52090ecfb8a0090c616abc7245b7b4498a289065db0f10" /src: (run (shell "opam exec -- dune build @install @check @runtest && rm -rf _build")) File "arrayjit/test/dune", line 6, characters 7-15: 6 | (pps ppx_jane))) ^^^^^^^^ Error: Library "ppx_jane" not found. -> required by _build/default/arrayjit/test/.merlin-conf/exe-test_numerical_types -> required by alias arrayjit/test/check (cd _build/default/test/config && ../../arrayjit/bin/read_config.exe --read=backend) Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/config/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file Wrote value of 'backend' to ocannl_backend.txt File "test_ppx/test_ppx_op_expected.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test_ppx/test_ppx_op_expected.ml _build/default/test_ppx/test_ppx_op_actual.ml diff --git a/_build/default/test_ppx/test_ppx_op_expected.ml b/_build/default/test_ppx/test_ppx_op_actual.ml index a6aa4ae..c971ef0 100644 --- a/_build/default/test_ppx/test_ppx_op_expected.ml +++ b/_build/default/test_ppx/test_ppx_op_actual.ml @@ -2,20 +2,20 @@ open Base open Ocannl module TDSL = Operation.TDSL let y0 = - let hey1 = TDSL.param ?values:None "hey1" in + let hey1 = TDSL.param ?value:None "hey1" in let open! TDSL.O in ((+) ?label:(Some ["y0"])) ((( *. ) ?label:None) (TDSL.number (Float.of_int 2)) hey1) (TDSL.number (Float.of_int 3)) let y1 = - let hey2 = TDSL.param ?values:None "hey2" in + let hey2 = TDSL.param ?value:None "hey2" in let open! TDSL.O in fun x -> ((+) ?label:(Some (List.concat [["y1"]; (x.Tensor.value).Ir.Tnode.label]))) ((( * ) ?label:None) hey2 (TDSL.number (Float.of_int 2))) x let y2 = - let hey3 = TDSL.param ?values:None "hey3" in + let hey3 = TDSL.param ?value:None "hey3" in let open! TDSL.O in fun x1 x2 -> ((+) ?label:(Some @@ -36,14 +36,14 @@ let b = ~output_dims:[2] [|(Float.of_int 7);(Float.of_int 8);(Float.of_int 9);(Float.of_int 10)|] let y = - let hey4 = TDSL.param ?values:None "hey4" in + let hey4 = TDSL.param ?value:None "hey4" in let open! TDSL.O in ((+) ?label:(Some ["y"])) ((( * ) ?label:None) hey4 (TDSL.number ?label:None ~axis_label:"q" 2.0)) (TDSL.number ?label:None ~axis_label:"p" 1.0) let z = - let hey5 = TDSL.param ?values:None "hey5" - and hey6 = TDSL.param ?values:None "hey6" in + let hey5 = TDSL.param ?value:None "hey5" + and hey6 = TDSL.param ?value:None "hey6" in let open! TDSL.O in ((+) ?label:(Some ["z"])) ((( * ) ?label:None) (TDSL.number ?label:None ~axis_label:"q" 2.0) hey5) @@ -51,8 +51,8 @@ let z = let stride = 2 and dilation = 3 let z2 = - let hey7 = TDSL.param ?values:None "hey7" - and hey8 = TDSL.param ?values:None "hey8" in + let hey7 = TDSL.param ?value:None "hey7" + and hey8 = TDSL.param ?value:None "hey8" in let open! TDSL.O in TDSL.einsum ?label:(Some ["z2"]) (String.concat ~sep:"" @@ -61,8 +61,8 @@ let z2 = let z3 = let s = 2 and d = 3 in - let hey10 = TDSL.param ?values:None "hey10" - and hey9 = TDSL.param ?values:None "hey9" in + let hey10 = TDSL.param ?value:None "hey10" + and hey9 = TDSL.param ?value:None "hey9" in let open! TDSL.O in TDSL.einsum ?label:(Some []) (String.concat ~sep:"" @@ -78,7 +78,7 @@ let mlp_layer = let b = (TDSL.param ~more_label:(config.label)) ~output_dims:[config.hid_dim] "b" - and w = (TDSL.param ~more_label:(config.label)) ?values:None "w" in + and w = (TDSL.param ~more_label:(config.label)) ?value:None "w" in fun x -> (relu ?label:(Some File "test/test_print_style.expected", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/test_print_style.expected _build/default/test/test_print_style.exe.output diff --git a/_build/default/test/test_print_style.expected b/_build/default/test/test_print_style.exe.output index e51ab37..c55c8ad 100644 --- a/_build/default/test/test_print_style.expected +++ b/_build/default/test/test_print_style.exe.output @@ -1,5 +1,5 @@ -Welcome to OCANNL! Reading configuration defaults from /Users/lukstafi/ocannl/_build/default/test/ocannl_config. +Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file Testing print_style functionality: File "test/einsum/test_conv_syntax.expected", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/einsum/test_conv_syntax.expected _build/default/test/einsum/test_conv_syntax.exe.output diff --git a/_build/default/test/einsum/test_conv_syntax.expected b/_build/default/test/einsum/test_conv_syntax.exe.output index 7098c49..02f1b08 100644 --- a/_build/default/test/einsum/test_conv_syntax.expected +++ b/_build/default/test/einsum/test_conv_syntax.exe.output @@ -1,5 +1,5 @@ -Welcome to OCANNL! Reading configuration defaults from /Users/lukstafi/ocannl/_build/default/test/einsum/ocannl_config. +Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/einsum/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file Testing conv syntax parsing... (cd _build/default/test_ppx && ./test_ppx_op.exe) Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test_ppx/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/default/test_ppx && ./test_ppx_op_expected.exe) Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test_ppx/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition '' -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition 'Found 0, in the config file' -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition 'Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/0e16ff772f933dca10af1f602582b417/default/test/ocannl_config.' -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition 'Retrieving commandline, environment, or config file variable ocannl_log_level' -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file File "test/dune", lines 30-40, characters 0-281: 30 | (rule 31 | (alias runtest) 32 | (target 33 | (dir log_files)) 34 | (action 35 | (run 36 | %{dep:micrograd_demo_logging.exe} 37 | "--ocannl_debug_backend=text" 38 | "--ocannl_log_file_stem=micrograd_demo_logging" 39 | "--ocannl_log_main_domain_to_stdout=false" 40 | "--ocannl_debug_log_to_stream_files=true"))) (cd _build/default/test && ./micrograd_demo_logging.exe --ocannl_debug_backend=text --ocannl_log_file_stem=micrograd_demo_logging --ocannl_log_main_domain_to_stdout=false --ocannl_debug_log_to_stream_files=true) Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file Retrieving commandline, environment, or config file variable ocannl_backend Found multicore_cc, in the config file Retrieving commandline, environment, or config file variable ocannl_limit_constant_fill_size Not found, using default 256 Retrieving commandline, environment, or config file variable ocannl_cd_ident_style Not found, using default heuristic Retrieving commandline, environment, or config file variable ocannl_ll_ident_style Not found, using default heuristic Retrieving commandline, environment, or config file variable ocannl_prefer_backend_uniformity Found true, in the config file Retrieving commandline, environment, or config file variable ocannl_debug_log_to_stream_files Found true, commandline --ocannl_debug_log_to_stream_files=true Retrieving commandline, environment, or config file variable ocannl_cc_backend_optimization_level Not found, using default 3 Retrieving commandline, environment, or config file variable ocannl_cc_backend_compiler_command Not found, using default gcc Retrieving commandline, environment, or config file variable ocannl_cc_backend_verify_codesign Not found, using default false Fatal error: exception File "src/printbox-text/PrintBox_text.ml", line 212, characters 6-12: Assertion failed Raised at PrintBox_text.Output.Make_out.to_buf_aux_ in file "src/printbox-text/PrintBox_text.ml", line 212, characters 6-50 Called from Stdlib__Map.Make.fold in file "map.ml", line 329, characters 19-42 Called from Stdlib__Map.Make.fold in file "map.ml", line 329, characters 26-41 Called from Stdlib__Map.Make.fold in file "map.ml", line 329, characters 26-41 Called from Stdlib__Map.Make.fold in file "map.ml", line 329, characters 26-41 Called from PrintBox_text.Output.Make_out.render in file "src/printbox-text/PrintBox_text.ml", line 242, characters 14-64 Called from PrintBox_text.output in file "src/printbox-text/PrintBox_text.ml", line 851, characters 2-31 Called from Minidebug_runtime.PrintBox.output_box in file "minidebug_runtime.ml", line 1527, characters 19-59 Called from Minidebug_runtime.PrintBox.close_log_impl.close_tree in file "minidebug_runtime.ml", line 1572, characters 6-38 Called from Backends.Add_buffer_retrieval_and_syncing.sync_routine in file "arrayjit/lib/backends.ml", lines 207-235, characters 31-82 Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 539-540, characters 4-92 Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 Called from Dune__exe__Micrograd_demo_logging in file "test/micrograd_demo_logging.ml", line 36, characters 13-76 (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition zero2hero_1of7.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition primitive_ops.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition micrograd_demo.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition hello_world_op.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file File "test/einsum/einsum_trivia_exec.expected", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/einsum/einsum_trivia_exec.expected _build/default/test/einsum/einsum_trivia_exec.exe.output diff --git a/_build/default/test/einsum/einsum_trivia_exec.expected b/_build/default/test/einsum/einsum_trivia_exec.exe.output index ab456b5..74ceb2e 100644 --- a/_build/default/test/einsum/einsum_trivia_exec.expected +++ b/_build/default/test/einsum/einsum_trivia_exec.exe.output @@ -1,5 +1,5 @@ -Welcome to OCANNL! Reading configuration defaults from /Users/lukstafi/ocannl/_build/default/test/einsum/ocannl_config. +Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/einsum/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file HERE: test/einsum/einsum_trivia_exec.ml:31:21 (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition einsum_trivia.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file File "test/einsum/moons_demo_variant.expected", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/einsum/moons_demo_variant.expected _build/default/test/einsum/moons_demo_variant.exe.output diff --git a/_build/default/test/einsum/moons_demo_variant.expected b/_build/default/test/einsum/moons_demo_variant.exe.output index 68171df..84e7a1e 100644 --- a/_build/default/test/einsum/moons_demo_variant.expected +++ b/_build/default/test/einsum/moons_demo_variant.exe.output @@ -1,121 +1,59 @@ -Welcome to OCANNL! Reading configuration defaults from /Users/lukstafi/ocannl/_build/default/test/einsum/ocannl_config. +Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/einsum/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file Half-moons scatterplot and decision boundary: ┌────────────────────────────────────────────────────────────────────────────────────────────────────┐ -│****************************#***********************************************************************│ -│****************************###***###***#***********************************************************│ -│***************************#********#***#**********************************************************.│ -│********************#*********************##*#*#************************************************....│ -│******************#***#*****##*#*#*#****#*#*#*##**********************************************......│ -│**************#*#**###**#*##**##********#**#****###**#*************************************.........│ -│***********#***##***#**##*****************#*#*##*###*************************************...........│ -│***************#**###******************************##*#********************************.............│ -│*************#*****#******************************#*#***#*#*************************................│ -│*********#***#************************************##*###**#***********************..................│ -│******#***#*#*#******************************************#*#*******************.....................│ -│*****###****#*******************************************#********************.......................│ -│*******#*###***********************************************#*#************..........................│ -│**********##**********************%%*********************#**##**********............................│ -│*****#***#**********************%**%%********************************.............................%.│ -│**####*************************%**%**************************##*#**............................%...%│ -│****#**************************%%**%*************************#*#*..............................%....│ -│*#****##************************%*...%*********************##*..#...........................%...%%..│ -│****#*#***********************.%..%.%........................#..#.............................%.%...│ -│##*##*********************.......%.............................#.##.........................%.%.%%%.│ -│**#####*****************.......%...%%.......................................................%..%.%..│ -│**********************...........%.%.%%.....................#.#.#.#.............................%%..│ -│***##****************............%..%..%........................#.#.........................%...%...│ -│##**#**************................%...%.....................#..###......................%.%%.......│ -│###***************...............%.%.%...........................##............................%....│ -│**##************.................................................#......................%%.%..%.....│ -│**#************....................%%...........................##......................%%.%........│ -│**************........................%......................#..#...........................%.......│ -│************........................%%.%.......................##.....................%..%..........│ -│***********............................%%..............................................%%.%%........│ -│**********.............................%...%............................................%%%.........│ -│********...................................%..%....................................%.....%..........│ -│*******.................................%%..%.%................................%.%.%.%%.............│ -│******.....................................%%.%...%%.%.......................%%%....%.%.%...........│ -│*****....................................%.....%.%..%....................%.....%%.%.%...............│ -│***.........................................%%...%.%%%%..%.......%.........%%.%%....................│ -│**................................................%..%.......%..%%%%...%.%..%.%.....................│ -│*..................................................%.%.%............%..%...%........................│ -│..................................................%....%.%......%%...%.%............................│ -│..........................................................%..%......%%%.............................│ +│...................................#................................................................│ +│..........................##..#..###....##..........................................................│ +│.........................#....#.....#...............................................................│ +│.......................#.##.#.......#####...#.#.....................................................│ +│..................#.#.#...#..#.##....#.#.##..#....#.................................................│ +│.............#.#......##.#..#..........###..#....#..................................................│ +│.............##..##.....#....................#.##.###...#...........................................│ +│.............#.##.#.#.#.......................#...##..#.............................................│ +│.........#.#....##..................................##..#...........................................│ +│...........#.#....................................#..##...#.#.......................................│ +│..........##..........................................#...#.#.......................................│ +│.......#...###............................................#.........................................│ +│....#....##..........................%..................#....##.....................................│ +│.........##................................................#..#..............................%......│ +│.....#..#........................%..%......................##.#...................................%.│ +│...##.#..................................................#....................................%...%.│ +│..###...........................%..%%......................###................................%.%...│ +│..#...#.#.......................%%.%%.......................#.##.#............................%.....│ +│..#...#.............................%.........................##..............................%...%.│ +│..##..#............................%%.........................#..##.........................%%..%.%%│ +│#...#..#.........................%.%%%.......................##...#............................%.%..│ +│##...#.............................%%.........................#.#..#................................│ +│...................................%%%%.......................##.#.#........................%%..%.%.│ +│###................................%%..............................#.......................%....%...│ +│..##................................%%%.%....................................................%%..%..│ +│.#.....................................%.......................#.###......................%.%%.%....│ +│.....#................................%...%................................................%...%....│ +│...................................%.....%%......................#..................................│ +│.....................................%%%..%...........................................%..%%..%......│ +│.......................................%%.....%......................................%..%.%.%.......│ +│.........................................%%............................................%..%%.%......│ +│..........................................%.%.....%...............................%.%.%%..%.........│ +│.........................................%%%..%%......................................%.............│ +│...........................................%.%..%..............................%%.%.%.%%.%..........│ +│..............................................%%%%..%.%.......................%%..%.%.%.............│ +│..............................................%..%%%.%%...................%..%%...%.................│ +│...............................................%...%.%.%..%...%.%%..%%.%%...%.......................│ +│......................................................%%.%%%..%.%..%%.%.%.%%........................│ +│.....................................................%...........%%%%......%%.......................│ +│.......................................................%.%%%%.%.......%...%.........................│ └────────────────────────────────────────────────────────────────────────────────────────────────────┘Loss: -┌────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ -│ 3.79e+1│- │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│l │ │ -│o │ │ -│s │ │ -│s │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ - │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ │ -│ │ - - - - │ -│ 0.00 │ - ---- --- --- ---- ---- ---- ---- -- - ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- │ -├────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ -│ │0.00 7.90e+1│ -│ │ step │ -└────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘Log-loss, for better visibility: ┌─────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ -│ 3.63 │- │ +│ 4.11e+31│ - │ │ │ │ │ │ │ │ │ │ -│ │ - │ │ │ │ │ │ │ │ │ │ │ │ │ -│ │ - │ -│ │ - │ -│ │ - - - │ -│ │ - - - - │ -│ │ -- - - - - - - - -- - - - - - -- │ -│ │ - - - - - - - -- - - -- - - - - │ -│ │ - - - │ -│l │ - - - -- - │ -│o │ - - │ -│g │ - - - - │ -│ │ - - - │ -│l │ - -- │ -│o │ - - - │ -│s │ - │ -│s │ - │ -│ │ - │ │ │ │ │ │ │ │ │ │ @@ -126,11 +64,73 @@ Half-moons scatterplot and decision boundary: │ │ │ │ │ │ │ │ │ +│l │ │ +│o │ │ +│s │ │ +│s │ │ │ │ │ │ │ │ │ │ │ │ │ │ -│ -1.00e+1│ - - - - - - - │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ 0.00 │---- ---- ---- ---- ---- --- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- ---- │ +├─────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ +│ │0.00 7.90e+1│ +│ │ step │ +└─────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘Log-loss, for better visibility: +┌─────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ +│ 7.27e+1 │ - │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│l │ │ +│o │ │ +│g │ │ +│ │ │ +│l │ │ +│o │ │ +│s │ │ +│s │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ │ +│ │ -- --- │ +│ │- --- -- -- ---- ---- -- - ---- ---- ---- - │ +│ │ - - -- -- │ +│ │ - │ +│ │ │ +│ -1.00e+1│ --- ---- ---- ---- --- ---- ---- ---- ---- - │ ├─────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ │ │0.00 7.90e+1│ │ │ step │ (cd _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test && .tutorials.inline-tests/inline-test-runner.exe inline-test-runner tutorials -partition moons_demo_parallel.ml -source-tree-root .. -diff-cmd -) Welcome to OCANNL! Reading configuration defaults from /src/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file File "test/zero2hero_1of7.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/zero2hero_1of7.ml _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/zero2hero_1of7.ml.corrected diff --git a/_build/default/test/zero2hero_1of7.ml b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/zero2hero_1of7.ml.corrected index 3b7cc00..394b619 100644 --- a/_build/default/test/zero2hero_1of7.ml +++ b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/zero2hero_1of7.ml.corrected @@ -30,24 +30,24 @@ let%expect_test "Graph drawing recompile" = Tensor.print_tree ~with_grad:true ~depth:9 f_nd; [%expect {| - #15 +_f_nd - 6.00e+1 - #16 grad_+_f_nd Virt/30 - <void> - #13 - Virt/152 │#2 5. Virt/40 - <void> │<void> - #14 grad_- Virt/30 │ - <void> │ - #11 *. Virt/152 │ #4 *. Virt/152 │ - <void> │ <void> │ - #12 grad_*. Virt/30 │ #5 grad_*. Virt/30 │ - <void> │ <void> │ - #10 3. Virt/40│#7 **. Virt/152 │#3 4. Virt/40│#0 x │ - <void> │<void> │<void> │ 5.00 │ - │#8 grad_**. Virt/30│ │#1 grad_x Local/30│ - │<void> │ │<void> │ - │[0]│ #6 2. Virt/40 │ │ │ - │ │ <void> │ │ │ + #15 +_f_nd + 5.00 + #16 grad_+_f_nd Virt/30 + <void> + #13 - Virt/152 │#2 5. Virt/40 + <void> │<void> + #14 grad_- Virt/30 │ + <void> │ + #11 *. Virt/152 │#4 *. Virt/152 │ + <void> │<void> │ + #12 grad_*. Virt/30 │#5 grad_*. Virt/30│ + <void> │<void> │ + #10 3. Virt/40│#7 **. Virt/152 │#3 4. Virt/40│[0] │ + <void> │<void> │<void> │ │ + │#8 grad_**. Virt/30│ │ │ + │<void> │ │ │ + │[0]│ #6 2. Virt/40 │ │ │ + │ │ <void> │ │ │ |}]; let%op f = (3 *. ("x" [ 5 ] **. 2)) - (4 *. x) + 5 in Train.every_non_literal_on_host f; @@ -55,27 +55,7 @@ let%expect_test "Graph drawing recompile" = let f_bprop = Train.to_routine (module Backend) ctx IDX.empty f_upd in Train.run f_bprop; Tensor.print_tree ~with_grad:true ~depth:9 f; - [%expect - {| - #32 +_f - 6.00e+1 - #33 grad_+_f - 1.00 - #30 - │#19 5. Virt/40 - 5.50e+1 │<void> - #31 grad_- │ - 1.00 │ - #28 *. │ #21 *. │ - 7.50e+1 │ 2.00e+1 │ - #29 grad_*. │ #22 grad_*. │ - 1.00 │ -1.00 │ - #27 3. Virt/40│ #24 **. │#20 4. Virt/40│#17 x │ - <void> │ 2.50e+1 │<void> │ 5.00 │ - │ #25 grad_**. │ │#18 grad_x│ - │ 3.00 │ │ 2.60e+1 │ - │[17]│#23 2. Virt/40│ │ │ - │ │<void> │ │ │ - |}]; + [%expect.unreachable]; let xs = Array.init 10 ~f:Float.(fun i -> of_int i - 5.) in let ys = Array.map xs ~f:(fun v -> @@ -94,54 +74,21 @@ let%expect_test "Graph drawing recompile" = [ Scatterplot { points = Array.zip_exn xs ys; content = PrintBox.line "#" } ] in PrintBox_text.output Stdio.stdout plot_box; - [%expect - {| - ┌────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │ 1.00e+2│# │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ # │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │f │ │ - │( │ │ - │x │ │ - │) │ │ - │ │ │ - │ │ # │ - │ │ │ - │ │ │ - │ │ #│ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ # │ - │ │ │ - │ │ # │ - │ │ │ - │ │ │ - │ │ # │ - │ │ # │ - │ │ │ - │ 4.00 │ # # │ - ├────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ │-5.00 4.00│ - │ │ x │ - └────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node x\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Zero2hero_1of7.(fun) in file "test/zero2hero_1of7.ml", line 55, characters 16-69 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "Graph drawing fetch" = Tensor.unsafe_reinitialize (); @@ -271,96 +218,39 @@ let%expect_test "Simple gradients hosted" = let sgd_routine = Train.to_routine (module Backend) grad_routine.context IDX.empty sgd in (* Check out the initial state without running a forward pass. *) Tensor.print_tree ~spy:true ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l Host&stream/41 - <not-in-yet> - #13 grad_*._l Host&stream/41 - <not-in-yet> - #8 +_d Host&stream/41 │#10 f Host&shared/39 - <not-in-yet> │<not-in-yet> - #9 grad_+_d Host&stream/41 │#11 grad_f Host&stream/41 - <not-in-yet> │<not-in-yet> - #4 *._e Host&stream/41 │#6 c Host&shared/39 │ - <not-in-yet> │<not-in-yet> │ - #5 grad_*._e Host&stream/41 │#7 grad_c Host&stream/41│ - <not-in-yet> │<not-in-yet> │ - #0 a Host&shared/39 │#2 b Host&shared/39 │ │ - <not-in-yet> │<not-in-yet> │ │ - #1 grad_a Host&stream/41│#3 grad_b Host&stream/41│ │ - <not-in-yet> │<not-in-yet> │ │ - |}]; + [%expect.unreachable]; (* Do not update the params: all values and gradients will be at initial points, which are specified in the tensor in the brackets. *) Train.run grad_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -8.00 - #13 grad_*._l - 1.00 - #8 +_d │#10 f - 4.00 │ -2.00 - #9 grad_+_d │#11 grad_f - -2.00 │ 4.00 - #4 *._e │#6 c │ - -6.00 │ 1.00e+1 │ - #5 grad_*._e │#7 grad_c│ - -2.00 │ -2.00 │ - #0 a │#2 b │ │ - 2.00 │ -3.00 │ │ - #1 grad_a│#3 grad_b│ │ - 6.00 │ -4.00 │ │ - |}]; + [%expect.unreachable]; (* Now we update the params, but we are not doing the forward and backward passes: only params values will change, compared to the above. The update is in the opposite direction of the gradient. *) Train.run sgd_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -8.00 - #13 grad_*._l - 1.00 - #8 +_d │#10 f - 4.00 │ -2.40 - #9 grad_+_d │#11 grad_f - -2.00 │ 4.00 - #4 *._e │#6 c │ - -6.00 │ 1.02e+1 │ - #5 grad_*._e │#7 grad_c│ - -2.00 │ -2.00 │ - #0 a │#2 b │ │ - 1.40 │ -2.60 │ │ - #1 grad_a│#3 grad_b│ │ - 6.00 │ -4.00 │ │ - |}]; + [%expect.unreachable]; (* Now the params will remain as above, but both param gradients and the values and gradients of other nodes will change thanks to the forward and backward passes. *) Train.run grad_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -1.57e+1 - #13 grad_*._l - 1.00 - #8 +_d │#10 f - 6.56 │ -2.40 - #9 grad_+_d │#11 grad_f - -2.40 │ 6.56 - #4 *._e │#6 c │ - -3.64 │ 1.02e+1 │ - #5 grad_*._e │#7 grad_c│ - -2.40 │ -2.40 │ - #0 a │#2 b │ │ - 1.40 │ -2.60 │ │ - #1 grad_a│#3 grad_b│ │ - 6.24 │ -3.36 │ │ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node a\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Zero2hero_1of7.(fun) in file "test/zero2hero_1of7.ml", line 270, characters 21-73 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "Simple gradients virtual" = Tensor.unsafe_reinitialize (); @@ -381,68 +271,29 @@ let%expect_test "Simple gradients virtual" = Tensor.print_tree ~spy:true ~with_grad:true ~depth:9 l; [%expect {| - #12 *._l Host&dev/41 - <not-in-yet> - #13 grad_*._l unknown - <not-in-yet> - #8 +_d unknown │#10 f Host-non-const/24 - <not-in-yet> │<not-in-yet> - #9 grad_+_d unknown │#11 grad_f Material/28 - <not-in-yet> │<not-in-yet> - #4 *._e unknown │#6 c Host-non-const/24│ - <not-in-yet> │<not-in-yet> │ - #5 grad_*._e unknown │#7 grad_c Material/28 │ - <not-in-yet> │<not-in-yet> │ - #0 a Host-non-const/24│#2 b Host-non-const/24│ │ - <not-in-yet> │<not-in-yet> │ │ - #1 grad_a Material/28 │#3 grad_b Material/28 │ │ - <not-in-yet> │<not-in-yet> │ │ + #12 *._l Host&dev/412 + <not-in-yet> + #13 grad_*._l unknown + <not-in-yet> + #8 +_d unknown │[10] + <not-in-yet> │ + #9 grad_+_d unknown │ + <not-in-yet> │ + #4 *._e unknown │[6]│ + <not-in-yet> │ │ + #5 grad_*._e unknown│ │ + <not-in-yet> │ │ + [0]│[2] │ │ |}]; let grad_routine = Train.to_routine (module Backend) ctx IDX.empty grad in (* Check out the state without running a forward pass or compiling the SGD update. *) Tensor.print_tree ~spy:true ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l Host&stream/41 - <not-in-yet> - #13 grad_*._l Virt/40 - <not-in-yet> - #8 +_d Local/46 │#10 f Host&shared/39 - <not-in-yet> │<not-in-yet> - #9 grad_+_d Virt/40 │#11 grad_f Dev-stream/41 - <not-in-yet> │<not-in-yet> - #4 *._e Virt/152 │#6 c Host&shared/39 │ - <not-in-yet> │<not-in-yet> │ - #5 grad_*._e Virt/40 │#7 grad_c Dev-stream/41│ - <not-in-yet> │<not-in-yet> │ - #0 a Host&shared/39 │#2 b Host&shared/39 │ │ - <not-in-yet> │<not-in-yet> │ │ - #1 grad_a Dev-stream/41│#3 grad_b Dev-stream/41│ │ - <not-in-yet> │<not-in-yet> │ │ - |}]; + [%expect.unreachable]; (* Do not update the params: all values and gradients will be at initial points, which are specified in the tensor in the brackets. *) Train.run grad_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -8.00 - #13 grad_*._l Virt/40 - <void> - #8 +_d Local/46 │#10 f - <void> │ -2.00 - #9 grad_+_d Virt/40 │#11 grad_f Dev-stream/41 - <void> │<void> - #4 *._e Virt/152 │#6 c │ - <void> │ 1.00e+1 │ - #5 grad_*._e Virt/40 │#7 grad_c Dev-stream/41│ - <void> │<void> │ - #0 a │#2 b │ │ - 2.00 │ -3.00 │ │ - #1 grad_a Dev-stream/41│#3 grad_b Dev-stream/41│ │ - <void> │<void> │ │ - |}]; + [%expect.unreachable]; (* Only now compile the SGD update. *) let sgd_routine = Train.to_routine (module Backend) grad_routine.context IDX.empty sgd in (* Now we update the params, but are not doing the forward and backward passes: only params values @@ -450,48 +301,28 @@ let%expect_test "Simple gradients virtual" = always be recomputed using the latest parameter state. *) Train.run sgd_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -8.00 - #13 grad_*._l Virt/40 - <void> - #8 +_d Local/46 │#10 f - <void> │ -2.40 - #9 grad_+_d Virt/40 │#11 grad_f Dev-stream/41 - <void> │<void> - #4 *._e Virt/152 │#6 c │ - <void> │ 1.02e+1 │ - #5 grad_*._e Virt/40 │#7 grad_c Dev-stream/41│ - <void> │<void> │ - #0 a │#2 b │ │ - 1.40 │ -2.60 │ │ - #1 grad_a Dev-stream/41│#3 grad_b Dev-stream/41│ │ - <void> │<void> │ │ - |}]; + [%expect.unreachable]; (* Now the params will remain as above, but both param gradients and the values and gradients of other nodes will change thanks to the forward and backward passes. *) Train.run grad_routine; Tensor.print_tree ~with_grad:true ~depth:9 l; - [%expect - {| - #12 *._l - -1.57e+1 - #13 grad_*._l Virt/40 - <void> - #8 +_d Local/46 │#10 f - <void> │ -2.40 - #9 grad_+_d Virt/40 │#11 grad_f Dev-stream/41 - <void> │<void> - #4 *._e Virt/152 │#6 c │ - <void> │ 1.02e+1 │ - #5 grad_*._e Virt/40 │#7 grad_c Dev-stream/41│ - <void> │<void> │ - #0 a │#2 b │ │ - 1.40 │ -2.60 │ │ - #1 grad_a Dev-stream/41│#3 grad_b Dev-stream/41│ │ - <void> │<void> │ │ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node a\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Zero2hero_1of7.(fun) in file "test/zero2hero_1of7.ml", line 401, characters 21-73 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "tanh plot" = Tensor.unsafe_reinitialize (); @@ -510,21 +341,22 @@ let%expect_test "2D neuron hosted" = let routine = Train.to_routine (module Backend) ctx IDX.empty update in Train.run routine; Tensor.print_tree ~with_grad:true ~depth:9 v; - [%expect - {| - #8 +_v - 7.00e-1 - #9 grad_+_v - 1.00 - #6 * │#0 b - -6.00 │ 6.70 - #7 grad_* │#1 grad_b - 1.00 │ 1.00 - #2 w │#4 x │ - -3.00 1.00 │ 2.00 0.00 │ - #3 grad_w │#5 grad_x │ - 2.00 0.00 │ -3.00 1.00 │ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node b\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 861, characters 8-11 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Zero2hero_1of7.(fun) in file "test/zero2hero_1of7.ml", line 510, characters 16-70 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "2D neuron virtual" = Tensor.unsafe_reinitialize (); @@ -537,18 +369,19 @@ let%expect_test "2D neuron virtual" = let routine = Train.to_routine (module Backend) ctx IDX.empty update in Train.run routine; Tensor.print_tree ~with_grad:true ~depth:9 v; - [%expect - {| - #8 +_v - 7.00e-1 - #9 grad_+_v Virt/40 - <void> - #6 * Local/46 │#0 b - <void> │ 6.70 - #7 grad_* Virt/40 │#1 grad_b Local/46 - <void> │<void> - #2 w │#4 x │ - -3.00 1.00 │ 2.00 0.00 │ - #3 grad_w Local/46│#5 grad_x Local/46│ - <void> │<void> │ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node b\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Zero2hero_1of7.(fun) in file "test/zero2hero_1of7.ml", line 537, characters 16-70 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] File "test/primitive_ops.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/primitive_ops.ml _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/primitive_ops.ml.corrected diff --git a/_build/default/test/primitive_ops.ml b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/primitive_ops.ml.corrected index c0aa497..44c6652 100644 --- a/_build/default/test/primitive_ops.ml +++ b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/primitive_ops.ml.corrected @@ -318,54 +318,129 @@ let%expect_test "sin(x)" = let%op f x = sin x in let plot_box = plot_unop ~f () in PrintBox_text.output Stdio.stdout plot_box; - [%expect - {| - ┌─────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │ 9.99e-1 │ # │ - │ │#### ## *** ** ### ### │ - │ │ # ** ** # # │ - │ │ # * * # # │ - │ │ # # # │ - │ │ # * * # # │ - │ │ * * # │ - │ │ # * #* │ - │ │ # # # │ - │ │ * * │ - │ │ # * # * # │ - │ │ # # │ - │ │ # * * │ - │ │ # # # │ - │ │* * * *│ - │ │ # # # │ - │ │ * * * * │ - │ │ # # # │ - │f │ * * * * │ - │( │ # # # │ - │x │- * - - - - - - *- - - - - - * - - - - - -* │ - │) │ # # # │ - │ │ * * * * │ - │ │ # # # │ - │ │ * * * * │ - │ │ # # # │ - │ │ * * * * │ - │ │ # # # │ - │ │ * * * # * │ - │ │ # * # * │ - │ │ * # # # * │ - │ │ * * * * │ - │ │ # # # │ - │ │ * * # * #* │ - │ │ * * * * │ - │ │ * # # * # │ - │ │ * * # # * * # │ - │ │ * * # # * * # │ - │ │ ** * # # * ** # │ - │ -9.99e-1│ * ***** ####### **** * ## ####│ - ├─────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ │-5.00 5.00│ - │ │ x │ - └─────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + "Assert_failure arrayjit/lib/c_syntax.ml:453:4" + Raised at Ir__C_syntax.C_syntax.pp_array_offset in file "arrayjit/lib/c_syntax.ml", line 453, characters 4-39 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 509, characters 25-53 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 479, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.pp_ll in file "arrayjit/lib/c_syntax.ml", line 480, characters 17-25 + Called from Ir__C_syntax.C_syntax.compile_main in file "arrayjit/lib/c_syntax.ml" (inlined), line 732, characters 43-52 + Called from Ir__C_syntax.C_syntax.compile_proc in file "arrayjit/lib/c_syntax.ml", line 855, characters 63-79 + Called from Backends__Cc_backend.compile in file "arrayjit/lib/cc_backend.ml", line 137, characters 25-69 + Called from Backends.Add_device.compile in file "arrayjit/lib/backends.ml", line 326, characters 15-45 + Called from Backends.Raise_backend.compile in file "arrayjit/lib/backends.ml", line 443, characters 29-59 + Re-raised at Backends.Raise_backend.compile in file "arrayjit/lib/backends.ml", line 443, characters 4-59 + Re-raised at Backends.Raise_backend.compile in file "arrayjit/lib/backends.ml", lines 439-447, characters 26-99 + Called from Ocannl__Train.to_routine in file "lib/train.ml", line 354, characters 26-82 + Called from Tutorials__Primitive_ops.plot_unop in file "test/primitive_ops.ml", line 35, characters 19-72 + Called from Tutorials__Primitive_ops.(fun) in file "test/primitive_ops.ml", line 319, characters 17-32 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "cos(x)" = let%op f x = cos x in File "test/micrograd_demo.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/micrograd_demo.ml _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/micrograd_demo.ml.corrected diff --git a/_build/default/test/micrograd_demo.ml b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/micrograd_demo.ml.corrected index b394efe..4062941 100644 --- a/_build/default/test/micrograd_demo.ml +++ b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/micrograd_demo.ml.corrected @@ -31,57 +31,26 @@ let%expect_test "Micrograd README basic example" = let step = Train.to_routine (module Backend) ctx IDX.empty update in Train.run step; Tensor.print ~with_code:false ~with_grad:false `Default g; - [%expect - {| - ┌────────────────────┐ - │[75]: +_g shape 0:1 │ - │┌┬─────────┐ │ - │││axis 0 │ │ - │├┼─────────┤ │ - │││ 2.47e+1 │ │ - │└┴─────────┘ │ - └────────────────────┘ - |}]; + [%expect.unreachable]; Tensor.print ~with_code:false ~with_grad:true `Default a; - [%expect - {| - ┌─────────────────┐ - │[0]: a shape 0:1 │ - │┌┬───────┐ │ - │││axis 0 │ │ - │├┼───────┤ │ - │││ -4.00 │ │ - │└┴───────┘ │ - └─────────────────┘ - ┌────────────────────────┐ - │[0]: a shape 0:1 grad_a│ - │┌┬─────────┐ │ - │││axis 0 │ │ - │├┼─────────┤ │ - │││ 1.38e+2 │ │ - │└┴─────────┘ │ - └────────────────────────┘ - |}]; + [%expect.unreachable]; Tensor.print ~with_code:false ~with_grad:true `Default b; - [%expect - {| - ┌─────────────────┐ - │[2]: b shape 0:1 │ - │┌┬──────┐ │ - │││axis 0│ │ - │├┼──────┤ │ - │││ 2.00 │ │ - │└┴──────┘ │ - └─────────────────┘ - ┌────────────────────────┐ - │[2]: b shape 0:1 grad_b│ - │┌┬─────────┐ │ - │││axis 0 │ │ - │├┼─────────┤ │ - │││ 6.45e+2 │ │ - │└┴─────────┘ │ - └────────────────────────┘ - |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node a\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Micrograd_demo.(fun) in file "test/micrograd_demo.ml", line 31, characters 13-67 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] let%expect_test "Micrograd half-moons example" = Tensor.unsafe_reinitialize (); @@ -168,223 +137,34 @@ let%expect_test "Micrograd half-moons example" = in Stdio.printf "Half-moons scatterplot and decision boundary:\n%!"; PrintBox_text.output Stdio.stdout plot_moons; - [%expect - {| - Half-moons scatterplot and decision boundary: - ┌────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │****************************#***********************************************************************│ - │****************************###***###***#***********************************************************│ - │***************************#********#***#***********************************************************│ - │********************#*********************##*#*#****************************************************│ - │******************#***#*****##*#*#*#****#*#*#*##****************************************************│ - │**************#*#**###**#*##**##********#**#****###**#**********************************************│ - │***********#***##***#**##*****************#*#*##*###**********************************************..│ - │***************#**###******************************##*#******************************************...│ - │*************#*****#******************************#*#***#*#*************************************....│ - │*********#***#******************......************##*###**#************************************.....│ - │******#***#*#*#****************........******************#*#*********************************.......│ - │*****###****#****************...........****************#***********************************........│ - │*******#*###****************.............******************#*#*****************************.........│ - │**********##***************.......%%......***************#**##****************************..........│ - │*****#***#****************......%..%%......*********************************************..........%.│ - │**####*******************......%..%..........****************##*#**********************........%...%│ - │****#*******************.......%%..%..........***************#*#**********************.........%....│ - │*#****##***************.........%....%.........************##***#********************.......%...%%..│ - │****#*#***************.........%..%.%...........*************#**#*******************..........%.%...│ - │##*##****************............%...............**************#*##****************.........%.%.%%%.│ - │**#####************............%...%%.............*******************************...........%..%.%..│ - │******************...............%.%.%%............*********#*#*#*#*************................%%..│ - │***##************................%..%..%............************#*#************.............%...%...│ - │##**#***********...................%...%.............********#**###***********...........%.%%.......│ - │###************..................%.%.%................***********##**********..................%....│ - │**##**********.........................................**********#*********.............%%.%..%.....│ - │**#**********......................%%...................********##********..............%%.%........│ - │************..........................%..................****#**#********...................%.......│ - │***********.........................%%.%..................*****##******...............%..%..........│ - │**********.............................%%..................***********.................%%.%%........│ - │*********..............................%...%................*******.....................%%%.........│ - │********...................................%..%....................................%.....%..........│ - │*******.................................%%..%.%................................%.%.%.%%.............│ - │******.....................................%%.%...%%.%.......................%%%....%.%.%...........│ - │*****....................................%.....%.%..%....................%.....%%.%.%...............│ - │****........................................%%...%.%%%%..%.......%.........%%.%%....................│ - │***...............................................%..%.......%..%%%%...%.%..%.%.....................│ - │**.................................................%.%.%............%..%...%........................│ - │*.................................................%....%.%......%%...%.%............................│ - │..........................................................%..%......%%%.............................│ - └────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + [%expect.unreachable]; Stdio.printf "Loss:\n%!"; let plot_loss = PrintBox_utils.plot ~x_label:"step" ~y_label:"loss" [ Line_plot { points = Array.of_list_rev !losses; content = PrintBox.line "-" } ] in PrintBox_text.output Stdio.stdout plot_loss; - [%expect - {| - Loss: - ┌────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │ 3.79e+1│- │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │l │ │ - │o │ │ - │s │ │ - │s │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │- │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │- - - │ - │ 0.00 │----------------------------------------------------------------------------------------------------│ - ├────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ │0.00 3.99e+2│ - │ │ step │ - └────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + [%expect.unreachable]; Stdio.printf "Log-loss, for better visibility:\n%!"; let plot_loss = PrintBox_utils.plot ~x_label:"step" ~y_label:"log loss" [ Line_plot { points = Array.of_list_rev !log_losses; content = PrintBox.line "-" } ] in PrintBox_text.output Stdio.stdout plot_loss; - [%expect - {| - Log-loss, for better visibility: - ┌─────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │ 3.63 │- │ - │ │ │ - │ │ │ - │ │ │ - │ │- │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │- │ - │ │ - - │ - │ │ - - - │ - │ │------ - │ - │ │ ---- - - --- - - - - - │ - │ │ -- - - - --- -- - -- - │ - │ │ - -- - - --- - - --- ----- - -- │ - │l │ -- - - - - --- - - - │ - │o │ - - - -- - -- - - - - - │ - │g │ - - - - -- - - - - - - - │ - │ │ - - - -- - - - -- - - - │ - │l │ - - -- -- - - - - -│ - │o │ - - - - - - - - - - - │ - │s │ - - - - -- - │ - │s │ - - │ - │ │ - - - - - - - - - │ - │ │ - - - - - │ - │ │ - - │ - │ │ - - - │ - │ │ - - │ - │ │ │ - │ │ - │ - │ │ - │ - │ │ │ - │ │ │ - │ │ │ - │ │ │ - │ │ - │ - │ │ │ - │ │ │ - │ -1.00e+1│ - - - - ----- -- --------------- ---------------------------------------------------------│ - ├─────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ │0.00 3.99e+2│ - │ │ step │ - └─────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + [%expect.unreachable]; Stdio.printf "\nLearning rate:\n%!"; let plot_lr = PrintBox_utils.plot ~x_label:"step" ~y_label:"learning rate" [ Line_plot { points = Array.of_list_rev !learning_rates; content = PrintBox.line "-" } ] in PrintBox_text.output Stdio.stdout plot_lr; - [%expect - {| - Learning rate: - ┌─────────┬────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │ -1.00e-1│ -│ - │ │ ---│ - │ │ ---- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ --- │ - │ │ --- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │l │ ---- │ - │e │ --- │ - │a │ --- │ - │r │ --- │ - │n │ --- │ - │i │ ---- │ - │n │ --- │ - │g │ ---- │ - │ │ --- │ - │r │ ---- │ - │a │ --- │ - │t │ --- │ - │e │ --- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ --- │ - │ │ --- │ - │ │ --- │ - │ │ ---- │ - │ │ --- │ - │ │ ---- │ - │ -2.00e-1│--- │ - ├─────────┼────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ │0.00 3.99e+2│ - │ │ step │ - └─────────┴────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + [%expect.unreachable]; (* Testing how the syntax extension %op creates labels for the resulting tensors: *) Stdio.printf "mlp_result's name: %s\n%!" @@ Tensor.debug_name mlp_result; (* Note: mlp_result is not included in the resulting tensor's label, because the identifier label does not propagate across function calls. *) - [%expect {| mlp_result's name: mlp_point |}]; + [%expect.unreachable]; (Stdio.printf "(mlp moons_input) name: %s\n%!" @@ Tensor.debug_name @@ @@ -398,4 +178,20 @@ let%expect_test "Micrograd half-moons example" = ] -> subtensor | _ -> assert false); - [%expect {| (mlp moons_input) name: mlp_moons_input |}] + [%expect.unreachable] +[@@expect.uncaught_exn {| + (* CR expect_test_collector: This test expectation appears to contain a backtrace. + This is strongly discouraged as backtraces are fragile. + Please change this test to not include a backtrace. *) + ("Utils.User_error(\"The linked context lacks node b1\")") + Raised at Backends.verify_prior_context.(fun) in file "arrayjit/lib/backends.ml", line 285, characters 11-93 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 861, characters 8-11 + Called from Base__Set.Tree0.iter.iter in file "src/set.ml", line 860, characters 8-14 + Called from Base__Set.Accessors.iter in file "src/set.ml" (inlined), line 1193, characters 18-38 + Called from Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 281-285, characters 2-94 + Re-raised at Backends.verify_prior_context in file "arrayjit/lib/backends.ml", lines 280-285, characters 37-94 + Called from Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 527-528, characters 4-49 + Re-raised at Backends.Raise_backend.link in file "arrayjit/lib/backends.ml", lines 526-540, characters 23-92 + Called from Tutorials__Micrograd_demo.(fun) in file "test/micrograd_demo.ml", line 125, characters 4-83 + Called from Ppx_expect_runtime__Test_block.Configured.dump_backtrace in file "runtime/test_block.ml", line 142, characters 10-28 + |}] File "test/hello_world_op.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/hello_world_op.ml _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/hello_world_op.ml.corrected diff --git a/_build/default/test/hello_world_op.ml b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/hello_world_op.ml.corrected index 6407bb4..718250d 100644 --- a/_build/default/test/hello_world_op.ml +++ b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/hello_world_op.ml.corrected @@ -36,11 +36,11 @@ let%expect_test "Pointwise multiplication dims 1" = {| ┌────────────────────┐ │[3]: *._y shape 0:1 │ - │┌┬─────────┐ │ - │││axis 0 │ │ - │├┼─────────┤ │ - │││ 1.40e+1 │ │ - │└┴─────────┘ │ + │┌┬──────┐ │ + │││axis 0│ │ + │├┼──────┤ │ + │││ 0.00 │ │ + │└┴──────┘ │ └────────────────────┘ |}] @@ -67,11 +67,11 @@ let%expect_test "Matrix multiplication dims 1x1" = {| ┌────────────────────────┐ │[0]: hey shape 1:1->0:1 │ - │┌──────┬──────┐ │ - ││ │axis 1│ │ - │├──────┼──────┤ │ - ││axis 0│ 7.00 │ │ - │└──────┴──────┘ │ + │┌──────┬──────────┐ │ + ││ │axis 1 │ │ + │├──────┼──────────┤ │ + ││axis 0│ 9.95e-39 │ │ + │└──────┴──────────┘ │ └────────────────────────┘ |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ y; @@ -79,11 +79,11 @@ let%expect_test "Matrix multiplication dims 1x1" = {| ┌───────────────────┐ │[6]: +_y shape 0:1 │ - │┌┬─────────┐ │ - │││axis 0 │ │ - │├┼─────────┤ │ - │││ 1.50e+1 │ │ - │└┴─────────┘ │ + │┌┬──────┐ │ + │││axis 0│ │ + │├┼──────┤ │ + │││ 1.00 │ │ + │└┴──────┘ │ └───────────────────┘ |}] @@ -108,7 +108,7 @@ let%expect_test "Print constant tensor" = Tensor.print ~with_code:false ~with_grad:false `Inline @@ hey; [%expect {| - [0]: [ 1.00 , 2.00 , 3.00 ; 4.00 , 5.00 , 6.00 ]_hey shape 1:3->0:2 [ + [0]: [ 0.00 , 0.00 , 0.00 ; 0.00 , 0.00 , 0.00 ]_hey shape 1:3->0:2 [ 1.00 , 2.00 , 3.00 ; 4.00 , 5.00 , 6.00 ] @@ -117,7 +117,7 @@ let%expect_test "Print constant tensor" = [%expect {| ┌────────────────────────────────────────────────────────────────────────┐ - │[0]: [ 1.00 , 2.00 , 3.00 ; 4.00 , 5.00 , 6.00 ]_hey shape 1:3->0:2 │ + │[0]: [ 0.00 , 0.00 , 0.00 ; 0.00 , 0.00 , 0.00 ]_hey shape 1:3->0:2 │ │┌──────┬──────────────────┐ │ ││ │axis 1 │ │ │├──────┼──────────────────┤ │ @@ -131,7 +131,7 @@ let%expect_test "Print constant tensor" = Tensor.print ~with_code:false ~with_grad:false `Inline @@ hoo; [%expect {| - [1]: [| [ 1.00 ; 2.00 ; 3.00 ] ; [ 4.00 ; 5.00 ; 6.00 ] |]_hoo shape 0:2|1:3 [| + [1]: c2x3_hoo shape 0:2|1:3 [| [ 1.00 ; 2.00 ; 3.00 ] ; [ 4.00 ; 5.00 ; 6.00 ] |] @@ -139,15 +139,15 @@ let%expect_test "Print constant tensor" = Tensor.print ~with_code:false ~with_grad:false `Default @@ hoo; [%expect {| - ┌─────────────────────────────────────────────────────────────────────────────┐ - │[1]: [| [ 1.00 ; 2.00 ; 3.00 ] ; [ 4.00 ; 5.00 ; 6.00 ] |]_hoo shape 0:2|1:3 │ - │┌──────┬──────────────────┐ │ - ││ │axis 1 │ │ - │├──────┼──────────────────┤ │ - ││axis 0│ 1.00 2.00 3.00 │ │ - ││ │ 4.00 5.00 6.00 │ │ - │└──────┴──────────────────┘ │ - └─────────────────────────────────────────────────────────────────────────────┘ + ┌────────────────────────────┐ + │[1]: c2x3_hoo shape 0:2|1:3 │ + │┌──────┬──────────────────┐ │ + ││ │axis 1 │ │ + │├──────┼──────────────────┤ │ + ││axis 0│ 1.00 2.00 3.00 │ │ + ││ │ 4.00 5.00 6.00 │ │ + │└──────┴──────────────────┘ │ + └────────────────────────────┘ |}]; let%op hey2 = [ @@ -471,28 +471,28 @@ let%expect_test "Matrix multiplication dims 2x3" = Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect {| - ┌────────────────────────┐ - │[0]: hey shape 1:2->0:3 │ - │┌──────┬────────────┐ │ - ││ │axis 1 │ │ - │├──────┼────────────┤ │ - ││axis 0│ 7.00 7.00 │ │ - ││ │ 7.00 7.00 │ │ - ││ │ 7.00 7.00 │ │ - │└──────┴────────────┘ │ - └────────────────────────┘ + ┌──────────────────────────────┐ + │[0]: hey shape 1:2->0:3 │ + │┌──────┬─────────────────────┐│ + ││ │axis 1 ││ + │├──────┼─────────────────────┤│ + ││axis 0│ -1.15e+22 4.45e-41 ││ + ││ │ 0.00 0.00 ││ + ││ │ -1.02e+12 4.45e-41 ││ + │└──────┴─────────────────────┘│ + └──────────────────────────────┘ |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ y; [%expect {| - ┌──────────────────────────────┐ - │[6]: +_y shape 0:3 │ - │┌┬───────────────────────────┐│ - │││axis 0 ││ - │├┼───────────────────────────┤│ - │││ 3.90e+1 4.00e+1 4.10e+1 ││ - │└┴───────────────────────────┘│ - └──────────────────────────────┘ + ┌───────────────────────────────┐ + │[6]: +_y shape 0:3 │ + │┌┬────────────────────────────┐│ + │││axis 0 ││ + │├┼────────────────────────────┤│ + │││ -2.31e+22 5.00 -2.04e+12 ││ + │└┴────────────────────────────┘│ + └───────────────────────────────┘ |}] let%expect_test "Big matrix" = @@ -515,71 +515,38 @@ let%expect_test "Big matrix" = let y = TDSL.O.((hey * zero_to_twenty) + zero_to_twenty) in Train.forward_and_forget backend ctx y; Tensor.print ~with_code:false ~with_grad:false `Inline zero_to_twenty; + [%expect + {| [2]: 0...20 shape 0:21 <virtual> |}]; + Tensor.print ~with_code:false ~with_grad:false `Default zero_to_twenty; + [%expect + {| [2]: 0...20 shape 0:21 <virtual> |}]; + Tensor.print ~with_code:false ~with_grad:false `Default hey; [%expect {| - [2]: 0...20 shape 0:21 [ - 0.00 - ; 1.00 - ; 2.00 - ; 3.00 - ; 4.00 - ; 5.00 - ; 6.00 - ; 7.00 - ; 8.00 - ; 9.00 - ; 10.00 - ; 11.00 - ; 12.00 - ; 13.00 - ; 14.00 - ; 15.00 - ; 16.00 - ; 17.00 - ; 18.00 - ; 19.00 - ; 20.00 - ] + ┌──────────────────────────────────────┐ + │[0]: hey shape 1:21->0:21 │ + │┌──────┬─────────────────────────────┐│ + ││ │axis 1 ││ + │├──────┼─────────────────────────────┤│ + ││axis 0│ 0.00 0.00 ... 0.00 0.00 ││ + ││ │ 0.00 0.00 ... 0.00 0.00 ││ + ││ │ ... ... ... ... ... ││ + ││ │ 0.00 0.00 ... 0.00 0.00 ││ + ││ │ 0.00 0.00 ... 0.00 0.00 ││ + │└──────┴─────────────────────────────┘│ + └──────────────────────────────────────┘ |}]; - Tensor.print ~with_code:false ~with_grad:false `Default zero_to_twenty; + Tensor.print ~with_code:false ~with_grad:false `Default y; [%expect {| ┌──────────────────────────────────────┐ - │[2]: 0...20 shape 0:21 │ + │[5]: + shape 0:21 │ │┌┬───────────────────────────────────┐│ │││axis 0 ││ │├┼───────────────────────────────────┤│ │││ 0.00 1.00 ... 1.90e+1 2.00e+1 ││ │└┴───────────────────────────────────┘│ └──────────────────────────────────────┘ - |}]; - Tensor.print ~with_code:false ~with_grad:false `Default hey; - [%expect - {| - ┌──────────────────────────────────────────────────┐ - │[0]: hey shape 1:21->0:21 │ - │┌──────┬─────────────────────────────────────────┐│ - ││ │axis 1 ││ - │├──────┼─────────────────────────────────────────┤│ - ││axis 0│ 5.00e-1 5.00e-1 ... 5.00e-1 5.00e-1 ││ - ││ │ 5.00e-1 5.00e-1 ... 5.00e-1 5.00e-1 ││ - ││ │ ... ... ... ... ... ││ - ││ │ 5.00e-1 5.00e-1 ... 5.00e-1 5.00e-1 ││ - ││ │ 5.00e-1 5.00e-1 ... 5.00e-1 5.00e-1 ││ - │└──────┴─────────────────────────────────────────┘│ - └──────────────────────────────────────────────────┘ - |}]; - Tensor.print ~with_code:false ~with_grad:false `Default y; - [%expect - {| - ┌────────────────────────────────────────────┐ - │[5]: + shape 0:21 │ - │┌┬─────────────────────────────────────────┐│ - │││axis 0 ││ - │├┼─────────────────────────────────────────┤│ - │││ 1.05e+2 1.06e+2 ... 1.24e+2 1.25e+2 ││ - │└┴─────────────────────────────────────────┘│ - └────────────────────────────────────────────┘ |}] let%expect_test "Very big tensor" = @@ -603,140 +570,7 @@ let%expect_test "Very big tensor" = Train.forward_and_forget backend ctx hoo; Tensor.print ~with_code:false ~with_grad:false `Default hey; [%expect - {| - ┌───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[0]: r6x10x11x7x8x9 shape 0:6|3:7,4:8,5:9->1:10,2:11 │ - │┌──────┬─────────────────────────────────────────┬─────────────────────────────────────────┬──────┬─────────────────────────────────────────┬─────────────────────────────────────────┐│ - ││0 @ 0 │0 @ 4 │1 @ 4 │~~~~~ │6 @ 4 │7 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5│axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││0 @ 1 │ 0.00 1.00 ... 7.00 8.00 │ 9.00 1.00e+1 ... 1.60e+1 1.70e+1 │ ... │ 5.40e+1 5.50e+1 ... 6.10e+1 6.20e+1 │ 6.30e+1 6.40e+1 ... 7.00e+1 7.10e+1 ││ - ││axis 2│ 5.04e+2 5.05e+2 ... 5.11e+2 5.12e+2 │ 5.13e+2 5.14e+2 ... 5.20e+2 5.21e+2 │ │ 5.58e+2 5.59e+2 ... 5.65e+2 5.66e+2 │ 5.67e+2 5.68e+2 ... 5.74e+2 5.75e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 4.53e+3 4.53e+3 ... 4.54e+3 4.54e+3 │ 4.54e+3 4.54e+3 ... 4.55e+3 4.55e+3 │ │ 4.59e+3 4.59e+3 ... 4.59e+3 4.59e+3 │ 4.59e+3 4.60e+3 ... 4.60e+3 4.60e+3 ││ - ││ │ 5.04e+3 5.04e+3 ... 5.04e+3 5.04e+3 │ 5.04e+3 5.05e+3 ... 5.05e+3 5.05e+3 │ │ 5.09e+3 5.09e+3 ... 5.10e+3 5.10e+3 │ 5.10e+3 5.10e+3 ... 5.11e+3 5.11e+3 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││1 @ 1 │ 5.54e+3 5.54e+3 ... 5.55e+3 5.55e+3 │ 5.55e+3 5.55e+3 ... 5.56e+3 5.56e+3 │ ... │ 5.59e+3 5.59e+3 ... 5.60e+3 5.60e+3 │ 5.60e+3 5.60e+3 ... 5.61e+3 5.61e+3 ││ - ││axis 2│ 6.04e+3 6.04e+3 ... 6.05e+3 6.05e+3 │ 6.05e+3 6.05e+3 ... 6.06e+3 6.06e+3 │ │ 6.10e+3 6.10e+3 ... 6.10e+3 6.11e+3 │ 6.11e+3 6.11e+3 ... 6.11e+3 6.11e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.00e+4 1.00e+4 ... 1.00e+4 1.00e+4 │ 1.00e+4 1.00e+4 ... 1.00e+4 1.00e+4 │ │ 1.01e+4 1.01e+4 ... 1.01e+4 1.01e+4 │ 1.01e+4 1.01e+4 ... 1.01e+4 1.01e+4 ││ - ││ │ 1.05e+4 1.05e+4 ... 1.05e+4 1.05e+4 │ 1.05e+4 1.05e+4 ... 1.06e+4 1.06e+4 │ │ 1.06e+4 1.06e+4 ... 1.06e+4 1.06e+4 │ 1.06e+4 1.06e+4 ... 1.06e+4 1.06e+4 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... │ ... ││ - ││axis 2│ │ │ │ │ ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││8 @ 1 │ 4.43e+4 4.43e+4 ... 4.43e+4 4.43e+4 │ 4.43e+4 4.43e+4 ... 4.43e+4 4.43e+4 │ ... │ 4.44e+4 4.44e+4 ... 4.44e+4 4.44e+4 │ 4.44e+4 4.44e+4 ... 4.44e+4 4.44e+4 ││ - ││axis 2│ 4.48e+4 4.48e+4 ... 4.48e+4 4.48e+4 │ 4.48e+4 4.48e+4 ... 4.48e+4 4.48e+4 │ │ 4.49e+4 4.49e+4 ... 4.49e+4 4.49e+4 │ 4.49e+4 4.49e+4 ... 4.49e+4 4.49e+4 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 4.88e+4 4.88e+4 ... 4.88e+4 4.88e+4 │ 4.88e+4 4.88e+4 ... 4.89e+4 4.89e+4 │ │ 4.89e+4 4.89e+4 ... 4.89e+4 4.89e+4 │ 4.89e+4 4.89e+4 ... 4.89e+4 4.89e+4 ││ - ││ │ 4.93e+4 4.93e+4 ... 4.93e+4 4.94e+4 │ 4.94e+4 4.94e+4 ... 4.94e+4 4.94e+4 │ │ 4.94e+4 4.94e+4 ... 4.94e+4 4.94e+4 │ 4.94e+4 4.94e+4 ... 4.94e+4 4.94e+4 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││9 @ 1 │ 4.98e+4 4.98e+4 ... 4.99e+4 4.99e+4 │ 4.99e+4 4.99e+4 ... 4.99e+4 4.99e+4 │ ... │ 4.99e+4 4.99e+4 ... 4.99e+4 4.99e+4 │ 4.99e+4 4.99e+4 ... 4.99e+4 4.99e+4 ││ - ││axis 2│ 5.04e+4 5.04e+4 ... 5.04e+4 5.04e+4 │ 5.04e+4 5.04e+4 ... 5.04e+4 5.04e+4 │ │ 5.04e+4 5.04e+4 ... 5.04e+4 5.04e+4 │ 5.04e+4 5.04e+4 ... 5.04e+4 5.04e+4 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 5.44e+4 5.44e+4 ... 5.44e+4 5.44e+4 │ 5.44e+4 5.44e+4 ... 5.44e+4 5.44e+4 │ │ 5.44e+4 5.44e+4 ... 5.44e+4 5.44e+4 │ 5.44e+4 5.44e+4 ... 5.45e+4 5.45e+4 ││ - ││ │ 5.49e+4 5.49e+4 ... 5.49e+4 5.49e+4 │ 5.49e+4 5.49e+4 ... 5.49e+4 5.49e+4 │ │ 5.49e+4 5.49e+4 ... 5.49e+4 5.49e+4 │ 5.49e+4 5.50e+4 ... 5.50e+4 5.50e+4 ││ - │└──────┴─────────────────────────────────────────┴─────────────────────────────────────────┴──────┴─────────────────────────────────────────┴─────────────────────────────────────────┘│ - ├───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │┌──────┬─────────────────────────────────────────┬─────────────────────────────────────────┬──────┬─────────────────────────────────────────┬─────────────────────────────────────────┐│ - ││1 @ 0 │0 @ 4 │1 @ 4 │~~~~~ │6 @ 4 │7 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5│axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││0 @ 1 │ 5.54e+4 5.54e+4 ... 5.54e+4 5.54e+4 │ 5.54e+4 5.54e+4 ... 5.54e+4 5.54e+4 │ ... │ 5.54e+4 5.54e+4 ... 5.55e+4 5.55e+4 │ 5.55e+4 5.55e+4 ... 5.55e+4 5.55e+4 ││ - ││axis 2│ 5.59e+4 5.59e+4 ... 5.59e+4 5.59e+4 │ 5.59e+4 5.59e+4 ... 5.59e+4 5.59e+4 │ │ 5.59e+4 5.59e+4 ... 5.60e+4 5.60e+4 │ 5.60e+4 5.60e+4 ... 5.60e+4 5.60e+4 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 5.99e+4 5.99e+4 ... 5.99e+4 5.99e+4 │ 5.99e+4 5.99e+4 ... 5.99e+4 5.99e+4 │ │ 6.00e+4 6.00e+4 ... 6.00e+4 6.00e+4 │ 6.00e+4 6.00e+4 ... 6.00e+4 6.00e+4 ││ - ││ │ 6.04e+4 6.04e+4 ... 6.04e+4 6.04e+4 │ 6.04e+4 6.04e+4 ... 6.04e+4 6.04e+4 │ │ 6.05e+4 6.05e+4 ... 6.05e+4 6.05e+4 │ 6.05e+4 6.05e+4 ... 6.05e+4 6.05e+4 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││1 @ 1 │ 6.09e+4 6.09e+4 ... 6.09e+4 6.09e+4 │ 6.09e+4 6.09e+4 ... 6.10e+4 6.10e+4 │ ... │ 6.10e+4 6.10e+4 ... 6.10e+4 6.10e+4 │ 6.10e+4 6.10e+4 ... 6.10e+4 6.10e+4 ││ - ││axis 2│ 6.14e+4 6.14e+4 ... 6.14e+4 6.14e+4 │ 6.14e+4 6.14e+4 ... 6.15e+4 6.15e+4 │ │ 6.15e+4 6.15e+4 ... 6.15e+4 6.15e+4 │ 6.15e+4 6.15e+4 ... 6.15e+4 6.15e+4 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 6.55e+4 6.55e+4 ... 6.55e+4 6.55e+4 │ 6.55e+4 6.55e+4 ... 6.55e+4 6.55e+4 │ │ 6.55e+4 6.55e+4 ... 6.55e+4 6.55e+4 │ 6.55e+4 6.55e+4 ... 6.55e+4 6.55e+4 ││ - ││ │ 6.60e+4 6.60e+4 ... 6.60e+4 6.60e+4 │ 6.60e+4 6.60e+4 ... 6.60e+4 6.60e+4 │ │ 6.60e+4 6.60e+4 ... 6.60e+4 6.60e+4 │ 6.60e+4 6.60e+4 ... 6.60e+4 6.60e+4 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... │ ... ││ - ││axis 2│ │ │ │ │ ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││8 @ 1 │ 9.97e+4 9.97e+4 ... 9.97e+4 9.98e+4 │ 9.98e+4 9.98e+4 ... 9.98e+4 9.98e+4 │ ... │ 9.98e+4 9.98e+4 ... 9.98e+4 9.98e+4 │ 9.98e+4 9.98e+4 ... 9.98e+4 9.98e+4 ││ - ││axis 2│ 1.00e+5 1.00e+5 ... 1.00e+5 1.00e+5 │ 1.00e+5 1.00e+5 ... 1.00e+5 1.00e+5 │ │ 1.00e+5 1.00e+5 ... 1.00e+5 1.00e+5 │ 1.00e+5 1.00e+5 ... 1.00e+5 1.00e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 ││ - ││ │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 │ 1.04e+5 1.04e+5 ... 1.04e+5 1.04e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││9 @ 1 │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ ... │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 ││ - ││axis 2│ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 │ 1.05e+5 1.05e+5 ... 1.05e+5 1.05e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.09e+5 1.09e+5 ... 1.09e+5 1.09e+5 │ 1.09e+5 1.09e+5 ... 1.09e+5 1.09e+5 │ │ 1.09e+5 1.09e+5 ... 1.09e+5 1.09e+5 │ 1.09e+5 1.09e+5 ... 1.09e+5 1.09e+5 ││ - ││ │ 1.10e+5 1.10e+5 ... 1.10e+5 1.10e+5 │ 1.10e+5 1.10e+5 ... 1.10e+5 1.10e+5 │ │ 1.10e+5 1.10e+5 ... 1.10e+5 1.10e+5 │ 1.10e+5 1.10e+5 ... 1.10e+5 1.10e+5 ││ - │└──────┴─────────────────────────────────────────┴─────────────────────────────────────────┴──────┴─────────────────────────────────────────┴─────────────────────────────────────────┘│ - ├───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │ ... │ - ├───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │┌──────┬─────────────────────────────────────────┬─────────────────────────────────────────┬──────┬─────────────────────────────────────────┬─────────────────────────────────────────┐│ - ││4 @ 0 │0 @ 4 │1 @ 4 │~~~~~ │6 @ 4 │7 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5│axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││0 @ 1 │ 2.21e+5 2.21e+5 ... 2.21e+5 2.21e+5 │ 2.21e+5 2.21e+5 ... 2.21e+5 2.21e+5 │ ... │ 2.21e+5 2.21e+5 ... 2.21e+5 2.21e+5 │ 2.21e+5 2.21e+5 ... 2.21e+5 2.21e+5 ││ - ││axis 2│ 2.22e+5 2.22e+5 ... 2.22e+5 2.22e+5 │ 2.22e+5 2.22e+5 ... 2.22e+5 2.22e+5 │ │ 2.22e+5 2.22e+5 ... 2.22e+5 2.22e+5 │ 2.22e+5 2.22e+5 ... 2.22e+5 2.22e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 ││ - ││ │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 │ 2.26e+5 2.26e+5 ... 2.26e+5 2.26e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││1 @ 1 │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ ... │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 ││ - ││axis 2│ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 │ 2.27e+5 2.27e+5 ... 2.27e+5 2.27e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.31e+5 2.31e+5 ... 2.31e+5 2.31e+5 │ 2.31e+5 2.31e+5 ... 2.31e+5 2.31e+5 │ │ 2.31e+5 2.31e+5 ... 2.31e+5 2.31e+5 │ 2.31e+5 2.31e+5 ... 2.31e+5 2.31e+5 ││ - ││ │ 2.32e+5 2.32e+5 ... 2.32e+5 2.32e+5 │ 2.32e+5 2.32e+5 ... 2.32e+5 2.32e+5 │ │ 2.32e+5 2.32e+5 ... 2.32e+5 2.32e+5 │ 2.32e+5 2.32e+5 ... 2.32e+5 2.32e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... │ ... ││ - ││axis 2│ │ │ │ │ ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││8 @ 1 │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ ... │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 ││ - ││axis 2│ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 │ 2.66e+5 2.66e+5 ... 2.66e+5 2.66e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.70e+5 2.70e+5 ... 2.70e+5 2.70e+5 │ 2.70e+5 2.70e+5 ... 2.70e+5 2.70e+5 │ │ 2.70e+5 2.70e+5 ... 2.70e+5 2.70e+5 │ 2.70e+5 2.70e+5 ... 2.70e+5 2.70e+5 ││ - ││ │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││9 @ 1 │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ ... │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 │ 2.71e+5 2.71e+5 ... 2.71e+5 2.71e+5 ││ - ││axis 2│ 2.72e+5 2.72e+5 ... 2.72e+5 2.72e+5 │ 2.72e+5 2.72e+5 ... 2.72e+5 2.72e+5 │ │ 2.72e+5 2.72e+5 ... 2.72e+5 2.72e+5 │ 2.72e+5 2.72e+5 ... 2.72e+5 2.72e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 ││ - ││ │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 │ 2.76e+5 2.76e+5 ... 2.76e+5 2.76e+5 ││ - │└──────┴─────────────────────────────────────────┴─────────────────────────────────────────┴──────┴─────────────────────────────────────────┴─────────────────────────────────────────┘│ - ├───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │┌──────┬─────────────────────────────────────────┬─────────────────────────────────────────┬──────┬─────────────────────────────────────────┬─────────────────────────────────────────┐│ - ││5 @ 0 │0 @ 4 │1 @ 4 │~~~~~ │6 @ 4 │7 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5│axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││0 @ 1 │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ ... │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 ││ - ││axis 2│ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 │ 2.77e+5 2.77e+5 ... 2.77e+5 2.77e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.81e+5 2.81e+5 ... 2.81e+5 2.81e+5 │ 2.81e+5 2.81e+5 ... 2.81e+5 2.81e+5 │ │ 2.81e+5 2.81e+5 ... 2.81e+5 2.81e+5 │ 2.81e+5 2.81e+5 ... 2.81e+5 2.81e+5 ││ - ││ │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││1 @ 1 │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ ... │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 │ 2.82e+5 2.82e+5 ... 2.82e+5 2.82e+5 ││ - ││axis 2│ 2.83e+5 2.83e+5 ... 2.83e+5 2.83e+5 │ 2.83e+5 2.83e+5 ... 2.83e+5 2.83e+5 │ │ 2.83e+5 2.83e+5 ... 2.83e+5 2.83e+5 │ 2.83e+5 2.83e+5 ... 2.83e+5 2.83e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 ││ - ││ │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 │ 2.87e+5 2.87e+5 ... 2.87e+5 2.87e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... │ ... ││ - ││axis 2│ │ │ │ │ ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││8 @ 1 │ 3.21e+5 3.21e+5 ... 3.21e+5 3.21e+5 │ 3.21e+5 3.21e+5 ... 3.21e+5 3.21e+5 │ ... │ 3.21e+5 3.21e+5 ... 3.21e+5 3.21e+5 │ 3.21e+5 3.21e+5 ... 3.21e+5 3.21e+5 ││ - ││axis 2│ 3.22e+5 3.22e+5 ... 3.22e+5 3.22e+5 │ 3.22e+5 3.22e+5 ... 3.22e+5 3.22e+5 │ │ 3.22e+5 3.22e+5 ... 3.22e+5 3.22e+5 │ 3.22e+5 3.22e+5 ... 3.22e+5 3.22e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 ││ - ││ │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 │ 3.26e+5 3.26e+5 ... 3.26e+5 3.26e+5 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼──────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││9 @ 1 │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ ... │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 ││ - ││axis 2│ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 │ 3.27e+5 3.27e+5 ... 3.27e+5 3.27e+5 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 3.31e+5 3.31e+5 ... 3.31e+5 3.31e+5 │ 3.31e+5 3.31e+5 ... 3.31e+5 3.31e+5 │ │ 3.31e+5 3.31e+5 ... 3.31e+5 3.31e+5 │ 3.31e+5 3.31e+5 ... 3.31e+5 3.31e+5 ││ - ││ │ 3.32e+5 3.32e+5 ... 3.32e+5 3.32e+5 │ 3.32e+5 3.32e+5 ... 3.32e+5 3.32e+5 │ │ 3.32e+5 3.32e+5 ... 3.32e+5 3.32e+5 │ 3.32e+5 3.32e+5 ... 3.32e+5 3.32e+5 ││ - │└──────┴─────────────────────────────────────────┴─────────────────────────────────────────┴──────┴─────────────────────────────────────────┴─────────────────────────────────────────┘│ - └───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r6x10x11x7x8x9 shape 0:6|3:7,4:8,5:9->1:10,2:11 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default hoo; (* Disable line wrapping for viewing the output. In VSCode: `View: Toggle Word Wrap`. *) [%expect File "test/einsum_trivia.ml", line 1, characters 0-0: /usr/bin/git --no-pager diff --no-index --color=always -u _build/default/test/einsum_trivia.ml _build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/einsum_trivia.ml.corrected diff --git a/_build/default/test/einsum_trivia.ml b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/einsum_trivia.ml.corrected index 35a36fa..b1207f6 100644 --- a/_build/default/test/einsum_trivia.ml +++ b/_build/.sandbox/c7be9de083554f9a10c5fa61456838c0/default/test/einsum_trivia.ml.corrected @@ -28,20 +28,7 @@ let%expect_test "einsum1 permute axes" = Train.forward_and_forget backend ctx ho; Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[0]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho; [%expect {| @@ -64,106 +51,7 @@ let%expect_test "einsum1 permute axes" = Train.forward_and_forget backend ctx ho2; Tensor.print ~with_code:false ~with_grad:false `Default @@ hey2; [%expect - {| - ┌────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[2]: r2x3x6x7x4x5 shape 0:2,1:3|4:4,5:5->2:6,3:7 │ - │┌──────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┐│ - ││0 @ 1 │0 @ 4 │1 @ 4 │2 @ 4 │3 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││0 @ 2 │ 0.00 1.00 2.00 3.00 4.00 │ 5.00 6.00 7.00 8.00 9.00 │ 1.00e+1 1.10e+1 1.20e+1 1.30e+1 1.40e+1 │ 1.50e+1 1.60e+1 1.70e+1 1.80e+1 1.90e+1 ││ - ││axis 3│ 2.00e+1 2.10e+1 2.20e+1 2.30e+1 2.40e+1 │ 2.50e+1 2.60e+1 2.70e+1 2.80e+1 2.90e+1 │ 3.00e+1 3.10e+1 3.20e+1 3.30e+1 3.40e+1 │ 3.50e+1 3.60e+1 3.70e+1 3.80e+1 3.90e+1 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.00e+2 1.01e+2 1.02e+2 1.03e+2 1.04e+2 │ 1.05e+2 1.06e+2 1.07e+2 1.08e+2 1.09e+2 │ 1.10e+2 1.11e+2 1.12e+2 1.13e+2 1.14e+2 │ 1.15e+2 1.16e+2 1.17e+2 1.18e+2 1.19e+2 ││ - ││ │ 1.20e+2 1.21e+2 1.22e+2 1.23e+2 1.24e+2 │ 1.25e+2 1.26e+2 1.27e+2 1.28e+2 1.29e+2 │ 1.30e+2 1.31e+2 1.32e+2 1.33e+2 1.34e+2 │ 1.35e+2 1.36e+2 1.37e+2 1.38e+2 1.39e+2 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││1 @ 2 │ 1.40e+2 1.41e+2 1.42e+2 1.43e+2 1.44e+2 │ 1.45e+2 1.46e+2 1.47e+2 1.48e+2 1.49e+2 │ 1.50e+2 1.51e+2 1.52e+2 1.53e+2 1.54e+2 │ 1.55e+2 1.56e+2 1.57e+2 1.58e+2 1.59e+2 ││ - ││axis 3│ 1.60e+2 1.61e+2 1.62e+2 1.63e+2 1.64e+2 │ 1.65e+2 1.66e+2 1.67e+2 1.68e+2 1.69e+2 │ 1.70e+2 1.71e+2 1.72e+2 1.73e+2 1.74e+2 │ 1.75e+2 1.76e+2 1.77e+2 1.78e+2 1.79e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.40e+2 2.41e+2 2.42e+2 2.43e+2 2.44e+2 │ 2.45e+2 2.46e+2 2.47e+2 2.48e+2 2.49e+2 │ 2.50e+2 2.51e+2 2.52e+2 2.53e+2 2.54e+2 │ 2.55e+2 2.56e+2 2.57e+2 2.58e+2 2.59e+2 ││ - ││ │ 2.60e+2 2.61e+2 2.62e+2 2.63e+2 2.64e+2 │ 2.65e+2 2.66e+2 2.67e+2 2.68e+2 2.69e+2 │ 2.70e+2 2.71e+2 2.72e+2 2.73e+2 2.74e+2 │ 2.75e+2 2.76e+2 2.77e+2 2.78e+2 2.79e+2 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... ││ - ││axis 3│ │ │ │ ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││4 @ 2 │ 5.60e+2 5.61e+2 5.62e+2 5.63e+2 5.64e+2 │ 5.65e+2 5.66e+2 5.67e+2 5.68e+2 5.69e+2 │ 5.70e+2 5.71e+2 5.72e+2 5.73e+2 5.74e+2 │ 5.75e+2 5.76e+2 5.77e+2 5.78e+2 5.79e+2 ││ - ││axis 3│ 5.80e+2 5.81e+2 5.82e+2 5.83e+2 5.84e+2 │ 5.85e+2 5.86e+2 5.87e+2 5.88e+2 5.89e+2 │ 5.90e+2 5.91e+2 5.92e+2 5.93e+2 5.94e+2 │ 5.95e+2 5.96e+2 5.97e+2 5.98e+2 5.99e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 6.60e+2 6.61e+2 6.62e+2 6.63e+2 6.64e+2 │ 6.65e+2 6.66e+2 6.67e+2 6.68e+2 6.69e+2 │ 6.70e+2 6.71e+2 6.72e+2 6.73e+2 6.74e+2 │ 6.75e+2 6.76e+2 6.77e+2 6.78e+2 6.79e+2 ││ - ││ │ 6.80e+2 6.81e+2 6.82e+2 6.83e+2 6.84e+2 │ 6.85e+2 6.86e+2 6.87e+2 6.88e+2 6.89e+2 │ 6.90e+2 6.91e+2 6.92e+2 6.93e+2 6.94e+2 │ 6.95e+2 6.96e+2 6.97e+2 6.98e+2 6.99e+2 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││5 @ 2 │ 7.00e+2 7.01e+2 7.02e+2 7.03e+2 7.04e+2 │ 7.05e+2 7.06e+2 7.07e+2 7.08e+2 7.09e+2 │ 7.10e+2 7.11e+2 7.12e+2 7.13e+2 7.14e+2 │ 7.15e+2 7.16e+2 7.17e+2 7.18e+2 7.19e+2 ││ - ││axis 3│ 7.20e+2 7.21e+2 7.22e+2 7.23e+2 7.24e+2 │ 7.25e+2 7.26e+2 7.27e+2 7.28e+2 7.29e+2 │ 7.30e+2 7.31e+2 7.32e+2 7.33e+2 7.34e+2 │ 7.35e+2 7.36e+2 7.37e+2 7.38e+2 7.39e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 8.00e+2 8.01e+2 8.02e+2 8.03e+2 8.04e+2 │ 8.05e+2 8.06e+2 8.07e+2 8.08e+2 8.09e+2 │ 8.10e+2 8.11e+2 8.12e+2 8.13e+2 8.14e+2 │ 8.15e+2 8.16e+2 8.17e+2 8.18e+2 8.19e+2 ││ - ││ │ 8.20e+2 8.21e+2 8.22e+2 8.23e+2 8.24e+2 │ 8.25e+2 8.26e+2 8.27e+2 8.28e+2 8.29e+2 │ 8.30e+2 8.31e+2 8.32e+2 8.33e+2 8.34e+2 │ 8.35e+2 8.36e+2 8.37e+2 8.38e+2 8.39e+2 ││ - │└──────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┘│ - ├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │┌──────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┐│ - ││1 @ 1 │0 @ 4 │1 @ 4 │2 @ 4 │3 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││0 @ 2 │ 8.40e+2 8.41e+2 8.42e+2 8.43e+2 8.44e+2 │ 8.45e+2 8.46e+2 8.47e+2 8.48e+2 8.49e+2 │ 8.50e+2 8.51e+2 8.52e+2 8.53e+2 8.54e+2 │ 8.55e+2 8.56e+2 8.57e+2 8.58e+2 8.59e+2 ││ - ││axis 3│ 8.60e+2 8.61e+2 8.62e+2 8.63e+2 8.64e+2 │ 8.65e+2 8.66e+2 8.67e+2 8.68e+2 8.69e+2 │ 8.70e+2 8.71e+2 8.72e+2 8.73e+2 8.74e+2 │ 8.75e+2 8.76e+2 8.77e+2 8.78e+2 8.79e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 9.40e+2 9.41e+2 9.42e+2 9.43e+2 9.44e+2 │ 9.45e+2 9.46e+2 9.47e+2 9.48e+2 9.49e+2 │ 9.50e+2 9.51e+2 9.52e+2 9.53e+2 9.54e+2 │ 9.55e+2 9.56e+2 9.57e+2 9.58e+2 9.59e+2 ││ - ││ │ 9.60e+2 9.61e+2 9.62e+2 9.63e+2 9.64e+2 │ 9.65e+2 9.66e+2 9.67e+2 9.68e+2 9.69e+2 │ 9.70e+2 9.71e+2 9.72e+2 9.73e+2 9.74e+2 │ 9.75e+2 9.76e+2 9.77e+2 9.78e+2 9.79e+2 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││1 @ 2 │ 9.80e+2 9.81e+2 9.82e+2 9.83e+2 9.84e+2 │ 9.85e+2 9.86e+2 9.87e+2 9.88e+2 9.89e+2 │ 9.90e+2 9.91e+2 9.92e+2 9.93e+2 9.94e+2 │ 9.95e+2 9.96e+2 9.97e+2 9.98e+2 9.99e+2 ││ - ││axis 3│ 1.00e+3 1.00e+3 1.00e+3 1.00e+3 1.00e+3 │ 1.00e+3 1.00e+3 1.00e+3 1.00e+3 1.00e+3 │ 1.01e+3 1.01e+3 1.01e+3 1.01e+3 1.01e+3 │ 1.01e+3 1.01e+3 1.01e+3 1.01e+3 1.01e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.08e+3 1.08e+3 1.08e+3 1.08e+3 1.08e+3 │ 1.08e+3 1.08e+3 1.08e+3 1.08e+3 1.08e+3 │ 1.09e+3 1.09e+3 1.09e+3 1.09e+3 1.09e+3 │ 1.09e+3 1.09e+3 1.09e+3 1.09e+3 1.09e+3 ││ - ││ │ 1.10e+3 1.10e+3 1.10e+3 1.10e+3 1.10e+3 │ 1.10e+3 1.10e+3 1.10e+3 1.10e+3 1.10e+3 │ 1.11e+3 1.11e+3 1.11e+3 1.11e+3 1.11e+3 │ 1.11e+3 1.11e+3 1.11e+3 1.11e+3 1.11e+3 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... ││ - ││axis 3│ │ │ │ ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││4 @ 2 │ 1.40e+3 1.40e+3 1.40e+3 1.40e+3 1.40e+3 │ 1.40e+3 1.40e+3 1.40e+3 1.40e+3 1.40e+3 │ 1.41e+3 1.41e+3 1.41e+3 1.41e+3 1.41e+3 │ 1.41e+3 1.41e+3 1.41e+3 1.41e+3 1.41e+3 ││ - ││axis 3│ 1.42e+3 1.42e+3 1.42e+3 1.42e+3 1.42e+3 │ 1.42e+3 1.42e+3 1.42e+3 1.42e+3 1.42e+3 │ 1.43e+3 1.43e+3 1.43e+3 1.43e+3 1.43e+3 │ 1.43e+3 1.43e+3 1.43e+3 1.43e+3 1.43e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.50e+3 1.50e+3 1.50e+3 1.50e+3 1.50e+3 │ 1.50e+3 1.50e+3 1.50e+3 1.50e+3 1.50e+3 │ 1.51e+3 1.51e+3 1.51e+3 1.51e+3 1.51e+3 │ 1.51e+3 1.51e+3 1.51e+3 1.51e+3 1.51e+3 ││ - ││ │ 1.52e+3 1.52e+3 1.52e+3 1.52e+3 1.52e+3 │ 1.52e+3 1.52e+3 1.52e+3 1.52e+3 1.52e+3 │ 1.53e+3 1.53e+3 1.53e+3 1.53e+3 1.53e+3 │ 1.53e+3 1.53e+3 1.53e+3 1.53e+3 1.53e+3 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││5 @ 2 │ 1.54e+3 1.54e+3 1.54e+3 1.54e+3 1.54e+3 │ 1.54e+3 1.54e+3 1.54e+3 1.54e+3 1.54e+3 │ 1.55e+3 1.55e+3 1.55e+3 1.55e+3 1.55e+3 │ 1.55e+3 1.55e+3 1.55e+3 1.55e+3 1.55e+3 ││ - ││axis 3│ 1.56e+3 1.56e+3 1.56e+3 1.56e+3 1.56e+3 │ 1.56e+3 1.56e+3 1.56e+3 1.56e+3 1.56e+3 │ 1.57e+3 1.57e+3 1.57e+3 1.57e+3 1.57e+3 │ 1.57e+3 1.57e+3 1.57e+3 1.57e+3 1.57e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.64e+3 1.64e+3 1.64e+3 1.64e+3 1.64e+3 │ 1.64e+3 1.64e+3 1.64e+3 1.64e+3 1.64e+3 │ 1.65e+3 1.65e+3 1.65e+3 1.65e+3 1.65e+3 │ 1.65e+3 1.65e+3 1.65e+3 1.65e+3 1.65e+3 ││ - ││ │ 1.66e+3 1.66e+3 1.66e+3 1.66e+3 1.66e+3 │ 1.66e+3 1.66e+3 1.66e+3 1.66e+3 1.66e+3 │ 1.67e+3 1.67e+3 1.67e+3 1.67e+3 1.67e+3 │ 1.67e+3 1.67e+3 1.67e+3 1.67e+3 1.67e+3 ││ - │└──────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┘│ - ├────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┤ - │┌──────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┐│ - ││2 @ 1 │0 @ 4 │1 @ 4 │2 @ 4 │3 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││0 @ 2 │ 1.68e+3 1.68e+3 1.68e+3 1.68e+3 1.68e+3 │ 1.68e+3 1.68e+3 1.68e+3 1.68e+3 1.68e+3 │ 1.69e+3 1.69e+3 1.69e+3 1.69e+3 1.69e+3 │ 1.69e+3 1.69e+3 1.69e+3 1.69e+3 1.69e+3 ││ - ││axis 3│ 1.70e+3 1.70e+3 1.70e+3 1.70e+3 1.70e+3 │ 1.70e+3 1.70e+3 1.70e+3 1.70e+3 1.70e+3 │ 1.71e+3 1.71e+3 1.71e+3 1.71e+3 1.71e+3 │ 1.71e+3 1.71e+3 1.71e+3 1.71e+3 1.71e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.78e+3 1.78e+3 1.78e+3 1.78e+3 1.78e+3 │ 1.78e+3 1.78e+3 1.78e+3 1.78e+3 1.78e+3 │ 1.79e+3 1.79e+3 1.79e+3 1.79e+3 1.79e+3 │ 1.79e+3 1.79e+3 1.79e+3 1.79e+3 1.79e+3 ││ - ││ │ 1.80e+3 1.80e+3 1.80e+3 1.80e+3 1.80e+3 │ 1.80e+3 1.80e+3 1.80e+3 1.80e+3 1.80e+3 │ 1.81e+3 1.81e+3 1.81e+3 1.81e+3 1.81e+3 │ 1.81e+3 1.81e+3 1.81e+3 1.81e+3 1.81e+3 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││1 @ 2 │ 1.82e+3 1.82e+3 1.82e+3 1.82e+3 1.82e+3 │ 1.82e+3 1.82e+3 1.82e+3 1.82e+3 1.82e+3 │ 1.83e+3 1.83e+3 1.83e+3 1.83e+3 1.83e+3 │ 1.83e+3 1.83e+3 1.83e+3 1.83e+3 1.83e+3 ││ - ││axis 3│ 1.84e+3 1.84e+3 1.84e+3 1.84e+3 1.84e+3 │ 1.84e+3 1.84e+3 1.84e+3 1.84e+3 1.84e+3 │ 1.85e+3 1.85e+3 1.85e+3 1.85e+3 1.85e+3 │ 1.85e+3 1.85e+3 1.85e+3 1.85e+3 1.85e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 1.92e+3 1.92e+3 1.92e+3 1.92e+3 1.92e+3 │ 1.92e+3 1.92e+3 1.92e+3 1.92e+3 1.92e+3 │ 1.93e+3 1.93e+3 1.93e+3 1.93e+3 1.93e+3 │ 1.93e+3 1.93e+3 1.93e+3 1.93e+3 1.93e+3 ││ - ││ │ 1.94e+3 1.94e+3 1.94e+3 1.94e+3 1.94e+3 │ 1.94e+3 1.94e+3 1.94e+3 1.94e+3 1.94e+3 │ 1.95e+3 1.95e+3 1.95e+3 1.95e+3 1.95e+3 │ 1.95e+3 1.95e+3 1.95e+3 1.95e+3 1.95e+3 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││~~~~~ │ ... │ ... │ ... │ ... ││ - ││axis 3│ │ │ │ ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││4 @ 2 │ 2.24e+3 2.24e+3 2.24e+3 2.24e+3 2.24e+3 │ 2.24e+3 2.24e+3 2.24e+3 2.24e+3 2.24e+3 │ 2.25e+3 2.25e+3 2.25e+3 2.25e+3 2.25e+3 │ 2.25e+3 2.25e+3 2.25e+3 2.25e+3 2.25e+3 ││ - ││axis 3│ 2.26e+3 2.26e+3 2.26e+3 2.26e+3 2.26e+3 │ 2.26e+3 2.26e+3 2.26e+3 2.26e+3 2.26e+3 │ 2.27e+3 2.27e+3 2.27e+3 2.27e+3 2.27e+3 │ 2.27e+3 2.27e+3 2.27e+3 2.27e+3 2.27e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.34e+3 2.34e+3 2.34e+3 2.34e+3 2.34e+3 │ 2.34e+3 2.34e+3 2.34e+3 2.34e+3 2.34e+3 │ 2.35e+3 2.35e+3 2.35e+3 2.35e+3 2.35e+3 │ 2.35e+3 2.35e+3 2.35e+3 2.35e+3 2.35e+3 ││ - ││ │ 2.36e+3 2.36e+3 2.36e+3 2.36e+3 2.36e+3 │ 2.36e+3 2.36e+3 2.36e+3 2.36e+3 2.36e+3 │ 2.37e+3 2.37e+3 2.37e+3 2.37e+3 2.37e+3 │ 2.37e+3 2.37e+3 2.37e+3 2.37e+3 2.37e+3 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ - ││5 @ 2 │ 2.38e+3 2.38e+3 2.38e+3 2.38e+3 2.38e+3 │ 2.38e+3 2.38e+3 2.38e+3 2.38e+3 2.38e+3 │ 2.39e+3 2.39e+3 2.39e+3 2.39e+3 2.39e+3 │ 2.39e+3 2.39e+3 2.39e+3 2.39e+3 2.39e+3 ││ - ││axis 3│ 2.40e+3 2.40e+3 2.40e+3 2.40e+3 2.40e+3 │ 2.40e+3 2.40e+3 2.40e+3 2.40e+3 2.40e+3 │ 2.41e+3 2.41e+3 2.41e+3 2.41e+3 2.41e+3 │ 2.41e+3 2.41e+3 2.41e+3 2.41e+3 2.41e+3 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 2.48e+3 2.48e+3 2.48e+3 2.48e+3 2.48e+3 │ 2.48e+3 2.48e+3 2.48e+3 2.48e+3 2.48e+3 │ 2.49e+3 2.49e+3 2.49e+3 2.49e+3 2.49e+3 │ 2.49e+3 2.49e+3 2.49e+3 2.49e+3 2.49e+3 ││ - ││ │ 2.50e+3 2.50e+3 2.50e+3 2.50e+3 2.50e+3 │ 2.50e+3 2.50e+3 2.50e+3 2.50e+3 2.50e+3 │ 2.51e+3 2.51e+3 2.51e+3 2.51e+3 2.51e+3 │ 2.51e+3 2.51e+3 2.51e+3 2.51e+3 2.51e+3 ││ - │└──────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┴─────────────────────────────────────────────┘│ - └────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [2]: r2x3x6x7x4x5 shape 0:2,1:3|4:4,5:5->2:6,3:7 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho2; [%expect {| @@ -292,20 +180,7 @@ let%expect_test "einsum1 sum out axes" = Train.forward_and_forget backend ctx ho; Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[0]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho; [%expect {| @@ -364,28 +239,10 @@ let%expect_test "einsum outer product" = Train.forward_and_forget backend ctx c; Tensor.print ~with_code:false ~with_grad:false `Default @@ a; [%expect - {| - ┌──────────────────┐ - │[0]: r2 shape 0:2 │ - │┌┬────────────┐ │ - │││axis 0 │ │ - │├┼────────────┤ │ - │││ 0.00 1.00 │ │ - │└┴────────────┘ │ - └──────────────────┘ - |}]; + {| [0]: r2 shape 0:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌─────────────────────┐ - │[1]: r3 shape 0:3 │ - │┌┬──────────────────┐│ - │││axis 0 ││ - │├┼──────────────────┤│ - │││ 0.00 1.00 2.00 ││ - │└┴──────────────────┘│ - └─────────────────────┘ - |}]; + {| [1]: r3 shape 0:3 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| @@ -406,37 +263,10 @@ let%expect_test "einsum outer product" = Train.forward_and_forget backend ctx c; Tensor.print ~with_code:false ~with_grad:false `Default @@ a; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[5]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [5]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[6]: r5x7x6 shape 0:5|2:6->1:7 │ - │┌──────┬─────────────────────────────────────────┬─────────────────────────────────────────┬─────────────────────────────────────────┬─────────────────────────────────────────┬─────────────────────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 │2 @ 0 │3 @ 0 │4 @ 0 ││ - ││ │axis 2 │axis 2 │axis 2 │axis 2 │axis 2 ││ - │├──────┼─────────────────────────────────────────┼─────────────────────────────────────────┼─────────────────────────────────────────┼─────────────────────────────────────────┼─────────────────────────────────────────┤│ - ││axis 1│ 0.00 1.00 ... 4.00 5.00 │ 4.20e+1 4.30e+1 ... 4.60e+1 4.70e+1 │ 8.40e+1 8.50e+1 ... 8.80e+1 8.90e+1 │ 1.26e+2 1.27e+2 ... 1.30e+2 1.31e+2 │ 1.68e+2 1.69e+2 ... 1.72e+2 1.73e+2 ││ - ││ │ 6.00 7.00 ... 1.00e+1 1.10e+1 │ 4.80e+1 4.90e+1 ... 5.20e+1 5.30e+1 │ 9.00e+1 9.10e+1 ... 9.40e+1 9.50e+1 │ 1.32e+2 1.33e+2 ... 1.36e+2 1.37e+2 │ 1.74e+2 1.75e+2 ... 1.78e+2 1.79e+2 ││ - ││ │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... │ ... ... ... ... ... ││ - ││ │ 3.00e+1 3.10e+1 ... 3.40e+1 3.50e+1 │ 7.20e+1 7.30e+1 ... 7.60e+1 7.70e+1 │ 1.14e+2 1.15e+2 ... 1.18e+2 1.19e+2 │ 1.56e+2 1.57e+2 ... 1.60e+2 1.61e+2 │ 1.98e+2 1.99e+2 ... 2.02e+2 2.03e+2 ││ - ││ │ 3.60e+1 3.70e+1 ... 4.00e+1 4.10e+1 │ 7.80e+1 7.90e+1 ... 8.20e+1 8.30e+1 │ 1.20e+2 1.21e+2 ... 1.24e+2 1.25e+2 │ 1.62e+2 1.63e+2 ... 1.66e+2 1.67e+2 │ 2.04e+2 2.05e+2 ... 2.08e+2 2.09e+2 ││ - │└──────┴─────────────────────────────────────────┴─────────────────────────────────────────┴─────────────────────────────────────────┴─────────────────────────────────────────┴─────────────────────────────────────────┘│ - └──────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [6]: r5x7x6 shape 0:5|2:6->1:7 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| @@ -723,20 +553,7 @@ let%expect_test "einsum1 broadcast or sum out prefix axes" = let ctx = Train.forward_and_ctx backend ctx ho in Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[0]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho; [%expect {| @@ -776,106 +593,7 @@ let%expect_test "einsum1 broadcast or sum out prefix axes" = let ctx = Train.forward_and_ctx backend ctx ho3 in Tensor.print ~with_code:false ~with_grad:false `Default @@ hey2; [%expect - {| - ┌────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[3]: r2x3x6x7x4x5 shape 0:2,1:3|4:4,5:5->2:6,3:7 │ - │┌──────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┬─────────────────────────────────────────────┐│ - ││0 @ 1 │0 @ 4 │1 @ 4 │2 @ 4 │3 @ 4 ││ - ││ │axis 5 │axis 5 │axis 5 │axis 5 ││ - │├──────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┼─────────────────────────────────────────────┤│ ...TRUNCATED BY DUNE... - │├──────┼────────────────────────┼────────────────────────────────────┼────────────────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 3.00 │ 8.00 9.00 1.00e+1 1.10e+1 │ 1.60e+1 1.70e+1 1.80e+1 1.90e+1 ││ - ││ │ 4.00 5.00 6.00 7.00 │ 1.20e+1 1.30e+1 1.40e+1 1.50e+1 │ 2.00e+1 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴────────────────────────────────────┴────────────────────────────────────┘│ - └───────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r3x2x4 shape 0:3|2:4->1:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌────────────────────────────────┐ - │[1]: r3x4x1 shape 0:3|2:1->1:4 │ - │┌──────┬──────┬──────┬─────────┐│ - ││ │0 @ 0 │1 @ 0 │2 @ 0 ││ - ││ │axis 2│axis 2│axis 2 ││ - │├──────┼──────┼──────┼─────────┤│ - ││axis 1│ 0.00 │ 4.00 │ 8.00 ││ - ││ │ 1.00 │ 5.00 │ 9.00 ││ - ││ │ 2.00 │ 6.00 │ 1.00e+1 ││ - ││ │ 3.00 │ 7.00 │ 1.10e+1 ││ - │└──────┴──────┴──────┴─────────┘│ - └────────────────────────────────┘ - |}]; + {| [1]: r3x4x1 shape 0:3|2:1->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| @@ -1188,32 +845,10 @@ let%expect_test "einsum broadcast or sum out prefix axes" = Train.forward_and_forget backend ctx f; Tensor.print ~with_code:false ~with_grad:false `Default @@ d; [%expect - {| - ┌─────────────────────────┐ - │[3]: r3x2 shape 1:2->0:3 │ - │┌──────┬────────────┐ │ - ││ │axis 1 │ │ - │├──────┼────────────┤ │ - ││axis 0│ 0.00 1.00 │ │ - ││ │ 2.00 3.00 │ │ - ││ │ 4.00 5.00 │ │ - │└──────┴────────────┘ │ - └─────────────────────────┘ - |}]; + {| [3]: r3x2 shape 1:2->0:3 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ e; [%expect - {| - ┌───────────────────────────────────────┐ - │[4]: r3x4 shape 1:4->0:3 │ - │┌──────┬──────────────────────────────┐│ - ││ │axis 1 ││ - │├──────┼──────────────────────────────┤│ - ││axis 0│ 0.00 1.00 2.00 3.00 ││ - ││ │ 4.00 5.00 6.00 7.00 ││ - ││ │ 8.00 9.00 1.00e+1 1.10e+1 ││ - │└──────┴──────────────────────────────┘│ - └───────────────────────────────────────┘ - |}]; + {| [4]: r3x4 shape 1:4->0:3 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ f; [%expect {| @@ -1253,20 +888,7 @@ let%expect_test "einsum1 fixed dim axis" = let ctx = Train.forward_and_ctx backend ctx ho in Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[0]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho; [%expect {| @@ -1284,20 +906,7 @@ let%expect_test "einsum1 fixed dim axis" = let ctx = Train.forward_and_ctx backend ctx ho2 in Tensor.print ~with_code:false ~with_grad:false `Default @@ hey; [%expect - {| - ┌─────────────────────────────────────────────────────────────┐ - │[0]: r2x4x3 shape 0:2|2:3->1:4 │ - │┌──────┬────────────────────────┬───────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 ││ - ││ │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 1.20e+1 1.30e+1 1.40e+1 ││ - ││ │ 3.00 4.00 5.00 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 6.00 7.00 8.00 │ 1.80e+1 1.90e+1 2.00e+1 ││ - ││ │ 9.00 1.00e+1 1.10e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴───────────────────────────┘│ - └─────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r2x4x3 shape 0:2|2:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho2; [%expect {| @@ -1316,18 +925,7 @@ let%expect_test "einsum1 fixed dim axis" = let ctx = Train.forward_and_ctx backend ctx ho3 in Tensor.print ~with_code:false ~with_grad:false `Default @@ hey2; [%expect - {| - ┌─────────────────────────┐ - │[3]: r3x2 shape 1:2->0:3 │ - │┌──────┬────────────┐ │ - ││ │axis 1 │ │ - │├──────┼────────────┤ │ - ││axis 0│ 0.00 1.00 │ │ - ││ │ 2.00 3.00 │ │ - ││ │ 4.00 5.00 │ │ - │└──────┴────────────┘ │ - └─────────────────────────┘ - |}]; + {| [3]: r3x2 shape 1:2->0:3 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ho3; [%expect {| @@ -1379,34 +977,10 @@ let%expect_test "einsum with fixed dim axes" = Train.forward_and_forget backend ctx c; Tensor.print ~with_code:false ~with_grad:false `Default @@ a; [%expect - {| - ┌───────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[0]: r3x2x4 shape 0:3|2:4->1:2 │ - │┌──────┬────────────────────────┬────────────────────────────────────┬────────────────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 │2 @ 0 ││ - ││ │axis 2 │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼────────────────────────────────────┼────────────────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 3.00 │ 8.00 9.00 1.00e+1 1.10e+1 │ 1.60e+1 1.70e+1 1.80e+1 1.90e+1 ││ - ││ │ 4.00 5.00 6.00 7.00 │ 1.20e+1 1.30e+1 1.40e+1 1.50e+1 │ 2.00e+1 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴────────────────────────────────────┴────────────────────────────────────┘│ - └───────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r3x2x4 shape 0:3|2:4->1:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌────────────────────────────────┐ - │[1]: r3x4x1 shape 0:3|2:1->1:4 │ - │┌──────┬──────┬──────┬─────────┐│ - ││ │0 @ 0 │1 @ 0 │2 @ 0 ││ - ││ │axis 2│axis 2│axis 2 ││ - │├──────┼──────┼──────┼─────────┤│ - ││axis 1│ 0.00 │ 4.00 │ 8.00 ││ - ││ │ 1.00 │ 5.00 │ 9.00 ││ - ││ │ 2.00 │ 6.00 │ 1.00e+1 ││ - ││ │ 3.00 │ 7.00 │ 1.10e+1 ││ - │└──────┴──────┴──────┴─────────┘│ - └────────────────────────────────┘ - |}]; + {| [1]: r3x4x1 shape 0:3|2:1->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| @@ -1450,144 +1024,144 @@ let%expect_test "outer_sum simulating axis concatenation" = Tensor.print ~with_code:false ~with_grad:false `Default @@ positions; [%expect {| - ┌────────────────────────────────┐ - │[9]: ;=>+ shape 0:4,1:5,2:6,3:3 │ - │┌──────┬──────────────────┐ │ - ││0 @ 0 │axis 3 │ │ - │├──────┼──────────────────┤ │ - ││0 @ 1 │ 0.00 0.00 0.00 │ │ - ││axis 2│ 0.00 0.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 0.00 0.00 4.00 │ │ - ││ │ 0.00 0.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││1 @ 1 │ 0.00 1.00 0.00 │ │ - ││axis 2│ 0.00 1.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 0.00 1.00 4.00 │ │ - ││ │ 0.00 1.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││2 @ 1 │ 0.00 2.00 0.00 │ │ - ││axis 2│ 0.00 2.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 0.00 2.00 4.00 │ │ - ││ │ 0.00 2.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││3 @ 1 │ 0.00 3.00 0.00 │ │ - ││axis 2│ 0.00 3.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 0.00 3.00 4.00 │ │ - ││ │ 0.00 3.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││4 @ 1 │ 0.00 4.00 0.00 │ │ - ││axis 2│ 0.00 4.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 0.00 4.00 4.00 │ │ - ││ │ 0.00 4.00 5.00 │ │ - │└──────┴──────────────────┘ │ - ├────────────────────────────────┤ - │┌──────┬──────────────────┐ │ - ││1 @ 0 │axis 3 │ │ - │├──────┼──────────────────┤ │ - ││0 @ 1 │ 1.00 0.00 0.00 │ │ - ││axis 2│ 1.00 0.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 1.00 0.00 4.00 │ │ - ││ │ 1.00 0.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││1 @ 1 │ 1.00 1.00 0.00 │ │ - ││axis 2│ 1.00 1.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 1.00 1.00 4.00 │ │ - ││ │ 1.00 1.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││2 @ 1 │ 1.00 2.00 0.00 │ │ - ││axis 2│ 1.00 2.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 1.00 2.00 4.00 │ │ - ││ │ 1.00 2.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││3 @ 1 │ 1.00 3.00 0.00 │ │ - ││axis 2│ 1.00 3.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 1.00 3.00 4.00 │ │ - ││ │ 1.00 3.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││4 @ 1 │ 1.00 4.00 0.00 │ │ - ││axis 2│ 1.00 4.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 1.00 4.00 4.00 │ │ - ││ │ 1.00 4.00 5.00 │ │ - │└──────┴──────────────────┘ │ - ├────────────────────────────────┤ - │┌──────┬──────────────────┐ │ - ││2 @ 0 │axis 3 │ │ - │├──────┼──────────────────┤ │ - ││0 @ 1 │ 2.00 0.00 0.00 │ │ - ││axis 2│ 2.00 0.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 2.00 0.00 4.00 │ │ - ││ │ 2.00 0.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││1 @ 1 │ 2.00 1.00 0.00 │ │ - ││axis 2│ 2.00 1.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 2.00 1.00 4.00 │ │ - ││ │ 2.00 1.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││2 @ 1 │ 2.00 2.00 0.00 │ │ - ││axis 2│ 2.00 2.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 2.00 2.00 4.00 │ │ - ││ │ 2.00 2.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││3 @ 1 │ 2.00 3.00 0.00 │ │ - ││axis 2│ 2.00 3.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 2.00 3.00 4.00 │ │ - ││ │ 2.00 3.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││4 @ 1 │ 2.00 4.00 0.00 │ │ - ││axis 2│ 2.00 4.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 2.00 4.00 4.00 │ │ - ││ │ 2.00 4.00 5.00 │ │ - │└──────┴──────────────────┘ │ - ├────────────────────────────────┤ - │┌──────┬──────────────────┐ │ - ││3 @ 0 │axis 3 │ │ - │├──────┼──────────────────┤ │ - ││0 @ 1 │ 3.00 0.00 0.00 │ │ - ││axis 2│ 3.00 0.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 3.00 0.00 4.00 │ │ - ││ │ 3.00 0.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││1 @ 1 │ 3.00 1.00 0.00 │ │ - ││axis 2│ 3.00 1.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 3.00 1.00 4.00 │ │ - ││ │ 3.00 1.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││2 @ 1 │ 3.00 2.00 0.00 │ │ - ││axis 2│ 3.00 2.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 3.00 2.00 4.00 │ │ - ││ │ 3.00 2.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││3 @ 1 │ 3.00 3.00 0.00 │ │ - ││axis 2│ 3.00 3.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 3.00 3.00 4.00 │ │ - ││ │ 3.00 3.00 5.00 │ │ - │├──────┼──────────────────┤ │ - ││4 @ 1 │ 3.00 4.00 0.00 │ │ - ││axis 2│ 3.00 4.00 1.00 │ │ - ││ │ ... ... ... │ │ - ││ │ 3.00 4.00 4.00 │ │ - ││ │ 3.00 4.00 5.00 │ │ - │└──────┴──────────────────┘ │ - └────────────────────────────────┘ + ┌───────────────────────────────────┐ + │[9]: ;=>+ shape 0:4,1:5,2:6,3:3 │ + │┌──────┬──────────────────────────┐│ + ││0 @ 0 │axis 3 ││ + │├──────┼──────────────────────────┤│ + ││0 @ 1 │ 5.10e+34 4.19e-41 0.00 ││ + ││axis 2│ 0.00 0.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 0.00 0.00 4.00 ││ + ││ │ 0.00 0.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││1 @ 1 │ 5.10e+34 1.00 0.00 ││ + ││axis 2│ 0.00 1.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 0.00 1.00 4.00 ││ + ││ │ 0.00 1.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││2 @ 1 │ 5.10e+34 2.00 0.00 ││ + ││axis 2│ 0.00 2.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 0.00 2.00 4.00 ││ + ││ │ 0.00 2.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││3 @ 1 │ 5.10e+34 3.00 0.00 ││ + ││axis 2│ 0.00 3.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 0.00 3.00 4.00 ││ + ││ │ 0.00 3.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││4 @ 1 │ 5.10e+34 4.00 1.68e-44 ││ + ││axis 2│ 9.49e-39 4.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 9.49e-39 4.00 4.00 ││ + ││ │ 9.49e-39 4.00 5.00 ││ + │└──────┴──────────────────────────┘│ + ├───────────────────────────────────┤ + │┌──────┬──────────────────────────┐│ + ││1 @ 0 │axis 3 ││ + │├──────┼──────────────────────────┤│ + ││0 @ 1 │ 5.10e+34 4.19e-41 0.00 ││ + ││axis 2│ 1.00 0.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 1.00 0.00 4.00 ││ + ││ │ 1.00 0.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││1 @ 1 │ 5.10e+34 1.00 0.00 ││ + ││axis 2│ 1.00 1.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 1.00 1.00 4.00 ││ + ││ │ 1.00 1.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││2 @ 1 │ 5.10e+34 2.00 0.00 ││ + ││axis 2│ 1.00 2.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 1.00 2.00 4.00 ││ + ││ │ 1.00 2.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││3 @ 1 │ 5.10e+34 3.00 0.00 ││ + ││axis 2│ 1.00 3.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 1.00 3.00 4.00 ││ + ││ │ 1.00 3.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││4 @ 1 │ 5.10e+34 4.00 1.68e-44 ││ + ││axis 2│ 1.00 4.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 1.00 4.00 4.00 ││ + ││ │ 1.00 4.00 5.00 ││ + │└──────┴──────────────────────────┘│ + ├───────────────────────────────────┤ + │┌──────┬──────────────────────────┐│ + ││2 @ 0 │axis 3 ││ + │├──────┼──────────────────────────┤│ + ││0 @ 1 │ 5.10e+34 4.19e-41 0.00 ││ + ││axis 2│ 2.00 0.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 2.00 0.00 4.00 ││ + ││ │ 2.00 0.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││1 @ 1 │ 5.10e+34 1.00 0.00 ││ + ││axis 2│ 2.00 1.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 2.00 1.00 4.00 ││ + ││ │ 2.00 1.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││2 @ 1 │ 5.10e+34 2.00 0.00 ││ + ││axis 2│ 2.00 2.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 2.00 2.00 4.00 ││ + ││ │ 2.00 2.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││3 @ 1 │ 5.10e+34 3.00 0.00 ││ + ││axis 2│ 2.00 3.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 2.00 3.00 4.00 ││ + ││ │ 2.00 3.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││4 @ 1 │ 5.10e+34 4.00 1.68e-44 ││ + ││axis 2│ 2.00 4.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 2.00 4.00 4.00 ││ + ││ │ 2.00 4.00 5.00 ││ + │└──────┴──────────────────────────┘│ + ├───────────────────────────────────┤ + │┌──────┬──────────────────────────┐│ + ││3 @ 0 │axis 3 ││ + │├──────┼──────────────────────────┤│ + ││0 @ 1 │ 5.10e+34 4.19e-41 0.00 ││ + ││axis 2│ 3.00 0.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 3.00 0.00 4.00 ││ + ││ │ 3.00 0.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││1 @ 1 │ 5.10e+34 1.00 0.00 ││ + ││axis 2│ 3.00 1.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 3.00 1.00 4.00 ││ + ││ │ 3.00 1.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││2 @ 1 │ 5.10e+34 2.00 0.00 ││ + ││axis 2│ 3.00 2.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 3.00 2.00 4.00 ││ + ││ │ 3.00 2.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││3 @ 1 │ 5.10e+34 3.00 0.00 ││ + ││axis 2│ 3.00 3.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 3.00 3.00 4.00 ││ + ││ │ 3.00 3.00 5.00 ││ + │├──────┼──────────────────────────┤│ + ││4 @ 1 │ 5.10e+34 4.00 1.68e-44 ││ + ││axis 2│ 3.00 4.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 3.00 4.00 4.00 ││ + ││ │ 3.00 4.00 5.00 ││ + │└──────┴──────────────────────────┘│ + └───────────────────────────────────┘ |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ ti; [%expect @@ -1607,18 +1181,18 @@ let%expect_test "outer_sum simulating axis concatenation" = Tensor.print ~with_code:false ~with_grad:false `Default @@ tk; [%expect {| - ┌───────────────────────────┐ - │[7]: =>_tk shape 0:6,1:3 │ - │┌──────┬──────────────────┐│ - ││ │axis 1 ││ - │├──────┼──────────────────┤│ - ││axis 0│ 0.00 0.00 0.00 ││ - ││ │ 0.00 0.00 1.00 ││ - ││ │ ... ... ... ││ - ││ │ 0.00 0.00 4.00 ││ - ││ │ 0.00 0.00 5.00 ││ - │└──────┴──────────────────┘│ - └───────────────────────────┘ + ┌───────────────────────────────────┐ + │[7]: =>_tk shape 0:6,1:3 │ + │┌──────┬──────────────────────────┐│ + ││ │axis 1 ││ + │├──────┼──────────────────────────┤│ + ││axis 0│ 5.10e+34 4.19e-41 0.00 ││ + ││ │ 0.00 0.00 1.00 ││ + ││ │ ... ... ... ││ + ││ │ 0.00 0.00 4.00 ││ + ││ │ 0.00 0.00 5.00 ││ + │└──────┴──────────────────────────┘│ + └───────────────────────────────────┘ |}] let%expect_test "einsum with a leftmost input axis preserved as output axis" = @@ -1644,54 +1218,10 @@ let%expect_test "einsum with a leftmost input axis preserved as output axis" = Train.forward_and_forget backend ctx c; Tensor.print ~with_code:false ~with_grad:false `Default @@ a; [%expect - {| - ┌───────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[0]: r3x2x4_a shape 0:3|2:4->1:2 │ - │┌──────┬────────────────────────┬────────────────────────────────────┬────────────────────────────────────┐│ - ││ │0 @ 0 │1 @ 0 │2 @ 0 ││ - ││ │axis 2 │axis 2 │axis 2 ││ - │├──────┼────────────────────────┼────────────────────────────────────┼────────────────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 3.00 │ 8.00 9.00 1.00e+1 1.10e+1 │ 1.60e+1 1.70e+1 1.80e+1 1.90e+1 ││ - ││ │ 4.00 5.00 6.00 7.00 │ 1.20e+1 1.30e+1 1.40e+1 1.50e+1 │ 2.00e+1 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴────────────────────────┴────────────────────────────────────┴────────────────────────────────────┘│ - └───────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [0]: r3x2x4_a shape 0:3|2:4->1:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌────────────────────────────────────────────────────────────────┐ - │[1]: r3x4x2x3_b shape 0:3|2:2,3:3->1:4 │ - │┌──────┬───────────────────────────┬───────────────────────────┐│ - ││0 @ 0 │0 @ 2 │1 @ 2 ││ - ││ │axis 3 │axis 3 ││ - │├──────┼───────────────────────────┼───────────────────────────┤│ - ││axis 1│ 0.00 1.00 2.00 │ 3.00 4.00 5.00 ││ - ││ │ 6.00 7.00 8.00 │ 9.00 1.00e+1 1.10e+1 ││ - ││ │ 1.20e+1 1.30e+1 1.40e+1 │ 1.50e+1 1.60e+1 1.70e+1 ││ - ││ │ 1.80e+1 1.90e+1 2.00e+1 │ 2.10e+1 2.20e+1 2.30e+1 ││ - │└──────┴───────────────────────────┴───────────────────────────┘│ - ├────────────────────────────────────────────────────────────────┤ - │┌──────┬───────────────────────────┬───────────────────────────┐│ - ││1 @ 0 │0 @ 2 │1 @ 2 ││ - ││ │axis 3 │axis 3 ││ - │├──────┼───────────────────────────┼───────────────────────────┤│ - ││axis 1│ 2.40e+1 2.50e+1 2.60e+1 │ 2.70e+1 2.80e+1 2.90e+1 ││ - ││ │ 3.00e+1 3.10e+1 3.20e+1 │ 3.30e+1 3.40e+1 3.50e+1 ││ - ││ │ 3.60e+1 3.70e+1 3.80e+1 │ 3.90e+1 4.00e+1 4.10e+1 ││ - ││ │ 4.20e+1 4.30e+1 4.40e+1 │ 4.50e+1 4.60e+1 4.70e+1 ││ - │└──────┴───────────────────────────┴───────────────────────────┘│ - ├────────────────────────────────────────────────────────────────┤ - │┌──────┬───────────────────────────┬───────────────────────────┐│ - ││2 @ 0 │0 @ 2 │1 @ 2 ││ - ││ │axis 3 │axis 3 ││ - │├──────┼───────────────────────────┼───────────────────────────┤│ - ││axis 1│ 4.80e+1 4.90e+1 5.00e+1 │ 5.10e+1 5.20e+1 5.30e+1 ││ - ││ │ 5.40e+1 5.50e+1 5.60e+1 │ 5.70e+1 5.80e+1 5.90e+1 ││ - ││ │ 6.00e+1 6.10e+1 6.20e+1 │ 6.30e+1 6.40e+1 6.50e+1 ││ - ││ │ 6.60e+1 6.70e+1 6.80e+1 │ 6.90e+1 7.00e+1 7.10e+1 ││ - │└──────┴───────────────────────────┴───────────────────────────┘│ - └────────────────────────────────────────────────────────────────┘ - |}]; + {| [1]: r3x4x2x3_b shape 0:3|2:2,3:3->1:4 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| @@ -1728,34 +1258,10 @@ let%expect_test "einsum permuting two leftmost input axes as output axes" = Train.forward_and_forget backend ctx c; Tensor.print ~with_code:false ~with_grad:false `Default @@ a; [%expect - {| - ┌───────────────────────────┐ - │[0]: r2x2_a shape 1:2->0:2 │ - │┌──────┬────────────┐ │ - ││ │axis 1 │ │ - │├──────┼────────────┤ │ - ││axis 0│ 0.00 1.00 │ │ - ││ │ 2.00 3.00 │ │ - │└──────┴────────────┘ │ - └───────────────────────────┘ - |}]; + {| [0]: r2x2_a shape 1:2->0:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ b; [%expect - {| - ┌───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┐ - │[1]: r2x2x3x4_b shape 1:2,2:3,3:4->0:2 │ - │┌──────┬────────────────────────────────────┬────────────────────────────────────┬────────────────────────────────────┐│ - ││ │0 @ 2 │1 @ 2 │2 @ 2 ││ - ││ │axis 3 │axis 3 │axis 3 ││ - │├──────┼────────────────────────────────────┼────────────────────────────────────┼────────────────────────────────────┤│ - ││0 @ 1 │ 0.00 1.00 2.00 3.00 │ 4.00 5.00 6.00 7.00 │ 8.00 9.00 1.00e+1 1.10e+1 ││ - ││axis 0│ 2.40e+1 2.50e+1 2.60e+1 2.70e+1 │ 2.80e+1 2.90e+1 3.00e+1 3.10e+1 │ 3.20e+1 3.30e+1 3.40e+1 3.50e+1 ││ - │├──────┼────────────────────────────────────┼────────────────────────────────────┼────────────────────────────────────┤│ - ││1 @ 1 │ 1.20e+1 1.30e+1 1.40e+1 1.50e+1 │ 1.60e+1 1.70e+1 1.80e+1 1.90e+1 │ 2.00e+1 2.10e+1 2.20e+1 2.30e+1 ││ - ││axis 0│ 3.60e+1 3.70e+1 3.80e+1 3.90e+1 │ 4.00e+1 4.10e+1 4.20e+1 4.30e+1 │ 4.40e+1 4.50e+1 4.60e+1 4.70e+1 ││ - │└──────┴────────────────────────────────────┴────────────────────────────────────┴────────────────────────────────────┘│ - └───────────────────────────────────────────────────────────────────────────────────────────────────────────────────────┘ - |}]; + {| [1]: r2x2x3x4_b shape 1:2,2:3,3:4->0:2 <virtual> |}]; Tensor.print ~with_code:false ~with_grad:false `Default @@ c; [%expect {| (cd _build/default/test && ./moons_demo_parallel_run.exe) Welcome to OCANNL! Reading configuration defaults from /src/_build/default/test/ocannl_config. Retrieving commandline, environment, or config file variable ocannl_log_level Found 0, in the config file ("Set log_level to" 1) └─{orphaned from #50} Retrieving commandline, environment, or config file variable ocannl_backend Found multicore_cc, in the config file Properties of devices: (multicore_devices (device ((device_name CPU) (device_ordinal 0) (num_domains 128)))) @!Retrieving commandline, environment, or config file variable ocannl_prefer_backend_uniformity Found true, in the config file Retrieving commandline, environment, or config file variable ocannl_debug_log_to_stream_files Not found, using default false Retrieving commandline, environment, or config file variable ocannl_ll_ident_style Not found, using default heuristic Retrieving commandline, environment, or config file variable ocannl_cc_backend_optimization_level Not found, using default 3 Retrieving commandline, environment, or config file variable ocannl_cc_backend_compiler_command Not found, using default gcc Retrieving commandline, environment, or config file variable ocannl_cc_backend_verify_codesign Not found, using default false Retrieving commandline, environment, or config file variable ocannl_never_capture_stdout Not found, using default false Batch=59, step=60, lr=0.200000, batch loss=0.000000, epoch loss=0.000000 Batch=119, step=120, lr=0.199750, batch loss=15116081563697152.000000, epoch loss=15116081563697152.000000 Batch=179, step=180, lr=0.199500, batch loss=11022092681015022.000000, epoch loss=26138174244712176.000000 Batch=239, step=240, lr=0.199250, batch loss=6933390903074270.000000, epoch loss=33071565147786448.000000 Batch=299, step=300, lr=0.199000, batch loss=2849975370881433.500000, epoch loss=35921540518667880.000000 Batch=359, step=360, lr=0.198750, batch loss=0.000000, epoch loss=35921540518667880.000000 Batch=419, step=420, lr=0.198500, batch loss=0.000000, epoch loss=35921540518667880.000000 Batch=479, step=480, lr=0.198250, batch loss=0.000000, epoch loss=35921540518667880.000000 Batch=539, step=540, lr=0.198000, batch loss=0.000000, epoch loss=35921540518667880.000000 Batch=599, step=600, lr=0.197750, batch loss=0.000000, epoch loss=35921540518667880.000000 Batch=659, step=660, lr=0.197500, batch loss=39654819149975824.000000, epoch loss=75576359668643712.000000 Batch=719, step=720, lr=0.197250, batch loss=35605975241204260.000000, epoch loss=111182334909847968.000000 Batch=779, step=780, lr=0.197000, batch loss=31562417005518300.000000, epoch loss=142744751915366272.000000 Batch=839, step=840, lr=0.196750, batch loss=27524143297593344.000000, epoch loss=170268895212959616.000000 Batch=899, step=900, lr=0.196500, batch loss=23491153258435924.000000, epoch loss=193760048471395552.000000 Batch=959, step=960, lr=0.196250, batch loss=19463446029052588.000000, epoch loss=213223494500448128.000000 Batch=1019, step=1020, lr=0.196000, batch loss=15441020750449868.000000, epoch loss=228664515250897984.000000 Batch=1079, step=1080, lr=0.195750, batch loss=11423877708958924.000000, epoch loss=240088392959856896.000000 Batch=1139, step=1140, lr=0.195500, batch loss=7412015186592836.000000, epoch loss=247500408146449728.000000 Batch=1199, step=1200, lr=0.195250, batch loss=3405433756013909.500000, epoch loss=250905841902463648.000000 Epoch=0, step=1200, lr=0.195250, epoch loss=250905841902463648.000000 Batch=59, step=1260, lr=0.195000, batch loss=39026585693655312.000000, epoch loss=39026585693655312.000000 Batch=119, step=1320, lr=0.194750, batch loss=35029017967774924.000000, epoch loss=74055603661430240.000000 Batch=179, step=1380, lr=0.194500, batch loss=31036731333681700.000000, epoch loss=105092334995111936.000000 Batch=239, step=1440, lr=0.194250, batch loss=27049724646051020.000000, epoch loss=132142059641162960.000000 Batch=299, step=1500, lr=0.194000, batch loss=23067997904882892.000000, epoch loss=155210057546045856.000000 Batch=359, step=1560, lr=0.193750, batch loss=19091549964852700.000000, epoch loss=174301607510898560.000000 Batch=419, step=1620, lr=0.193500, batch loss=15120380253298142.000000, epoch loss=189421987764196704.000000 Batch=479, step=1680, lr=0.193250, batch loss=11154488770219212.000000, epoch loss=200576476534415904.000000 Batch=539, step=1740, lr=0.193000, batch loss=7193874370291302.000000, epoch loss=207770350904707200.000000 Batch=599, step=1800, lr=0.192750, batch loss=3238535908189798.500000, epoch loss=211008886812896992.000000 Batch=659, step=1860, lr=0.192500, batch loss=39142242863651224.000000, epoch loss=250151129676548224.000000 Batch=719, step=1920, lr=0.192250, batch loss=35195921542222096.000000, epoch loss=285347051218770304.000000 Batch=779, step=1980, lr=0.192000, batch loss=31254877876606292.000000, epoch loss=316601929095376576.000000 Batch=839, step=2040, lr=0.191750, batch loss=27319109576154588.000000, epoch loss=343921038671531136.000000 Batch=899, step=2100, lr=0.191500, batch loss=23388616927198140.000000, epoch loss=367309655598729280.000000 Batch=959, step=2160, lr=0.191250, batch loss=19463399643405788.000000, epoch loss=386773055242135040.000000 Batch=1019, step=2220, lr=0.191000, batch loss=15543456865784082.000000, epoch loss=402316512107919104.000000 Batch=1079, step=2280, lr=0.190750, batch loss=11628787735339554.000000, epoch loss=413945299843258688.000000 Batch=1139, step=2340, lr=0.190500, batch loss=7719391679409903.000000, epoch loss=421664691522668608.000000 Batch=1199, step=2400, lr=0.190250, batch loss=3815268697995127.500000, epoch loss=425479960220663744.000000 Epoch=1, step=2400, lr=0.190250, epoch loss=425479960220663744.000000 Batch=59, step=2460, lr=0.190000, batch loss=38514298029133008.000000, epoch loss=38514298029133008.000000 Batch=119, step=2520, lr=0.189750, batch loss=34619251745270444.000000, epoch loss=73133549774403456.000000 Batch=179, step=2580, lr=0.189500, batch loss=30729478535922756.000000, epoch loss=103863028310326208.000000 Batch=239, step=2640, lr=0.189250, batch loss=26844977255765332.000000, epoch loss=130708005566091536.000000 Batch=299, step=2700, lr=0.189000, batch loss=22965747904798172.000000, epoch loss=153673753470889696.000000 Batch=359, step=2760, lr=0.188750, batch loss=19091789910358972.000000, epoch loss=172765543381248672.000000 Batch=419, step=2820, lr=0.188500, batch loss=15223102413454268.000000, epoch loss=187988645794702944.000000 Batch=479, step=2880, lr=0.188250, batch loss=11359684555090602.000000, epoch loss=199348330349793536.000000 Batch=539, step=2940, lr=0.188000, batch loss=7501536048936823.000000, epoch loss=206849866398730368.000000 Batch=599, step=3000, lr=0.187750, batch loss=3648656035999471.000000, epoch loss=210498522434729824.000000 Batch=659, step=3060, lr=0.187500, batch loss=38629673449274296.000000, epoch loss=249128195884004128.000000 Batch=719, step=3120, lr=0.187250, batch loss=34785873569862996.000000, epoch loss=283914069453867136.000000 Batch=779, step=3180, lr=0.187000, batch loss=30947342183668120.000000, epoch loss=314861411637535232.000000 Batch=839, step=3240, lr=0.186750, batch loss=27114079863351980.000000, epoch loss=341975491500887232.000000 Batch=899, step=3300, lr=0.186500, batch loss=23286085177258804.000000, epoch loss=365261576678146048.000000 Batch=959, step=3360, lr=0.186250, batch loss=19463358125388596.000000, epoch loss=384724934803534656.000000 Batch=1019, step=3420, lr=0.186000, batch loss=15645897276085590.000000, epoch loss=400370832079620224.000000 Batch=1079, step=3480, lr=0.185750, batch loss=11833702915680938.000000, epoch loss=412204534995301184.000000 Batch=1139, step=3540, lr=0.185500, batch loss=8026773898850031.000000, epoch loss=420231308894151232.000000 Batch=1199, step=3600, lr=0.185250, batch loss=4225109366599407.000000, epoch loss=424456418260750656.000000 Epoch=2, step=3600, lr=0.185250, epoch loss=424456418260750656.000000 Batch=59, step=3660, lr=0.185000, batch loss=38002006928636856.000000, epoch loss=38002006928636856.000000 Batch=119, step=3720, lr=0.184750, batch loss=34209480941467512.000000, epoch loss=72211487870104368.000000 Batch=179, step=3780, lr=0.184500, batch loss=30422220011540752.000000, epoch loss=102633707881645120.000000 Batch=239, step=3840, lr=0.184250, batch loss=26640224138856584.000000, epoch loss=129273932020501696.000000 Batch=299, step=3900, lr=0.184000, batch loss=22863492178090392.000000, epoch loss=152137424198592096.000000 Batch=359, step=3960, lr=0.183750, batch loss=19092023270248720.000000, epoch loss=171229447468840832.000000 Batch=419, step=4020, lr=0.183500, batch loss=15325817415331566.000000, epoch loss=186555264884172384.000000 Batch=479, step=4080, lr=0.183250, batch loss=11564874040676626.000000, epoch loss=198120138924849024.000000 Batch=539, step=4140, lr=0.183000, batch loss=7809192000959283.000000, epoch loss=205929330925808320.000000 Batch=599, step=4200, lr=0.182750, batch loss=4058771582510694.500000, epoch loss=209988102508319008.000000 Batch=659, step=4260, lr=0.182500, batch loss=38117105180221984.000000, epoch loss=248105207688540992.000000 Batch=719, step=4320, lr=0.182250, batch loss=34375825597503896.000000, epoch loss=282481033286044896.000000 Batch=779, step=4380, lr=0.182000, batch loss=30639806490729948.000000, epoch loss=313120839776774848.000000 Batch=839, step=4440, lr=0.181750, batch loss=26909049005224756.000000, epoch loss=340029888781999616.000000 Batch=899, step=4500, lr=0.181500, batch loss=23183551709332548.000000, epoch loss=363213440491332160.000000 Batch=959, step=4560, lr=0.181250, batch loss=19463314316722176.000000, epoch loss=382676754808054336.000000 Batch=1019, step=4620, lr=0.181000, batch loss=15748335682069026.000000, epoch loss=398425090490123392.000000 Batch=1079, step=4680, lr=0.180750, batch loss=12038615805373098.000000, epoch loss=410463706295496512.000000 Batch=1139, step=4740, lr=0.180500, batch loss=8334153827640934.000000, epoch loss=418797860123137472.000000 Batch=1199, step=4800, lr=0.180250, batch loss=4634950035203686.000000, epoch loss=423432810158341184.000000 Epoch=3, step=4800, lr=0.180250, epoch loss=423432810158341184.000000 Batch=59, step=4860, lr=0.180000, batch loss=37489714682816104.000000, epoch loss=37489714682816104.000000 Batch=119, step=4920, lr=0.179750, batch loss=33799707847015356.000000, epoch loss=71289422529831456.000000 Batch=179, step=4980, lr=0.179500, batch loss=30114959196509524.000000, epoch loss=101404381726340976.000000 Batch=239, step=5040, lr=0.179250, batch loss=26435467013311692.000000, epoch loss=127839848739652672.000000 Batch=299, step=5100, lr=0.179000, batch loss=22761231583753012.000000, epoch loss=150601080323405696.000000 Batch=359, step=5160, lr=0.178750, batch loss=19092251762508868.000000, epoch loss=169693332085914560.000000 Batch=419, step=5220, lr=0.178500, batch loss=15428527549579264.000000, epoch loss=185121859635493824.000000 Batch=479, step=5280, lr=0.178250, batch loss=11770058944964198.000000, epoch loss=196891918580458016.000000 Batch=539, step=5340, lr=0.178000, batch loss=8116844517007906.000000, epoch loss=205008763097465920.000000 Batch=599, step=5400, lr=0.177750, batch loss=4468883693048081.000000, epoch loss=209477646790514016.000000 Batch=659, step=5460, lr=0.177500, batch loss=37604541492468120.000000, epoch loss=247082188282982144.000000 Batch=719, step=5520, lr=0.177250, batch loss=33965781061118636.000000, epoch loss=281047969344100768.000000 Batch=779, step=5580, lr=0.177000, batch loss=30332274233765616.000000, epoch loss=311380243577866368.000000 Batch=839, step=5640, lr=0.176750, batch loss=26704021010409064.000000, epoch loss=338084264588275456.000000 Batch=899, step=5700, lr=0.176500, batch loss=23081020245724364.000000, epoch loss=361165284833999808.000000 Batch=959, step=5760, lr=0.176250, batch loss=19463271653380368.000000, epoch loss=380628556487380160.000000 Batch=1019, step=5820, lr=0.176000, batch loss=15850775233377076.000000, epoch loss=396479331720757248.000000 Batch=1079, step=5880, lr=0.175750, batch loss=12243530413052178.000000, epoch loss=408722862133809408.000000 Batch=1139, step=5940, lr=0.175500, batch loss=8641536047081062.000000, epoch loss=417364398180890496.000000 Batch=1199, step=6000, lr=0.175250, batch loss=5044791849132578.000000, epoch loss=422409190030023104.000000 Epoch=4, step=6000, lr=0.175250, epoch loss=422409190030023104.000000 Batch=59, step=6060, lr=0.175000, batch loss=36977420146346120.000000, epoch loss=36977420146346120.000000 Batch=119, step=6120, lr=0.174750, batch loss=33389931316589364.000000, epoch loss=70367351462935488.000000 Batch=179, step=6180, lr=0.174500, batch loss=29807692654855236.000000, epoch loss=100175044117790720.000000 Batch=239, step=6240, lr=0.174250, batch loss=26230704161143740.000000, epoch loss=126405748278934464.000000 Batch=299, step=6300, lr=0.174000, batch loss=22658964403799108.000000, epoch loss=149064712682733568.000000 Batch=359, step=6360, lr=0.173750, batch loss=19092473955483648.000000, epoch loss=168157186638217216.000000 Batch=419, step=6420, lr=0.173500, batch loss=15531231670872746.000000, epoch loss=183688418309089952.000000 Batch=479, step=6480, lr=0.173250, batch loss=11975236977304098.000000, epoch loss=195663655286394048.000000 Batch=539, step=6540, lr=0.173000, batch loss=8424489015784243.000000, epoch loss=204088144302178304.000000 Batch=599, step=6600, lr=0.172750, batch loss=4878987786313182.000000, epoch loss=208967132088491488.000000 Batch=659, step=6660, lr=0.172500, batch loss=37091983531337320.000000, epoch loss=246059115619828800.000000 Batch=719, step=6720, lr=0.172250, batch loss=33555741106031820.000000, epoch loss=279614856725860608.000000 Batch=779, step=6780, lr=0.172000, batch loss=30024745412775116.000000, epoch loss=309639602138635712.000000 Batch=839, step=6840, lr=0.171750, batch loss=26498995306242596.000000, epoch loss=336138597444878336.000000 Batch=899, step=6900, lr=0.171500, batch loss=22978491072765404.000000, epoch loss=359117088517643712.000000 Batch=959, step=6960, lr=0.171250, batch loss=19463232139681244.000000, epoch loss=378580320657324928.000000 Batch=1019, step=7020, lr=0.171000, batch loss=15953217934327808.000000, epoch loss=394533538591652736.000000 Batch=1079, step=7080, lr=0.170750, batch loss=12448447884042786.000000, epoch loss=406981986475695552.000000 Batch=1139, step=7140, lr=0.170500, batch loss=8948921702495027.000000, epoch loss=415930908178190592.000000 Batch=1199, step=7200, lr=0.170250, batch loss=5454638244359919.000000, epoch loss=421385546422550528.000000 Epoch=5, step=7200, lr=0.170250, epoch loss=421385546422550528.000000 Batch=59, step=7260, lr=0.170000, batch loss=36465119883253080.000000, epoch loss=36465119883253080.000000 Batch=119, step=7320, lr=0.169750, batch loss=32980147914215696.000000, epoch loss=69445267797468776.000000 Batch=179, step=7380, lr=0.169500, batch loss=29500419241253272.000000, epoch loss=98945687038722048.000000 Batch=239, step=7440, lr=0.169250, batch loss=26025933291703500.000000, epoch loss=124971620330425552.000000 Batch=299, step=7500, lr=0.169000, batch loss=22556689779235224.000000, epoch loss=147528310109660768.000000 Batch=359, step=7560, lr=0.168750, batch loss=19092687844854988.000000, epoch loss=166620997954515744.000000 Batch=419, step=7620, lr=0.168500, batch loss=15633927202231638.000000, epoch loss=182254925156747392.000000 Batch=479, step=7680, lr=0.168250, batch loss=12180406992371712.000000, epoch loss=194435332149119104.000000 Batch=539, step=7740, lr=0.168000, batch loss=8732126642612907.000000, epoch loss=203167458791732000.000000 Batch=599, step=7800, lr=0.167750, batch loss=5289086152955221.000000, epoch loss=208456544944687232.000000 Batch=659, step=7860, lr=0.167500, batch loss=36579432442154192.000000, epoch loss=245035977386841408.000000 Batch=719, step=7920, lr=0.167250, batch loss=33145706877568068.000000, epoch loss=278181684264409472.000000 Batch=779, step=7980, lr=0.167000, batch loss=29717221173083068.000000, epoch loss=307898905437492544.000000 Batch=839, step=8040, lr=0.166750, batch loss=26293974183374576.000000, epoch loss=334192879620867136.000000 Batch=899, step=8100, lr=0.166500, batch loss=22875966194773744.000000, epoch loss=357068845815640896.000000 Batch=959, step=8160, lr=0.166250, batch loss=19463196348287112.000000, epoch loss=376532042163928000.000000 Batch=1019, step=8220, lr=0.166000, batch loss=16055664071252378.000000, epoch loss=392587706235180352.000000 Batch=1079, step=8280, lr=0.165750, batch loss=12653368791007232.000000, epoch loss=405241075026187584.000000 Batch=1139, step=8340, lr=0.165500, batch loss=9256310793882828.000000, epoch loss=414497385820070400.000000 Batch=1199, step=8400, lr=0.165250, batch loss=5864489220885709.000000, epoch loss=420361875040956096.000000 Epoch=6, step=8400, lr=0.165250, epoch loss=420361875040956096.000000 Batch=59, step=8460, lr=0.165000, batch loss=35952813893536972.000000, epoch loss=35952813893536972.000000 Batch=119, step=8520, lr=0.164750, batch loss=32570358785218968.000000, epoch loss=68523172678755936.000000 Batch=179, step=8580, lr=0.164500, batch loss=29193140101028248.000000, epoch loss=97716312779784192.000000 Batch=239, step=8640, lr=0.164250, batch loss=25821156695640200.000000, epoch loss=123537469475424384.000000 Batch=299, step=8700, lr=0.164000, batch loss=22454408569054824.000000, epoch loss=145991878044479200.000000 Batch=359, step=8760, lr=0.163750, batch loss=19092895148609808.000000, epoch loss=165084773193089024.000000 Batch=419, step=8820, lr=0.163500, batch loss=15736616434305160.000000, epoch loss=180821389627394176.000000 Batch=479, step=8880, lr=0.163250, batch loss=12385571280816264.000000, epoch loss=193206960908210432.000000 Batch=539, step=8940, lr=0.163000, batch loss=9039759688143122.000000, epoch loss=202246720596353568.000000 Batch=599, step=9000, lr=0.162750, batch loss=5699181083623424.000000, epoch loss=207945901679976992.000000 Batch=659, step=9060, lr=0.162500, batch loss=36066882498295672.000000, epoch loss=244012784178272672.000000 Batch=719, step=9120, lr=0.162250, batch loss=32735672649104316.000000, epoch loss=276748456827376992.000000 Batch=779, step=9180, lr=0.162000, batch loss=29409695788066408.000000, epoch loss=306158152615443392.000000 Batch=839, step=9240, lr=0.161750, batch loss=26088951342519636.000000, epoch loss=332247103957963008.000000 Batch=899, step=9300, lr=0.161500, batch loss=22773439026132856.000000, epoch loss=355020542984095872.000000 Batch=959, step=9360, lr=0.161250, batch loss=19463157979912600.000000, epoch loss=374483700964008448.000000 Batch=1019, step=9420, lr=0.161000, batch loss=16158108203858876.000000, epoch loss=390641809167867328.000000 Batch=1079, step=9480, lr=0.160750, batch loss=12858289125309372.000000, epoch loss=403500098293176704.000000 Batch=1139, step=9540, lr=0.160500, batch loss=9563699885270630.000000, epoch loss=413063798178447360.000000 Batch=1199, step=9600, lr=0.160250, batch loss=6274340197411499.000000, epoch loss=419338138375858880.000000 Epoch=7, step=9600, lr=0.160250, epoch loss=419338138375858880.000000 Batch=59, step=9660, lr=0.160000, batch loss=35440507903820868.000000, epoch loss=35440507903820868.000000 Batch=119, step=9720, lr=0.159750, batch loss=32160567365573016.000000, epoch loss=67601075269393888.000000 Batch=179, step=9780, lr=0.159500, batch loss=28885856379504776.000000, epoch loss=96486931648898656.000000 Batch=239, step=9840, lr=0.159250, batch loss=25616374945616144.000000, epoch loss=122103306594514800.000000 Batch=299, step=9900, lr=0.159000, batch loss=22352121918582512.000000, epoch loss=144455428513097312.000000 Batch=359, step=9960, lr=0.158750, batch loss=19093097298403876.000000, epoch loss=163548525811501184.000000 Batch=419, step=10020, lr=0.158500, batch loss=15839300512417928.000000, epoch loss=179387826323919104.000000 Batch=479, step=10080, lr=0.158250, batch loss=12590730987962368.000000, epoch loss=191978557311881472.000000 Batch=539, step=10140, lr=0.158000, batch loss=9347388152374886.000000, epoch loss=201325945464256352.000000 Batch=599, step=10200, lr=0.157750, batch loss=6109271432993178.000000, epoch loss=207435216897249536.000000 Batch=659, step=10260, lr=0.157500, batch loss=35554335990410992.000000, epoch loss=242989552887660544.000000 Batch=719, step=10320, lr=0.157250, batch loss=32325640711289788.000000, epoch loss=275315193598950336.000000 Batch=779, step=10380, lr=0.157000, batch loss=29102171548374356.000000, epoch loss=304417365147324672.000000 Batch=839, step=10440, lr=0.156750, batch loss=25883928501664700.000000, epoch loss=330301293648989376.000000 Batch=899, step=10500, lr=0.156500, batch loss=22670911284829660.000000, epoch loss=352972204933819008.000000 Batch=959, step=10560, lr=0.156250, batch loss=19463119038875784.000000, epoch loss=372435323972694784.000000 Batch=1019, step=10620, lr=0.156000, batch loss=16260551191140762.000000, epoch loss=388695875163835520.000000 Batch=1079, step=10680, lr=0.155750, batch loss=13063207168962286.000000, epoch loss=401759082332797824.000000 Batch=1139, step=10740, lr=0.155500, batch loss=9871086686009208.000000, epoch loss=411630169018807040.000000 Batch=1199, step=10800, lr=0.155250, batch loss=6684190028612676.000000, epoch loss=418314359047419712.000000 Epoch=8, step=10800, lr=0.155250, epoch loss=418314359047419712.000000 Batch=59, step=10860, lr=0.155000, batch loss=34928201914104764.000000, epoch loss=34928201914104764.000000 Batch=119, step=10920, lr=0.154750, batch loss=31750775945927064.000000, epoch loss=66678977860031824.000000 Batch=179, step=10980, lr=0.154500, batch loss=28578572657981304.000000, epoch loss=95257550518013120.000000 Batch=239, step=11040, lr=0.154250, batch loss=25411592050267476.000000, epoch loss=120669142568280592.000000 Batch=299, step=11100, lr=0.154000, batch loss=22249833836454436.000000, epoch loss=142918976404735024.000000 Batch=359, step=11160, lr=0.153750, batch loss=19093297443879868.000000, epoch loss=162012273848614880.000000 Batch=419, step=11220, lr=0.153500, batch loss=15941982299881472.000000, epoch loss=177954256148496352.000000 Batch=479, step=11280, lr=0.153250, batch loss=12795887831796940.000000, epoch loss=190750143980293280.000000 Batch=539, step=11340, lr=0.153000, batch loss=9655013180632814.000000, epoch loss=200405157160926080.000000 Batch=599, step=11400, lr=0.152750, batch loss=6519359491713707.000000, epoch loss=206924516652639776.000000 Batch=659, step=11460, lr=0.152500, batch loss=35041791773175536.000000, epoch loss=241966308425815296.000000 Batch=719, step=11520, lr=0.152250, batch loss=31915611064124484.000000, epoch loss=273881919489939776.000000 Batch=779, step=11580, lr=0.152000, batch loss=28794649599331532.000000, epoch loss=302676569089271296.000000 Batch=839, step=11640, lr=0.151750, batch loss=25678908524121292.000000, epoch loss=328355477613392576.000000 Batch=899, step=11700, lr=0.151500, batch loss=22568386406838000.000000, epoch loss=350923864020230592.000000 Batch=959, step=11760, lr=0.151250, batch loss=19463082674819344.000000, epoch loss=370386946695049920.000000 Batch=1019, step=11820, lr=0.151000, batch loss=16362997041734178.000000, epoch loss=386749943736784128.000000 Batch=1079, step=11880, lr=0.150750, batch loss=13268129221251346.000000, epoch loss=400018072958035456.000000 Batch=1139, step=11940, lr=0.150500, batch loss=10178478640708540.000000, epoch loss=410196551598744000.000000 Batch=1199, step=12000, lr=0.150250, batch loss=7094044441112303.000000, epoch loss=417290596039856320.000000 Epoch=9, step=12000, lr=0.150250, epoch loss=417290596039856320.000000 Batch=59, step=12060, lr=0.150000, batch loss=34415890197765596.000000, epoch loss=34415890197765596.000000 Batch=119, step=12120, lr=0.149750, batch loss=31340976509008828.000000, epoch loss=65756866706774424.000000 Batch=179, step=12180, lr=0.149500, batch loss=28271279773860932.000000, epoch loss=94028146480635360.000000 Batch=239, step=12240, lr=0.149250, batch loss=25206799992321912.000000, epoch loss=119234946472957280.000000 Batch=299, step=12300, lr=0.149000, batch loss=22147536019067152.000000, epoch loss=141382482492024432.000000 Batch=359, step=12360, lr=0.148750, batch loss=19093487567765504.000000, epoch loss=160475970059789952.000000 Batch=419, step=12420, lr=0.148500, batch loss=16044654065754658.000000, epoch loss=176520624125544608.000000 Batch=479, step=12480, lr=0.148250, batch loss=13001035513034616.000000, epoch loss=189521659638579232.000000 Batch=539, step=12540, lr=0.148000, batch loss=9962631336943070.000000, epoch loss=199484290975522304.000000 Batch=599, step=12600, lr=0.147750, batch loss=6929440678486562.000000, epoch loss=206413731654008864.000000 Batch=659, step=12660, lr=0.147500, batch loss=34529253282563140.000000, epoch loss=240942984936572000.000000 Batch=719, step=12720, lr=0.147250, batch loss=31505584852933016.000000, epoch loss=272448569789505024.000000 Batch=779, step=12780, lr=0.147000, batch loss=28487130513600240.000000, epoch loss=300935700303105280.000000 Batch=839, step=12840, lr=0.146750, batch loss=25473889691902500.000000, epoch loss=326409589995007808.000000 Batch=899, step=12900, lr=0.146500, batch loss=22465861815177488.000000, epoch loss=348875451810185280.000000 Batch=959, step=12960, lr=0.146250, batch loss=19463046310762904.000000, epoch loss=368338498120948160.000000 Batch=1019, step=13020, lr=0.146000, batch loss=16465442892327594.000000, epoch loss=384803941013275776.000000 Batch=1079, step=13080, lr=0.145750, batch loss=13473051273540404.000000, epoch loss=398276992286816192.000000 Batch=1139, step=13140, lr=0.145500, batch loss=10485870595407872.000000, epoch loss=408762862882224064.000000 Batch=1199, step=13200, lr=0.145250, batch loss=7503901144261154.000000, epoch loss=416266764026485248.000000 Epoch=10, step=13200, lr=0.145250, epoch loss=416266764026485248.000000 Batch=59, step=13260, lr=0.145000, batch loss=33903575045452596.000000, epoch loss=33903575045452596.000000 Batch=119, step=13320, lr=0.144750, batch loss=30931173636116752.000000, epoch loss=64834748681569344.000000 Batch=179, step=13380, lr=0.144500, batch loss=27963982881104420.000000, epoch loss=92798731562673760.000000 Batch=239, step=13440, lr=0.144250, batch loss=25002002207753284.000000, epoch loss=117800733770427040.000000 Batch=299, step=13500, lr=0.144000, batch loss=22045231902394504.000000, epoch loss=139845965672821536.000000 Batch=359, step=13560, lr=0.143750, batch loss=19093671106034620.000000, epoch loss=158939636778856160.000000 Batch=419, step=13620, lr=0.143500, batch loss=16147319818673630.000000, epoch loss=175086956597529792.000000 Batch=479, step=13680, lr=0.143250, batch loss=13206176894986922.000000, epoch loss=188293133492516704.000000 Batch=539, step=13740, lr=0.143000, batch loss=10270242048643346.000000, epoch loss=198563375541160064.000000 Batch=599, step=13800, lr=0.142750, batch loss=7339514993311744.000000, epoch loss=205902890534471808.000000 Batch=659, step=13860, lr=0.142500, batch loss=34016721663898420.000000, epoch loss=239919612198370240.000000 Batch=719, step=13920, lr=0.142250, batch loss=31095565513689224.000000, epoch loss=271015177712059456.000000 Batch=779, step=13980, lr=0.142000, batch loss=28179617154492008.000000, epoch loss=299194794866551488.000000 Batch=839, step=14040, lr=0.141750, batch loss=25268876013644460.000000, epoch loss=324463670880195968.000000 Batch=899, step=14100, lr=0.141500, batch loss=22363341804815428.000000, epoch loss=346827012685011392.000000 Batch=959, step=14160, lr=0.141250, batch loss=19463014528004916.000000, epoch loss=366290027213016320.000000 Batch=1019, step=14220, lr=0.141000, batch loss=16567893324219460.000000, epoch loss=382857920537235776.000000 Batch=1079, step=14280, lr=0.140750, batch loss=13677977907127910.000000, epoch loss=396535898444363712.000000 Batch=1139, step=14340, lr=0.140500, batch loss=10793267704067960.000000, epoch loss=407329166148431680.000000 Batch=1199, step=14400, lr=0.140250, batch loss=7913762428708454.000000, epoch loss=415242928577140160.000000 Epoch=11, step=14400, lr=0.140250, epoch loss=415242928577140160.000000 Batch=59, step=14460, lr=0.140000, batch loss=33391255311841144.000000, epoch loss=33391255311841144.000000 Batch=119, step=14520, lr=0.139750, batch loss=30521365036601616.000000, epoch loss=63912620348442760.000000 Batch=179, step=14580, lr=0.139500, batch loss=27656679116400232.000000, epoch loss=91569299464842992.000000 Batch=239, step=14640, lr=0.139250, batch loss=24797198123899288.000000, epoch loss=116366497588742272.000000 Batch=299, step=14700, lr=0.139000, batch loss=21942921486436488.000000, epoch loss=138309419075178752.000000 Batch=359, step=14760, lr=0.138750, batch loss=19093848345018368.000000, epoch loss=157403267420197120.000000 Batch=419, step=14820, lr=0.138500, batch loss=16249978699644928.000000, epoch loss=173653246119842048.000000 Batch=479, step=14880, lr=0.138250, batch loss=13411311404991556.000000, epoch loss=187064557524833600.000000 Batch=539, step=14940, lr=0.138000, batch loss=10577847033720558.000000, epoch loss=197642404558554144.000000 Batch=599, step=15000, lr=0.137750, batch loss=7749584726838477.000000, epoch loss=205391989285392608.000000 Batch=659, step=15060, lr=0.137500, batch loss=33504192335882924.000000, epoch loss=238896181621275520.000000 Batch=719, step=15120, lr=0.137250, batch loss=30685546174445432.000000, epoch loss=269581727795720960.000000 Batch=779, step=15180, lr=0.137000, batch loss=27872102650059160.000000, epoch loss=297453830445780096.000000 Batch=839, step=15240, lr=0.136750, batch loss=25063861190061808.000000, epoch loss=322517691635841920.000000 Batch=899, step=15300, lr=0.136500, batch loss=22260820649128756.000000, epoch loss=344778512284970688.000000 Batch=959, step=15360, lr=0.136250, batch loss=19462981027260008.000000, epoch loss=364241493312230720.000000 Batch=1019, step=15420, lr=0.136000, batch loss=16670341751793254.000000, epoch loss=380911835064024000.000000 Batch=1079, step=15480, lr=0.135750, batch loss=13882902250066194.000000, epoch loss=394794737314090176.000000 Batch=1139, step=15540, lr=0.135500, batch loss=11100663094741128.000000, epoch loss=405895400408831296.000000 Batch=1199, step=15600, lr=0.135250, batch loss=8323623713155755.000000, epoch loss=414219024121987072.000000 Epoch=12, step=15600, lr=0.135250, epoch loss=414219024121987072.000000 Batch=59, step=15660, lr=0.135000, batch loss=32878934432905080.000000, epoch loss=32878934432905080.000000 Batch=119, step=15720, lr=0.134750, batch loss=30111553001112644.000000, epoch loss=62990487434017728.000000 Batch=179, step=15780, lr=0.134500, batch loss=27349371343059900.000000, epoch loss=90339858777077632.000000 Batch=239, step=15840, lr=0.134250, batch loss=24592389458746844.000000, epoch loss=114932248235824480.000000 Batch=299, step=15900, lr=0.134000, batch loss=21840605630186564.000000, epoch loss=136772853866011040.000000 Batch=359, step=15960, lr=0.133750, batch loss=19094019857379056.000000, epoch loss=155866873723390080.000000 Batch=419, step=16020, lr=0.133500, batch loss=16352631853993164.000000, epoch loss=172219505577383232.000000 Batch=479, step=16080, lr=0.133250, batch loss=13616441333697740.000000, epoch loss=185835946911080960.000000 Batch=539, step=16140, lr=0.133000, batch loss=10885447437499324.000000, epoch loss=196721394348580288.000000 Batch=599, step=16200, lr=0.132750, batch loss=8159649879066761.000000, epoch loss=204881044227647040.000000 Batch=659, step=16260, lr=0.132500, batch loss=32991668734490488.000000, epoch loss=237872712962137536.000000 Batch=719, step=16320, lr=0.132250, batch loss=30275532561824700.000000, epoch loss=268148245523962240.000000 Batch=779, step=16380, lr=0.132000, batch loss=27564593299587072.000000, epoch loss=295712838823549312.000000 Batch=839, step=16440, lr=0.131750, batch loss=24858850375115300.000000, epoch loss=320571689198664640.000000 Batch=899, step=16500, lr=0.131500, batch loss=22158302929415920.000000, epoch loss=342729992128080576.000000 Batch=959, step=16560, lr=0.131250, batch loss=19462950962488936.000000, epoch loss=362192943090569536.000000 Batch=1019, step=16620, lr=0.131000, batch loss=16772794188003192.000000, epoch loss=378965737278572736.000000 Batch=1079, step=16680, lr=0.130750, batch loss=14087831746965230.000000, epoch loss=393053569025537984.000000 Batch=1139, step=16740, lr=0.130500, batch loss=11408064212037358.000000, epoch loss=404461633237575360.000000 Batch=1199, step=16800, lr=0.130250, batch loss=8733489578901504.000000, epoch loss=413195122816476864.000000 Epoch=13, step=16800, lr=0.130250, epoch loss=413195122816476864.000000 Batch=59, step=16860, lr=0.130000, batch loss=32366607827345956.000000, epoch loss=32366607827345956.000000 Batch=119, step=16920, lr=0.129750, batch loss=29701736384325224.000000, epoch loss=62068344211671184.000000 Batch=179, step=16980, lr=0.129500, batch loss=27042058415758812.000000, epoch loss=89110402627430000.000000 Batch=239, step=17040, lr=0.129250, batch loss=24387574494309036.000000, epoch loss=113497977121739040.000000 Batch=299, step=17100, lr=0.129000, batch loss=21738282901988968.000000, epoch loss=135236260023728000.000000 Batch=359, step=17160, lr=0.128750, batch loss=19094184211460916.000000, epoch loss=154330444235188928.000000 Batch=419, step=17220, lr=0.128500, batch loss=16455278136393728.000000, epoch loss=170785722371582656.000000 Batch=479, step=17280, lr=0.128250, batch loss=13821563817793946.000000, epoch loss=184607286189376608.000000 Batch=539, step=17340, lr=0.128000, batch loss=11193040969330414.000000, epoch loss=195800327158707008.000000 Batch=599, step=17400, lr=0.127750, batch loss=8569709304671983.000000, epoch loss=204370036463378976.000000 Batch=659, step=17460, lr=0.127500, batch loss=32479148569071888.000000, epoch loss=236849185032450880.000000 Batch=719, step=17520, lr=0.127250, batch loss=29865521239853192.000000, epoch loss=266714706272304064.000000 Batch=779, step=17580, lr=0.127000, batch loss=27257085667101900.000000, epoch loss=293971791939405952.000000 Batch=839, step=17640, lr=0.126750, batch loss=24653840705493400.000000, epoch loss=318625632644899328.000000 Batch=899, step=17700, lr=0.126500, batch loss=22055786068696540.000000, epoch loss=340681418713595840.000000 Batch=959, step=17760, lr=0.126250, batch loss=19462921756711320.000000, epoch loss=360144340470307136.000000 Batch=1019, step=17820, lr=0.126000, batch loss=16875246910544282.000000, epoch loss=377019587380851392.000000 Batch=1079, step=17880, lr=0.125750, batch loss=14292761243864268.000000, epoch loss=391312348624715648.000000 Batch=1139, step=17940, lr=0.125500, batch loss=11715464756671284.000000, epoch loss=403027813381386944.000000 Batch=1199, step=18000, lr=0.125250, batch loss=9143356589971866.000000, epoch loss=412171169971358784.000000 Epoch=14, step=18000, lr=0.125250, epoch loss=412171169971358784.000000 Batch=59, step=18060, lr=0.125000, batch loss=31854280076462216.000000, epoch loss=31854280076462216.000000 Batch=119, step=18120, lr=0.124750, batch loss=29291916331563964.000000, epoch loss=61146196408026176.000000 Batch=179, step=18180, lr=0.124500, batch loss=26734740907159280.000000, epoch loss=87880937315185456.000000 Batch=239, step=18240, lr=0.124250, batch loss=24182753803248164.000000, epoch loss=112063691118433616.000000 Batch=299, step=18300, lr=0.124000, batch loss=21635954733499460.000000, epoch loss=133699645851933072.000000 Batch=359, step=18360, lr=0.123750, batch loss=19094343411582020.000000, epoch loss=152793989263515104.000000 Batch=419, step=18420, lr=0.123500, batch loss=16557918978502382.000000, epoch loss=169351908242017472.000000 Batch=479, step=18480, lr=0.123250, batch loss=14026681147929396.000000, epoch loss=183378589389946880.000000 Batch=539, step=18540, lr=0.123000, batch loss=11500629919863058.000000, epoch loss=194879219309809952.000000 Batch=599, step=18600, lr=0.122750, batch loss=8979764148978756.000000, epoch loss=203858983458788704.000000 Batch=659, step=18660, lr=0.122500, batch loss=31966632984951740.000000, epoch loss=235825616443740448.000000 Batch=719, step=18720, lr=0.122250, batch loss=29455513353855524.000000, epoch loss=265281129797595968.000000 Batch=779, step=18780, lr=0.122000, batch loss=26949579752603648.000000, epoch loss=292230709550199616.000000 Batch=839, step=18840, lr=0.121750, batch loss=24448832181196116.000000, epoch loss=316679541731395712.000000 Batch=899, step=18900, lr=0.121500, batch loss=21953270066970624.000000, epoch loss=338632811798366336.000000 Batch=959, step=18960, lr=0.121250, batch loss=19462892837264860.000000, epoch loss=358095704635631168.000000 Batch=1019, step=19020, lr=0.121000, batch loss=16977700205747678.000000, epoch loss=375073404841378816.000000 Batch=1079, step=19080, lr=0.120750, batch loss=14497691886087918.000000, epoch loss=389571096727466752.000000 Batch=1139, step=19140, lr=0.120500, batch loss=12022867591954432.000000, epoch loss=401593964319421184.000000 Batch=1199, step=19200, lr=0.120250, batch loss=9553227037016064.000000, epoch loss=411147191356437248.000000 Epoch=15, step=19200, lr=0.120250, epoch loss=411147191356437248.000000 Batch=59, step=19260, lr=0.120000, batch loss=31341947744280028.000000, epoch loss=31341947744280028.000000 Batch=119, step=19320, lr=0.119750, batch loss=28882090552179644.000000, epoch loss=60224038296459672.000000 Batch=179, step=19380, lr=0.119500, batch loss=26427417099274376.000000, epoch loss=86651455395734048.000000 Batch=239, step=19440, lr=0.119250, batch loss=23977926812901924.000000, epoch loss=110629382208635968.000000 Batch=299, step=19500, lr=0.119000, batch loss=21533619406731128.000000, epoch loss=132163001615367104.000000 Batch=359, step=19560, lr=0.118750, batch loss=19094494594430840.000000, epoch loss=151257496209797952.000000 Batch=419, step=19620, lr=0.118500, batch loss=16660552089669906.000000, epoch loss=167918048299467872.000000 Batch=479, step=19680, lr=0.118250, batch loss=14231791033454866.000000, epoch loss=182149839332922752.000000 Batch=539, step=19740, lr=0.118000, batch loss=11808211425785720.000000, epoch loss=193958050758708480.000000 Batch=599, step=19800, lr=0.117750, batch loss=9389813266662468.000000, epoch loss=203347864025370944.000000 Batch=659, step=19860, lr=0.117500, batch loss=31454120836805428.000000, epoch loss=234801984862176384.000000 Batch=719, step=19920, lr=0.117250, batch loss=29045507758507076.000000, epoch loss=263847492620683456.000000 Batch=779, step=19980, lr=0.117000, batch loss=26642076128754620.000000, epoch loss=290489568749438080.000000 Batch=839, step=20040, lr=0.116750, batch loss=24243825947548056.000000, epoch loss=314733394696986112.000000 Batch=899, step=20100, lr=0.116500, batch loss=21850755783231624.000000, epoch loss=336584150480217728.000000 Batch=959, step=20160, lr=0.116250, batch loss=19462865778970896.000000, epoch loss=356047016259188608.000000 Batch=1019, step=20220, lr=0.116000, batch loss=17080155362103570.000000, epoch loss=373127171621292160.000000 Batch=1079, step=20280, lr=0.115750, batch loss=14702624246298488.000000, epoch loss=387829795867590656.000000 Batch=1139, step=20340, lr=0.115500, batch loss=12330272145224500.000000, epoch loss=400160068012815168.000000 Batch=1199, step=20400, lr=0.115250, batch loss=9963098629384874.000000, epoch loss=410123166642200064.000000 Epoch=16, step=20400, lr=0.115250, epoch loss=410123166642200064.000000 Batch=59, step=20460, lr=0.115000, batch loss=30829613121448620.000000, epoch loss=30829613121448620.000000 Batch=119, step=20520, lr=0.114750, batch loss=28472261909483792.000000, epoch loss=59301875030932416.000000 Batch=179, step=20580, lr=0.114500, batch loss=26120089855415636.000000, epoch loss=85421964886348048.000000 Batch=239, step=20640, lr=0.114250, batch loss=23773095813919540.000000, epoch loss=109195060700267584.000000 Batch=299, step=20700, lr=0.114000, batch loss=21431279784995500.000000, epoch loss=130626340485263088.000000 Batch=359, step=20760, lr=0.113750, batch loss=19094641625477940.000000, epoch loss=149720982110741024.000000 Batch=419, step=20820, lr=0.113500, batch loss=16763180619538978.000000, epoch loss=166484162730280000.000000 Batch=479, step=20880, lr=0.113250, batch loss=14436896910344192.000000, epoch loss=180921059640624192.000000 Batch=539, step=20940, lr=0.113000, batch loss=12115789495734546.000000, epoch loss=193036849136358752.000000 Batch=599, step=21000, lr=0.112750, batch loss=9799858948372344.000000, epoch loss=202836708084731104.000000 Batch=659, step=21060, lr=0.112500, batch loss=30941613269957564.000000, epoch loss=233778321354688672.000000 Batch=719, step=21120, lr=0.112250, batch loss=28635506171794772.000000, epoch loss=262413827526483456.000000 Batch=779, step=21180, lr=0.112000, batch loss=26334575940879428.000000, epoch loss=288748403467362880.000000 Batch=839, step=21240, lr=0.111750, batch loss=24038821431886916.000000, epoch loss=312787224899249792.000000 Batch=899, step=21300, lr=0.111500, batch loss=21748242644817236.000000, epoch loss=334535467544067008.000000 Batch=959, step=21360, lr=0.111250, batch loss=19462839436504816.000000, epoch loss=353998306980571840.000000 Batch=1019, step=21420, lr=0.111000, batch loss=17182611234287342.000000, epoch loss=371180918214859200.000000 Batch=1079, step=21480, lr=0.110750, batch loss=14907557751833668.000000, epoch loss=386088475966692864.000000 Batch=1139, step=21540, lr=0.110500, batch loss=12637678416481484.000000, epoch loss=398726154383174336.000000 Batch=1199, step=21600, lr=0.110250, batch loss=10372973657727522.000000, epoch loss=409099128040901888.000000 Epoch=17, step=21600, lr=0.110250, epoch loss=409099128040901888.000000 Batch=59, step=21660, lr=0.110000, batch loss=30317273917318760.000000, epoch loss=30317273917318760.000000 Batch=119, step=21720, lr=0.109750, batch loss=28062427540164880.000000, epoch loss=58379701457483640.000000 Batch=179, step=21780, lr=0.109500, batch loss=25812755166946920.000000, epoch loss=84192456624430560.000000 Batch=239, step=21840, lr=0.109250, batch loss=23568256797664872.000000, epoch loss=107760713422095424.000000 Batch=299, step=21900, lr=0.109000, batch loss=21328931716490856.000000, epoch loss=129089645138586272.000000 Batch=359, step=21960, lr=0.108750, batch loss=19094779637093716.000000, epoch loss=148184424775680000.000000 Batch=419, step=22020, lr=0.108500, batch loss=16865800702639036.000000, epoch loss=165050225478319040.000000 Batch=479, step=22080, lr=0.108250, batch loss=14641993910967774.000000, epoch loss=179692219389286816.000000 Batch=539, step=22140, lr=0.108000, batch loss=12423359548411084.000000, epoch loss=192115578937697888.000000 Batch=599, step=22200, lr=0.107750, batch loss=10209896612809932.000000, epoch loss=202325475550507808.000000 Batch=659, step=22260, lr=0.107500, batch loss=30429111429732760.000000, epoch loss=232754586980240576.000000 Batch=719, step=22320, lr=0.107250, batch loss=28225509166380920.000000, epoch loss=260980096146621504.000000 Batch=779, step=22380, lr=0.107000, batch loss=26027078616315768.000000, epoch loss=287007174762937280.000000 Batch=839, step=22440, lr=0.106750, batch loss=23833819779537304.000000, epoch loss=310840994542474560.000000 Batch=899, step=22500, lr=0.106500, batch loss=21645732083383228.000000, epoch loss=332486726625857792.000000 Batch=959, step=22560, lr=0.106250, batch loss=19462815527853532.000000, epoch loss=351949542153711296.000000 Batch=1019, step=22620, lr=0.106000, batch loss=17285069397120342.000000, epoch loss=369234611550831616.000000 Batch=1079, step=22680, lr=0.105750, batch loss=15112493834349226.000000, epoch loss=384347105385180864.000000 Batch=1139, step=22740, lr=0.105500, batch loss=12945088123712308.000000, epoch loss=397292193508893184.000000 Batch=1199, step=22800, lr=0.105250, batch loss=10782852122044006.000000, epoch loss=408075045630937216.000000 Epoch=18, step=22800, lr=0.105250, epoch loss=408075045630937216.000000 Batch=59, step=22860, lr=0.105000, batch loss=29804931277215060.000000, epoch loss=29804931277215060.000000 Batch=119, step=22920, lr=0.104750, batch loss=27652588589547520.000000, epoch loss=57457519866762576.000000 Batch=179, step=22980, lr=0.104500, batch loss=25505415897179752.000000, epoch loss=82962935763942336.000000 Batch=239, step=23040, lr=0.104250, batch loss=23363412341118292.000000, epoch loss=106326348105060624.000000 Batch=299, step=23100, lr=0.104000, batch loss=21226577921363148.000000, epoch loss=127552926026423776.000000 Batch=359, step=23160, lr=0.103750, batch loss=19094912065252012.000000, epoch loss=146647838091675776.000000 Batch=419, step=23220, lr=0.103500, batch loss=16968414915950456.000000, epoch loss=163616253007626240.000000 Batch=479, step=23280, lr=0.103250, batch loss=14847085757630600.000000, epoch loss=178463338765256832.000000 Batch=539, step=23340, lr=0.103000, batch loss=12730924447126870.000000, epoch loss=191194263212383712.000000 Batch=599, step=23400, lr=0.102750, batch loss=10619930841273686.000000, epoch loss=201814194053657408.000000 Batch=659, step=23460, lr=0.102500, batch loss=29916613025481796.000000, epoch loss=231730807079139200.000000 Batch=719, step=23520, lr=0.102250, batch loss=27815514451616292.000000, epoch loss=259546321530755488.000000 Batch=779, step=23580, lr=0.102000, batch loss=25719583582401332.000000, epoch loss=285265905113156832.000000 Batch=839, step=23640, lr=0.101750, batch loss=23628819845174612.000000, epoch loss=308894724958331456.000000 Batch=899, step=23700, lr=0.101500, batch loss=21543223096770560.000000, epoch loss=330437948055102016.000000 Batch=959, step=23760, lr=0.101250, batch loss=19462792907692440.000000, epoch loss=349900740962794432.000000 Batch=1019, step=23820, lr=0.101000, batch loss=17387528991609106.000000, epoch loss=367288269954403520.000000 Batch=1079, step=23880, lr=0.100750, batch loss=15317431062189398.000000, epoch loss=382605701016592896.000000 Batch=1139, step=23940, lr=0.100500, batch loss=13252498976267742.000000, epoch loss=395858199992860608.000000 Batch=1199, step=24000, lr=0.100250, batch loss=11192732304347410.000000, epoch loss=407050932297208000.000000 Epoch=19, step=24000, lr=0.100250, epoch loss=407050932297208000.000000 Half-moons scatterplot and decision boundary: ┌────────────────────────────────────────────────────────────────────────────────────────────────────┐ │...................................#................................................................│ │.......................#.#################.#........................................................│ │......................##########################....................................................│ │.................#################################..................................................│ │..............######################################................................................│ │.............#########################################..............................................│ │...........################.............#################...........................................│ │..........##############...................###############..........................................│ │........##############..........................###########.........................................│ │.......#############............................##.##########.......................................│ │......###########..................................############.....................................│ │.....###########....................................###########.....................................│ │....##########.....................%.................###########....................................│ │...#########.#...................%%%%.................#########..............................%%%.%%.│ │...##########..................%%%.%%...................########.#...........................%%.%%%.│ │..#######......................%%%%%%%..................##########...........................%%%%%%%│ │.#########....................%%%%%%%%....................########...........................%%%%%%.│ │.########......................%%%%%%%.....................########..........................%%%%%%.│ │#########.....................%%.%%%%%......................########........................%%%%%%%.│ │.#######.......................%%%%%%%%....................#########........................%%%%%%%.│ │########.......................%%%%%%%%.....................########.......................%%%%%%%%.│ │#######.........................%%%%%%%......................#######.......................%%%%%%%%.│ │#######.........................%%%%%%%%......................#######.....................%%%%%%%%..│ │#######.........................%%%%%%%%.....................#######.....................%%%%%%%%%..│ │######............................%%%%%%%%...................#######....................%%%%%%%%%...│ │######...........................%%%%%%%%....................########..................%%%%%%%%%%...│ │######............................%%%%%%%%%..................#######..................%.%%%%%%%%....│ │###.#.............................%.%%%%%%%%%.................#.#.##..................%%.%%%%%%%....│ │....................................%%%%%%%%%.......................................%%%%%%%%%%%.....│ │......................................%%%%%%%.%.....................................%%%%%%%%%%%.....│ │.....................................%%%%%%%%%%%.................................%%%%%%%%%%%%.......│ │......................................%%.%%%%%%%%...............................%%%%%%%%%%%.........│ │........................................%%%%%%%%%%%%........................%..%%%%%%%%%%%%.........│ │........................................%%%%%%%%%%%%%%.%....................%%%%%%%%%%%%%...........│ │..........................................%%%%%%%%%%%%%%%%%..............%%%%%%%%%%%%%%%............│ │............................................%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%...............│ │................................................%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%.%..%..............│ │................................................%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%.%..................│ │..................................................%%%%%%%%%%%%%%%%%%%%%%%%%%%%.%....................│ │.....................................................%%.%..%%%%%%%%%%%%%%%..........................│ └────────────────────────────────────────────────────────────────────────────────────────────────────┘ "/usr/bin/env" "bash" "-c" "opam exec -- dune build @install @check @runtest && rm -rf _build" failed with exit status 1 2025-07-12 13:40.25: Job failed: Failed: Build failed