1
0
Fork 0
mirror of https://github.com/ton-blockchain/ton synced 2025-02-12 19:22:37 +00:00
ton/crypto/vm/stack.cpp
EmelyanenkoK 9f93888cf4
TVM Upgrade (#686)
* New TVM instructions

* Remove PREVBLOCKS

* Separate target ton_crypto into TVM-related and -unrelared code

* Add fine for failed "send message"; rework SENDMSG

* Fix include

* Fix bugs, improve action fines

* Disable fines for special accounts

* Handle msg_balance_remaining.grams == null in transaction.cpp

* Bugfixes in SENDMSG

* Fix fee calculation in SENDMSG

* Fix CellStorageStat and transaction.cpp after merge

* SETBOUNCEONACTIONPHASEFAIL instruction

* ADDDIVMOD instructions

* RUNVM, RUNVMX instructions

* Changes in RUNVM

* Tests for adddiv and runvm

* HASHEXT instruction

* Improve opcode-timing

More iterations
Don't measure preliminary run
Remove logs and other excessive operations
Add "error" to output

* Increase RUNVM gas price

* Optimize HASHEXT, adjust gas price

* Add "bounce of action fail" flag to actions

* Stack operations with unlimited arguments

* Ristretto255 instructions

* Adjust gas consumption

* Optional fixed number of return values in RUNVM, fix exception handling

* Adjust gas consumption

* Simplify gas consumption logic

* Support of secp256k1 and sodium libraries in builds (#11)

* add support of secp256k1 library to the builds (linux, win)

* add support of secp256k1 library to the builds (linux, win)

* install secp256k1 via brew

* install libsodium via brew;
change sodium to upper case in FindSodium.cmake

* install libsodium via brew;
change sodium to upper case in FindSodium.cmake

* simplify FindSodium.cmake

* bug fixing

* bug fixing

* bug fixing

* add macro SODIUM_STATIC

* adjust build command for windows

* put back original FindSodium.cmake

* put back original FindSodium.cmake

* fix sodium unzipped path for windows;
add ninja

* fix sodium unzipped path for windows;
add ninja

* fix sodium unzipped path for windows;
add ninja

* Win32 github build for secp256k1

* x64 architecture github build for secp256k1

* fix sodium linking on linux

* enable docker buildx arm64 builds from forked repos

* enable docker buildx arm64 builds from forked repos

* enable docker buildx arm64 builds from forked repos

* adjust mac builds for secp2561k and sodium

* fix tonlib jni generation

* minor fix

* sync fixes across platforms

* add libsodium build script for android and precompiled static libraries

* build tonlib for android (fails)

* FindSodium uppercase

* remove system libsodium for android, use precompiled instead;
specify SECP256K1_INCLUDE_DIR fir mac 12.6

* uppercase sodium

* simplify FindSodium

* fix windows build sodium path;
use ninja for windows

* simplify sodium 2

* adjust windows sodium paths;
add paths to android jni

* add ninja build windows

* add ninja build windows

* add ninja build windows 2

* remove win ninja

* fix 1

* fix 2

* fix win 3

* fix linux compile 3

* fix jni 1

* fix jni 2 and mac

* fix jni 3

* fix jni 4

* fix jni 5

* fix mac 6

* fix mac 7 and jni paths

* fix jni 8

* rework sodium for android

* rework sodium for android

* rework sodium for android 2

* fixed sodium for android 2

* fixed sodium for android 3

* static secp256k1 for android

* add precompiled arm secp256k1

* add precompiled arm secp256k1

* build native-lib with secp256k1 x86-64 (non arm)

* update precompiled with NDK libsecp256k1.a

* update precompiled with NDK libsecp256k1.a

* update precompiled with NDK libsecp256k1.a

* refactor llvm-strip location

* refactor llvm-strip location

* add native-lib.so for armv7a, armv8a

* add native-lib.so for armv7a, armv8a

* test  armv7a, armv8a

* armv7a - fails linking on sodium, test -> armv8a

* works x86-64, armv7a - fails linking on sodium, armv8a - fails linking secp256k1 (incompatible with aarch64linux)

* update libpsec256k1, sodium static libs

* test x86 android native-lib

* test armv7 android native-lib

* test armv8 android native-lib

* x86_64 and arm64 android native-lib works

* x86_64 and arm64 android native-lib works

* x86_64 and arm64 android native-lib works

* test armv7 android native-lib

* test all android native-libs

* test all android native-libs

* test all android native-libs

* test all android native-libs - without SodiumAndroid

* test all android native-libs - with FindSodiumAndroid.cmake

* win, with Sodium via SODIUM_DIR

* win, with Sodium via SODIUM_DIR env

* win, with Sodium via SODIUM_DIR env

* win, with Sodium via SODIUM_DIR env and SODIUM_USE_STATIC_LIBS

* win, with Sodium via SODIUM_DIR, SODIUM_USE_STATIC_LIBS and SODIUM_INCLUDE_DIR

* android, with FindSodium

* android, with FindSodium with SODIUM_USE_STATIC_LIBS

* remove if not apple

* target_link_libraries(ton_crypto_core PUBLIC secp256k1)

* android SECP256K1_INCLUDE_DIRS

* android SECP256K1_INCLUDE_DIR

* add libsecp256k1.a/so pre-compiled with ubuntu 22 x86-64

* add libsecp256k1.a/so pre-compiled with ubuntu 22 x86-64

* sodium dirs

* sodium dirs

* sodium dirs

* remove NOT APPLE and SodiumAndroid

* add NOT APPLE and remove SodiumAndroid

* add NOT APPLE and remove SodiumAndroid

* remove build scripts for 18.04, reduce CMakeLists.txt

* remove build scripts for 18.04, reduce CMakeLists.txt

* Fix cas consumption during library load

* Fix fetch_config_params after merge

* Add all ADDDIVMOD ops to Asm.fif

* Save unpaid storage fee to due_payment

* Add "set prev blocks info" to emulator

* Adjusted builds  (#13)

* Update flake.nix

Add libsodium

* add libsecp256k1-dev and libsodium-dev into wasm build

* make back emulator a shared library;
put emulator to artifacts;
compile wasm artifacts with sodium and secp256k1.

* add secp256k1 to nix

* compile emulator statically with nix

* compile emulator statically with nix

* compile emulator lib statically with nix

* compile emulator lib statically with nix

* add libemulator to artifacts

* add shared libemulator library to artifacts

* minor release fix

* update set-output commands;
add recent_changelog.md

* releases fixes

* releases fixes, multiline

* releases fixes, multiline

* releases fixes, multiline

* put back multiline changelog

* put back multiline changelog

* ConfigParam 19 (global-id) and GLOBALID instruction

* Fix gas consumption in HASHEXT

* Add blst library

* Add bls instructions

* Allow passing long code to opcode-timing

* Add bls testcase

* More BLS instructions

* Fix tests, add bls tests

* Add more bls tests

* Improve some bls operations

* Adjust some BLS gas prices

* Adjust BLS gas prices

* Enable __BLST_PORTABLE__ flag only if PORTABLE flag is set

* Add tests for BLS_PAIRING

* GASCONSUMED instruction

* Fix compilation against docker with blst library; (#14)

* fix compilation against docker with blst library;
add precompiled libblst.a to android builds

* minor fix

* Adjust BLKSWX gas

* Fix comparison with NAN

* Allow arbitrary integers for scalars in ristretto multiplication, fix test

* Adjust nix builds according to PR 694 (#15)

* integrate and test PR-694

* integrate and test PR-694, test 2

* Add P256_CHKSIGN (secp256r1)

---------

Co-authored-by: SpyCheese <mikle98@yandex.ru>
Co-authored-by: neodiX42 <namlem@gmail.com>
2023-05-24 21:14:13 +03:00

1004 lines
27 KiB
C++

/*
This file is part of TON Blockchain Library.
TON Blockchain Library is free software: you can redistribute it and/or modify
it under the terms of the GNU Lesser General Public License as published by
the Free Software Foundation, either version 2 of the License, or
(at your option) any later version.
TON Blockchain Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public License
along with TON Blockchain Library. If not, see <http://www.gnu.org/licenses/>.
Copyright 2017-2020 Telegram Systems LLP
*/
#include "vm/stack.hpp"
#include "vm/continuation.h"
#include "vm/box.hpp"
#include "vm/atom.h"
#include "vm/vmstate.h"
#include "vm/boc.h"
#include "td/utils/misc.h"
namespace td {
template class td::Cnt<std::string>;
template class td::Ref<td::Cnt<std::string>>;
template class td::Cnt<std::vector<vm::StackEntry>>;
template class td::Ref<td::Cnt<std::vector<vm::StackEntry>>>;
} // namespace td
namespace vm {
// from_object_t from_object{};
const char* exception_messages[(int)(Excno::total)] = {
"normal termination", "alternative termination", "stack underflow", "stack overflow", "integer overflow",
"integer out of range", "invalid opcode", "type check error", "cell overflow", "cell underflow",
"dictionary error", "unknown error", "fatal error"};
const char* get_exception_msg(Excno exc_no) {
if (exc_no >= Excno::none && exc_no < Excno::total) {
return exception_messages[static_cast<int>(exc_no)];
} else {
return "unknown vm exception";
}
}
bool StackEntry::is_list(const StackEntry* se) {
Ref<Tuple> tuple;
while (!se->empty()) {
tuple = se->as_tuple_range(2, 2);
if (tuple.is_null()) {
return false;
}
se = &tuple->at(1);
}
return true;
}
static const char HEX_digits[] = "0123456789ABCDEF";
std::string str_to_hex(std::string data, std::string prefix) {
prefix.reserve(prefix.size() + data.size() * 2);
for (char c : data) {
prefix += HEX_digits[(c >> 4) & 15];
prefix += HEX_digits[c & 15];
}
return prefix;
}
std::string StackEntry::to_string() const {
std::ostringstream os;
dump(os);
return std::move(os).str();
}
std::string StackEntry::to_lisp_string() const {
std::ostringstream os;
print_list(os);
return std::move(os).str();
}
void StackEntry::dump(std::ostream& os, bool verbose) const {
switch (tp) {
case t_null:
os << "(null)";
break;
case t_int:
os << dec_string(as_int());
break;
case t_cell:
if (ref.not_null()) {
if (verbose) {
std::string serialized = "???";
auto boc = vm::std_boc_serialize(as_cell());
if (boc.is_ok()) {
serialized = td::buffer_to_hex(boc.move_as_ok().as_slice());
}
os << "C{" << serialized << "}";
} else {
os << "C{" << *as_cell() << "}";
}
} else {
os << "C{null}";
}
break;
case t_builder:
if (ref.not_null()) {
os << "BC{" << *as_builder() << "}";
} else {
os << "BC{null}";
}
break;
case t_slice: {
if (ref.not_null()) {
os << "CS{";
static_cast<Ref<CellSlice>>(ref)->dump(os, 1, false);
os << '}';
} else {
os << "CS{null}";
}
break;
}
case t_string:
os << "\"" << as_string() << "\"";
break;
case t_bytes:
os << "BYTES:" << str_to_hex(as_bytes());
break;
case t_box: {
os << "Box{" << (const void*)&*ref << "}";
break;
}
case t_atom:
os << as_atom();
break;
case t_tuple: {
const auto& tuple = *static_cast<Ref<Tuple>>(ref);
auto n = tuple.size();
if (!n) {
os << "[]";
} else if (n == 1) {
os << "[ ";
tuple[0].dump(os);
os << " ]";
} else {
os << "[ ";
for (const auto& entry : tuple) {
entry.dump(os);
os << ' ';
}
os << ']';
}
break;
}
case t_object: {
os << "Object{" << (const void*)&*ref << "}";
break;
}
case t_vmcont: {
if (ref.not_null()) {
if (verbose) {
os << "Cont{" << *as_cont() << "}";
} else {
os << "Cont{" << as_cont()->type() << "}";
}
} else {
os << "Cont{null}";
}
break;
}
default:
os << "???";
}
}
void StackEntry::print_list(std::ostream& os, bool verbose) const {
switch (tp) {
case t_null:
os << "()";
break;
case t_tuple: {
const auto& tuple = *static_cast<Ref<Tuple>>(ref);
if (is_list()) {
os << '(';
tuple[0].print_list(os, verbose);
print_list_tail(os, &tuple[1]);
break;
}
auto n = tuple.size();
if (!n) {
os << "[]";
} else if (n == 1) {
os << "[";
tuple[0].print_list(os, verbose);
os << "]";
} else {
os << "[";
unsigned c = 0;
for (const auto& entry : tuple) {
if (c++) {
os << " ";
}
entry.print_list(os, verbose);
}
os << ']';
}
break;
}
default:
dump(os, verbose);
}
}
void StackEntry::print_list_tail(std::ostream& os, const StackEntry* se) {
Ref<Tuple> tuple;
while (!se->empty()) {
tuple = se->as_tuple_range(2, 2);
if (tuple.is_null()) {
os << " . ";
se->print_list(os);
break;
}
os << ' ';
tuple->at(0).print_list(os);
se = &tuple->at(1);
}
os << ')';
}
StackEntry StackEntry::make_list(std::vector<StackEntry>&& elems) {
StackEntry tail;
std::size_t n = elems.size();
while (n > 0) {
--n;
tail = StackEntry{vm::make_tuple_ref(std::move(elems[n]), tail)};
}
return tail;
}
StackEntry StackEntry::make_list(const std::vector<StackEntry>& elems) {
StackEntry tail;
std::size_t n = elems.size();
while (n > 0) {
--n;
tail = StackEntry{vm::make_tuple_ref(elems[n], tail)};
}
return tail;
}
StackEntry::StackEntry(Ref<Stack> stack_ref) : ref(std::move(stack_ref)), tp(t_stack) {
}
StackEntry::StackEntry(Ref<Continuation> cont_ref) : ref(std::move(cont_ref)), tp(t_vmcont) {
}
Ref<Continuation> StackEntry::as_cont() const& {
return as<Continuation, t_vmcont>();
}
Ref<Continuation> StackEntry::as_cont() && {
return move_as<Continuation, t_vmcont>();
}
StackEntry::StackEntry(Ref<Box> box_ref) : ref(std::move(box_ref)), tp(t_box) {
}
Ref<Box> StackEntry::as_box() const& {
return as<Box, t_box>();
}
Ref<Box> StackEntry::as_box() && {
return move_as<Box, t_box>();
}
StackEntry::StackEntry(Ref<Tuple> tuple_ref) : ref(std::move(tuple_ref)), tp(t_tuple) {
}
StackEntry::StackEntry(const std::vector<StackEntry>& tuple_components)
: ref(Ref<Tuple>{true, tuple_components}), tp(t_tuple) {
}
StackEntry::StackEntry(std::vector<StackEntry>&& tuple_components)
: ref(Ref<Tuple>{true, std::move(tuple_components)}), tp(t_tuple) {
}
Ref<Tuple> StackEntry::as_tuple() const& {
return as<Tuple, t_tuple>();
}
Ref<Tuple> StackEntry::as_tuple() && {
return move_as<Tuple, t_tuple>();
}
Ref<Tuple> StackEntry::as_tuple_range(unsigned max_len, unsigned min_len) const& {
auto t = as<Tuple, t_tuple>();
if (t.not_null() && t->size() <= max_len && t->size() >= min_len) {
return t;
} else {
return {};
}
}
Ref<Tuple> StackEntry::as_tuple_range(unsigned max_len, unsigned min_len) && {
auto t = move_as<Tuple, t_tuple>();
if (t.not_null() && t->size() <= max_len && t->size() >= min_len) {
return t;
} else {
return {};
}
}
StackEntry::StackEntry(Ref<Atom> atom_ref) : ref(std::move(atom_ref)), tp(t_atom) {
}
Ref<Atom> StackEntry::as_atom() const& {
return as<Atom, t_atom>();
}
Ref<Atom> StackEntry::as_atom() && {
return move_as<Atom, t_atom>();
}
bool StackEntry::for_each_scalar(const std::function<bool(const StackEntry&)>& func) const {
auto t = as<Tuple, t_tuple>();
if (t.not_null()) {
for (const auto& entry : *t) {
if (!entry.for_each_scalar(func)) {
return false;
}
}
return true;
} else {
return func(*this);
}
}
void StackEntry::for_each_scalar(const std::function<void(const StackEntry&)>& func) const {
auto t = as<Tuple, t_tuple>();
if (t.not_null()) {
for (const auto& entry : *t) {
entry.for_each_scalar(func);
}
} else {
func(*this);
}
}
const StackEntry& tuple_index(const Ref<Tuple>& tup, unsigned idx) {
if (idx >= tup->size()) {
throw VmError{Excno::range_chk, "tuple index out of range"};
}
return (*tup)[idx];
}
StackEntry tuple_extend_index(const Ref<Tuple>& tup, unsigned idx) {
if (tup.is_null() || idx >= tup->size()) {
return {};
} else {
return tup->at(idx);
}
}
unsigned tuple_extend_set_index(Ref<Tuple>& tup, unsigned idx, StackEntry&& value, bool force) {
if (tup.is_null()) {
if (value.empty() && !force) {
return 0;
}
tup = Ref<Tuple>{true, idx + 1};
tup.unique_write().at(idx) = std::move(value);
return idx + 1;
}
if (tup->size() <= idx) {
if (value.empty() && !force) {
return 0;
}
auto& tuple = tup.write();
tuple.resize(idx + 1);
tuple.at(idx) = std::move(value);
return idx + 1;
} else {
tup.write().at(idx) = std::move(value);
return (unsigned)tup->size();
}
}
Stack::Stack(const Stack& old_stack, unsigned copy_elem, unsigned skip_top) {
push_from_stack(old_stack, copy_elem, skip_top);
}
Stack::Stack(Stack&& old_stack, unsigned copy_elem, unsigned skip_top) {
push_from_stack(old_stack, copy_elem, skip_top);
}
void Stack::push_from_stack(const Stack& old_stack, unsigned copy_elem, unsigned skip_top) {
unsigned n = old_stack.depth();
if (skip_top > n || copy_elem > n - skip_top) {
throw VmError{Excno::stk_und, "cannot construct stack from another one: not enough elements"};
}
stack.reserve(stack.size() + copy_elem);
auto it = old_stack.stack.cend() - skip_top;
std::copy(it - copy_elem, it, std::back_inserter(stack));
}
void Stack::push_from_stack(Stack&& old_stack, unsigned copy_elem, unsigned skip_top) {
unsigned n = old_stack.depth();
if (skip_top > n || copy_elem > n - skip_top) {
throw VmError{Excno::stk_und, "cannot construct stack from another one: not enough elements"};
}
stack.reserve(stack.size() + copy_elem);
auto it = old_stack.stack.cend() - skip_top;
std::move(it - copy_elem, it, std::back_inserter(stack));
}
void Stack::move_from_stack(Stack& old_stack, unsigned copy_elem) {
unsigned n = old_stack.depth();
if (copy_elem > n) {
throw VmError{Excno::stk_und, "cannot construct stack from another one: not enough elements"};
}
LOG(DEBUG) << "moving " << copy_elem << " top elements to another stack\n";
stack.reserve(stack.size() + copy_elem);
auto it = old_stack.stack.cend();
std::move(it - copy_elem, it, std::back_inserter(stack));
old_stack.pop_many(copy_elem);
}
void Stack::pop_null() {
check_underflow(1);
if (!pop().empty()) {
throw VmError{Excno::type_chk, "not an null"};
}
}
td::RefInt256 Stack::pop_int() {
check_underflow(1);
td::RefInt256 res = pop().as_int();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not an integer"};
}
return res;
}
td::RefInt256 Stack::pop_int_finite() {
auto res = pop_int();
if (!res->is_valid()) {
throw VmError{Excno::int_ov};
}
return res;
}
bool Stack::pop_bool() {
return sgn(pop_int_finite()) != 0;
}
long long Stack::pop_long() {
return pop_int()->to_long();
}
long long Stack::pop_long_range(long long max, long long min) {
auto res = pop_long();
if (res > max || res < min) {
throw VmError{Excno::range_chk};
}
return res;
}
int Stack::pop_smallint_range(int max, int min) {
return (int)pop_long_range(max, min);
}
Ref<Cell> Stack::pop_cell() {
check_underflow(1);
auto res = pop().as_cell();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a cell"};
}
return res;
}
Ref<Cell> Stack::pop_maybe_cell() {
check_underflow(1);
auto tmp = pop();
if (tmp.empty()) {
return {};
}
auto res = std::move(tmp).as_cell();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a cell"};
}
return res;
}
Ref<CellBuilder> Stack::pop_builder() {
check_underflow(1);
auto res = pop().as_builder();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a cell builder"};
}
return res;
}
Ref<CellSlice> Stack::pop_cellslice() {
check_underflow(1);
auto res = pop().as_slice();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a cell slice"};
}
return res;
}
std::string Stack::pop_string() {
check_underflow(1);
auto res = pop().as_string_ref();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a string"};
}
return *res;
}
std::string Stack::pop_bytes() {
check_underflow(1);
auto res = pop().as_bytes_ref();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a bytes chunk"};
}
return *res;
}
Ref<Continuation> Stack::pop_cont() {
check_underflow(1);
auto res = pop().as_cont();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a continuation"};
}
return res;
}
Ref<Box> Stack::pop_box() {
check_underflow(1);
auto res = pop().as_box();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a box"};
}
return res;
}
Ref<Tuple> Stack::pop_tuple() {
check_underflow(1);
auto res = pop().as_tuple();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a tuple"};
}
return res;
}
Ref<Tuple> Stack::pop_tuple_range(unsigned max_len, unsigned min_len) {
check_underflow(1);
auto res = pop().as_tuple();
if (res.is_null() || res->size() > max_len || res->size() < min_len) {
throw VmError{Excno::type_chk, "not a tuple of valid size"};
}
return res;
}
Ref<Tuple> Stack::pop_maybe_tuple() {
check_underflow(1);
auto val = pop();
if (val.empty()) {
return {};
}
auto res = std::move(val).as_tuple();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not a tuple"};
}
return res;
}
Ref<Tuple> Stack::pop_maybe_tuple_range(unsigned max_len) {
check_underflow(1);
auto val = pop();
if (val.empty()) {
return {};
}
auto res = std::move(val).as_tuple();
if (res.is_null() || res->size() > max_len) {
throw VmError{Excno::type_chk, "not a tuple of valid size"};
}
return res;
}
Ref<Atom> Stack::pop_atom() {
check_underflow(1);
auto res = pop().as_atom();
if (res.is_null()) {
throw VmError{Excno::type_chk, "not an atom"};
}
return res;
}
void Stack::push_null() {
push({});
}
void Stack::push_int(td::RefInt256 val) {
if (!val->signed_fits_bits(257)) {
throw VmError{Excno::int_ov};
}
push(std::move(val));
}
void Stack::push_int_quiet(td::RefInt256 val, bool quiet) {
if (!val->signed_fits_bits(257)) {
if (!quiet) {
throw VmError{Excno::int_ov};
} else if (val->is_valid()) {
push(td::make_refint());
return;
}
}
push(std::move(val));
}
void Stack::push_string(std::string str) {
push(std::move(str));
}
void Stack::push_string(td::Slice slice) {
push(slice.str());
}
void Stack::push_bytes(std::string str) {
push(std::move(str), true);
}
void Stack::push_bytes(td::Slice slice) {
push(slice.str(), true);
}
void Stack::push_cell(Ref<Cell> cell) {
push(std::move(cell));
}
void Stack::push_maybe_cell(Ref<Cell> cell) {
push_maybe(std::move(cell));
}
void Stack::push_builder(Ref<CellBuilder> cb) {
push(std::move(cb));
}
void Stack::push_smallint(long long val) {
push(td::make_refint(val));
}
void Stack::push_bool(bool val) {
push_smallint(val ? -1 : 0);
}
void Stack::push_cont(Ref<Continuation> cont) {
push(std::move(cont));
}
void Stack::push_box(Ref<Box> box) {
push(std::move(box));
}
void Stack::push_tuple(Ref<Tuple> tuple) {
push(std::move(tuple));
}
void Stack::push_maybe_tuple(Ref<Tuple> tuple) {
if (tuple.not_null()) {
push(std::move(tuple));
} else {
push_null();
}
}
void Stack::push_tuple(const std::vector<StackEntry>& components) {
push(components);
}
void Stack::push_tuple(std::vector<StackEntry>&& components) {
push(std::move(components));
}
void Stack::push_atom(Ref<Atom> atom) {
push(std::move(atom));
}
Ref<Stack> Stack::split_top(unsigned top_cnt, unsigned drop_cnt) {
unsigned n = depth();
if (top_cnt > n || drop_cnt > n - top_cnt) {
return Ref<Stack>{};
}
Ref<Stack> new_stk = Ref<Stack>{true};
if (top_cnt) {
new_stk.unique_write().move_from_stack(*this, top_cnt);
}
if (drop_cnt) {
pop_many(drop_cnt);
}
return new_stk;
}
void Stack::dump(std::ostream& os, int mode) const {
os << " [ ";
if (mode & 2) {
for (const auto& x : stack) {
x.print_list(os, mode & 4);
os << ' ';
}
} else {
for (const auto& x : stack) {
x.dump(os, mode & 4);
os << ' ';
}
}
os << "] ";
if (mode & 1) {
os << std::endl;
}
}
void Stack::push_cellslice(Ref<CellSlice> cs) {
push(std::move(cs));
}
void Stack::push_maybe_cellslice(Ref<CellSlice> cs) {
push_maybe(std::move(cs));
}
bool Stack::for_each_scalar(const std::function<bool(const StackEntry&)>& func) const {
for (const auto& v : stack) {
if (!v.for_each_scalar(func)) {
return false;
}
}
return true;
}
void Stack::for_each_scalar(const std::function<void(const StackEntry&)>& func) const {
for (const auto& v : stack) {
v.for_each_scalar(func);
}
}
/*
*
* SERIALIZE/DESERIALIZE STACK VALUES
*
*/
bool StackEntry::serialize(vm::CellBuilder& cb, int mode) const {
auto* vsi = VmStateInterface::get();
if (vsi && !vsi->register_op()) {
return false;
}
switch (tp) {
case t_null:
return cb.store_long_bool(0, 8); // vm_stk_null#00 = VmStackValue;
case t_int: {
auto val = as_int();
if (!val->is_valid()) {
// vm_stk_nan#02ff = VmStackValue;
return cb.store_long_bool(0x02ff, 16);
} else if (!(mode & 1) && val->signed_fits_bits(64)) {
// vm_stk_tinyint#01 value:int64 = VmStackValue;
return cb.store_long_bool(1, 8) && cb.store_int256_bool(std::move(val), 64);
} else {
// vm_stk_int#0201_ value:int257 = VmStackValue;
return cb.store_long_bool(0x0200 / 2, 15) && cb.store_int256_bool(std::move(val), 257);
}
}
case t_cell:
// vm_stk_cell#03 cell:^Cell = VmStackValue;
return cb.store_long_bool(3, 8) && cb.store_ref_bool(as_cell());
case t_slice: {
// _ cell:^Cell st_bits:(## 10) end_bits:(## 10) { st_bits <= end_bits }
// st_ref:(#<= 4) end_ref:(#<= 4) { st_ref <= end_ref } = VmCellSlice;
const auto& cs = *static_cast<Ref<CellSlice>>(ref);
return ((mode & 0x1000) || cb.store_long_bool(4, 8)) // vm_stk_slice#04 _:VmCellSlice = VmStackValue;
&& cb.store_ref_bool(cs.get_base_cell()) // _ cell:^Cell
&& cb.store_long_bool(cs.cur_pos(), 10) // st_bits:(## 10)
&& cb.store_long_bool(cs.cur_pos() + cs.size(), 10) // end_bits:(## 10)
&& cb.store_long_bool(cs.cur_ref(), 3) // st_ref:(#<= 4)
&& cb.store_long_bool(cs.cur_ref() + cs.size_refs(), 3); // end_ref:(#<= 4)
}
case t_builder:
// vm_stk_builder#05 cell:^Cell = VmStackValue;
return cb.store_long_bool(5, 8) && cb.store_ref_bool(as_builder()->finalize_copy());
case t_vmcont:
// vm_stk_cont#06 cont:VmCont = VmStackValue;
return !(mode & 2) && cb.store_long_bool(6, 8) && as_cont()->serialize(cb);
case t_tuple: {
const auto& tuple = *static_cast<Ref<Tuple>>(ref);
auto n = tuple.size();
// vm_stk_tuple#07 len:(## 16) data:(VmTuple len) = VmStackValue;
Ref<Cell> head, tail;
vm::CellBuilder cb2;
for (std::size_t i = 0; i < n; i++) {
std::swap(head, tail);
if (i > 1 &&
!(cb2.store_ref_bool(std::move(tail)) && cb2.store_ref_bool(std::move(head)) && cb2.finalize_to(head))) {
return false;
}
if (!(tuple[i].serialize(cb2, mode) && cb2.finalize_to(tail))) {
return false;
}
}
return cb.store_long_bool(7, 8) && cb.store_long_bool(n, 16) && (head.is_null() || cb.store_ref_bool(head)) &&
(tail.is_null() || cb.store_ref_bool(tail));
}
default:
return false;
}
}
bool StackEntry::deserialize(CellSlice& cs, int mode) {
auto* vsi = VmStateInterface::get();
if (vsi && !vsi->register_op()) {
return false;
}
clear();
int t = (mode & 0xf000) ? ((mode >> 12) & 15) : (int)cs.prefetch_ulong(8);
switch (t) {
case 0:
// vm_stk_null#00 = VmStackValue;
return cs.advance(8);
case 1: {
// vm_stk_tinyint#01 value:int64 = VmStackValue;
td::RefInt256 val;
return !(mode & 1) && cs.advance(8) && cs.fetch_int256_to(64, val) && set_int(std::move(val));
}
case 2: {
t = (int)cs.prefetch_ulong(16) & 0x1ff;
if (t == 0xff) {
// vm_stk_nan#02ff = VmStackValue;
return cs.advance(16) && set_int(td::make_refint());
} else {
// vm_stk_int#0201_ value:int257 = VmStackValue;
td::RefInt256 val;
return cs.fetch_ulong(15) == 0x0200 / 2 && cs.fetch_int256_to(257, val) && set_int(std::move(val));
}
}
case 3: {
// vm_stk_cell#03 cell:^Cell = VmStackValue;
return cs.have_refs() && cs.advance(8) && set(t_cell, cs.fetch_ref());
}
case 4: {
// _ cell:^Cell st_bits:(## 10) end_bits:(## 10) { st_bits <= end_bits }
// st_ref:(#<= 4) end_ref:(#<= 4) { st_ref <= end_ref } = VmCellSlice;
// vm_stk_slice#04 _:VmCellSlice = VmStackValue;
unsigned st_bits, end_bits, st_ref, end_ref;
Ref<Cell> cell;
Ref<CellSlice> csr;
return ((mode & 0xf000) || cs.advance(8)) // vm_stk_slice#04
&& cs.fetch_ref_to(cell) // cell:^Cell
&& cs.fetch_uint_to(10, st_bits) // st_bits:(## 10)
&& cs.fetch_uint_to(10, end_bits) // end_bits:(## 10)
&& st_bits <= end_bits // { st_bits <= end_bits }
&& cs.fetch_uint_to(3, st_ref) // st_ref:(#<= 4)
&& cs.fetch_uint_to(3, end_ref) // end_ref:(#<= 4)
&& st_ref <= end_ref && end_ref <= 4 // { st_ref <= end_ref }
&& (csr = load_cell_slice_ref(std::move(cell))).not_null() // load cell slice
&& csr->have(end_bits, end_ref) &&
csr.write().skip_last(csr->size() - end_bits, csr->size_refs() - end_ref) &&
csr.write().skip_first(st_bits, st_ref) && set(t_slice, std::move(csr));
}
case 5: {
// vm_stk_builder#05 cell:^Cell = VmStackValue;
Ref<Cell> cell;
Ref<CellSlice> csr;
Ref<CellBuilder> cb{true};
return cs.advance(8) && cs.fetch_ref_to(cell) && (csr = load_cell_slice_ref(std::move(cell))).not_null() &&
cb.write().append_cellslice_bool(std::move(csr)) && set(t_builder, std::move(cb));
}
case 6: {
// vm_stk_cont#06 cont:VmCont = VmStackValue;
Ref<Continuation> cont;
return !(mode & 2) && cs.advance(8) && Continuation::deserialize_to(cs, cont, mode) &&
set(t_vmcont, std::move(cont));
}
case 7: {
// vm_stk_tuple#07 len:(## 16) data:(VmTuple len) = VmStackValue;
int n;
if (!(cs.advance(8) && cs.fetch_uint_to(16, n))) {
return false;
}
Ref<Tuple> tuple{true, n};
auto& t = tuple.write();
if (n > 1) {
Ref<Cell> head, tail;
n--;
if (!(cs.fetch_ref_to(head) && cs.fetch_ref_to(tail) && t[n].deserialize(std::move(tail), mode))) {
return false;
}
vm::CellSlice cs2;
while (--n > 0) {
if (!(cs2.load(std::move(head)) && cs2.fetch_ref_to(head) && cs2.fetch_ref_to(tail) && cs2.empty_ext() &&
t[n].deserialize(std::move(tail), mode))) {
return false;
}
}
if (!t[0].deserialize(std::move(head), mode)) {
return false;
}
} else if (n == 1) {
return cs.have_refs() && t[0].deserialize(cs.fetch_ref(), mode) && set(t_tuple, std::move(tuple));
}
return set(t_tuple, std::move(tuple));
}
default:
return false;
}
}
bool StackEntry::deserialize(Ref<Cell> cell, int mode) {
if (cell.is_null()) {
clear();
return false;
}
CellSlice cs = load_cell_slice(std::move(cell));
return deserialize(cs, mode) && cs.empty_ext();
}
bool Stack::serialize(vm::CellBuilder& cb, int mode) const {
auto* vsi = VmStateInterface::get();
if (vsi && !vsi->register_op()) {
return false;
}
try {
// vm_stack#_ depth:(## 24) stack:(VmStackList depth) = VmStack;
unsigned n = depth();
if (!cb.store_ulong_rchk_bool(n, 24)) { // vm_stack#_ depth:(## 24)
return false;
}
if (!n) {
return true;
}
vm::CellBuilder cb2;
Ref<vm::Cell> rest = cb2.finalize(); // vm_stk_nil#_ = VmStackList 0;
for (unsigned i = 0; i < n - 1; i++) {
// vm_stk_cons#_ {n:#} rest:^(VmStackList n) tos:VmStackValue = VmStackList (n + 1);
if (!(cb2.store_ref_bool(std::move(rest)) && stack[i].serialize(cb2, mode) && cb2.finalize_to(rest))) {
return false;
}
}
return cb.store_ref_bool(std::move(rest)) && stack[n - 1].serialize(cb, mode);
} catch (CellBuilder::CellCreateError) {
return false;
} catch (CellBuilder::CellWriteError) {
return false;
}
}
bool Stack::deserialize(vm::CellSlice& cs, int mode) {
auto* vsi = VmStateInterface::get();
if (vsi && !vsi->register_op()) {
return false;
}
clear();
// vm_stack#_ depth:(## 24) stack:(VmStackList depth) = VmStack;
int n;
if (!cs.fetch_uint_to(24, n)) {
return false;
}
if (!n) {
return true;
}
stack.resize(n);
Ref<Cell> rest;
if (!(cs.fetch_ref_to(rest) && stack[n - 1].deserialize(cs, mode))) {
clear();
return false;
}
for (int i = n - 2; i >= 0; --i) {
// vm_stk_cons#_ {n:#} rest:^(VmStackList n) tos:VmStackValue = VmStackList (n + 1);
vm::CellSlice cs2 = load_cell_slice(std::move(rest));
if (!(cs2.fetch_ref_to(rest) && stack[i].deserialize(cs2, mode) && cs2.empty_ext())) {
clear();
return false;
}
}
if (!load_cell_slice(std::move(rest)).empty_ext()) {
clear();
return false;
}
return true;
}
bool Stack::deserialize_to(vm::CellSlice& cs, Ref<Stack>& stack, int mode) {
stack = Ref<Stack>{true};
if (stack.unique_write().deserialize(cs, mode)) {
return true;
} else {
stack.clear();
return false;
}
}
} // namespace vm