mirror of
				https://github.com/ton-blockchain/ton
				synced 2025-03-09 15:40:10 +00:00 
			
		
		
		
	Merge branch 'testnet' into block-generation
This commit is contained in:
		
						commit
						e4e77c16c5
					
				
					 463 changed files with 29976 additions and 2517 deletions
				
			
		|  | @ -74,7 +74,7 @@ void ArchiveManager::add_handle(BlockHandle handle, td::Promise<td::Unit> promis | |||
| } | ||||
| 
 | ||||
| void ArchiveManager::update_handle(BlockHandle handle, td::Promise<td::Unit> promise) { | ||||
|   FileDescription *f; | ||||
|   const FileDescription *f; | ||||
|   if (handle->handle_moved_to_archive()) { | ||||
|     CHECK(handle->inited_unix_time()); | ||||
|     if (!handle->need_flush()) { | ||||
|  | @ -439,15 +439,15 @@ void ArchiveManager::check_persistent_state(BlockIdExt block_id, BlockIdExt mast | |||
| 
 | ||||
| void ArchiveManager::get_block_by_unix_time(AccountIdPrefixFull account_id, UnixTime ts, | ||||
|                                             td::Promise<ConstBlockHandle> promise) { | ||||
|   auto f = get_file_desc_by_unix_time(account_id, ts, false); | ||||
|   if (f) { | ||||
|     auto n = f; | ||||
|     do { | ||||
|       n = get_next_file_desc(n); | ||||
|     } while (n != nullptr && !n->has_account_prefix(account_id)); | ||||
|   auto f1 = get_file_desc_by_unix_time(account_id, ts, false); | ||||
|   auto f2 = get_next_file_desc(f1, account_id, false); | ||||
|   if (!f1) { | ||||
|     std::swap(f1, f2); | ||||
|   } | ||||
|   if (f1) { | ||||
|     td::actor::ActorId<ArchiveSlice> aid; | ||||
|     if (n) { | ||||
|       aid = n->file_actor_id(); | ||||
|     if (f2) { | ||||
|       aid = f2->file_actor_id(); | ||||
|     } | ||||
|     auto P = td::PromiseCreator::lambda( | ||||
|         [aid, account_id, ts, promise = std::move(promise)](td::Result<ConstBlockHandle> R) mutable { | ||||
|  | @ -457,7 +457,7 @@ void ArchiveManager::get_block_by_unix_time(AccountIdPrefixFull account_id, Unix | |||
|             td::actor::send_closure(aid, &ArchiveSlice::get_block_by_unix_time, account_id, ts, std::move(promise)); | ||||
|           } | ||||
|         }); | ||||
|     td::actor::send_closure(f->file_actor_id(), &ArchiveSlice::get_block_by_unix_time, account_id, ts, std::move(P)); | ||||
|     td::actor::send_closure(f1->file_actor_id(), &ArchiveSlice::get_block_by_unix_time, account_id, ts, std::move(P)); | ||||
|   } else { | ||||
|     promise.set_error(td::Status::Error(ErrorCode::notready, "ts not in db")); | ||||
|   } | ||||
|  | @ -465,15 +465,15 @@ void ArchiveManager::get_block_by_unix_time(AccountIdPrefixFull account_id, Unix | |||
| 
 | ||||
| void ArchiveManager::get_block_by_lt(AccountIdPrefixFull account_id, LogicalTime lt, | ||||
|                                      td::Promise<ConstBlockHandle> promise) { | ||||
|   auto f = get_file_desc_by_lt(account_id, lt, false); | ||||
|   if (f) { | ||||
|     auto n = f; | ||||
|     do { | ||||
|       n = get_next_file_desc(n); | ||||
|     } while (n != nullptr && !n->has_account_prefix(account_id)); | ||||
|   auto f1 = get_file_desc_by_lt(account_id, lt, false); | ||||
|   auto f2 = get_next_file_desc(f1, account_id, false); | ||||
|   if (!f1) { | ||||
|     std::swap(f1, f2); | ||||
|   } | ||||
|   if (f1) { | ||||
|     td::actor::ActorId<ArchiveSlice> aid; | ||||
|     if (n) { | ||||
|       aid = n->file_actor_id(); | ||||
|     if (f2) { | ||||
|       aid = f2->file_actor_id(); | ||||
|     } | ||||
|     auto P = td::PromiseCreator::lambda( | ||||
|         [aid, account_id, lt, promise = std::move(promise)](td::Result<ConstBlockHandle> R) mutable { | ||||
|  | @ -483,7 +483,7 @@ void ArchiveManager::get_block_by_lt(AccountIdPrefixFull account_id, LogicalTime | |||
|             td::actor::send_closure(aid, &ArchiveSlice::get_block_by_lt, account_id, lt, std::move(promise)); | ||||
|           } | ||||
|         }); | ||||
|     td::actor::send_closure(f->file_actor_id(), &ArchiveSlice::get_block_by_lt, account_id, lt, std::move(P)); | ||||
|     td::actor::send_closure(f1->file_actor_id(), &ArchiveSlice::get_block_by_lt, account_id, lt, std::move(P)); | ||||
|   } else { | ||||
|     promise.set_error(td::Status::Error(ErrorCode::notready, "lt not in db")); | ||||
|   } | ||||
|  | @ -558,11 +558,16 @@ void ArchiveManager::deleted_package(PackageId id, td::Promise<td::Unit> promise | |||
|   auto it = m.find(id); | ||||
|   CHECK(it != m.end()); | ||||
|   CHECK(it->second.deleted); | ||||
|   it->second.clear_actor_id(); | ||||
|   it->second.file.reset(); | ||||
|   promise.set_value(td::Unit()); | ||||
| } | ||||
| 
 | ||||
| void ArchiveManager::load_package(PackageId id) { | ||||
|   auto &m = get_file_map(id); | ||||
|   if (m.count(id)) { | ||||
|     LOG(WARNING) << "Duplicate id " << id.name(); | ||||
|     return; | ||||
|   } | ||||
|   auto key = create_serialize_tl_object<ton_api::db_files_package_key>(id.id, id.key, id.temp); | ||||
| 
 | ||||
|   std::string value; | ||||
|  | @ -595,11 +600,11 @@ void ArchiveManager::load_package(PackageId id) { | |||
| 
 | ||||
|   desc.file = td::actor::create_actor<ArchiveSlice>("slice", id.id, id.key, id.temp, false, db_root_); | ||||
| 
 | ||||
|   get_file_map(id).emplace(id, std::move(desc)); | ||||
|   m.emplace(id, std::move(desc)); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, | ||||
|                                                                UnixTime ts, LogicalTime lt, bool force) { | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, | ||||
|                                                                      UnixTime ts, LogicalTime lt, bool force) { | ||||
|   auto &f = get_file_map(id); | ||||
|   auto it = f.find(id); | ||||
|   if (it != f.end()) { | ||||
|  | @ -607,7 +612,7 @@ ArchiveManager::FileDescription *ArchiveManager::get_file_desc(ShardIdFull shard | |||
|       return nullptr; | ||||
|     } | ||||
|     if (force && !id.temp) { | ||||
|       update_desc(it->second, shard, seqno, ts, lt); | ||||
|       update_desc(f, it->second, shard, seqno, ts, lt); | ||||
|     } | ||||
|     return &it->second; | ||||
|   } | ||||
|  | @ -618,17 +623,18 @@ ArchiveManager::FileDescription *ArchiveManager::get_file_desc(ShardIdFull shard | |||
|   return add_file_desc(shard, id, seqno, ts, lt); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::add_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, | ||||
|                                                                UnixTime ts, LogicalTime lt) { | ||||
| const ArchiveManager::FileDescription *ArchiveManager::add_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, | ||||
|                                                                      UnixTime ts, LogicalTime lt) { | ||||
|   auto &f = get_file_map(id); | ||||
|   CHECK(f.count(id) == 0); | ||||
| 
 | ||||
|   FileDescription desc{id, false}; | ||||
|   FileDescription new_desc{id, false}; | ||||
|   td::mkdir(db_root_ + id.path()).ensure(); | ||||
|   std::string prefix = PSTRING() << db_root_ << id.path() << id.name(); | ||||
|   desc.file = td::actor::create_actor<ArchiveSlice>("slice", id.id, id.key, id.temp, false, db_root_); | ||||
|   new_desc.file = td::actor::create_actor<ArchiveSlice>("slice", id.id, id.key, id.temp, false, db_root_); | ||||
|   const FileDescription &desc = f.emplace(id, std::move(new_desc)); | ||||
|   if (!id.temp) { | ||||
|     update_desc(desc, shard, seqno, ts, lt); | ||||
|     update_desc(f, desc, shard, seqno, ts, lt); | ||||
|   } | ||||
| 
 | ||||
|   std::vector<tl_object_ptr<ton_api::db_files_package_firstBlock>> vec; | ||||
|  | @ -652,7 +658,6 @@ ArchiveManager::FileDescription *ArchiveManager::add_file_desc(ShardIdFull shard | |||
|     for (auto &e : temp_files_) { | ||||
|       tt.push_back(e.first.id); | ||||
|     } | ||||
|     (id.temp ? tt : (id.key ? tk : t)).push_back(id.id); | ||||
|     index_ | ||||
|         ->set(create_serialize_tl_object<ton_api::db_files_index_key>().as_slice(), | ||||
|               create_serialize_tl_object<ton_api::db_files_index_value>(std::move(t), std::move(tk), std::move(tt)) | ||||
|  | @ -668,17 +673,16 @@ ArchiveManager::FileDescription *ArchiveManager::add_file_desc(ShardIdFull shard | |||
|         .ensure(); | ||||
|   } | ||||
|   index_->commit_transaction().ensure(); | ||||
| 
 | ||||
|   return &f.emplace(id, std::move(desc)).first->second; | ||||
|   return &desc; | ||||
| } | ||||
| 
 | ||||
| void ArchiveManager::update_desc(FileDescription &desc, ShardIdFull shard, BlockSeqno seqno, UnixTime ts, | ||||
|                                  LogicalTime lt) { | ||||
| void ArchiveManager::update_desc(FileMap &f, const FileDescription &desc, ShardIdFull shard, BlockSeqno seqno, | ||||
|                                  UnixTime ts, LogicalTime lt) { | ||||
|   auto it = desc.first_blocks.find(shard); | ||||
|   if (it != desc.first_blocks.end() && it->second.seqno <= seqno) { | ||||
|     return; | ||||
|   } | ||||
|   desc.first_blocks[shard] = FileDescription::Desc{seqno, ts, lt}; | ||||
|   f.set_shard_first_block(desc, shard, FileDescription::Desc{seqno, ts, lt}); | ||||
|   std::vector<tl_object_ptr<ton_api::db_files_package_firstBlock>> vec; | ||||
|   for (auto &e : desc.first_blocks) { | ||||
|     vec.push_back(create_tl_object<ton_api::db_files_package_firstBlock>(e.first.workchain, e.first.shard, | ||||
|  | @ -694,150 +698,91 @@ void ArchiveManager::update_desc(FileDescription &desc, ShardIdFull shard, Block | |||
|   index_->commit_transaction().ensure(); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_seqno(ShardIdFull shard, BlockSeqno seqno, | ||||
|                                                                         bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     auto i = it->second.first_blocks.find(shard); | ||||
|     if (i != it->second.first_blocks.end() && i->second.seqno <= seqno) { | ||||
|       if (it->second.deleted) { | ||||
|         return nullptr; | ||||
|       } else { | ||||
|         return &it->second; | ||||
|       } | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_seqno(ShardIdFull shard, BlockSeqno seqno, | ||||
|                                                                               bool key_block) { | ||||
|   return get_file_map(PackageId{0, key_block, false}).get_file_desc_by_seqno(shard, seqno); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_unix_time(ShardIdFull shard, UnixTime ts, | ||||
|                                                                             bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     auto i = it->second.first_blocks.find(shard); | ||||
|     if (i != it->second.first_blocks.end() && i->second.ts <= ts) { | ||||
|       if (it->second.deleted) { | ||||
|         return nullptr; | ||||
|       } else { | ||||
|         return &it->second; | ||||
|       } | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_unix_time(ShardIdFull shard, UnixTime ts, | ||||
|                                                                                   bool key_block) { | ||||
|   return get_file_map(PackageId{0, key_block, false}).get_file_desc_by_unix_time(shard, ts); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_lt(ShardIdFull shard, LogicalTime lt, | ||||
|                                                                      bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     auto i = it->second.first_blocks.find(shard); | ||||
|     if (i != it->second.first_blocks.end() && i->second.lt <= lt) { | ||||
|       if (it->second.deleted) { | ||||
|         return nullptr; | ||||
|       } else { | ||||
|         return &it->second; | ||||
|       } | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_lt(ShardIdFull shard, LogicalTime lt, | ||||
|                                                                            bool key_block) { | ||||
|   return get_file_map(PackageId{0, key_block, false}).get_file_desc_by_lt(shard, lt); | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_seqno(AccountIdPrefixFull account, BlockSeqno seqno, | ||||
|                                                                         bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_seqno(AccountIdPrefixFull account, | ||||
|                                                                               BlockSeqno seqno, bool key_block) { | ||||
|   if (account.is_masterchain()) { | ||||
|     return get_file_desc_by_seqno(ShardIdFull{masterchainId}, seqno, key_block); | ||||
|   } | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     if (it->second.deleted) { | ||||
|       continue; | ||||
|     } | ||||
|     bool found = false; | ||||
|     for (int i = 0; i < 60; i++) { | ||||
|       auto shard = shard_prefix(account, i); | ||||
|       auto it2 = it->second.first_blocks.find(shard); | ||||
|       if (it2 != it->second.first_blocks.end()) { | ||||
|         if (it2->second.seqno <= seqno) { | ||||
|           return &it->second; | ||||
|         } | ||||
|         found = true; | ||||
|       } else if (found) { | ||||
|         break; | ||||
|       } | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   const FileDescription *result = nullptr; | ||||
|   for (int i = 0; i <= 60; i++) { | ||||
|     const FileDescription *desc = f.get_file_desc_by_seqno(shard_prefix(account, i), seqno); | ||||
|     if (desc && (!result || result->id < desc->id)) { | ||||
|       result = desc; | ||||
|     } else if (result && (!desc || desc->id < result->id)) { | ||||
|       break; | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
|   return result; | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_unix_time(AccountIdPrefixFull account, UnixTime ts, | ||||
|                                                                             bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_unix_time(AccountIdPrefixFull account, | ||||
|                                                                                   UnixTime ts, bool key_block) { | ||||
|   if (account.is_masterchain()) { | ||||
|     return get_file_desc_by_unix_time(ShardIdFull{masterchainId}, ts, key_block); | ||||
|   } | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     if (it->second.deleted) { | ||||
|       continue; | ||||
|     } | ||||
|     bool found = false; | ||||
|     for (int i = 0; i < 60; i++) { | ||||
|       auto shard = shard_prefix(account, i); | ||||
|       auto it2 = it->second.first_blocks.find(shard); | ||||
|       if (it2 != it->second.first_blocks.end()) { | ||||
|         if (it2->second.ts <= ts) { | ||||
|           return &it->second; | ||||
|         } | ||||
|         found = true; | ||||
|       } else if (found) { | ||||
|         break; | ||||
|       } | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   const FileDescription *result = nullptr; | ||||
|   for (int i = 0; i <= 60; i++) { | ||||
|     const FileDescription *desc = f.get_file_desc_by_unix_time(shard_prefix(account, i), ts); | ||||
|     if (desc && (!result || result->id < desc->id)) { | ||||
|       result = desc; | ||||
|     } else if (result && (!desc || desc->id < result->id)) { | ||||
|       break; | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
|   return result; | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_lt(AccountIdPrefixFull account, LogicalTime lt, | ||||
|                                                                      bool key_block) { | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_file_desc_by_lt(AccountIdPrefixFull account, LogicalTime lt, | ||||
|                                                                            bool key_block) { | ||||
|   if (account.is_masterchain()) { | ||||
|     return get_file_desc_by_lt(ShardIdFull{masterchainId}, lt, key_block); | ||||
|   } | ||||
|   for (auto it = f.rbegin(); it != f.rend(); it++) { | ||||
|     if (it->second.deleted) { | ||||
|       continue; | ||||
|     } | ||||
|     bool found = false; | ||||
|     for (int i = 0; i < 60; i++) { | ||||
|       auto shard = shard_prefix(account, i); | ||||
|       auto it2 = it->second.first_blocks.find(shard); | ||||
|       if (it2 != it->second.first_blocks.end()) { | ||||
|         if (it2->second.lt <= lt) { | ||||
|           return &it->second; | ||||
|         } | ||||
|         found = true; | ||||
|       } else if (found) { | ||||
|         break; | ||||
|       } | ||||
|   auto &f = get_file_map(PackageId{0, key_block, false}); | ||||
|   const FileDescription *result = nullptr; | ||||
|   for (int i = 0; i <= 60; i++) { | ||||
|     const FileDescription *desc = f.get_file_desc_by_lt(shard_prefix(account, i), lt); | ||||
|     if (desc && (!result || result->id < desc->id)) { | ||||
|       result = desc; | ||||
|     } else if (result && (!desc || desc->id < result->id)) { | ||||
|       break; | ||||
|     } | ||||
|   } | ||||
|   return nullptr; | ||||
|   return result; | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_next_file_desc(FileDescription *f) { | ||||
|   auto &m = get_file_map(f->id); | ||||
|   auto it = m.find(f->id); | ||||
|   CHECK(it != m.end()); | ||||
|   while (true) { | ||||
|     it++; | ||||
|     if (it == m.end()) { | ||||
|       return nullptr; | ||||
|     } else if (!it->second.deleted) { | ||||
|       return &it->second; | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_next_file_desc(const FileDescription *f, | ||||
|                                                                           AccountIdPrefixFull shard, bool key_block) { | ||||
|   auto &m = get_file_map(PackageId{0, key_block, false}); | ||||
|   const FileDescription *result = nullptr; | ||||
|   for (int i = 0; i <= 60; i++) { | ||||
|     const FileDescription *desc = m.get_next_file_desc(shard_prefix(shard, i), f); | ||||
|     if (desc && (!result || desc->id < result->id)) { | ||||
|       result = desc; | ||||
|     } else if (result && (!desc || result->id < desc->id)) { | ||||
|       break; | ||||
|     } | ||||
|   } | ||||
|   return result; | ||||
| } | ||||
| 
 | ||||
| ArchiveManager::FileDescription *ArchiveManager::get_temp_file_desc_by_idx(PackageId idx) { | ||||
| const ArchiveManager::FileDescription *ArchiveManager::get_temp_file_desc_by_idx(PackageId idx) { | ||||
|   auto it = temp_files_.find(idx); | ||||
|   if (it != temp_files_.end()) { | ||||
|     if (it->second.deleted) { | ||||
|  | @ -1257,14 +1202,100 @@ void ArchiveManager::truncate(BlockSeqno masterchain_seqno, ConstBlockHandle han | |||
|   } | ||||
| } | ||||
| 
 | ||||
| bool ArchiveManager::FileDescription::has_account_prefix(AccountIdPrefixFull account_id) const { | ||||
|   for (int i = 0; i < 60; i++) { | ||||
|     auto shard = shard_prefix(account_id, i); | ||||
|     if (first_blocks.count(shard)) { | ||||
|       return true; | ||||
|     } | ||||
| void ArchiveManager::FileMap::shard_index_add(const FileDescription &desc) { | ||||
|   for (const auto &p : desc.first_blocks) { | ||||
|     ShardIndex &s = shards_[p.first]; | ||||
|     s.seqno_index_[p.second.seqno] = &desc; | ||||
|     s.lt_index_[p.second.lt] = &desc; | ||||
|     s.unix_time_index_[p.second.ts] = &desc; | ||||
|     s.packages_index_[desc.id] = &desc; | ||||
|   } | ||||
|   return false; | ||||
| } | ||||
| 
 | ||||
| void ArchiveManager::FileMap::shard_index_del(const FileDescription &desc) { | ||||
|   for (const auto &p : desc.first_blocks) { | ||||
|     ShardIndex &s = shards_[p.first]; | ||||
|     s.seqno_index_.erase(p.second.seqno); | ||||
|     s.lt_index_.erase(p.second.lt); | ||||
|     s.unix_time_index_.erase(p.second.ts); | ||||
|     s.packages_index_.erase(desc.id); | ||||
|   } | ||||
| } | ||||
| 
 | ||||
| void ArchiveManager::FileMap::set_shard_first_block(const FileDescription &desc, ShardIdFull shard, | ||||
|                                                     FileDescription::Desc v) { | ||||
|   ShardIndex &s = shards_[shard]; | ||||
|   auto &d = const_cast<FileDescription &>(desc); | ||||
|   auto it = d.first_blocks.find(shard); | ||||
|   if (it != d.first_blocks.end()) { | ||||
|     s.seqno_index_.erase(it->second.seqno); | ||||
|     s.lt_index_.erase(it->second.lt); | ||||
|     s.unix_time_index_.erase(it->second.ts); | ||||
|   } | ||||
|   d.first_blocks[shard] = v; | ||||
|   s.seqno_index_[v.seqno] = &d; | ||||
|   s.lt_index_[v.lt] = &d; | ||||
|   s.unix_time_index_[v.ts] = &d; | ||||
|   s.packages_index_[d.id] = &d; | ||||
| } | ||||
| 
 | ||||
| const ArchiveManager::FileDescription *ArchiveManager::FileMap::get_file_desc_by_seqno(ShardIdFull shard, | ||||
|                                                                                        BlockSeqno seqno) const { | ||||
|   auto it = shards_.find(shard); | ||||
|   if (it == shards_.end()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   const auto &map = it->second.seqno_index_; | ||||
|   auto it2 = map.upper_bound(seqno); | ||||
|   if (it2 == map.begin()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   --it2; | ||||
|   return it2->second->deleted ? nullptr : it2->second; | ||||
| } | ||||
| 
 | ||||
| const ArchiveManager::FileDescription *ArchiveManager::FileMap::get_file_desc_by_lt(ShardIdFull shard, | ||||
|                                                                                     LogicalTime lt) const { | ||||
|   auto it = shards_.find(shard); | ||||
|   if (it == shards_.end()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   const auto &map = it->second.lt_index_; | ||||
|   auto it2 = map.upper_bound(lt); | ||||
|   if (it2 == map.begin()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   --it2; | ||||
|   return it2->second->deleted ? nullptr : it2->second; | ||||
| } | ||||
| 
 | ||||
| const ArchiveManager::FileDescription *ArchiveManager::FileMap::get_file_desc_by_unix_time(ShardIdFull shard, | ||||
|                                                                                            UnixTime ts) const { | ||||
|   auto it = shards_.find(shard); | ||||
|   if (it == shards_.end()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   const auto &map = it->second.unix_time_index_; | ||||
|   auto it2 = map.upper_bound(ts); | ||||
|   if (it2 == map.begin()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   --it2; | ||||
|   return it2->second->deleted ? nullptr : it2->second; | ||||
| } | ||||
| 
 | ||||
| const ArchiveManager::FileDescription *ArchiveManager::FileMap::get_next_file_desc(ShardIdFull shard, | ||||
|                                                                                    const FileDescription *desc) const { | ||||
|   auto it = shards_.find(shard); | ||||
|   if (it == shards_.end()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   const auto &map = it->second.packages_index_; | ||||
|   auto it2 = desc ? map.upper_bound(desc->id) : map.begin(); | ||||
|   if (it2 == map.end()) { | ||||
|     return nullptr; | ||||
|   } | ||||
|   return it2->second->deleted ? nullptr : it2->second; | ||||
| } | ||||
| 
 | ||||
| }  // namespace validator
 | ||||
|  |  | |||
|  | @ -71,7 +71,6 @@ class ArchiveManager : public td::actor::Actor { | |||
| 
 | ||||
|   void start_up() override; | ||||
| 
 | ||||
|   void begin_transaction(); | ||||
|   void commit_transaction(); | ||||
|   void set_async_mode(bool mode, td::Promise<td::Unit> promise); | ||||
| 
 | ||||
|  | @ -94,26 +93,83 @@ class ArchiveManager : public td::actor::Actor { | |||
|     auto file_actor_id() const { | ||||
|       return file.get(); | ||||
|     } | ||||
|     void clear_actor_id() { | ||||
|       file.reset(); | ||||
|     } | ||||
|     bool has_account_prefix(AccountIdPrefixFull account_id) const; | ||||
|     PackageId id; | ||||
|     bool deleted; | ||||
|     mutable bool deleted; | ||||
| 
 | ||||
|     std::map<ShardIdFull, Desc> first_blocks; | ||||
|     td::actor::ActorOwn<ArchiveSlice> file; | ||||
|     mutable td::actor::ActorOwn<ArchiveSlice> file; | ||||
|   }; | ||||
| 
 | ||||
|   std::map<PackageId, FileDescription> files_; | ||||
|   std::map<PackageId, FileDescription> key_files_; | ||||
|   std::map<PackageId, FileDescription> temp_files_; | ||||
|   class FileMap { | ||||
|    public: | ||||
|     std::map<PackageId, FileDescription>::const_iterator begin() const { | ||||
|       return files_.cbegin(); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_iterator end() const { | ||||
|       return files_.cend(); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_reverse_iterator rbegin() const { | ||||
|       return files_.crbegin(); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_reverse_iterator rend() const { | ||||
|       return files_.crend(); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_iterator find(PackageId x) const { | ||||
|       return files_.find(x); | ||||
|     } | ||||
|     size_t count(const PackageId &x) const { | ||||
|       return files_.count(x); | ||||
|     } | ||||
|     size_t size() const { | ||||
|       return files_.size(); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_iterator lower_bound(const PackageId &x) const { | ||||
|       return files_.lower_bound(x); | ||||
|     } | ||||
|     std::map<PackageId, FileDescription>::const_iterator upper_bound(const PackageId &x) const { | ||||
|       return files_.upper_bound(x); | ||||
|     } | ||||
|     void clear() { | ||||
|       files_.clear(); | ||||
|       shards_.clear(); | ||||
|     } | ||||
|     const FileDescription &emplace(const PackageId &id, FileDescription desc) { | ||||
|       auto it = files_.emplace(id, std::move(desc)); | ||||
|       if (it.second) { | ||||
|         shard_index_add(it.first->second); | ||||
|       } | ||||
|       return it.first->second; | ||||
|     } | ||||
|     void erase(std::map<PackageId, FileDescription>::const_iterator it) { | ||||
|       shard_index_del(it->second); | ||||
|       files_.erase(it); | ||||
|     } | ||||
|     void set_shard_first_block(const FileDescription &desc, ShardIdFull shard, FileDescription::Desc v); | ||||
|     const FileDescription *get_file_desc_by_seqno(ShardIdFull shard, BlockSeqno seqno) const; | ||||
|     const FileDescription *get_file_desc_by_lt(ShardIdFull shard, LogicalTime lt) const; | ||||
|     const FileDescription *get_file_desc_by_unix_time(ShardIdFull shard, UnixTime ts) const; | ||||
|     const FileDescription *get_next_file_desc(ShardIdFull shard, const FileDescription *desc) const; | ||||
| 
 | ||||
|    private: | ||||
|     std::map<PackageId, FileDescription> files_; | ||||
|     struct ShardIndex { | ||||
|       std::map<BlockSeqno, const FileDescription *> seqno_index_; | ||||
|       std::map<LogicalTime, const FileDescription *> lt_index_; | ||||
|       std::map<UnixTime, const FileDescription *> unix_time_index_; | ||||
|       std::map<PackageId, const FileDescription *> packages_index_; | ||||
|     }; | ||||
|     std::map<ShardIdFull, ShardIndex> shards_; | ||||
| 
 | ||||
|     void shard_index_add(const FileDescription &desc); | ||||
|     void shard_index_del(const FileDescription &desc); | ||||
|   }; | ||||
|   FileMap files_, key_files_, temp_files_; | ||||
|   BlockSeqno finalized_up_to_{0}; | ||||
|   bool async_mode_ = false; | ||||
|   bool huge_transaction_started_ = false; | ||||
|   td::uint32 huge_transaction_size_ = 0; | ||||
| 
 | ||||
|   auto &get_file_map(const PackageId &p) { | ||||
|   FileMap &get_file_map(const PackageId &p) { | ||||
|     return p.key ? key_files_ : p.temp ? temp_files_ : files_; | ||||
|   } | ||||
| 
 | ||||
|  | @ -126,18 +182,19 @@ class ArchiveManager : public td::actor::Actor { | |||
|   void get_handle_finish(BlockHandle handle, td::Promise<BlockHandle> promise); | ||||
|   void get_file_short_cont(FileReference ref_id, PackageId idx, td::Promise<td::BufferSlice> promise); | ||||
| 
 | ||||
|   FileDescription *get_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, UnixTime ts, LogicalTime lt, | ||||
|                                  bool force); | ||||
|   FileDescription *add_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, UnixTime ts, LogicalTime lt); | ||||
|   void update_desc(FileDescription &desc, ShardIdFull shard, BlockSeqno seqno, UnixTime ts, LogicalTime lt); | ||||
|   FileDescription *get_file_desc_by_seqno(ShardIdFull shard, BlockSeqno seqno, bool key_block); | ||||
|   FileDescription *get_file_desc_by_lt(ShardIdFull shard, LogicalTime lt, bool key_block); | ||||
|   FileDescription *get_file_desc_by_unix_time(ShardIdFull shard, UnixTime ts, bool key_block); | ||||
|   FileDescription *get_file_desc_by_seqno(AccountIdPrefixFull shard, BlockSeqno seqno, bool key_block); | ||||
|   FileDescription *get_file_desc_by_lt(AccountIdPrefixFull shard, LogicalTime lt, bool key_block); | ||||
|   FileDescription *get_file_desc_by_unix_time(AccountIdPrefixFull shard, UnixTime ts, bool key_block); | ||||
|   FileDescription *get_next_file_desc(FileDescription *f); | ||||
|   FileDescription *get_temp_file_desc_by_idx(PackageId idx); | ||||
|   const FileDescription *get_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, UnixTime ts, LogicalTime lt, | ||||
|                                        bool force); | ||||
|   const FileDescription *add_file_desc(ShardIdFull shard, PackageId id, BlockSeqno seqno, UnixTime ts, LogicalTime lt); | ||||
|   void update_desc(FileMap &f, const FileDescription &desc, ShardIdFull shard, BlockSeqno seqno, UnixTime ts, | ||||
|                    LogicalTime lt); | ||||
|   const FileDescription *get_file_desc_by_seqno(ShardIdFull shard, BlockSeqno seqno, bool key_block); | ||||
|   const FileDescription *get_file_desc_by_lt(ShardIdFull shard, LogicalTime lt, bool key_block); | ||||
|   const FileDescription *get_file_desc_by_unix_time(ShardIdFull shard, UnixTime ts, bool key_block); | ||||
|   const FileDescription *get_file_desc_by_seqno(AccountIdPrefixFull shard, BlockSeqno seqno, bool key_block); | ||||
|   const FileDescription *get_file_desc_by_lt(AccountIdPrefixFull shard, LogicalTime lt, bool key_block); | ||||
|   const FileDescription *get_file_desc_by_unix_time(AccountIdPrefixFull shard, UnixTime ts, bool key_block); | ||||
|   const FileDescription *get_next_file_desc(const FileDescription *f, AccountIdPrefixFull shard, bool key_block); | ||||
|   const FileDescription *get_temp_file_desc_by_idx(PackageId idx); | ||||
|   PackageId get_max_temp_file_desc_idx(); | ||||
|   PackageId get_prev_temp_file_desc_idx(PackageId id); | ||||
| 
 | ||||
|  |  | |||
|  | @ -32,6 +32,7 @@ | |||
| #include "vm/cells/MerkleUpdate.h" | ||||
| #include <map> | ||||
| #include <queue> | ||||
| #include "common/global-version.h" | ||||
| 
 | ||||
| namespace ton { | ||||
| 
 | ||||
|  | @ -40,7 +41,7 @@ using td::Ref; | |||
| 
 | ||||
| class Collator final : public td::actor::Actor { | ||||
|   static constexpr int supported_version() { | ||||
|     return 3; | ||||
|     return SUPPORTED_VERSION; | ||||
|   } | ||||
|   static constexpr long long supported_capabilities() { | ||||
|     return ton::capCreateStatsEnabled | ton::capBounceMsgBody | ton::capReportVersion | ton::capShortDequeue; | ||||
|  | @ -72,7 +73,7 @@ class Collator final : public td::actor::Actor { | |||
|   Ref<ValidatorSet> validator_set_; | ||||
|   td::actor::ActorId<ValidatorManager> manager; | ||||
|   td::Timestamp timeout; | ||||
|   td::Timestamp soft_timeout_, medium_timeout_; | ||||
|   td::Timestamp queue_cleanup_timeout_, soft_timeout_, medium_timeout_; | ||||
|   td::Promise<BlockCandidate> main_promise; | ||||
|   ton::BlockSeqno last_block_seqno{0}; | ||||
|   ton::BlockSeqno prev_mc_block_seqno{0}; | ||||
|  |  | |||
|  | @ -66,6 +66,8 @@ Collator::Collator(ShardIdFull shard, bool is_hardfork, BlockIdExt min_mastercha | |||
|     , validator_set_(std::move(validator_set)) | ||||
|     , manager(manager) | ||||
|     , timeout(timeout) | ||||
|     // default timeout is 10 seconds, declared in validator/validator-group.cpp:generate_block_candidate:run_collate_query
 | ||||
|     , queue_cleanup_timeout_(td::Timestamp::at(timeout.at() - 5.0)) | ||||
|     , soft_timeout_(td::Timestamp::at(timeout.at() - 3.0)) | ||||
|     , medium_timeout_(td::Timestamp::at(timeout.at() - 1.5)) | ||||
|     , main_promise(std::move(promise)) | ||||
|  | @ -535,6 +537,7 @@ bool Collator::unpack_last_mc_state() { | |||
|       mc_state_root, | ||||
|       block::ConfigInfo::needShardHashes | block::ConfigInfo::needLibraries | block::ConfigInfo::needValidatorSet | | ||||
|           block::ConfigInfo::needWorkchainInfo | block::ConfigInfo::needCapabilities | | ||||
|           block::ConfigInfo::needPrevBlocks | | ||||
|           (is_masterchain() ? block::ConfigInfo::needAccountsRoot | block::ConfigInfo::needSpecialSmc : 0)); | ||||
|   if (res.is_error()) { | ||||
|     td::Status err = res.move_as_error(); | ||||
|  | @ -1424,7 +1427,13 @@ bool Collator::import_new_shard_top_blocks() { | |||
|   } | ||||
|   LOG(INFO) << "total fees_imported = " << value_flow_.fees_imported.to_str() | ||||
|             << " ; out of them, total fees_created = " << import_created_.to_str(); | ||||
|   value_flow_.fees_collected += value_flow_.fees_imported; | ||||
|   block::CurrencyCollection burned = | ||||
|       config_->get_burning_config().calculate_burned_fees(value_flow_.fees_imported - import_created_); | ||||
|   if (!burned.is_valid()) { | ||||
|     return fatal_error("cannot calculate amount of burned imported fees"); | ||||
|   } | ||||
|   value_flow_.burned += burned; | ||||
|   value_flow_.fees_collected += value_flow_.fees_imported - burned; | ||||
|   return true; | ||||
| } | ||||
| 
 | ||||
|  | @ -1847,6 +1856,11 @@ bool Collator::out_msg_queue_cleanup() { | |||
|   auto res = out_msg_queue_->filter([&](vm::CellSlice& cs, td::ConstBitPtr key, int n) -> int { | ||||
|     assert(n == 352); | ||||
|     // LOG(DEBUG) << "key is " << key.to_hex(n);
 | ||||
|     if (queue_cleanup_timeout_.is_in_past(td::Timestamp::now())) { | ||||
|       LOG(WARNING) << "cleaning up outbound queue takes too long, ending"; | ||||
|       outq_cleanup_partial_ = true; | ||||
|       return (1 << 30) + 1;  // retain all remaining outbound queue entries including this one without processing
 | ||||
|     } | ||||
|     if (block_full_) { | ||||
|       LOG(WARNING) << "BLOCK FULL while cleaning up outbound queue, cleanup completed only partially"; | ||||
|       outq_cleanup_partial_ = true; | ||||
|  | @ -2234,6 +2248,7 @@ Ref<vm::Cell> Collator::create_ordinary_transaction(Ref<vm::Cell> msg_root) { | |||
| 
 | ||||
|   register_new_msgs(*trans); | ||||
|   update_max_lt(acc->last_trans_end_lt_); | ||||
|   value_flow_.burned += trans->blackhole_burned; | ||||
|   return trans_root; | ||||
| } | ||||
| 
 | ||||
|  | @ -2307,7 +2322,8 @@ td::Result<std::unique_ptr<block::transaction::Transaction>> Collator::impl_crea | |||
|     return td::Status::Error( | ||||
|         -669, "cannot create action phase of a new transaction for smart contract "s + acc->addr.to_hex()); | ||||
|   } | ||||
|   if (trans->bounce_enabled && (!trans->compute_phase->success || trans->action_phase->state_exceeds_limits) && | ||||
|   if (trans->bounce_enabled && | ||||
|       (!trans->compute_phase->success || trans->action_phase->state_exceeds_limits || trans->action_phase->bounce) && | ||||
|       !trans->prepare_bounce_phase(*action_phase_cfg)) { | ||||
|     return td::Status::Error( | ||||
|         -669, "cannot create bounce phase of a new transaction for smart contract "s + acc->addr.to_hex()); | ||||
|  | @ -2590,7 +2606,7 @@ bool Collator::precheck_inbound_message(Ref<vm::CellSlice> enq_msg, ton::Logical | |||
|     return false; | ||||
|   } | ||||
|   if (!block::tlb::t_MsgEnvelope.validate_ref(msg_env)) { | ||||
|     LOG(ERROR) << "inbound internal MsgEnvelope is invalid according to automated checks"; | ||||
|     LOG(ERROR) << "inbound internal MsgEnvelope is invalid according to hand-written checks"; | ||||
|     return false; | ||||
|   } | ||||
|   return true; | ||||
|  | @ -2622,6 +2638,10 @@ bool Collator::process_inbound_message(Ref<vm::CellSlice> enq_msg, ton::LogicalT | |||
|                   "its contents"; | ||||
|     return false; | ||||
|   } | ||||
|   if (!block::tlb::validate_message_libs(env.msg)) { | ||||
|     LOG(ERROR) << "inbound internal message has invalid StateInit"; | ||||
|     return false; | ||||
|   } | ||||
|   // 2.0. update last_proc_int_msg
 | ||||
|   if (!update_last_proc_int_msg(std::pair<ton::LogicalTime, ton::Bits256>(lt, env.msg->get_hash().bits()))) { | ||||
|     return fatal_error("processing a message AFTER a newer message has been processed"); | ||||
|  | @ -2997,6 +3017,7 @@ bool Collator::process_new_messages(bool enqueue_only) { | |||
|   while (!new_msgs.empty()) { | ||||
|     block::NewOutMsg msg = new_msgs.top(); | ||||
|     new_msgs.pop(); | ||||
|     block_limit_status_->extra_out_msgs--; | ||||
|     if (block_full_ && !enqueue_only) { | ||||
|       LOG(INFO) << "BLOCK FULL, enqueue all remaining new messages"; | ||||
|       enqueue_only = true; | ||||
|  | @ -3018,6 +3039,7 @@ void Collator::register_new_msg(block::NewOutMsg new_msg) { | |||
|     min_new_msg_lt = new_msg.lt; | ||||
|   } | ||||
|   new_msgs.push(std::move(new_msg)); | ||||
|   block_limit_status_->extra_out_msgs++; | ||||
| } | ||||
| 
 | ||||
| void Collator::register_new_msgs(block::transaction::Transaction& trans) { | ||||
|  | @ -3767,7 +3789,16 @@ bool Collator::compute_total_balance() { | |||
|     LOG(ERROR) << "cannot unpack CurrencyCollection from the root of OutMsgDescr"; | ||||
|     return false; | ||||
|   } | ||||
|   value_flow_.fees_collected += new_transaction_fees + new_import_fees; | ||||
|   block::CurrencyCollection total_fees = new_transaction_fees + new_import_fees; | ||||
|   value_flow_.fees_collected += total_fees; | ||||
|   if (is_masterchain()) { | ||||
|     block::CurrencyCollection burned = config_->get_burning_config().calculate_burned_fees(total_fees); | ||||
|     if (!burned.is_valid()) { | ||||
|       return fatal_error("cannot calculate amount of burned masterchain fees"); | ||||
|     } | ||||
|     value_flow_.fees_collected -= burned; | ||||
|     value_flow_.burned += burned; | ||||
|   } | ||||
|   // 3. compute total_validator_fees
 | ||||
|   total_validator_fees_ += value_flow_.fees_collected; | ||||
|   total_validator_fees_ -= value_flow_.recovered; | ||||
|  | @ -3916,7 +3947,7 @@ bool Collator::create_block() { | |||
|   } | ||||
|   if (verify >= 1) { | ||||
|     LOG(INFO) << "verifying new Block"; | ||||
|     if (!block::gen::t_Block.validate_ref(1000000, new_block)) { | ||||
|     if (!block::gen::t_Block.validate_ref(10000000, new_block)) { | ||||
|       return fatal_error("new Block failed to pass automatic validity tests"); | ||||
|     } | ||||
|   } | ||||
|  | @ -4077,6 +4108,18 @@ bool Collator::create_block_candidate() { | |||
|       ton::BlockIdExt{ton::BlockId{shard_, new_block_seqno}, new_block->get_hash().bits(), | ||||
|                       block::compute_file_hash(blk_slice.as_slice())}, | ||||
|       block::compute_file_hash(cdata_slice.as_slice()), blk_slice.clone(), cdata_slice.clone()); | ||||
|   // 3.1 check block and collated data size
 | ||||
|   auto consensus_config = config_->get_consensus_config(); | ||||
|   if (block_candidate->data.size() > consensus_config.max_block_size) { | ||||
|     return fatal_error(PSTRING() << "block size (" << block_candidate->data.size() | ||||
|                                  << ") exceeds the limit in consensus config (" << consensus_config.max_block_size | ||||
|                                  << ")"); | ||||
|   } | ||||
|   if (block_candidate->collated_data.size() > consensus_config.max_collated_data_size) { | ||||
|     return fatal_error(PSTRING() << "collated data size (" << block_candidate->collated_data.size() | ||||
|                                  << ") exceeds the limit in consensus config (" | ||||
|                                  << consensus_config.max_collated_data_size << ")"); | ||||
|   } | ||||
|   // 4. save block candidate
 | ||||
|   LOG(INFO) << "saving new BlockCandidate"; | ||||
|   td::actor::send_closure_later(manager, &ValidatorManager::set_block_candidate, block_candidate->id, | ||||
|  | @ -4139,6 +4182,9 @@ td::Result<bool> Collator::register_external_message_cell(Ref<vm::Cell> ext_msg, | |||
|   if (!block::tlb::t_Message.validate_ref(256, ext_msg)) { | ||||
|     return td::Status::Error("external message is not a (Message Any) according to hand-written checks"); | ||||
|   } | ||||
|   if (!block::tlb::validate_message_libs(ext_msg)) { | ||||
|     return td::Status::Error("external message has invalid libs in StateInit"); | ||||
|   } | ||||
|   block::gen::CommonMsgInfo::Record_ext_in_msg_info info; | ||||
|   if (!tlb::unpack_cell_inexact(ext_msg, info)) { | ||||
|     return td::Status::Error("cannot unpack external message header"); | ||||
|  |  | |||
|  | @ -1192,6 +1192,7 @@ void LiteQuery::finish_getAccountState(td::BufferSlice shard_proof) { | |||
|       return; | ||||
|     } | ||||
|     auto rconfig = config.move_as_ok(); | ||||
|     rconfig->set_block_id_ext(mc_state_->get_block_id()); | ||||
|     acc_state_promise_.set_value(std::make_tuple( | ||||
|                                   std::move(acc_csr), sstate.gen_utime, sstate.gen_lt, std::move(rconfig) | ||||
|                                  )); | ||||
|  | @ -1687,13 +1688,23 @@ void LiteQuery::continue_getConfigParams(int mode, std::vector<int> param_list) | |||
|     } | ||||
|   } | ||||
| 
 | ||||
|   auto res = keyblk ? block::Config::extract_from_key_block(mpb.root(), mode) | ||||
|                     : block::Config::extract_from_state(mpb.root(), mode); | ||||
|   if (res.is_error()) { | ||||
|     fatal_error(res.move_as_error()); | ||||
|     return; | ||||
|   std::unique_ptr<block::Config> cfg; | ||||
|   if (keyblk || !(mode & block::ConfigInfo::needPrevBlocks)) { | ||||
|     auto res = keyblk ? block::Config::extract_from_key_block(mpb.root(), mode) | ||||
|                       : block::Config::extract_from_state(mpb.root(), mode); | ||||
|     if (res.is_error()) { | ||||
|       fatal_error(res.move_as_error()); | ||||
|       return; | ||||
|     } | ||||
|     cfg = res.move_as_ok(); | ||||
|   } else { | ||||
|     auto res = block::ConfigInfo::extract_config(mpb.root(), mode); | ||||
|     if (res.is_error()) { | ||||
|       fatal_error(res.move_as_error()); | ||||
|       return; | ||||
|     } | ||||
|     cfg = res.move_as_ok(); | ||||
|   } | ||||
|   auto cfg = res.move_as_ok(); | ||||
|   if (!cfg) { | ||||
|     fatal_error("cannot extract configuration from last mc state"); | ||||
|     return; | ||||
|  | @ -1706,6 +1717,9 @@ void LiteQuery::continue_getConfigParams(int mode, std::vector<int> param_list) | |||
|         visit(cfg->get_config_param(i)); | ||||
|       } | ||||
|     } | ||||
|     if (!keyblk && mode & block::ConfigInfo::needPrevBlocks) { | ||||
|       ((block::ConfigInfo*)cfg.get())->get_prev_blocks_info(); | ||||
|     } | ||||
|   } catch (vm::VmError& err) { | ||||
|     fatal_error("error while traversing required configuration parameters: "s + err.get_msg()); | ||||
|     return; | ||||
|  |  | |||
|  | @ -62,12 +62,10 @@ td::Status ShardTopBlockDescrQ::unpack_one_proof(BlockIdExt& cur_id, Ref<vm::Cel | |||
|   block::gen::Block::Record blk; | ||||
|   block::gen::BlockInfo::Record info; | ||||
|   block::gen::BlockExtra::Record extra; | ||||
|   block::gen::ValueFlow::Record flow; | ||||
|   block::CurrencyCollection fees_collected, funds_created; | ||||
|   block::ValueFlow flow; | ||||
|   if (!(tlb::unpack_cell(virt_root, blk) && tlb::unpack_cell(blk.info, info) && !info.version && | ||||
|         block::gen::t_ValueFlow.force_validate_ref(blk.value_flow) && tlb::unpack_cell(blk.value_flow, flow) && | ||||
|         /*tlb::unpack_cell(blk.extra, extra) &&*/ fees_collected.unpack(flow.fees_collected) && | ||||
|         funds_created.unpack(flow.r2.created))) { | ||||
|         flow.unpack(vm::load_cell_slice_ref(blk.value_flow))) | ||||
|       /*&& tlb::unpack_cell(blk.extra, extra)*/) { | ||||
|     return td::Status::Error(-666, "cannot unpack block header in link for block "s + cur_id.to_str()); | ||||
|   } | ||||
|   // remove this "try ... catch ..." later and uncomment tlb::unpack_cell(blk.extra, extra) in the previous condition
 | ||||
|  | @ -131,7 +129,7 @@ td::Status ShardTopBlockDescrQ::unpack_one_proof(BlockIdExt& cur_id, Ref<vm::Cel | |||
|   } | ||||
|   chain_mc_blk_ids_.push_back(mc_blkid); | ||||
|   chain_blk_ids_.push_back(cur_id); | ||||
|   chain_fees_.emplace_back(std::move(fees_collected), std::move(funds_created)); | ||||
|   chain_fees_.emplace_back(std::move(flow.fees_collected), std::move(flow.created)); | ||||
|   creators_.push_back(extra.created_by); | ||||
|   if (!is_head) { | ||||
|     if (info.after_split || info.after_merge) { | ||||
|  |  | |||
|  | @ -682,7 +682,7 @@ bool ValidateQuery::try_unpack_mc_state() { | |||
|         mc_state_root_, | ||||
|         block::ConfigInfo::needShardHashes | block::ConfigInfo::needLibraries | block::ConfigInfo::needValidatorSet | | ||||
|             block::ConfigInfo::needWorkchainInfo | block::ConfigInfo::needStateExtraRoot | | ||||
|             block::ConfigInfo::needCapabilities | | ||||
|             block::ConfigInfo::needCapabilities | block::ConfigInfo::needPrevBlocks | | ||||
|             (is_masterchain() ? block::ConfigInfo::needAccountsRoot | block::ConfigInfo::needSpecialSmc : 0)); | ||||
|     if (res.is_error()) { | ||||
|       return fatal_error(-666, "cannot extract configuration from reference masterchain state "s + mc_blkid_.to_str() + | ||||
|  | @ -785,10 +785,20 @@ bool ValidateQuery::fetch_config_params() { | |||
|                                                   storage_phase_cfg_.delete_due_limit)) { | ||||
|       return fatal_error("cannot unpack current gas prices and limits from masterchain configuration"); | ||||
|     } | ||||
|     storage_phase_cfg_.enable_due_payment = config_->get_global_version() >= 4; | ||||
|     compute_phase_cfg_.block_rand_seed = rand_seed_; | ||||
|     compute_phase_cfg_.libraries = std::make_unique<vm::Dictionary>(config_->get_libraries_root(), 256); | ||||
|     compute_phase_cfg_.max_vm_data_depth = size_limits.max_vm_data_depth; | ||||
|     compute_phase_cfg_.global_config = config_->get_root_cell(); | ||||
|     compute_phase_cfg_.global_version = config_->get_global_version(); | ||||
|     if (compute_phase_cfg_.global_version >= 4) { | ||||
|       auto prev_blocks_info = config_->get_prev_blocks_info(); | ||||
|       if (prev_blocks_info.is_error()) { | ||||
|         return fatal_error(prev_blocks_info.move_as_error_prefix( | ||||
|             "cannot fetch prev blocks info from masterchain configuration: ")); | ||||
|       } | ||||
|       compute_phase_cfg_.prev_blocks_info = prev_blocks_info.move_as_ok(); | ||||
|     } | ||||
|     compute_phase_cfg_.suspended_addresses = config_->get_suspended_addresses(now_); | ||||
|   } | ||||
|   { | ||||
|  | @ -811,6 +821,9 @@ bool ValidateQuery::fetch_config_params() { | |||
|     action_phase_cfg_.workchains = &config_->get_workchain_list(); | ||||
|     action_phase_cfg_.bounce_msg_body = (config_->has_capability(ton::capBounceMsgBody) ? 256 : 0); | ||||
|     action_phase_cfg_.size_limits = size_limits; | ||||
|     action_phase_cfg_.action_fine_enabled = config_->get_global_version() >= 4; | ||||
|     action_phase_cfg_.bounce_on_fail_enabled = config_->get_global_version() >= 4; | ||||
|     action_phase_cfg_.mc_blackhole_addr = config_->get_burning_config().blackhole_addr; | ||||
|   } | ||||
|   { | ||||
|     // fetch block_grams_created
 | ||||
|  | @ -2193,13 +2206,13 @@ bool ValidateQuery::unpack_block_data() { | |||
|   auto outmsg_cs = vm::load_cell_slice_ref(std::move(extra.out_msg_descr)); | ||||
|   // run some hand-written checks from block::tlb::
 | ||||
|   // (automatic tests from block::gen:: have been already run for the entire block)
 | ||||
|   if (!block::tlb::t_InMsgDescr.validate_upto(1000000, *inmsg_cs)) { | ||||
|   if (!block::tlb::t_InMsgDescr.validate_upto(10000000, *inmsg_cs)) { | ||||
|     return reject_query("InMsgDescr of the new block failed to pass handwritten validity tests"); | ||||
|   } | ||||
|   if (!block::tlb::t_OutMsgDescr.validate_upto(1000000, *outmsg_cs)) { | ||||
|   if (!block::tlb::t_OutMsgDescr.validate_upto(10000000, *outmsg_cs)) { | ||||
|     return reject_query("OutMsgDescr of the new block failed to pass handwritten validity tests"); | ||||
|   } | ||||
|   if (!block::tlb::t_ShardAccountBlocks.validate_ref(1000000, extra.account_blocks)) { | ||||
|   if (!block::tlb::t_ShardAccountBlocks.validate_ref(10000000, extra.account_blocks)) { | ||||
|     return reject_query("ShardAccountBlocks of the new block failed to pass handwritten validity tests"); | ||||
|   } | ||||
|   in_msg_dict_ = std::make_unique<vm::AugmentedDictionary>(std::move(inmsg_cs), 256, block::tlb::aug_InMsgDescr); | ||||
|  | @ -2243,6 +2256,11 @@ bool ValidateQuery::unpack_precheck_value_flow(Ref<vm::Cell> value_flow_root) { | |||
|     return reject_query("ValueFlow of block "s + id_.to_str() + | ||||
|                         " is invalid (non-zero recovered value in a non-masterchain block)"); | ||||
|   } | ||||
|   if (!is_masterchain() && !value_flow_.burned.is_zero()) { | ||||
|     LOG(INFO) << "invalid value flow: " << os.str(); | ||||
|     return reject_query("ValueFlow of block "s + id_.to_str() + | ||||
|                         " is invalid (non-zero burned value in a non-masterchain block)"); | ||||
|   } | ||||
|   if (!value_flow_.recovered.is_zero() && recover_create_msg_.is_null()) { | ||||
|     return reject_query("ValueFlow of block "s + id_.to_str() + | ||||
|                         " has a non-zero recovered fees value, but there is no recovery InMsg"); | ||||
|  | @ -2325,15 +2343,10 @@ bool ValidateQuery::unpack_precheck_value_flow(Ref<vm::Cell> value_flow_root) { | |||
|         "cannot unpack CurrencyCollection with total transaction fees from the augmentation of the ShardAccountBlocks " | ||||
|         "dictionary"); | ||||
|   } | ||||
|   auto expected_fees = value_flow_.fees_imported + value_flow_.created + transaction_fees_ + import_fees_; | ||||
|   if (value_flow_.fees_collected != expected_fees) { | ||||
|     return reject_query(PSTRING() << "ValueFlow for " << id_.to_str() << " declares fees_collected=" | ||||
|                                   << value_flow_.fees_collected.to_str() << " but the total message import fees are " | ||||
|                                   << import_fees_ << ", the total transaction fees are " << transaction_fees_.to_str() | ||||
|                                   << ", creation fee for this block is " << value_flow_.created.to_str() | ||||
|                                   << " and the total imported fees from shards are " | ||||
|                                   << value_flow_.fees_imported.to_str() << " with a total of " | ||||
|                                   << expected_fees.to_str()); | ||||
|   if (is_masterchain()) { | ||||
|     auto x = config_->get_burning_config().calculate_burned_fees(transaction_fees_ + import_fees_); | ||||
|     fees_burned_ += x; | ||||
|     total_burned_ += x; | ||||
|   } | ||||
|   return true; | ||||
| } | ||||
|  | @ -4588,7 +4601,8 @@ bool ValidateQuery::check_one_transaction(block::Account& account, ton::LogicalT | |||
|     return reject_query(PSTRING() << "cannot re-create action phase of transaction " << lt << " for smart contract " | ||||
|                                   << addr.to_hex()); | ||||
|   } | ||||
|   if (trs->bounce_enabled && (!trs->compute_phase->success || trs->action_phase->state_exceeds_limits) && | ||||
|   if (trs->bounce_enabled && | ||||
|       (!trs->compute_phase->success || trs->action_phase->state_exceeds_limits || trs->action_phase->bounce) && | ||||
|       !trs->prepare_bounce_phase(action_phase_cfg_)) { | ||||
|     return reject_query(PSTRING() << "cannot re-create bounce phase of  transaction " << lt << " for smart contract " | ||||
|                                   << addr.to_hex()); | ||||
|  | @ -4646,6 +4660,7 @@ bool ValidateQuery::check_one_transaction(block::Account& account, ton::LogicalT | |||
|         << "transaction " << lt << " of " << addr.to_hex() | ||||
|         << " is invalid: it has produced a set of outbound messages different from that listed in the transaction"); | ||||
|   } | ||||
|   total_burned_ += trs->blackhole_burned; | ||||
|   // check new balance and value flow
 | ||||
|   auto new_balance = account.get_balance(); | ||||
|   block::CurrencyCollection total_fees; | ||||
|  | @ -4653,12 +4668,14 @@ bool ValidateQuery::check_one_transaction(block::Account& account, ton::LogicalT | |||
|     return reject_query(PSTRING() << "transaction " << lt << " of " << addr.to_hex() | ||||
|                                   << " has an invalid total_fees value"); | ||||
|   } | ||||
|   if (old_balance + money_imported != new_balance + money_exported + total_fees) { | ||||
|     return reject_query(PSTRING() << "transaction " << lt << " of " << addr.to_hex() | ||||
|                                   << " violates the currency flow condition: old balance=" << old_balance.to_str() | ||||
|                                   << " + imported=" << money_imported.to_str() << " does not equal new balance=" | ||||
|                                   << new_balance.to_str() << " + exported=" << money_exported.to_str() | ||||
|                                   << " + total_fees=" << total_fees.to_str()); | ||||
|   if (old_balance + money_imported != new_balance + money_exported + total_fees + trs->blackhole_burned) { | ||||
|     return reject_query( | ||||
|         PSTRING() << "transaction " << lt << " of " << addr.to_hex() | ||||
|                   << " violates the currency flow condition: old balance=" << old_balance.to_str() | ||||
|                   << " + imported=" << money_imported.to_str() << " does not equal new balance=" << new_balance.to_str() | ||||
|                   << " + exported=" << money_exported.to_str() << " + total_fees=" << total_fees.to_str() | ||||
|                   << (trs->blackhole_burned.is_zero() ? "" | ||||
|                                                       : PSTRING() << " burned=" << trs->blackhole_burned.to_str())); | ||||
|   } | ||||
|   return true; | ||||
| } | ||||
|  | @ -5529,6 +5546,9 @@ bool ValidateQuery::check_mc_block_extra() { | |||
|     return reject_query("invalid fees_imported in value flow: declared "s + value_flow_.fees_imported.to_str() + | ||||
|                         ", correct value is " + fees_imported.to_str()); | ||||
|   } | ||||
|   auto x = config_->get_burning_config().calculate_burned_fees(fees_imported - import_created_); | ||||
|   total_burned_ += x; | ||||
|   fees_burned_ += x; | ||||
|   // ^[ prev_blk_signatures:(HashmapE 16 CryptoSignaturePair)
 | ||||
|   if (prev_signatures_.not_null() && id_.seqno() == 1) { | ||||
|     return reject_query("block contains non-empty signature set for the zero state of the masterchain"); | ||||
|  | @ -5546,6 +5566,26 @@ bool ValidateQuery::check_mc_block_extra() { | |||
|   return true; | ||||
| } | ||||
| 
 | ||||
| bool ValidateQuery::postcheck_value_flow() { | ||||
|   auto expected_fees = | ||||
|       value_flow_.fees_imported + value_flow_.created + transaction_fees_ + import_fees_ - fees_burned_; | ||||
|   if (value_flow_.fees_collected != expected_fees) { | ||||
|     return reject_query(PSTRING() << "ValueFlow for " << id_.to_str() << " declares fees_collected=" | ||||
|                                   << value_flow_.fees_collected.to_str() << " but the total message import fees are " | ||||
|                                   << import_fees_ << ", the total transaction fees are " << transaction_fees_.to_str() | ||||
|                                   << ", creation fee for this block is " << value_flow_.created.to_str() | ||||
|                                   << ", the total imported fees from shards are " << value_flow_.fees_imported.to_str() | ||||
|                                   << " and the burned fees are " << fees_burned_.to_str() | ||||
|                                   << " with a total of " << expected_fees.to_str()); | ||||
|   } | ||||
|   if (total_burned_ != value_flow_.burned) { | ||||
|     return reject_query(PSTRING() << "invalid burned in value flow: " << id_.to_str() << " declared " | ||||
|                                   << value_flow_.burned.to_str() << ", correct value is " | ||||
|                                   << total_burned_.to_str()); | ||||
|   } | ||||
|   return true; | ||||
| } | ||||
| 
 | ||||
| Ref<vm::Cell> ValidateQuery::get_virt_state_root(td::Bits256 block_root_hash) { | ||||
|   auto it = virt_roots_.find(block_root_hash); | ||||
|   if (it == virt_roots_.end()) { | ||||
|  | @ -5607,7 +5647,7 @@ bool ValidateQuery::try_validate() { | |||
|       } | ||||
|     } | ||||
|     LOG(INFO) << "running automated validity checks for block candidate " << id_.to_str(); | ||||
|     if (!block::gen::t_Block.validate_ref(1000000, block_root_)) { | ||||
|     if (!block::gen::t_Block.validate_ref(10000000, block_root_)) { | ||||
|       return reject_query("block "s + id_.to_str() + " failed to pass automated validity checks"); | ||||
|     } | ||||
|     if (!fix_all_processed_upto()) { | ||||
|  | @ -5640,9 +5680,10 @@ bool ValidateQuery::try_validate() { | |||
|     if (!check_in_queue()) { | ||||
|       return reject_query("cannot check inbound message queues"); | ||||
|     } | ||||
|     if (!check_delivered_dequeued()) { | ||||
|     // Excessive check: validity of message in queue is checked elsewhere
 | ||||
|     /*if (!check_delivered_dequeued()) {
 | ||||
|       return reject_query("cannot check delivery status of all outbound messages"); | ||||
|     } | ||||
|     }*/ | ||||
|     if (!check_transactions()) { | ||||
|       return reject_query("invalid collection of account transactions in ShardAccountBlocks"); | ||||
|     } | ||||
|  | @ -5664,6 +5705,9 @@ bool ValidateQuery::try_validate() { | |||
|     if (!check_mc_state_extra()) { | ||||
|       return reject_query("new McStateExtra is invalid"); | ||||
|     } | ||||
|     if (!postcheck_value_flow()) { | ||||
|       return reject_query("new ValueFlow is invalid"); | ||||
|     } | ||||
|   } catch (vm::VmError& err) { | ||||
|     return fatal_error(-666, err.get_msg()); | ||||
|   } catch (vm::VmVirtError& err) { | ||||
|  |  | |||
|  | @ -28,6 +28,7 @@ | |||
| #include <vector> | ||||
| #include <string> | ||||
| #include <map> | ||||
| #include "common/global-version.h" | ||||
| 
 | ||||
| namespace ton { | ||||
| 
 | ||||
|  | @ -108,7 +109,7 @@ inline ErrorCtxSet ErrorCtx::set_guard(std::vector<std::string> str_list) { | |||
| 
 | ||||
| class ValidateQuery : public td::actor::Actor { | ||||
|   static constexpr int supported_version() { | ||||
|     return 3; | ||||
|     return SUPPORTED_VERSION; | ||||
|   } | ||||
|   static constexpr long long supported_capabilities() { | ||||
|     return ton::capCreateStatsEnabled | ton::capBounceMsgBody | ton::capReportVersion | ton::capShortDequeue; | ||||
|  | @ -217,7 +218,7 @@ class ValidateQuery : public td::actor::Actor { | |||
| 
 | ||||
|   std::unique_ptr<vm::AugmentedDictionary> in_msg_dict_, out_msg_dict_, account_blocks_dict_; | ||||
|   block::ValueFlow value_flow_; | ||||
|   block::CurrencyCollection import_created_, transaction_fees_; | ||||
|   block::CurrencyCollection import_created_, transaction_fees_, total_burned_{0}, fees_burned_{0}; | ||||
|   td::RefInt256 import_fees_; | ||||
| 
 | ||||
|   ton::LogicalTime proc_lt_{0}, claimed_proc_lt_{0}, min_enq_lt_{~0ULL}; | ||||
|  | @ -362,6 +363,7 @@ class ValidateQuery : public td::actor::Actor { | |||
|   bool check_one_prev_dict_update(ton::BlockSeqno seqno, Ref<vm::CellSlice> old_val_extra, | ||||
|                                   Ref<vm::CellSlice> new_val_extra); | ||||
|   bool check_mc_state_extra(); | ||||
|   bool postcheck_value_flow(); | ||||
|   td::Status check_counter_update(const block::DiscountedCounter& oc, const block::DiscountedCounter& nc, | ||||
|                                   unsigned expected_incr); | ||||
|   bool check_one_block_creator_update(td::ConstBitPtr key, Ref<vm::CellSlice> old_val, Ref<vm::CellSlice> new_val); | ||||
|  |  | |||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue