aboutsummaryrefslogblamecommitdiff
path: root/src/blockchain_utilities/blockchain_import.cpp
blob: fc00cdcbab1c1263b6d2b42b45ada182afcc09d5 (plain) (tree)

































                                                                                          

                                    
                                                    
                                                                        

                                                    
 
                                          
 

                      

                                             

         
         



                                                                                               


                                                                      












                                                                                          
 
 





                                      


























                                                                                             
                              
                                
                                  
                              
                                
                              
                                

                                 









                                                                













































                                                                                                                             
                            
                                                                                                 










                                                                                                                 
                                                 





                                                          
                                                                


                 







                                                                                                                                               




                                                                                
                            




                                            





                                                           









                                                                         
                                                                                

                                                        
                   
   
                                         
   

                                                                               

                                                                         
                  












                                                                                                                             
                                                            







                                                                               


                                                  
                        
                                  
                                          


               






                                                                         
                                               
 






                                                                                             
                                                                     
     

                                             



                                               
                                                                             










                                                                      
                       
     
                                                    
                              


                                                                                                            





               



                                                                      


                                  

                                                                                                           
       

                                          
         
                                      
                                                       


            
                                                       
         
                     

                                                             
                                           
         
                                                        
                                  



                                     
                                              
 
















                                                                                
         





                                                                                                              
              



                                                                   










                                                                                       




















                                                                                                                      
                              










                                                                                                   
                                                  








                                                                              
                                                  










                                                           


                                                           
 



                                                                              






                                                                                              
                                        




                                                                               
                       


                      
                                         
           
                                        

                                                                                      



                                                         
                                                        






                                   
                                  
                                                                          

















                                                          
                                                
      
                                                               
              

                                                                                               














                                                          
                          
                          
                      
                         





                                                                                     



                                                                                                                                    






                                                                   
                                                 














                                                                                                               
                                                       
                                                      
                                                     

                                                       

                                                         
                                                       


























                                                                           
                                                            


















                                                                                











                                                                             





                                                                                                    
                                                       









                                                                       

                                                  

                                             


            










                                                              



















                                                                                     

                                                              














                                                                               
                                                                           
                                                                



                                                       
                                                                














                                                                                            
                                                                         



                                                     








                                                                                      
                                                              



















                                                                                                           
// Copyright (c) 2014-2015, The Monero Project
//
// All rights reserved.
//
// Redistribution and use in source and binary forms, with or without modification, are
// permitted provided that the following conditions are met:
//
// 1. Redistributions of source code must retain the above copyright notice, this list of
//    conditions and the following disclaimer.
//
// 2. Redistributions in binary form must reproduce the above copyright notice, this list
//    of conditions and the following disclaimer in the documentation and/or other
//    materials provided with the distribution.
//
// 3. Neither the name of the copyright holder nor the names of its contributors may be
//    used to endorse or promote products derived from this software without specific
//    prior written permission.
//
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
// MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
// THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
// INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
// STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF
// THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

#include <atomic>
#include <cstdio>
#include <algorithm>
#include <fstream>

#include <boost/filesystem.hpp>
#include "bootstrap_file.h"
#include "bootstrap_serialization.h"
#include "cryptonote_core/cryptonote_format_utils.h"
#include "serialization/binary_utils.h" // dump_binary(), parse_binary()
#include "serialization/json_utils.h" // dump_json()
#include "include_base_utils.h"

#include <lmdb.h> // for db flag arguments

#include "fake_core.h"

unsigned int epee::g_test_dbg_lock_sleep = 0;

namespace
{
// CONFIG
bool opt_batch   = true;
bool opt_verify  = true; // use add_new_block, which does verification before calling add_block
bool opt_resume  = true;
bool opt_testnet = true;

// number of blocks per batch transaction
// adjustable through command-line argument according to available RAM
#if !defined(WIN32)
uint64_t db_batch_size = 20000;
#else
// set a lower default batch size, pending possible LMDB issue with large transaction size
uint64_t db_batch_size = 1000;
#endif

// when verifying, use a smaller default batch size so progress is more
// frequently saved
uint64_t db_batch_size_verify = 5000;

std::string refresh_string = "\r                                    \r";
}



namespace po = boost::program_options;

using namespace cryptonote;
using namespace epee;


int parse_db_arguments(const std::string& db_arg_str, std::string& db_engine, int& mdb_flags)
{
  std::vector<std::string> db_args;
  boost::split(db_args, db_arg_str, boost::is_any_of("#"));
  db_engine = db_args.front();
  boost::algorithm::trim(db_engine);

  if (db_args.size() == 1)
  {
    return 0;
  }
  else if (db_args.size() > 2)
  {
    std::cerr << "unrecognized database argument format: " << db_arg_str << ENDL;
    return 1;
  }

  std::string db_arg_str2 = db_args[1];
  boost::split(db_args, db_arg_str2, boost::is_any_of(","));
  for (auto& it : db_args)
  {
    boost::algorithm::trim(it);
    if (it.empty())
      continue;
    LOG_PRINT_L1("LMDB flag: " << it);
    if (it == "nosync")
      mdb_flags |= MDB_NOSYNC;
    else if (it == "nometasync")
      mdb_flags |= MDB_NOMETASYNC;
    else if (it == "writemap")
      mdb_flags |= MDB_WRITEMAP;
    else if (it == "mapasync")
      mdb_flags |= MDB_MAPASYNC;
    else if (it == "nordahead")
      mdb_flags |= MDB_NORDAHEAD;
    else
    {
      std::cerr << "unrecognized database flag: " << it << ENDL;
      return 1;
    }
  }
  return 0;
}


template <typename FakeCore>
int pop_blocks(FakeCore& simple_core, int num_blocks)
{
  bool use_batch = false;
  if (opt_batch)
  {
    if (simple_core.support_batch)
      use_batch = true;
    else
      LOG_PRINT_L0("WARNING: batch transactions enabled but unsupported or unnecessary for this database engine - ignoring");
  }

  if (use_batch)
    simple_core.batch_start();

  int quit = 0;
  block popped_block;
  std::vector<transaction> popped_txs;
  for (int i=0; i < num_blocks; ++i)
  {
    // simple_core.m_storage.pop_block_from_blockchain() is private, so call directly through db
    simple_core.m_storage.get_db().pop_block(popped_block, popped_txs);
    quit = 1;
  }



  if (use_batch)
  {
    if (quit > 1)
    {
      // There was an error, so don't commit pending data.
      // Destructor will abort write txn.
    }
    else
    {
      simple_core.batch_stop();
    }
#if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB)
    simple_core.m_storage.get_db().show_stats();
#endif
  }

  return num_blocks;
}

template <typename FakeCore>
int import_from_file(FakeCore& simple_core, std::string& import_file_path, uint64_t block_stop=0)
{
#if !defined(BLOCKCHAIN_DB)
  static_assert(std::is_same<fake_core_memory, FakeCore>::value || std::is_same<fake_core_lmdb, FakeCore>::value,
      "FakeCore constraint error");
#endif
#if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB)
  if (std::is_same<fake_core_lmdb, FakeCore>::value)
  {
    // Reset stats, in case we're using newly created db, accumulating stats
    // from addition of genesis block.
    // This aligns internal db counts with importer counts.
    simple_core.m_storage.get_db().reset_stats();
  }
#endif
  boost::filesystem::path raw_file_path(import_file_path);
  boost::system::error_code ec;
  if (!boost::filesystem::exists(raw_file_path, ec))
  {
    LOG_PRINT_L0("bootstrap file not found: " << raw_file_path);
    return false;
  }

  BootstrapFile bootstrap;
  // BootstrapFile bootstrap(import_file_path);
  uint64_t total_source_blocks = bootstrap.count_blocks(import_file_path);
  LOG_PRINT_L0("bootstrap file last block number: " << total_source_blocks-1 << " (zero-based height)  total blocks: " << total_source_blocks);

  std::cout << ENDL;
  std::cout << "Preparing to read blocks..." << ENDL;
  std::cout << ENDL;

  std::ifstream import_file;
  import_file.open(import_file_path, std::ios_base::binary | std::ifstream::in);

  uint64_t h = 0;
  uint64_t num_imported = 0;
  if (import_file.fail())
  {
    LOG_PRINT_L0("import_file.open() fail");
    return false;
  }

  // 4 byte magic + (currently) 1024 byte header structures
  bootstrap.seek_to_first_chunk(import_file);

  std::string str1;
  char buffer1[1024];
  char buffer_block[BUFFER_SIZE];
  block b;
  transaction tx;
  int quit = 0;
  uint64_t bytes_read = 0;

  uint64_t start_height = 1;
  if (opt_resume)
    start_height = simple_core.m_storage.get_current_blockchain_height();

  // Note that a new blockchain will start with block number 0 (total blocks: 1)
  // due to genesis block being added at initialization.

  if (! block_stop)
  {
    block_stop = total_source_blocks - 1;
  }

  // These are what we'll try to use, and they don't have to be a determination
  // from source and destination blockchains, but those are the defaults.
  LOG_PRINT_L0("start block: " << start_height << "  stop block: " <<
      block_stop);

  bool use_batch = false;
  if (opt_batch)
  {
    if (simple_core.support_batch)
      use_batch = true;
    else
      LOG_PRINT_L0("WARNING: batch transactions enabled but unsupported or unnecessary for this database engine - ignoring");
  }

  if (use_batch)
    simple_core.batch_start();

  LOG_PRINT_L0("Reading blockchain from bootstrap file...");
  std::cout << ENDL;

  // Within the loop, we skip to start_height before we start adding.
  // TODO: Not a bottleneck, but we can use what's done in count_blocks() and
  // only do the chunk size reads, skipping the chunk content reads until we're
  // at start_height.
  while (! quit)
  {
    uint32_t chunk_size;
    import_file.read(buffer1, sizeof(chunk_size));
    // TODO: bootstrap.read_chunk();
    if (! import_file) {
      std::cout << refresh_string;
      LOG_PRINT_L0("End of file reached");
      quit = 1;
      break;
    }
    bytes_read += sizeof(chunk_size);

    str1.assign(buffer1, sizeof(chunk_size));
    if (! ::serialization::parse_binary(str1, chunk_size))
    {
      throw std::runtime_error("Error in deserialization of chunk size");
    }
    LOG_PRINT_L3("chunk_size: " << chunk_size);

    if (chunk_size > BUFFER_SIZE)
    {
      LOG_PRINT_L0("WARNING: chunk_size " << chunk_size << " > BUFFER_SIZE " << BUFFER_SIZE);
      throw std::runtime_error("Aborting: chunk size exceeds buffer size");
    }
    if (chunk_size > 100000)
    {
      LOG_PRINT_L0("NOTE: chunk_size " << chunk_size << " > 100000");
    }
    else if (chunk_size == 0) {
      LOG_PRINT_L0("ERROR: chunk_size == 0");
      return 2;
    }
    import_file.read(buffer_block, chunk_size);
    if (! import_file) {
      LOG_PRINT_L0("ERROR: unexpected end of file: bytes read before error: "
          << import_file.gcount() << " of chunk_size " << chunk_size);
      return 2;
    }
    bytes_read += chunk_size;
    LOG_PRINT_L3("Total bytes read: " << bytes_read);

    if (h + NUM_BLOCKS_PER_CHUNK < start_height + 1)
    {
      h += NUM_BLOCKS_PER_CHUNK;
      continue;
    }
    if (h > block_stop)
    {
      std::cout << refresh_string << "block " << h-1
        << " / " << block_stop
        << std::flush;
      std::cout << ENDL << ENDL;
      LOG_PRINT_L0("Specified block number reached - stopping.  block: " << h-1 << "  total blocks: " << h);
      quit = 1;
      break;
    }

    try
    {
      str1.assign(buffer_block, chunk_size);
      bootstrap::block_package bp;
      if (! ::serialization::parse_binary(str1, bp))
        throw std::runtime_error("Error in deserialization of chunk");

      int display_interval = 1000;
      int progress_interval = 10;
      // NOTE: use of NUM_BLOCKS_PER_CHUNK is a placeholder in case multi-block chunks are later supported.
      for (int chunk_ind = 0; chunk_ind < NUM_BLOCKS_PER_CHUNK; ++chunk_ind)
      {
        ++h;
        if ((h-1) % display_interval == 0)
        {
          std::cout << refresh_string;
          LOG_PRINT_L0("loading block number " << h-1);
        }
        else
        {
          LOG_PRINT_L3("loading block number " << h-1);
        }
        b = bp.block;
        LOG_PRINT_L2("block prev_id: " << b.prev_id << ENDL);

        if ((h-1) % progress_interval == 0)
        {
          std::cout << refresh_string << "block " << h-1
            << " / " << block_stop
            << std::flush;
        }

        std::vector<transaction> txs;
        std::vector<transaction> archived_txs;

        archived_txs = bp.txs;

        // std::cout << refresh_string;
        // LOG_PRINT_L1("txs: " << archived_txs.size());

        // if archived_txs is invalid
        // {
        //   std::cout << refresh_string;
        //   LOG_PRINT_RED_L0("exception while de-archiving txs, height=" << h);
        //   quit = 1;
        //   break;
        // }

        // tx number 1: coinbase tx
        // tx number 2 onwards: archived_txs
        unsigned int tx_num = 1;
        for (const transaction& tx : archived_txs)
        {
          ++tx_num;
          // if tx is invalid
          // {
          //   LOG_PRINT_RED_L0("exception while indexing tx from txs, height=" << h <<", tx_num=" << tx_num);
          //   quit = 1;
          //   break;
          // }

          // std::cout << refresh_string;
          // LOG_PRINT_L1("tx hash: " << get_transaction_hash(tx));

          // crypto::hash hsh = null_hash;
          // size_t blob_size = 0;
          // NOTE: all tx hashes except for coinbase tx are available in the block data
          // get_transaction_hash(tx, hsh, blob_size);
          // LOG_PRINT_L0("tx " << tx_num << "  " << hsh << " : " << ENDL);
          // LOG_PRINT_L0(obj_to_json_str(tx) << ENDL);

          // add blocks with verification.
          // for Blockchain and blockchain_storage add_new_block().
          if (opt_verify)
          {
            // crypto::hash hsh = null_hash;
            // size_t blob_size = 0;
            // get_transaction_hash(tx, hsh, blob_size);
            tx_verification_context tvc = AUTO_VAL_INIT(tvc);
            bool r = true;
            r = simple_core.m_pool.add_tx(tx, tvc, true);
            if (!r)
            {
              LOG_PRINT_RED_L0("failed to add transaction to transaction pool, height=" << h <<", tx_num=" << tx_num);
              quit = 1;
              break;
            }
          }
          else
          {
            // for add_block() method, without (much) processing.
            // don't add coinbase transaction to txs.
            //
            // because add_block() calls
            // add_transaction(blk_hash, blk.miner_tx) first, and
            // then a for loop for the transactions in txs.
            txs.push_back(tx);
          }
        }

        if (opt_verify)
        {
          block_verification_context bvc = boost::value_initialized<block_verification_context>();
          simple_core.m_storage.add_new_block(b, bvc);

          if (bvc.m_verifivation_failed)
          {
            LOG_PRINT_L0("Failed to add block to blockchain, verification failed, height = " << h);
            LOG_PRINT_L0("skipping rest of file");
            // ok to commit previously batched data because it failed only in
            // verification of potential new block with nothing added to batch
            // yet
            quit = 1;
            break;
          }
          if (! bvc.m_added_to_main_chain)
          {
            LOG_PRINT_L0("Failed to add block to blockchain, height = " << h);
            LOG_PRINT_L0("skipping rest of file");
            // make sure we don't commit partial block data
            quit = 2;
            break;
          }
        }
        else
        {
          size_t block_size;
          difficulty_type cumulative_difficulty;
          uint64_t coins_generated;

          block_size = bp.block_size;
          cumulative_difficulty = bp.cumulative_difficulty;
          coins_generated = bp.coins_generated;

          // std::cout << refresh_string;
          // LOG_PRINT_L2("block_size: " << block_size);
          // LOG_PRINT_L2("cumulative_difficulty: " << cumulative_difficulty);
          // LOG_PRINT_L2("coins_generated: " << coins_generated);

          try
          {
            simple_core.add_block(b, block_size, cumulative_difficulty, coins_generated, txs);
          }
          catch (const std::exception& e)
          {
            std::cout << refresh_string;
            LOG_PRINT_RED_L0("Error adding block to blockchain: " << e.what());
            quit = 2; // make sure we don't commit partial block data
            break;
          }
        }
        ++num_imported;

        if (use_batch)
        {
          if ((h-1) % db_batch_size == 0)
          {
            std::cout << refresh_string;
            // zero-based height
            std::cout << ENDL << "[- batch commit at height " << h-1 << " -]" << ENDL;
            simple_core.batch_stop();
            simple_core.batch_start();
            std::cout << ENDL;
#if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB)
            simple_core.m_storage.get_db().show_stats();
#endif
          }
        }
      }
    }
    catch (const std::exception& e)
    {
      std::cout << refresh_string;
      LOG_PRINT_RED_L0("exception while reading from file, height=" << h);
      return 2;
    }
  } // while

  import_file.close();

  if (use_batch)
  {
    if (quit > 1)
    {
      // There was an error, so don't commit pending data.
      // Destructor will abort write txn.
    }
    else
    {
      simple_core.batch_stop();
    }
#if !defined(BLOCKCHAIN_DB) || (BLOCKCHAIN_DB == DB_LMDB)
    simple_core.m_storage.get_db().show_stats();
#endif
    LOG_PRINT_L0("Number of blocks imported: " << num_imported)
    if (h > 0)
      // TODO: if there was an error, the last added block is probably at zero-based height h-2
      LOG_PRINT_L0("Finished at block: " << h-1 << "  total blocks: " << h);
  }
  std::cout << ENDL;
  return 0;
}

int main(int argc, char* argv[])
{
  std::string import_filename = BLOCKCHAIN_RAW;
#if defined(BLOCKCHAIN_DB) && (BLOCKCHAIN_DB == DB_MEMORY)
  std::string default_db_engine = "memory";
#else
  std::string default_db_engine = "lmdb";
#endif

  uint32_t log_level = LOG_LEVEL_0;
  uint64_t num_blocks = 0;
  uint64_t block_stop = 0;
  std::string dirname;
  std::string db_arg_str;

  boost::filesystem::path default_data_path {tools::get_default_data_dir()};
  boost::filesystem::path default_testnet_data_path {default_data_path / "testnet"};

  po::options_description desc_cmd_only("Command line options");
  po::options_description desc_cmd_sett("Command line options and settings options");
  const command_line::arg_descriptor<uint32_t> arg_log_level   = {"log-level",  "", log_level};
  const command_line::arg_descriptor<uint64_t> arg_block_stop  = {"block-stop", "Stop at block number", block_stop};
  const command_line::arg_descriptor<uint64_t> arg_batch_size  = {"batch-size", "", db_batch_size};
  const command_line::arg_descriptor<uint64_t> arg_pop_blocks  = {"pop-blocks", "Remove blocks from end of blockchain", num_blocks};
  const command_line::arg_descriptor<bool>     arg_testnet_on  = {
    "testnet"
      , "Run on testnet."
      , false
  };
  const command_line::arg_descriptor<bool>     arg_count_blocks = {
    "count-blocks"
      , "Count blocks in bootstrap file and exit"
      , false
  };
  const command_line::arg_descriptor<std::string> arg_database = {
    "database", "available: memory, lmdb"
      , default_db_engine
  };
  const command_line::arg_descriptor<bool> arg_verify =  {"verify",
    "Verify blocks and transactions during import", true};
  const command_line::arg_descriptor<bool> arg_batch  =  {"batch",
    "Batch transactions for faster import", true};
  const command_line::arg_descriptor<bool> arg_resume =  {"resume",
    "Resume from current height if output database already exists", true};

  command_line::add_arg(desc_cmd_sett, command_line::arg_data_dir, default_data_path.string());
  command_line::add_arg(desc_cmd_sett, command_line::arg_testnet_data_dir, default_testnet_data_path.string());
  command_line::add_arg(desc_cmd_sett, arg_testnet_on);
  command_line::add_arg(desc_cmd_sett, arg_log_level);
  command_line::add_arg(desc_cmd_sett, arg_database);
  command_line::add_arg(desc_cmd_sett, arg_batch_size);
  command_line::add_arg(desc_cmd_sett, arg_block_stop);

  command_line::add_arg(desc_cmd_only, arg_count_blocks);
  command_line::add_arg(desc_cmd_only, arg_pop_blocks);
  command_line::add_arg(desc_cmd_only, command_line::arg_help);

  // call add_options() directly for these arguments since
  // command_line helpers support only boolean switch, not boolean argument
  desc_cmd_sett.add_options()
    (arg_verify.name, make_semantic(arg_verify), arg_verify.description)
    (arg_batch.name,  make_semantic(arg_batch),  arg_batch.description)
    (arg_resume.name, make_semantic(arg_resume), arg_resume.description)
    ;

  po::options_description desc_options("Allowed options");
  desc_options.add(desc_cmd_only).add(desc_cmd_sett);

  po::variables_map vm;
  bool r = command_line::handle_error_helper(desc_options, [&]()
  {
    po::store(po::parse_command_line(argc, argv, desc_options), vm);
    po::notify(vm);
    return true;
  });
  if (! r)
    return 1;

  log_level     = command_line::get_arg(vm, arg_log_level);
  opt_verify    = command_line::get_arg(vm, arg_verify);
  opt_batch     = command_line::get_arg(vm, arg_batch);
  opt_resume    = command_line::get_arg(vm, arg_resume);
  block_stop    = command_line::get_arg(vm, arg_block_stop);
  db_batch_size = command_line::get_arg(vm, arg_batch_size);

  if (command_line::get_arg(vm, command_line::arg_help))
  {
    std::cout << CRYPTONOTE_NAME << " v" << MONERO_VERSION_FULL << ENDL << ENDL;
    std::cout << desc_options << std::endl;
    return 1;
  }

  if (! opt_batch && ! vm["batch-size"].defaulted())
  {
    std::cerr << "Error: batch-size set, but batch option not enabled" << ENDL;
    exit(1);
  }
  if (! db_batch_size)
  {
    std::cerr << "Error: batch-size must be > 0" << ENDL;
    exit(1);
  }
  if (opt_verify && vm["batch-size"].defaulted())
  {
    // usually want batch size default lower if verify on, so progress can be
    // frequently saved.
    //
    // currently, with Windows, default batch size is low, so ignore
    // default db_batch_size_verify unless it's even lower
    if (db_batch_size > db_batch_size_verify)
    {
      db_batch_size = db_batch_size_verify;
    }
  }

  std::vector<std::string> db_engines {"memory", "lmdb"};

  opt_testnet = command_line::get_arg(vm, arg_testnet_on);
  auto data_dir_arg = opt_testnet ? command_line::arg_testnet_data_dir : command_line::arg_data_dir;
  dirname = command_line::get_arg(vm, data_dir_arg);
  db_arg_str = command_line::get_arg(vm, arg_database);

  log_space::get_set_log_detalisation_level(true, log_level);
  log_space::log_singletone::add_logger(LOGGER_CONSOLE, NULL, NULL);
  LOG_PRINT_L0("Starting...");
  LOG_PRINT_L0("Setting log level = " << log_level);

  boost::filesystem::path file_path {dirname};
  std::string import_file_path;

  import_file_path = (file_path / "export" / import_filename).string();
  if (command_line::has_arg(vm, arg_count_blocks))
  {
    BootstrapFile bootstrap;
    bootstrap.count_blocks(import_file_path);
    exit(0);
  }


  std::string db_engine;
  int mdb_flags = 0;
  int res = 0;
  res = parse_db_arguments(db_arg_str, db_engine, mdb_flags);
  if (res)
  {
    std::cerr << "Error parsing database argument(s)" << ENDL;
    exit(1);
  }

  if (std::find(db_engines.begin(), db_engines.end(), db_engine) == db_engines.end())
  {
    std::cerr << "Invalid database engine: " << db_engine << std::endl;
    exit(1);
  }

  LOG_PRINT_L0("database: " << db_engine);
  LOG_PRINT_L0("verify:  " << std::boolalpha << opt_verify << std::noboolalpha);
  if (opt_batch)
  {
    LOG_PRINT_L0("batch:   " << std::boolalpha << opt_batch << std::noboolalpha
        << "  batch size: " << db_batch_size);
  }
  else
  {
    LOG_PRINT_L0("batch:   " << std::boolalpha << opt_batch << std::noboolalpha);
  }
  LOG_PRINT_L0("resume:  " << std::boolalpha << opt_resume  << std::noboolalpha);
  LOG_PRINT_L0("testnet: " << std::boolalpha << opt_testnet << std::noboolalpha);

  LOG_PRINT_L0("bootstrap file path: " << import_file_path);
  LOG_PRINT_L0("database path:       " << file_path.string());

  try
  {

  // fake_core needed for verification to work when enabled.
  //
  // NOTE: don't need fake_core method of doing things when we're going to call
  // BlockchainDB add_block() directly and have available the 3 block
  // properties to do so. Both ways work, but fake core isn't necessary in that
  // circumstance.

  // for multi_db_runtime:
#if !defined(BLOCKCHAIN_DB)
  if (db_engine == "lmdb")
  {
    fake_core_lmdb simple_core(dirname, opt_testnet, opt_batch, mdb_flags);
    import_from_file(simple_core, import_file_path, block_stop);
  }
  else if (db_engine == "memory")
  {
    fake_core_memory simple_core(dirname, opt_testnet);
    import_from_file(simple_core, import_file_path, block_stop);
  }
  else
  {
    std::cerr << "database engine unrecognized" << ENDL;
    exit(1);
  }

  // for multi_db_compile:
#else
  if (db_engine != default_db_engine)
  {
    std::cerr << "Invalid database engine for compiled version: " << db_engine << std::endl;
    exit(1);
  }
#if BLOCKCHAIN_DB == DB_LMDB
  fake_core_lmdb simple_core(dirname, opt_testnet, opt_batch, mdb_flags);
#else
  fake_core_memory simple_core(dirname, opt_testnet);
#endif

  if (! vm["pop-blocks"].defaulted())
  {
    num_blocks = command_line::get_arg(vm, arg_pop_blocks);
    LOG_PRINT_L0("height: " << simple_core.m_storage.get_current_blockchain_height());
    pop_blocks(simple_core, num_blocks);
    LOG_PRINT_L0("height: " << simple_core.m_storage.get_current_blockchain_height());
    exit(0);
  }

  import_from_file(simple_core, import_file_path, block_stop);
#endif

  }
  catch (const DB_ERROR& e)
  {
    std::cout << std::string("Error loading blockchain db: ") + e.what() + " -- shutting down now" << ENDL;
    exit(1);
  }

  // destructors called at exit:
  //
  // ensure db closed
  //   - transactions properly checked and handled
  //   - disk sync if needed
  //
  // fake_core object's destructor is called when it goes out of scope. For an
  // LMDB fake_core, it calls Blockchain::deinit() on its object, which in turn
  // calls delete on its BlockchainDB derived class' object, which closes its
  // files.
}