/* * Copyright (c) Meta Platforms, Inc. and affiliates. * All rights reserved. * * This source code is licensed under the BSD-style license found in the * LICENSE file in the root directory of this source tree. */ #include #include #if defined(ET_USE_THREADPOOL) #include #include #endif DEFINE_string( model_path, "llama2.pte", "Model serialized in flatbuffer format."); DEFINE_string(tokenizer_path, "tokenizer.bin", "Tokenizer stuff."); DEFINE_string(prompt, "The answer to the ultimate question is", "Prompt."); DEFINE_double( temperature, 0.8f, "Temperature; Default is 0.8f. 0 = greedy argmax sampling (deterministic). Lower temperature = more deterministic"); DEFINE_int32( seq_len, 128, "Total number of tokens to generate (prompt + output). Defaults to max_seq_len. If the number of input tokens + seq_len > max_seq_len, the output will be truncated to max_seq_len tokens."); DEFINE_int32( cpu_threads, -1, "Number of CPU threads for inference. Defaults to -1, which implies we'll use a heuristic to derive the # of performant cores for a specific device."); DEFINE_bool(warmup, false, "Whether to run a warmup run."); int32_t main(int32_t argc, char** argv) { gflags::ParseCommandLineFlags(&argc, &argv, true); // Create a loader to get the data of the program file. There are other // DataLoaders that use mmap() or point32_t to data that's already in memory, // and users can create their own DataLoaders to load from arbitrary sources. const char* model_path = FLAGS_model_path.c_str(); const char* tokenizer_path = FLAGS_tokenizer_path.c_str(); const char* prompt = FLAGS_prompt.c_str(); double temperature = FLAGS_temperature; int32_t seq_len = FLAGS_seq_len; int32_t cpu_threads = FLAGS_cpu_threads; bool warmup = FLAGS_warmup; #if defined(ET_USE_THREADPOOL) uint32_t num_performant_cores = cpu_threads == -1 ? ::executorch::extension::cpuinfo::get_num_performant_cores() : static_cast(cpu_threads); ET_LOG( Info, "Resetting threadpool with num threads = %d", num_performant_cores); if (num_performant_cores > 0) { ::executorch::extension::threadpool::get_threadpool() ->_unsafe_reset_threadpool(num_performant_cores); } #endif // create llama runner example::Runner runner(model_path, tokenizer_path, temperature); if (warmup) { runner.warmup(prompt, seq_len); } // generate runner.generate(prompt, seq_len); return 0; }