/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #if defined _WIN32 || defined __APPLE__ #else #define _LINUX #endif #include "paddle/fluid/framework/data_feed.h" #include "paddle/fluid/framework/fleet/ps_gpu_wrapper.h" #ifdef _LINUX #include #include #include #endif #include "io/fs.h" #include "paddle/fluid/platform/monitor.h" #include "paddle/fluid/platform/timer.h" USE_INT_STAT(STAT_total_feasign_num_in_mem); DECLARE_bool(enable_ins_parser_file); namespace paddle { namespace framework { DLManager& global_dlmanager_pool() { static DLManager manager; return manager; } class BufferedLineFileReader { typedef std::function SampleFunc; static const int MAX_FILE_BUFF_SIZE = 4 * 1024 * 1024; class FILEReader { public: explicit FILEReader(FILE* fp) : fp_(fp) {} int read(char* buf, int len) { return fread(buf, sizeof(char), len, fp_); } private: FILE* fp_; }; public: typedef std::function LineFunc; private: template int read_lines(T* reader, LineFunc func, int skip_lines) { int lines = 0; size_t ret = 0; char* ptr = NULL; char* eol = NULL; total_len_ = 0; error_line_ = 0; SampleFunc spfunc = get_sample_func(); std::string x; while (!is_error() && (ret = reader->read(buff_, MAX_FILE_BUFF_SIZE)) > 0) { total_len_ += ret; ptr = buff_; eol = reinterpret_cast(memchr(ptr, '\n', ret)); while (eol != NULL) { int size = static_cast((eol - ptr) + 1); x.append(ptr, size - 1); ++lines; if (lines > skip_lines && spfunc()) { if (!func(x)) { ++error_line_; } } x.clear(); ptr += size; ret -= size; eol = reinterpret_cast(memchr(ptr, '\n', ret)); } if (ret > 0) { x.append(ptr, ret); } } if (!is_error() && !x.empty()) { ++lines; if (lines > skip_lines && spfunc()) { if (!func(x)) { ++error_line_; } } } return lines; } public: BufferedLineFileReader() : random_engine_(std::random_device()()), uniform_distribution_(0.0f, 1.0f) { total_len_ = 0; sample_line_ = 0; buff_ = reinterpret_cast(calloc(MAX_FILE_BUFF_SIZE + 1, sizeof(char))); } ~BufferedLineFileReader() { free(buff_); } int read_file(FILE* fp, LineFunc func, int skip_lines) { FILEReader reader(fp); return read_lines(&reader, func, skip_lines); } uint64_t file_size(void) { return total_len_; } void set_sample_rate(float r) { sample_rate_ = r; } size_t get_sample_line() { return sample_line_; } bool is_error(void) { return (error_line_ > 10); } private: SampleFunc get_sample_func() { if (std::abs(sample_rate_ - 1.0f) < 1e-5f) { return [this](void) { return true; }; } return [this](void) { return (uniform_distribution_(random_engine_) < sample_rate_); }; } private: char* buff_ = nullptr; uint64_t total_len_ = 0; std::default_random_engine random_engine_; std::uniform_real_distribution uniform_distribution_; float sample_rate_ = 1.0f; size_t sample_line_ = 0; size_t error_line_ = 0; }; void RecordCandidateList::ReSize(size_t length) { mutex_.lock(); capacity_ = length; CHECK(capacity_ > 0); // NOLINT candidate_list_.clear(); candidate_list_.resize(capacity_); full_ = false; cur_size_ = 0; total_size_ = 0; mutex_.unlock(); } void RecordCandidateList::ReInit() { mutex_.lock(); full_ = false; cur_size_ = 0; total_size_ = 0; mutex_.unlock(); } void RecordCandidateList::AddAndGet(const Record& record, RecordCandidate* result) { mutex_.lock(); size_t index = 0; ++total_size_; auto fleet_ptr = FleetWrapper::GetInstance(); if (!full_) { candidate_list_[cur_size_++] = record; full_ = (cur_size_ == capacity_); } else { CHECK(cur_size_ == capacity_); index = fleet_ptr->LocalRandomEngine()() % total_size_; if (index < capacity_) { candidate_list_[index] = record; } } index = fleet_ptr->LocalRandomEngine()() % cur_size_; *result = candidate_list_[index]; mutex_.unlock(); } void DataFeed::AddFeedVar(Variable* var, const std::string& name) { CheckInit(); for (size_t i = 0; i < use_slots_.size(); ++i) { if (name == use_slots_[i]) { if (var == nullptr) { feed_vec_[i] = nullptr; } else { feed_vec_[i] = var->GetMutable(); } } } } bool DataFeed::SetFileList(const std::vector& files) { std::unique_lock lock(*mutex_for_pick_file_); CheckInit(); // Do not set finish_set_filelist_ flag, // since a user may set file many times after init reader filelist_.assign(files.begin(), files.end()); finish_set_filelist_ = true; return true; } void DataFeed::SetBatchSize(int batch_size) { PADDLE_ENFORCE_GT(batch_size, 0, platform::errors::InvalidArgument( "Batch size %d is illegal.", batch_size)); default_batch_size_ = batch_size; } bool DataFeed::PickOneFile(std::string* filename) { PADDLE_ENFORCE_NOT_NULL( mutex_for_pick_file_, platform::errors::PreconditionNotMet( "You should call SetFileListMutex before PickOneFile")); PADDLE_ENFORCE_NOT_NULL( file_idx_, platform::errors::PreconditionNotMet( "You should call SetFileListIndex before PickOneFile")); std::unique_lock lock(*mutex_for_pick_file_); VLOG(4) << "filelist_ size: " << filelist_.size(); if (*file_idx_ == filelist_.size()) { VLOG(3) << "DataFeed::PickOneFile no more file to pick"; return false; } VLOG(3) << "file_idx_=" << *file_idx_; *filename = filelist_[(*file_idx_)++]; return true; } void DataFeed::CheckInit() { PADDLE_ENFORCE_EQ( finish_init_, true, platform::errors::PreconditionNotMet("DataFeed initialization failed.")); } void DataFeed::CheckSetFileList() { PADDLE_ENFORCE_EQ( finish_set_filelist_, true, platform::errors::PreconditionNotMet("DataFeed set filelist failed.")); } void DataFeed::CheckStart() { PADDLE_ENFORCE_EQ(finish_start_, true, platform::errors::PreconditionNotMet( "Datafeed has not started running yet.")); } void DataFeed::AssignFeedVar(const Scope& scope) { CheckInit(); for (size_t i = 0; i < use_slots_.size(); ++i) { feed_vec_[i] = scope.FindVar(use_slots_[i])->GetMutable(); } } void DataFeed::CopyToFeedTensor(void* dst, const void* src, size_t size) { if (platform::is_cpu_place(this->place_)) { memcpy(dst, src, size); } else { #ifdef PADDLE_WITH_CUDA cudaMemcpy(dst, src, size, cudaMemcpyHostToDevice); #elif defined(PADDLE_WITH_HIP) hipMemcpy(dst, src, size, hipMemcpyHostToDevice); #elif defined(PADDLE_WITH_XPU_KP) xpu_memcpy(dst, src, size, XPUMemcpyKind::XPU_HOST_TO_DEVICE); #else PADDLE_THROW(platform::errors::Unimplemented( "Not supported GPU/ROCM, please compile with option WITH_GPU=ON or " "WITH_ROCM=ON.")); #endif } } template void PrivateQueueDataFeed::SetQueueSize(int queue_size) { PADDLE_ENFORCE_GT( queue_size, 0, platform::errors::InvalidArgument( "Queue size %d is illegal in PrivateQueueDataFeed.", queue_size)); queue_size_ = queue_size; queue_ = paddle::framework::MakeChannel(); queue_->SetCapacity(queue_size); } template bool PrivateQueueDataFeed::Start() { VLOG(4) << "entering PrivateQueueDataFeed::Start()"; CheckSetFileList(); read_thread_ = std::thread(&PrivateQueueDataFeed::ReadThread, this); read_thread_.detach(); finish_start_ = true; return true; } template void PrivateQueueDataFeed::ReadThread() { #ifdef _LINUX VLOG(4) << "entering PrivateQueueDataFeed::ReadThread()"; std::string filename; while (PickOneFile(&filename)) { int err_no = 0; fp_ = fs_open_read(filename, &err_no, pipe_command_, true); __fsetlocking(&*fp_, FSETLOCKING_BYCALLER); T instance; while (ParseOneInstanceFromPipe(&instance)) { queue_->Put(instance); } } queue_->Close(); #endif } template int PrivateQueueDataFeed::Next() { #ifdef _LINUX CheckStart(); int index = 0; T ins_vec; while (index < default_batch_size_) { T instance; if (!queue_->Get(instance)) { break; } AddInstanceToInsVec(&ins_vec, instance, index++); } batch_size_ = index; if (batch_size_ != 0) { PutToFeedVec(ins_vec); } return batch_size_; #else return 0; #endif } // explicit instantiation template class PrivateQueueDataFeed>; template InMemoryDataFeed::InMemoryDataFeed() { this->file_idx_ = nullptr; this->mutex_for_pick_file_ = nullptr; this->fp_ = nullptr; this->thread_id_ = 0; this->thread_num_ = 1; this->parse_ins_id_ = false; this->parse_uid_ = false; this->parse_content_ = false; this->parse_logkey_ = false; this->enable_pv_merge_ = false; this->current_phase_ = 1; // 1:join ;0:update this->input_channel_ = nullptr; this->output_channel_ = nullptr; this->consume_channel_ = nullptr; } template bool InMemoryDataFeed::Start() { #ifdef _LINUX VLOG(4) << "entering InMemoryDataFeed::Start()"; this->CheckSetFileList(); if (output_channel_->Size() == 0 && input_channel_->Size() != 0) { std::vector data; input_channel_->Read(data); output_channel_->Write(std::move(data)); } #endif if (batch_offsets_.size() > 0) { VLOG(3) << "batch_size offsets: " << batch_offsets_.size(); enable_heterps_ = true; this->offset_index_ = 0; } this->finish_start_ = true; return true; } template int InMemoryDataFeed::Next() { #ifdef _LINUX this->CheckStart(); if (!enable_heterps_) { CHECK(output_channel_ != nullptr); CHECK(consume_channel_ != nullptr); VLOG(3) << "output_channel_ size=" << output_channel_->Size() << ", consume_channel_ size=" << consume_channel_->Size() << ", thread_id=" << thread_id_; int index = 0; T instance; std::vector ins_vec; ins_vec.reserve(this->default_batch_size_); while (index < this->default_batch_size_) { if (output_channel_->Size() == 0) { break; } output_channel_->Get(instance); ins_vec.push_back(instance); ++index; consume_channel_->Put(std::move(instance)); } this->batch_size_ = index; VLOG(3) << "batch_size_=" << this->batch_size_ << ", thread_id=" << thread_id_; if (this->batch_size_ != 0) { PutToFeedVec(ins_vec); } else { VLOG(3) << "finish reading, output_channel_ size=" << output_channel_->Size() << ", consume_channel_ size=" << consume_channel_->Size() << ", thread_id=" << thread_id_; } } else { VLOG(3) << "enable heter next: " << offset_index_ << " batch_offsets: " << batch_offsets_.size(); if (offset_index_ >= batch_offsets_.size()) { VLOG(3) << "offset_index: " << offset_index_ << " batch_offsets: " << batch_offsets_.size(); return 0; } auto& batch = batch_offsets_[offset_index_++]; this->batch_size_ = batch.second; VLOG(3) << "batch_size_=" << this->batch_size_ << ", thread_id=" << thread_id_; if (this->batch_size_ != 0) { PutToFeedVec(&records_[batch.first], this->batch_size_); } else { VLOG(3) << "finish reading for heterps, batch size zero, thread_id=" << thread_id_; } VLOG(3) << "enable heter next: " << offset_index_ << " batch_offsets: " << batch_offsets_.size() << " baych_size: " << this->batch_size_; } return this->batch_size_; #else return 0; #endif } template void InMemoryDataFeed::SetInputChannel(void* channel) { input_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetOutputChannel(void* channel) { output_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetConsumeChannel(void* channel) { consume_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetInputPvChannel(void* channel) { input_pv_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetOutputPvChannel(void* channel) { output_pv_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetConsumePvChannel(void* channel) { consume_pv_channel_ = static_cast*>(channel); } template void InMemoryDataFeed::SetThreadId(int thread_id) { thread_id_ = thread_id; } template void InMemoryDataFeed::SetThreadNum(int thread_num) { thread_num_ = thread_num; } template void InMemoryDataFeed::SetParseContent(bool parse_content) { parse_content_ = parse_content; } template void InMemoryDataFeed::SetParseLogKey(bool parse_logkey) { parse_logkey_ = parse_logkey; } template void InMemoryDataFeed::SetEnablePvMerge(bool enable_pv_merge) { enable_pv_merge_ = enable_pv_merge; } template void InMemoryDataFeed::SetCurrentPhase(int current_phase) { current_phase_ = current_phase; } template void InMemoryDataFeed::SetParseInsId(bool parse_ins_id) { parse_ins_id_ = parse_ins_id; } template void InMemoryDataFeed::SetParseUid(bool parse_uid) { parse_uid_ = parse_uid; } template void InMemoryDataFeed::LoadIntoMemory() { #ifdef _LINUX if (!so_parser_name_.empty()) { LoadIntoMemoryFromSo(); return; } VLOG(3) << "LoadIntoMemory() begin, thread_id=" << thread_id_; std::string filename; while (this->PickOneFile(&filename)) { VLOG(3) << "PickOneFile, filename=" << filename << ", thread_id=" << thread_id_; #ifdef PADDLE_WITH_BOX_PS if (BoxWrapper::GetInstance()->UseAfsApi()) { this->fp_ = BoxWrapper::GetInstance()->afs_manager->GetFile( filename, this->pipe_command_); } else { #endif int err_no = 0; this->fp_ = fs_open_read(filename, &err_no, this->pipe_command_, true); #ifdef PADDLE_WITH_BOX_PS } #endif CHECK(this->fp_ != nullptr); __fsetlocking(&*(this->fp_), FSETLOCKING_BYCALLER); paddle::framework::ChannelWriter writer(input_channel_); T instance; platform::Timer timeline; timeline.Start(); while (ParseOneInstanceFromPipe(&instance)) { writer << std::move(instance); instance = T(); } STAT_ADD(STAT_total_feasign_num_in_mem, fea_num_); { std::lock_guard flock(*mutex_for_fea_num_); *total_fea_num_ += fea_num_; fea_num_ = 0; } writer.Flush(); timeline.Pause(); VLOG(3) << "LoadIntoMemory() read all lines, file=" << filename << ", cost time=" << timeline.ElapsedSec() << " seconds, thread_id=" << thread_id_; } VLOG(3) << "LoadIntoMemory() end, thread_id=" << thread_id_; #endif } template void InMemoryDataFeed::LoadIntoMemoryFromSo() { #if (defined _LINUX) && (defined PADDLE_WITH_HETERPS) && \ (defined PADDLE_WITH_PSLIB) VLOG(3) << "LoadIntoMemoryFromSo() begin, thread_id=" << thread_id_; int buf_len = 1024 * 1024 * 10; char* buf = reinterpret_cast(malloc(buf_len + 10)); auto ps_gpu_ptr = PSGPUWrapper::GetInstance(); paddle::framework::CustomParser* parser = global_dlmanager_pool().Load(so_parser_name_, slot_conf_); std::string filename; while (this->PickOneFile(&filename)) { VLOG(3) << "PickOneFile, filename=" << filename << ", thread_id=" << thread_id_; platform::Timer timeline; timeline.Start(); if (ps_gpu_ptr->UseAfsApi()) { auto afs_reader = ps_gpu_ptr->OpenReader(filename); int read_len = 0; char* cursor = buf; int remain = 0; while ((read_len = afs_reader->read(cursor, buf_len - remain)) > 0) { std::vector instances; read_len += remain; remain = ParseInstanceFromSo(read_len, buf, &instances, parser); input_channel_->Write(std::move(instances)); instances = std::vector(); if (remain) { memmove(buf, buf + read_len - remain, remain); } cursor = buf + remain; } } else { VLOG(0) << "Should Call InitAfsApi First"; } timeline.Pause(); VLOG(3) << "LoadIntoMemoryFromSo() read all lines, file=" << filename << ", cost time=" << timeline.ElapsedSec() << " seconds, thread_id=" << thread_id_; } free(buf); VLOG(3) << "LoadIntoMemoryFromSo() end, thread_id=" << thread_id_; #endif } // explicit instantiation template class InMemoryDataFeed; void MultiSlotDataFeed::Init( const paddle::framework::DataFeedDesc& data_feed_desc) { finish_init_ = false; finish_set_filelist_ = false; finish_start_ = false; PADDLE_ENFORCE_EQ( data_feed_desc.has_multi_slot_desc(), true, platform::errors::PreconditionNotMet( "Multi_slot_desc has not been set in MultiSlotDataFeed.")); paddle::framework::MultiSlotDesc multi_slot_desc = data_feed_desc.multi_slot_desc(); SetBatchSize(data_feed_desc.batch_size()); // temporarily set queue size = batch size * 100 SetQueueSize(data_feed_desc.batch_size() * 100); size_t all_slot_num = multi_slot_desc.slots_size(); all_slots_.resize(all_slot_num); all_slots_type_.resize(all_slot_num); use_slots_index_.resize(all_slot_num); total_dims_without_inductive_.resize(all_slot_num); inductive_shape_index_.resize(all_slot_num); use_slots_.clear(); use_slots_is_dense_.clear(); for (size_t i = 0; i < all_slot_num; ++i) { const auto& slot = multi_slot_desc.slots(i); all_slots_[i] = slot.name(); all_slots_type_[i] = slot.type(); use_slots_index_[i] = slot.is_used() ? use_slots_.size() : -1; total_dims_without_inductive_[i] = 1; inductive_shape_index_[i] = -1; if (slot.is_used()) { use_slots_.push_back(all_slots_[i]); use_slots_is_dense_.push_back(slot.is_dense()); std::vector local_shape; if (slot.is_dense()) { for (int j = 0; j < slot.shape_size(); ++j) { if (slot.shape(j) > 0) { total_dims_without_inductive_[i] *= slot.shape(j); } if (slot.shape(j) == -1) { inductive_shape_index_[i] = j; } } } for (int j = 0; j < slot.shape_size(); ++j) { local_shape.push_back(slot.shape(j)); } use_slots_shape_.push_back(local_shape); } } feed_vec_.resize(use_slots_.size()); pipe_command_ = data_feed_desc.pipe_command(); finish_init_ = true; } void MultiSlotDataFeed::ReadThread() { #ifdef _LINUX VLOG(4) << "entering MultiSlotDataFeed::ReadThread()"; std::string filename; while (PickOneFile(&filename)) { int err_no = 0; fp_ = fs_open_read(filename, &err_no, pipe_command_, true); CHECK(fp_ != nullptr); __fsetlocking(&*fp_, FSETLOCKING_BYCALLER); std::vector instance; int ins_num = 0; while (ParseOneInstanceFromPipe(&instance)) { ins_num++; queue_->Put(instance); } VLOG(3) << "filename: " << filename << " inst num: " << ins_num; } queue_->Close(); #endif } bool MultiSlotDataFeed::CheckFile(const char* filename) { #ifdef _LINUX CheckInit(); // get info of slots std::ifstream fin(filename); if (!fin.good()) { VLOG(1) << "error: open file<" << filename << "> fail"; return false; } std::string line; int instance_cout = 0; std::string all_slots_alias = ""; for (const auto& alias : all_slots_) { all_slots_alias += alias + " "; } std::string use_slots_alias = ""; for (const auto& alias : use_slots_) { use_slots_alias += alias + " "; } VLOG(3) << "total slots num: " << all_slots_.size(); VLOG(3) << "total slots alias: " << all_slots_alias; VLOG(3) << "used slots num: " << use_slots_.size(); VLOG(3) << "used slots alias: " << use_slots_alias; while (getline(fin, line)) { ++instance_cout; const char* str = line.c_str(); char* endptr = const_cast(str); int len = line.length(); for (size_t i = 0; i < all_slots_.size(); ++i) { auto num = strtol(endptr, &endptr, 10); if (num < 0) { VLOG(0) << "error: the number of ids is a negative number: " << num; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); return false; } else if (num == 0) { VLOG(0) << "error: the number of ids can not be zero, you need " "padding it in data generator; or if there is something wrong" " with the data, please check if the data contains unresolvable " "characters."; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); return false; } else if (errno == ERANGE || num > INT_MAX) { VLOG(0) << "error: the number of ids greater than INT_MAX"; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); return false; } if (all_slots_type_[i] == "float") { for (int j = 0; j < num; ++j) { strtof(endptr, &endptr); if (errno == ERANGE) { VLOG(0) << "error: the value is out of the range of " "representable values for float"; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); VLOG(0) << "and in this slot: " << j << " th id with total id number: " << num; return false; } if (j + 1 != num && endptr - str == len) { VLOG(0) << "error: there is a wrong with the number of ids."; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); VLOG(0) << "and in this slot: " << j << " th id with total id number: " << num; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; return false; } } } else if (all_slots_type_[i] == "uint64") { for (int j = 0; j < num; ++j) { strtoull(endptr, &endptr, 10); if (errno == ERANGE) { VLOG(0) << "error: the value is out of the range of " "representable values for uint64_t"; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); VLOG(0) << "and in this slot: " << j << " th id with total id number: " << num; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; return false; } if (j + 1 != num && endptr - str == len) { VLOG(0) << "error: there is a wrong with the number of ids."; VLOG(0) << "Error occurred when parsing " << i << " th slot with total slots number: " << all_slots_.size(); VLOG(0) << "and in this slot: " << j << " th id with total id number: " << num; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; return false; } } } else { VLOG(0) << "error: this type<" << all_slots_type_[i] << "> is not supported"; return false; } } // It may be added '\t' character to the end of the output of reduce // task when processes data by Hadoop(when the output of the reduce // task of Hadoop has only one field, it will add a '\t' at the end // of the line by default, and you can use this option to avoid it: // `-D mapred.textoutputformat.ignoreseparator=true`), which does // not affect the correctness of the data. Therefore, it should be // judged that the data is not normal when the end of each line of // data contains characters which are not spaces. while (endptr - str != len) { if (!isspace(*(endptr++))) { VLOG(0) << "error: there is some extra characters at the end of the line."; VLOG(0) << "please check line<" << instance_cout << "> in file<" << filename << ">"; return false; } } } VLOG(3) << "instances cout: " << instance_cout; VLOG(3) << "The file format is correct"; #endif return true; } bool MultiSlotDataFeed::ParseOneInstanceFromPipe( std::vector* instance) { #ifdef _LINUX thread_local string::LineFileReader reader; if (!reader.getline(&*(fp_.get()))) { return false; } else { int use_slots_num = use_slots_.size(); instance->resize(use_slots_num); const char* str = reader.get(); std::string line = std::string(str); char* endptr = const_cast(str); int pos = 0; for (size_t i = 0; i < use_slots_index_.size(); ++i) { int idx = use_slots_index_[i]; int num = strtol(&str[pos], &endptr, 10); if (num <= 0) { std::stringstream ss; ss << "\n\nGot unexpected input, maybe something wrong with it.\n"; ss << "\n----------------------\n"; ss << "The Origin Input Data:\n"; ss << "----------------------\n"; ss << line << "\n"; ss << "\n----------------------\n"; ss << "Some Possible Errors:\n"; ss << "----------------------\n"; ss << "1. The number of ids can not be zero, you need padding.\n"; ss << "2. The input data contains unresolvable characters.\n"; ss << "3. We detect the slot " << i << "'s feasign number is " << num << " which is illegal.\n"; ss << "\n"; PADDLE_THROW(platform::errors::InvalidArgument(ss.str())); } if (idx != -1) { (*instance)[idx].Init(all_slots_type_[i]); if ((*instance)[idx].GetType()[0] == 'f') { // float for (int j = 0; j < num; ++j) { float feasign = strtof(endptr, &endptr); (*instance)[idx].AddValue(feasign); } } else if ((*instance)[idx].GetType()[0] == 'u') { // uint64 for (int j = 0; j < num; ++j) { uint64_t feasign = (uint64_t)strtoull(endptr, &endptr, 10); (*instance)[idx].AddValue(feasign); } } pos = endptr - str; } else { for (int j = 0; j <= num; ++j) { // pos = line.find_first_of(' ', pos + 1); while (line[pos + 1] != ' ') { pos++; } } } } return true; } #else return true; #endif } bool MultiSlotDataFeed::ParseOneInstance(std::vector* instance) { #ifdef _LINUX std::string line; if (getline(file_, line)) { int use_slots_num = use_slots_.size(); instance->resize(use_slots_num); // parse line const char* str = line.c_str(); char* endptr = const_cast(str); int pos = 0; for (size_t i = 0; i < use_slots_index_.size(); ++i) { int idx = use_slots_index_[i]; int num = strtol(&str[pos], &endptr, 10); PADDLE_ENFORCE_NE( num, 0, platform::errors::InvalidArgument( "The number of ids can not be zero, you need padding " "it in data generator; or if there is something wrong with " "the data, please check if the data contains unresolvable " "characters.\nplease check this error line: %s, \n Specifically, " "something wrong happened(the length of this slot's feasign is 0)" "when we parse the %d th slots." "Maybe something wrong around this slot" "\nWe detect the feasign number of this slot is %d, " "which is illegal.", str, i, num)); if (idx != -1) { (*instance)[idx].Init(all_slots_type_[i]); if ((*instance)[idx].GetType()[0] == 'f') { // float for (int j = 0; j < num; ++j) { float feasign = strtof(endptr, &endptr); (*instance)[idx].AddValue(feasign); } } else if ((*instance)[idx].GetType()[0] == 'u') { // uint64 for (int j = 0; j < num; ++j) { uint64_t feasign = (uint64_t)strtoull(endptr, &endptr, 10); (*instance)[idx].AddValue(feasign); } } pos = endptr - str; } else { for (int j = 0; j <= num; ++j) { pos = line.find_first_of(' ', pos + 1); } } } } else { return false; } #endif return false; } void MultiSlotDataFeed::AddInstanceToInsVec( std::vector* ins_vec, const std::vector& instance, int index) { #ifdef _LINUX if (index == 0) { ins_vec->resize(instance.size()); for (size_t i = 0; i < instance.size(); ++i) { (*ins_vec)[i].Init(instance[i].GetType()); (*ins_vec)[i].InitOffset(); } } for (size_t i = 0; i < instance.size(); ++i) { (*ins_vec)[i].AddIns(instance[i]); } #endif } void MultiSlotDataFeed::PutToFeedVec( const std::vector& ins_vec) { #ifdef _LINUX for (size_t i = 0; i < use_slots_.size(); ++i) { if (feed_vec_[i] == nullptr) { continue; } VLOG(4) << "MultiSlotDataFeed::PutToFeedVec i: " << i; const auto& type = ins_vec[i].GetType(); const auto& offset = ins_vec[i].GetOffset(); int total_instance = static_cast(offset.back()); VLOG(4) << "total_instance: " << total_instance; // platform::CPUPlace() VLOG(4) << "this->place_: " << this->place_; if (type[0] == 'f') { // float const auto& feasign = ins_vec[i].GetFloatData(); float* tensor_ptr = feed_vec_[i]->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, &feasign[0], total_instance * sizeof(float)); } else if (type[0] == 'u') { // uint64 // no uint64_t type in paddlepaddle const auto& feasign = ins_vec[i].GetUint64Data(); int64_t* tensor_ptr = feed_vec_[i]->mutable_data( {total_instance, 1}, this->place_); CopyToFeedTensor( tensor_ptr, &feasign[0], total_instance * sizeof(int64_t)); } if (!use_slots_is_dense_[i]) { LoD data_lod{offset}; feed_vec_[i]->set_lod(data_lod); } if (use_slots_is_dense_[i]) { if (inductive_shape_index_[i] != -1) { use_slots_shape_[i][inductive_shape_index_[i]] = total_instance / total_dims_without_inductive_[i]; } feed_vec_[i]->Resize(phi::make_ddim(use_slots_shape_[i])); } } #endif } void MultiSlotInMemoryDataFeed::Init( const paddle::framework::DataFeedDesc& data_feed_desc) { finish_init_ = false; finish_set_filelist_ = false; finish_start_ = false; PADDLE_ENFORCE_EQ( data_feed_desc.has_multi_slot_desc(), true, platform::errors::PreconditionNotMet( "Multi_slot_desc has not been set in MultiSlotInMemoryDataFeed.")); paddle::framework::MultiSlotDesc multi_slot_desc = data_feed_desc.multi_slot_desc(); SetBatchSize(data_feed_desc.batch_size()); size_t all_slot_num = multi_slot_desc.slots_size(); all_slots_.resize(all_slot_num); all_slots_type_.resize(all_slot_num); use_slots_index_.resize(all_slot_num); total_dims_without_inductive_.resize(all_slot_num); inductive_shape_index_.resize(all_slot_num); use_slots_.clear(); use_slots_is_dense_.clear(); slot_conf_.resize(all_slot_num); for (size_t i = 0; i < all_slot_num; ++i) { const auto& slot = multi_slot_desc.slots(i); all_slots_[i] = slot.name(); all_slots_type_[i] = slot.type(); use_slots_index_[i] = slot.is_used() ? use_slots_.size() : -1; slot_conf_[i].name = slot.name(); slot_conf_[i].type = slot.type(); slot_conf_[i].use_slots_index = use_slots_index_[i]; total_dims_without_inductive_[i] = 1; inductive_shape_index_[i] = -1; if (slot.is_used()) { use_slots_.push_back(all_slots_[i]); use_slots_is_dense_.push_back(slot.is_dense()); slot_conf_[i].use_slots_is_dense = slot.is_dense(); std::vector local_shape; if (slot.is_dense()) { for (int j = 0; j < slot.shape_size(); ++j) { if (slot.shape(j) > 0) { total_dims_without_inductive_[i] *= slot.shape(j); } if (slot.shape(j) == -1) { inductive_shape_index_[i] = j; } } } for (int j = 0; j < slot.shape_size(); ++j) { local_shape.push_back(slot.shape(j)); } use_slots_shape_.push_back(local_shape); } } uid_slot_ = multi_slot_desc.uid_slot(); feed_vec_.resize(use_slots_.size()); const int kEstimatedFeasignNumPerSlot = 5; // Magic Number for (size_t i = 0; i < all_slot_num; i++) { batch_float_feasigns_.push_back(std::vector()); batch_uint64_feasigns_.push_back(std::vector()); batch_float_feasigns_[i].reserve(default_batch_size_ * kEstimatedFeasignNumPerSlot); batch_uint64_feasigns_[i].reserve(default_batch_size_ * kEstimatedFeasignNumPerSlot); offset_.push_back(std::vector()); offset_[i].reserve(default_batch_size_ + 1); // Each lod info will prepend a zero } visit_.resize(all_slot_num, false); pipe_command_ = data_feed_desc.pipe_command(); so_parser_name_ = data_feed_desc.so_parser_name(); finish_init_ = true; input_type_ = data_feed_desc.input_type(); } void MultiSlotInMemoryDataFeed::GetMsgFromLogKey(const std::string& log_key, uint64_t* search_id, uint32_t* cmatch, uint32_t* rank) { std::string searchid_str = log_key.substr(16, 16); *search_id = (uint64_t)strtoull(searchid_str.c_str(), NULL, 16); std::string cmatch_str = log_key.substr(11, 3); *cmatch = (uint32_t)strtoul(cmatch_str.c_str(), NULL, 16); std::string rank_str = log_key.substr(14, 2); *rank = (uint32_t)strtoul(rank_str.c_str(), NULL, 16); } int MultiSlotInMemoryDataFeed::ParseInstanceFromSo( int len, const char* str, std::vector* instances, CustomParser* parser) { // VLOG(0) << "parser: " << parser; return parser->ParseInstance(len, str, instances); } bool MultiSlotInMemoryDataFeed::ParseOneInstanceFromPipe(Record* instance) { #ifdef _LINUX thread_local string::LineFileReader reader; if (!reader.getline(&*(fp_.get()))) { return false; } else { const char* str = reader.get(); std::string line = std::string(str); // VLOG(3) << line; char* endptr = const_cast(str); int pos = 0; if (parse_ins_id_) { int num = strtol(&str[pos], &endptr, 10); CHECK(num == 1); // NOLINT pos = endptr - str + 1; size_t len = 0; while (str[pos + len] != ' ') { ++len; } instance->ins_id_ = std::string(str + pos, len); pos += len + 1; VLOG(3) << "ins_id " << instance->ins_id_; } if (parse_content_) { int num = strtol(&str[pos], &endptr, 10); CHECK(num == 1); // NOLINT pos = endptr - str + 1; size_t len = 0; while (str[pos + len] != ' ') { ++len; } instance->content_ = std::string(str + pos, len); pos += len + 1; VLOG(3) << "content " << instance->content_; } if (parse_logkey_) { int num = strtol(&str[pos], &endptr, 10); CHECK(num == 1); // NOLINT pos = endptr - str + 1; size_t len = 0; while (str[pos + len] != ' ') { ++len; } // parse_logkey std::string log_key = std::string(str + pos, len); uint64_t search_id; uint32_t cmatch; uint32_t rank; GetMsgFromLogKey(log_key, &search_id, &cmatch, &rank); instance->ins_id_ = log_key; instance->search_id = search_id; instance->cmatch = cmatch; instance->rank = rank; pos += len + 1; } for (size_t i = 0; i < use_slots_index_.size(); ++i) { int idx = use_slots_index_[i]; int num = strtol(&str[pos], &endptr, 10); PADDLE_ENFORCE_NE( num, 0, platform::errors::InvalidArgument( "The number of ids can not be zero, you need padding " "it in data generator; or if there is something wrong with " "the data, please check if the data contains unresolvable " "characters.\nplease check this error line: %s, \n Specifically, " "something wrong happened(the length of this slot's feasign is 0)" "when we parse the %d th slots." "Maybe something wrong around this slot" "\nWe detect the feasign number of this slot is %d, " "which is illegal.", str, i, num)); #ifdef PADDLE_WITH_PSLIB if (parse_uid_ && all_slots_[i] == uid_slot_) { PADDLE_ENFORCE(num == 1 && all_slots_type_[i][0] == 'u', platform::errors::PreconditionNotMet( "The uid has to be uint64 and single.\n" "please check this error line: %s", str)); char* uidptr = endptr; uint64_t feasign = (uint64_t)strtoull(uidptr, &uidptr, 10); instance->uid_ = feasign; } #endif if (idx != -1) { if (all_slots_type_[i][0] == 'f') { // float for (int j = 0; j < num; ++j) { float feasign = strtof(endptr, &endptr); // if float feasign is equal to zero, ignore it // except when slot is dense if (fabs(feasign) < 1e-6 && !use_slots_is_dense_[i]) { continue; } FeatureFeasign f; f.float_feasign_ = feasign; instance->float_feasigns_.push_back(FeatureItem(f, idx)); } } else if (all_slots_type_[i][0] == 'u') { // uint64 for (int j = 0; j < num; ++j) { uint64_t feasign = (uint64_t)strtoull(endptr, &endptr, 10); // if uint64 feasign is equal to zero, ignore it // except when slot is dense if (feasign == 0 && !use_slots_is_dense_[i]) { continue; } FeatureFeasign f; f.uint64_feasign_ = feasign; instance->uint64_feasigns_.push_back(FeatureItem(f, idx)); } } pos = endptr - str; } else { for (int j = 0; j <= num; ++j) { // pos = line.find_first_of(' ', pos + 1); while (line[pos + 1] != ' ') { pos++; } } } } instance->float_feasigns_.shrink_to_fit(); instance->uint64_feasigns_.shrink_to_fit(); fea_num_ += instance->uint64_feasigns_.size(); return true; } #else return false; #endif } bool MultiSlotInMemoryDataFeed::ParseOneInstance(Record* instance) { #ifdef _LINUX std::string line; if (getline(file_, line)) { VLOG(3) << line; // parse line const char* str = line.c_str(); char* endptr = const_cast(str); int pos = 0; for (size_t i = 0; i < use_slots_index_.size(); ++i) { int idx = use_slots_index_[i]; int num = strtol(&str[pos], &endptr, 10); PADDLE_ENFORCE_NE( num, 0, platform::errors::InvalidArgument( "The number of ids can not be zero, you need padding " "it in data generator; or if there is something wrong with " "the data, please check if the data contains unresolvable " "characters.\nplease check this error line: %s, \n Specifically, " "something wrong happened(the length of this slot's feasign is 0)" "when we parse the %d th slots." "Maybe something wrong around this slot" "\nWe detect the feasign number of this slot is %d, " "which is illegal.", str, i, num)); if (idx != -1) { if (all_slots_type_[i][0] == 'f') { // float for (int j = 0; j < num; ++j) { float feasign = strtof(endptr, &endptr); if (fabs(feasign) < 1e-6) { continue; } FeatureFeasign f; f.float_feasign_ = feasign; instance->float_feasigns_.push_back(FeatureItem(f, idx)); } } else if (all_slots_type_[i][0] == 'u') { // uint64 for (int j = 0; j < num; ++j) { uint64_t feasign = (uint64_t)strtoull(endptr, &endptr, 10); if (feasign == 0) { continue; } FeatureFeasign f; f.uint64_feasign_ = feasign; instance->uint64_feasigns_.push_back(FeatureItem(f, idx)); } } pos = endptr - str; } else { for (int j = 0; j <= num; ++j) { pos = line.find_first_of(' ', pos + 1); } } } instance->float_feasigns_.shrink_to_fit(); instance->uint64_feasigns_.shrink_to_fit(); return true; } else { return false; } #endif return false; } void MultiSlotInMemoryDataFeed::PutToFeedVec(const Record* ins_vec, int num) { #ifdef _LINUX for (size_t i = 0; i < batch_float_feasigns_.size(); ++i) { batch_float_feasigns_[i].clear(); batch_uint64_feasigns_[i].clear(); offset_[i].clear(); offset_[i].push_back(0); } ins_content_vec_.clear(); ins_content_vec_.reserve(num); ins_id_vec_.clear(); ins_id_vec_.reserve(num); for (int i = 0; i < num; ++i) { auto& r = ins_vec[i]; ins_id_vec_.push_back(r.ins_id_); ins_content_vec_.push_back(r.content_); for (auto& item : r.float_feasigns_) { batch_float_feasigns_[item.slot()].push_back(item.sign().float_feasign_); visit_[item.slot()] = true; } for (auto& item : r.uint64_feasigns_) { batch_uint64_feasigns_[item.slot()].push_back( item.sign().uint64_feasign_); visit_[item.slot()] = true; } for (size_t j = 0; j < use_slots_.size(); ++j) { const auto& type = all_slots_type_[j]; if (visit_[j]) { visit_[j] = false; } else { // fill slot value with default value 0 if (type[0] == 'f') { // float batch_float_feasigns_[j].push_back(0.0); } else if (type[0] == 'u') { // uint64 batch_uint64_feasigns_[j].push_back(0); } } // get offset of this ins in this slot if (type[0] == 'f') { // float offset_[j].push_back(batch_float_feasigns_[j].size()); } else if (type[0] == 'u') { // uint64 offset_[j].push_back(batch_uint64_feasigns_[j].size()); } } } for (size_t i = 0; i < use_slots_.size(); ++i) { if (feed_vec_[i] == nullptr) { continue; } int total_instance = offset_[i].back(); const auto& type = all_slots_type_[i]; if (type[0] == 'f') { // float float* feasign = batch_float_feasigns_[i].data(); float* tensor_ptr = feed_vec_[i]->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(float)); } else if (type[0] == 'u') { // uint64 // no uint64_t type in paddlepaddle uint64_t* feasign = batch_uint64_feasigns_[i].data(); int64_t* tensor_ptr = feed_vec_[i]->mutable_data( {total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(int64_t)); } auto& slot_offset = offset_[i]; if (this->input_type_ == 0) { LoD data_lod{slot_offset}; feed_vec_[i]->set_lod(data_lod); } else if (this->input_type_ == 1) { if (!use_slots_is_dense_[i]) { std::vector tmp_offset; PADDLE_ENFORCE_EQ(slot_offset.size(), 2, platform::errors::InvalidArgument( "In batch reader, the sparse tensor lod size " "must be 2, but received %d.", slot_offset.size())); const auto& max_size = slot_offset[1]; tmp_offset.reserve(max_size + 1); for (unsigned int k = 0; k <= max_size; k++) { tmp_offset.emplace_back(k); } slot_offset = tmp_offset; LoD data_lod{slot_offset}; feed_vec_[i]->set_lod(data_lod); } } if (use_slots_is_dense_[i]) { if (inductive_shape_index_[i] != -1) { use_slots_shape_[i][inductive_shape_index_[i]] = total_instance / total_dims_without_inductive_[i]; } feed_vec_[i]->Resize(phi::make_ddim(use_slots_shape_[i])); } } #endif } void MultiSlotInMemoryDataFeed::PutToFeedVec( const std::vector& ins_vec) { #ifdef _LINUX for (size_t i = 0; i < batch_float_feasigns_.size(); ++i) { batch_float_feasigns_[i].clear(); batch_uint64_feasigns_[i].clear(); offset_[i].clear(); offset_[i].push_back(0); } ins_content_vec_.clear(); ins_content_vec_.reserve(ins_vec.size()); ins_id_vec_.clear(); ins_id_vec_.reserve(ins_vec.size()); for (size_t i = 0; i < ins_vec.size(); ++i) { auto& r = ins_vec[i]; ins_id_vec_.push_back(r.ins_id_); ins_content_vec_.push_back(r.content_); for (auto& item : r.float_feasigns_) { batch_float_feasigns_[item.slot()].push_back(item.sign().float_feasign_); visit_[item.slot()] = true; } for (auto& item : r.uint64_feasigns_) { batch_uint64_feasigns_[item.slot()].push_back( item.sign().uint64_feasign_); visit_[item.slot()] = true; } for (size_t j = 0; j < use_slots_.size(); ++j) { const auto& type = all_slots_type_[j]; if (visit_[j]) { visit_[j] = false; } else { // fill slot value with default value 0 if (type[0] == 'f') { // float batch_float_feasigns_[j].push_back(0.0); } else if (type[0] == 'u') { // uint64 batch_uint64_feasigns_[j].push_back(0); } } // get offset of this ins in this slot if (type[0] == 'f') { // float offset_[j].push_back(batch_float_feasigns_[j].size()); } else if (type[0] == 'u') { // uint64 offset_[j].push_back(batch_uint64_feasigns_[j].size()); } } } for (size_t i = 0; i < use_slots_.size(); ++i) { if (feed_vec_[i] == nullptr) { continue; } int total_instance = offset_[i].back(); const auto& type = all_slots_type_[i]; if (type[0] == 'f') { // float float* feasign = batch_float_feasigns_[i].data(); float* tensor_ptr = feed_vec_[i]->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(float)); } else if (type[0] == 'u') { // uint64 // no uint64_t type in paddlepaddle uint64_t* feasign = batch_uint64_feasigns_[i].data(); int64_t* tensor_ptr = feed_vec_[i]->mutable_data( {total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(int64_t)); } auto& slot_offset = offset_[i]; if (this->input_type_ == 0) { if (!use_slots_is_dense_[i]) { LoD data_lod{slot_offset}; feed_vec_[i]->set_lod(data_lod); } } else if (this->input_type_ == 1) { if (!use_slots_is_dense_[i]) { std::vector tmp_offset; PADDLE_ENFORCE_EQ(slot_offset.size(), 2, platform::errors::InvalidArgument( "In batch reader, the sparse tensor lod size " "must be 2, but received %d.", slot_offset.size())); const auto& max_size = slot_offset[1]; tmp_offset.reserve(max_size + 1); for (unsigned int k = 0; k <= max_size; k++) { tmp_offset.emplace_back(k); } slot_offset = tmp_offset; LoD data_lod{slot_offset}; feed_vec_[i]->set_lod(data_lod); } } if (use_slots_is_dense_[i]) { if (inductive_shape_index_[i] != -1) { use_slots_shape_[i][inductive_shape_index_[i]] = total_instance / total_dims_without_inductive_[i]; } feed_vec_[i]->Resize(phi::make_ddim(use_slots_shape_[i])); } } #endif } #if (defined(PADDLE_WITH_CUDA) || defined(PADDLE_WITH_HIP)) && !defined(_WIN32) template void PrivateInstantDataFeed::PutToFeedVec() { for (size_t i = 0; i < use_slots_.size(); ++i) { const auto& type = ins_vec_[i].GetType(); const auto& offset = ins_vec_[i].GetOffset(); int total_instance = static_cast(offset.back()); if (type[0] == 'f') { // float const auto& feasign = ins_vec_[i].GetFloatData(); float* tensor_ptr = feed_vec_[i]->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, &feasign[0], total_instance * sizeof(float)); } else if (type[0] == 'u') { // uint64 // no uint64_t type in paddlepaddle const auto& feasign = ins_vec_[i].GetUint64Data(); int64_t* tensor_ptr = feed_vec_[i]->mutable_data( {total_instance, 1}, this->place_); CopyToFeedTensor( tensor_ptr, &feasign[0], total_instance * sizeof(int64_t)); } LoD data_lod{offset}; feed_vec_[i]->set_lod(data_lod); if (use_slots_is_dense_[i]) { int64_t total_dims = 1; for (const auto e : use_slots_shape_[i]) { total_dims *= e; } PADDLE_ENFORCE_EQ( total_dims, total_instance, platform::errors::InvalidArgument( "The actual data size of slot[%s] doesn't match its declaration. " "The actual data size of slot is %lld" ", and its declaration is %lld.", use_slots_[i].c_str(), total_dims, total_instance)); feed_vec_[i]->Resize(phi::make_ddim(use_slots_shape_[i])); } } } template int PrivateInstantDataFeed::Next() { if (ParseOneMiniBatch()) { PutToFeedVec(); return ins_vec_[0].GetBatchSize(); } Postprocess(); std::string filename; if (!PickOneFile(&filename)) { return -1; } if (!Preprocess(filename)) { return -1; } PADDLE_ENFORCE_EQ( true, ParseOneMiniBatch(), platform::errors::InvalidArgument("Fail to parse mini-batch data.")); PutToFeedVec(); return ins_vec_[0].GetBatchSize(); } template void PrivateInstantDataFeed::Init(const DataFeedDesc& data_feed_desc) { finish_init_ = false; finish_set_filelist_ = false; finish_start_ = false; PADDLE_ENFORCE_EQ( data_feed_desc.has_multi_slot_desc(), true, platform::errors::PreconditionNotMet( "Multi_slot_desc has not been set in PrivateInstantDataFeed.")); paddle::framework::MultiSlotDesc multi_slot_desc = data_feed_desc.multi_slot_desc(); SetBatchSize(data_feed_desc.batch_size()); size_t all_slot_num = multi_slot_desc.slots_size(); all_slots_.resize(all_slot_num); all_slots_type_.resize(all_slot_num); use_slots_index_.resize(all_slot_num); multi_inductive_shape_index_.resize(all_slot_num); use_slots_.clear(); use_slots_is_dense_.clear(); for (size_t i = 0; i < all_slot_num; ++i) { const auto& slot = multi_slot_desc.slots(i); all_slots_[i] = slot.name(); all_slots_type_[i] = slot.type(); use_slots_index_[i] = slot.is_used() ? use_slots_.size() : -1; if (slot.is_used()) { use_slots_.push_back(all_slots_[i]); use_slots_is_dense_.push_back(slot.is_dense()); std::vector local_shape; if (slot.is_dense()) { for (int j = 0; j < slot.shape_size(); ++j) { if (slot.shape(j) == -1) { multi_inductive_shape_index_[i].push_back(j); } } } for (int j = 0; j < slot.shape_size(); ++j) { local_shape.push_back(slot.shape(j)); } use_slots_shape_.push_back(local_shape); } } feed_vec_.resize(use_slots_.size()); ins_vec_.resize(use_slots_.size()); finish_init_ = true; } template class PrivateInstantDataFeed>; bool MultiSlotFileInstantDataFeed::Preprocess(const std::string& filename) { fd_ = open(filename.c_str(), O_RDONLY); PADDLE_ENFORCE_NE( fd_, -1, platform::errors::Unavailable( "Fail to open file: %s in MultiSlotFileInstantDataFeed.", filename.c_str())); struct stat sb; fstat(fd_, &sb); end_ = static_cast(sb.st_size); buffer_ = reinterpret_cast(mmap(NULL, end_, PROT_READ, MAP_PRIVATE, fd_, 0)); PADDLE_ENFORCE_NE( buffer_, MAP_FAILED, platform::errors::Unavailable( "Memory map failed when create shared memory, error number is %s.", strerror(errno))); offset_ = 0; return true; } bool MultiSlotFileInstantDataFeed::Postprocess() { if (buffer_ != nullptr) { munmap(buffer_, end_); buffer_ = nullptr; } if (fd_ != -1) { close(fd_); fd_ = -1; end_ = 0; offset_ = 0; } return true; } bool MultiSlotFileInstantDataFeed::ParseOneMiniBatch() { if (offset_ == end_) { return false; } batch_size_ = 0; while (batch_size_ < default_batch_size_ && offset_ < end_) { for (size_t i = 0; i < use_slots_index_.size(); ++i) { int idx = use_slots_index_[i]; char type = all_slots_type_[i][0]; uint16_t num = *reinterpret_cast(buffer_ + offset_); PADDLE_ENFORCE_NE( num, 0, platform::errors::InvalidArgument( "The number of ids can not be zero, you need padding " "it in data generator; or if there is something wrong with " "the data, please check if the data contains unresolvable " "characters.")); offset_ += sizeof(uint16_t); if (idx != -1) { int inductive_size = multi_inductive_shape_index_[i].size(); if (UNLIKELY(batch_size_ == 0)) { ins_vec_[idx].Init(all_slots_type_[i], default_batch_size_ * num); ins_vec_[idx].InitOffset(default_batch_size_); uint64_t* inductive_shape = reinterpret_cast(buffer_ + offset_); for (int inductive_id = 0; inductive_id < inductive_size; ++inductive_id) { use_slots_shape_[i][multi_inductive_shape_index_[i][inductive_id]] = static_cast(*(inductive_shape + inductive_id)); } } num -= inductive_size; offset_ += sizeof(uint64_t) * inductive_size; if (type == 'f') { ins_vec_[idx].AppendValues( reinterpret_cast(buffer_ + offset_), num); offset_ += num * sizeof(float); } else if (type == 'u') { ins_vec_[idx].AppendValues( reinterpret_cast(buffer_ + offset_), num); offset_ += num * sizeof(uint64_t); } } else { if (type == 'f') { offset_ += num * sizeof(float); } else if (type == 'u') { offset_ += num * sizeof(uint64_t); } } } ++batch_size_; // OPTIMIZE: It is better to insert check codes between instances for format // checking } PADDLE_ENFORCE(batch_size_ == default_batch_size_ || offset_ == end_, platform::errors::InvalidArgument( "The batch size id not equal to default batch size, or " "the offset is not equal to end index." "The batch size is %d, default batcch size is %d, offset " "is %d, end index is %d.", batch_size_, default_batch_size_, offset_, end_)); return true; } #endif bool PaddleBoxDataFeed::Start() { #ifdef _LINUX int phase = GetCurrentPhase(); // join: 1, update: 0 this->CheckSetFileList(); if (enable_pv_merge_ && phase == 1) { // join phase : input_pv_channel to output_pv_channel if (output_pv_channel_->Size() == 0 && input_pv_channel_->Size() != 0) { std::vector data; input_pv_channel_->Read(data); output_pv_channel_->Write(std::move(data)); } } else { // input_channel to output if (output_channel_->Size() == 0 && input_channel_->Size() != 0) { std::vector data; input_channel_->Read(data); output_channel_->Write(std::move(data)); } } #endif this->finish_start_ = true; return true; } int PaddleBoxDataFeed::Next() { #ifdef _LINUX int phase = GetCurrentPhase(); // join: 1, update: 0 this->CheckStart(); if (enable_pv_merge_ && phase == 1) { // join phase : output_pv_channel to consume_pv_channel CHECK(output_pv_channel_ != nullptr); CHECK(consume_pv_channel_ != nullptr); VLOG(3) << "output_pv_channel_ size=" << output_pv_channel_->Size() << ", consume_pv_channel_ size=" << consume_pv_channel_->Size() << ", thread_id=" << thread_id_; int index = 0; PvInstance pv_instance; std::vector pv_vec; pv_vec.reserve(this->pv_batch_size_); while (index < this->pv_batch_size_) { if (output_pv_channel_->Size() == 0) { break; } output_pv_channel_->Get(pv_instance); pv_vec.push_back(pv_instance); ++index; consume_pv_channel_->Put(std::move(pv_instance)); } this->batch_size_ = index; VLOG(3) << "pv_batch_size_=" << this->batch_size_ << ", thread_id=" << thread_id_; if (this->batch_size_ != 0) { PutToFeedVec(pv_vec); } else { VLOG(3) << "finish reading, output_pv_channel_ size=" << output_pv_channel_->Size() << ", consume_pv_channel_ size=" << consume_pv_channel_->Size() << ", thread_id=" << thread_id_; } return this->batch_size_; } else { this->batch_size_ = MultiSlotInMemoryDataFeed::Next(); return this->batch_size_; } #else return 0; #endif } void PaddleBoxDataFeed::Init(const DataFeedDesc& data_feed_desc) { MultiSlotInMemoryDataFeed::Init(data_feed_desc); rank_offset_name_ = data_feed_desc.rank_offset(); pv_batch_size_ = data_feed_desc.pv_batch_size(); } void PaddleBoxDataFeed::GetRankOffset(const std::vector& pv_vec, int ins_number) { int index = 0; int max_rank = 3; // the value is setting int row = ins_number; int col = max_rank * 2 + 1; int pv_num = pv_vec.size(); std::vector rank_offset_mat(row * col, -1); rank_offset_mat.shrink_to_fit(); for (int i = 0; i < pv_num; i++) { auto pv_ins = pv_vec[i]; int ad_num = pv_ins->ads.size(); int index_start = index; for (int j = 0; j < ad_num; ++j) { auto ins = pv_ins->ads[j]; int rank = -1; if ((ins->cmatch == 222 || ins->cmatch == 223) && ins->rank <= static_cast(max_rank) && ins->rank != 0) { rank = ins->rank; } rank_offset_mat[index * col] = rank; if (rank > 0) { for (int k = 0; k < ad_num; ++k) { auto cur_ins = pv_ins->ads[k]; int fast_rank = -1; if ((cur_ins->cmatch == 222 || cur_ins->cmatch == 223) && cur_ins->rank <= static_cast(max_rank) && cur_ins->rank != 0) { fast_rank = cur_ins->rank; } if (fast_rank > 0) { int m = fast_rank - 1; rank_offset_mat[index * col + 2 * m + 1] = cur_ins->rank; rank_offset_mat[index * col + 2 * m + 2] = index_start + k; } } } index += 1; } } int* rank_offset = rank_offset_mat.data(); int* tensor_ptr = rank_offset_->mutable_data({row, col}, this->place_); CopyToFeedTensor(tensor_ptr, rank_offset, row * col * sizeof(int)); } void PaddleBoxDataFeed::AssignFeedVar(const Scope& scope) { MultiSlotInMemoryDataFeed::AssignFeedVar(scope); // set rank offset memory int phase = GetCurrentPhase(); // join: 1, update: 0 if (enable_pv_merge_ && phase == 1) { rank_offset_ = scope.FindVar(rank_offset_name_)->GetMutable(); } } void PaddleBoxDataFeed::PutToFeedVec(const std::vector& pv_vec) { #ifdef _LINUX int ins_number = 0; std::vector ins_vec; for (auto& pv : pv_vec) { ins_number += pv->ads.size(); for (auto ins : pv->ads) { ins_vec.push_back(ins); } } GetRankOffset(pv_vec, ins_number); PutToFeedVec(ins_vec); #endif } int PaddleBoxDataFeed::GetCurrentPhase() { #ifdef PADDLE_WITH_BOX_PS auto box_ptr = paddle::framework::BoxWrapper::GetInstance(); if (box_ptr->Mode() == 1) { // For AucRunner return 1; } else { return box_ptr->Phase(); } #else LOG(WARNING) << "It should be complied with BOX_PS..."; return current_phase_; #endif } void PaddleBoxDataFeed::PutToFeedVec(const std::vector& ins_vec) { #ifdef _LINUX for (size_t i = 0; i < batch_float_feasigns_.size(); ++i) { batch_float_feasigns_[i].clear(); batch_uint64_feasigns_[i].clear(); offset_[i].clear(); offset_[i].push_back(0); } ins_content_vec_.clear(); ins_content_vec_.reserve(ins_vec.size()); ins_id_vec_.clear(); ins_id_vec_.reserve(ins_vec.size()); for (size_t i = 0; i < ins_vec.size(); ++i) { auto r = ins_vec[i]; ins_id_vec_.push_back(r->ins_id_); ins_content_vec_.push_back(r->content_); for (auto& item : r->float_feasigns_) { batch_float_feasigns_[item.slot()].push_back(item.sign().float_feasign_); visit_[item.slot()] = true; } for (auto& item : r->uint64_feasigns_) { batch_uint64_feasigns_[item.slot()].push_back( item.sign().uint64_feasign_); visit_[item.slot()] = true; } for (size_t j = 0; j < use_slots_.size(); ++j) { const auto& type = all_slots_type_[j]; if (visit_[j]) { visit_[j] = false; } else { // fill slot value with default value 0 if (type[0] == 'f') { // float batch_float_feasigns_[j].push_back(0.0); } else if (type[0] == 'u') { // uint64 batch_uint64_feasigns_[j].push_back(0); } } // get offset of this ins in this slot if (type[0] == 'f') { // float offset_[j].push_back(batch_float_feasigns_[j].size()); } else if (type[0] == 'u') { // uint64 offset_[j].push_back(batch_uint64_feasigns_[j].size()); } } } for (size_t i = 0; i < use_slots_.size(); ++i) { if (feed_vec_[i] == nullptr) { continue; } int total_instance = offset_[i].back(); const auto& type = all_slots_type_[i]; if (type[0] == 'f') { // float float* feasign = batch_float_feasigns_[i].data(); float* tensor_ptr = feed_vec_[i]->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(float)); } else if (type[0] == 'u') { // uint64 // no uint64_t type in paddlepaddle uint64_t* feasign = batch_uint64_feasigns_[i].data(); int64_t* tensor_ptr = feed_vec_[i]->mutable_data( {total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(int64_t)); } auto& slot_offset = offset_[i]; LoD data_lod{slot_offset}; feed_vec_[i]->set_lod(data_lod); if (use_slots_is_dense_[i]) { if (inductive_shape_index_[i] != -1) { use_slots_shape_[i][inductive_shape_index_[i]] = total_instance / total_dims_without_inductive_[i]; } feed_vec_[i]->Resize(phi::make_ddim(use_slots_shape_[i])); } } #endif } template class InMemoryDataFeed; void SlotRecordInMemoryDataFeed::Init(const DataFeedDesc& data_feed_desc) { finish_init_ = false; finish_set_filelist_ = false; finish_start_ = false; PADDLE_ENFORCE(data_feed_desc.has_multi_slot_desc(), platform::errors::PreconditionNotMet( "Multi_slot_desc has not been set in data_feed_desc")); paddle::framework::MultiSlotDesc multi_slot_desc = data_feed_desc.multi_slot_desc(); SetBatchSize(data_feed_desc.batch_size()); size_t all_slot_num = multi_slot_desc.slots_size(); all_slots_.resize(all_slot_num); all_slots_info_.resize(all_slot_num); used_slots_info_.resize(all_slot_num); use_slot_size_ = 0; use_slots_.clear(); float_total_dims_size_ = 0; float_total_dims_without_inductives_.clear(); for (size_t i = 0; i < all_slot_num; ++i) { const auto& slot = multi_slot_desc.slots(i); all_slots_[i] = slot.name(); AllSlotInfo& all_slot = all_slots_info_[i]; all_slot.slot = slot.name(); all_slot.type = slot.type(); all_slot.used_idx = slot.is_used() ? use_slot_size_ : -1; all_slot.slot_value_idx = -1; if (slot.is_used()) { UsedSlotInfo& info = used_slots_info_[use_slot_size_]; info.idx = i; info.slot = slot.name(); info.type = slot.type(); info.dense = slot.is_dense(); info.total_dims_without_inductive = 1; info.inductive_shape_index = -1; // record float value and uint64_t value pos if (info.type[0] == 'u') { info.slot_value_idx = uint64_use_slot_size_; all_slot.slot_value_idx = uint64_use_slot_size_; ++uint64_use_slot_size_; } else if (info.type[0] == 'f') { info.slot_value_idx = float_use_slot_size_; all_slot.slot_value_idx = float_use_slot_size_; ++float_use_slot_size_; } use_slots_.push_back(slot.name()); if (slot.is_dense()) { for (int j = 0; j < slot.shape_size(); ++j) { if (slot.shape(j) > 0) { info.total_dims_without_inductive *= slot.shape(j); } if (slot.shape(j) == -1) { info.inductive_shape_index = j; } } } if (info.type[0] == 'f') { float_total_dims_without_inductives_.push_back( info.total_dims_without_inductive); float_total_dims_size_ += info.total_dims_without_inductive; } info.local_shape.clear(); for (int j = 0; j < slot.shape_size(); ++j) { info.local_shape.push_back(slot.shape(j)); } ++use_slot_size_; } } used_slots_info_.resize(use_slot_size_); feed_vec_.resize(used_slots_info_.size()); const int kEstimatedFeasignNumPerSlot = 5; // Magic Number for (size_t i = 0; i < all_slot_num; i++) { batch_float_feasigns_.push_back(std::vector()); batch_uint64_feasigns_.push_back(std::vector()); batch_float_feasigns_[i].reserve(default_batch_size_ * kEstimatedFeasignNumPerSlot); batch_uint64_feasigns_[i].reserve(default_batch_size_ * kEstimatedFeasignNumPerSlot); offset_.push_back(std::vector()); offset_[i].reserve(default_batch_size_ + 1); // Each lod info will prepend a zero } visit_.resize(all_slot_num, false); pipe_command_ = data_feed_desc.pipe_command(); finish_init_ = true; input_type_ = data_feed_desc.input_type(); size_t pos = pipe_command_.find(".so"); if (pos != std::string::npos) { pos = pipe_command_.rfind('|'); if (pos == std::string::npos) { so_parser_name_ = pipe_command_; pipe_command_.clear(); } else { so_parser_name_ = pipe_command_.substr(pos + 1); pipe_command_ = pipe_command_.substr(0, pos); } so_parser_name_ = paddle::string::erase_spaces(so_parser_name_); } else { so_parser_name_.clear(); } #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) gpu_graph_data_generator_.SetConfig(data_feed_desc); #endif if (gpu_graph_mode_) { train_mode_ = true; } else { train_mode_ = data_feed_desc.graph_config().gpu_graph_training(); } } #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) void SlotRecordInMemoryDataFeed::InitGraphResource() { #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) gpu_graph_data_generator_.AllocResource(thread_id_, feed_vec_); #endif } void SlotRecordInMemoryDataFeed::InitGraphTrainResource() { #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) gpu_graph_data_generator_.AllocTrainResource(thread_id_); #endif } #endif void SlotRecordInMemoryDataFeed::LoadIntoMemory() { VLOG(3) << "SlotRecord LoadIntoMemory() begin, thread_id=" << thread_id_; if (!so_parser_name_.empty()) { LoadIntoMemoryByLib(); } else { LoadIntoMemoryByCommand(); } } void SlotRecordInMemoryDataFeed::LoadIntoMemoryByLib(void) { if (true) { // user defined file format analysis LoadIntoMemoryByFile(); } else { LoadIntoMemoryByLine(); } } void SlotRecordInMemoryDataFeed::LoadIntoMemoryByFile(void) { #if (defined _LINUX) && (defined PADDLE_WITH_HETERPS) && \ (defined PADDLE_WITH_PSLIB) paddle::framework::CustomParser* parser = global_dlmanager_pool().Load(so_parser_name_, all_slots_info_); CHECK(parser != nullptr); // get slotrecord object auto pull_record_func = [this](std::vector& record_vec, int max_fetch_num, int offset) { if (offset > 0) { input_channel_->WriteMove(offset, &record_vec[0]); if (max_fetch_num > 0) { SlotRecordPool().get(&record_vec[0], offset); } else { // free all max_fetch_num = static_cast(record_vec.size()); if (max_fetch_num > offset) { SlotRecordPool().put(&record_vec[offset], (max_fetch_num - offset)); } } } else if (max_fetch_num > 0) { SlotRecordPool().get(&record_vec, max_fetch_num); } else { SlotRecordPool().put(&record_vec); } }; std::string filename; while (this->PickOneFile(&filename)) { VLOG(3) << "PickOneFile, filename=" << filename << ", thread_id=" << thread_id_; platform::Timer timeline; timeline.Start(); int lines = 0; bool is_ok = true; auto ps_gpu_ptr = PSGPUWrapper::GetInstance(); do { if (ps_gpu_ptr->UseAfsApi()) { auto afs_reader = ps_gpu_ptr->OpenReader(filename); is_ok = parser->ParseFileInstance( [this, afs_reader](char* buf, int len) { return afs_reader->read(buf, len); }, pull_record_func, lines); } else { int err_no = 0; this->fp_ = fs_open_read(filename, &err_no, this->pipe_command_, true); CHECK(this->fp_ != nullptr); __fsetlocking(&*(this->fp_), FSETLOCKING_BYCALLER); is_ok = parser->ParseFileInstance( [this](char* buf, int len) { return fread(buf, sizeof(char), len, this->fp_.get()); }, pull_record_func, lines); if (!is_ok) { LOG(WARNING) << "parser error, filename=" << filename << ", lines=" << lines; } } } while (!is_ok); timeline.Pause(); VLOG(3) << "LoadIntoMemoryByLib() read all file, file=" << filename << ", cost time=" << timeline.ElapsedSec() << " seconds, thread_id=" << thread_id_ << ", lines=" << lines; } #endif } void SlotRecordInMemoryDataFeed::LoadIntoMemoryByLine(void) { #ifdef _LINUX paddle::framework::CustomParser* parser = global_dlmanager_pool().Load(so_parser_name_, all_slots_info_); std::string filename; BufferedLineFileReader line_reader; line_reader.set_sample_rate(sample_rate_); BufferedLineFileReader::LineFunc line_func = nullptr; while (this->PickOneFile(&filename)) { VLOG(3) << "PickOneFile, filename=" << filename << ", thread_id=" << thread_id_; std::vector record_vec; platform::Timer timeline; timeline.Start(); int offset = 0; int old_offset = 0; SlotRecordPool().get(&record_vec, OBJPOOL_BLOCK_SIZE); // get slotrecord object function auto record_func = [this, &offset, &record_vec, &old_offset]( std::vector& vec, int num) { vec.resize(num); if (offset + num > OBJPOOL_BLOCK_SIZE) { input_channel_->WriteMove(offset, &record_vec[0]); SlotRecordPool().get(&record_vec[0], offset); record_vec.resize(OBJPOOL_BLOCK_SIZE); offset = 0; old_offset = 0; } for (int i = 0; i < num; ++i) { auto& ins = record_vec[offset + i]; ins->reset(); vec[i] = ins; } offset = offset + num; }; line_func = [this, &parser, &record_vec, &offset, &filename, &record_func, &old_offset](const std::string& line) { old_offset = offset; if (!parser->ParseOneInstance(line, record_func)) { offset = old_offset; LOG(WARNING) << "read file:[" << filename << "] item error, line:[" << line << "]"; return false; } if (offset >= OBJPOOL_BLOCK_SIZE) { input_channel_->Write(std::move(record_vec)); record_vec.clear(); SlotRecordPool().get(&record_vec, OBJPOOL_BLOCK_SIZE); offset = 0; } return true; }; int lines = 0; do { int err_no = 0; this->fp_ = fs_open_read(filename, &err_no, this->pipe_command_, true); CHECK(this->fp_ != nullptr); __fsetlocking(&*(this->fp_), FSETLOCKING_BYCALLER); lines = line_reader.read_file(this->fp_.get(), line_func, lines); } while (line_reader.is_error()); if (offset > 0) { input_channel_->WriteMove(offset, &record_vec[0]); if (offset < OBJPOOL_BLOCK_SIZE) { SlotRecordPool().put(&record_vec[offset], (OBJPOOL_BLOCK_SIZE - offset)); } } else { SlotRecordPool().put(&record_vec); } record_vec.clear(); record_vec.shrink_to_fit(); timeline.Pause(); VLOG(3) << "LoadIntoMemoryByLib() read all lines, file=" << filename << ", cost time=" << timeline.ElapsedSec() << " seconds, thread_id=" << thread_id_ << ", lines=" << lines << ", sample lines=" << line_reader.get_sample_line() << ", filesize=" << line_reader.file_size() / 1024.0 / 1024.0 << "MB"; } VLOG(3) << "LoadIntoMemoryByLib() end, thread_id=" << thread_id_ << ", total size: " << line_reader.file_size(); #endif } void SlotRecordInMemoryDataFeed::LoadIntoMemoryByCommand(void) { #ifdef _LINUX std::string filename; BufferedLineFileReader line_reader; line_reader.set_sample_rate(sample_rate_); while (this->PickOneFile(&filename)) { VLOG(3) << "PickOneFile, filename=" << filename << ", thread_id=" << thread_id_; int lines = 0; std::vector record_vec; platform::Timer timeline; timeline.Start(); SlotRecordPool().get(&record_vec, OBJPOOL_BLOCK_SIZE); int offset = 0; do { int err_no = 0; this->fp_ = fs_open_read(filename, &err_no, this->pipe_command_, true); CHECK(this->fp_ != nullptr); __fsetlocking(&*(this->fp_), FSETLOCKING_BYCALLER); lines = line_reader.read_file( this->fp_.get(), [this, &record_vec, &offset, &filename](const std::string& line) { if (ParseOneInstance(line, &record_vec[offset])) { ++offset; } else { LOG(WARNING) << "read file:[" << filename << "] item error, line:[" << line << "]"; return false; } if (offset >= OBJPOOL_BLOCK_SIZE) { input_channel_->Write(std::move(record_vec)); record_vec.clear(); SlotRecordPool().get(&record_vec, OBJPOOL_BLOCK_SIZE); offset = 0; } return true; }, lines); } while (line_reader.is_error()); if (offset > 0) { input_channel_->WriteMove(offset, &record_vec[0]); if (offset < OBJPOOL_BLOCK_SIZE) { SlotRecordPool().put(&record_vec[offset], (OBJPOOL_BLOCK_SIZE - offset)); } } else { SlotRecordPool().put(&record_vec); } record_vec.clear(); record_vec.shrink_to_fit(); timeline.Pause(); VLOG(3) << "LoadIntoMemory() read all lines, file=" << filename << ", lines=" << lines << ", sample lines=" << line_reader.get_sample_line() << ", cost time=" << timeline.ElapsedSec() << " seconds, thread_id=" << thread_id_; } VLOG(3) << "LoadIntoMemory() end, thread_id=" << thread_id_ << ", total size: " << line_reader.file_size(); #endif } static void parser_log_key(const std::string& log_key, uint64_t* search_id, uint32_t* cmatch, uint32_t* rank) { std::string searchid_str = log_key.substr(16, 16); *search_id = static_cast(strtoull(searchid_str.c_str(), NULL, 16)); std::string cmatch_str = log_key.substr(11, 3); *cmatch = static_cast(strtoul(cmatch_str.c_str(), NULL, 16)); std::string rank_str = log_key.substr(14, 2); *rank = static_cast(strtoul(rank_str.c_str(), NULL, 16)); } bool SlotRecordInMemoryDataFeed::ParseOneInstance(const std::string& line, SlotRecord* ins) { SlotRecord& rec = (*ins); // parse line const char* str = line.c_str(); char* endptr = const_cast(str); int pos = 0; thread_local std::vector> slot_float_feasigns; thread_local std::vector> slot_uint64_feasigns; slot_float_feasigns.resize(float_use_slot_size_); slot_uint64_feasigns.resize(uint64_use_slot_size_); if (parse_ins_id_) { int num = strtol(&str[pos], &endptr, 10); CHECK(num == 1); // NOLINT pos = endptr - str + 1; size_t len = 0; while (str[pos + len] != ' ') { ++len; } rec->ins_id_ = std::string(str + pos, len); pos += len + 1; } if (parse_logkey_) { int num = strtol(&str[pos], &endptr, 10); CHECK(num == 1); // NOLINT pos = endptr - str + 1; size_t len = 0; while (str[pos + len] != ' ') { ++len; } // parse_logkey std::string log_key = std::string(str + pos, len); uint64_t search_id; uint32_t cmatch; uint32_t rank; parser_log_key(log_key, &search_id, &cmatch, &rank); rec->ins_id_ = log_key; rec->search_id = search_id; rec->cmatch = cmatch; rec->rank = rank; pos += len + 1; } int float_total_slot_num = 0; int uint64_total_slot_num = 0; for (size_t i = 0; i < all_slots_info_.size(); ++i) { auto& info = all_slots_info_[i]; int num = strtol(&str[pos], &endptr, 10); PADDLE_ENFORCE(num, "The number of ids can not be zero, you need padding " "it in data generator; or if there is something wrong with " "the data, please check if the data contains unresolvable " "characters.\nplease check this error line: %s", str); if (info.used_idx != -1) { if (info.type[0] == 'f') { // float auto& slot_fea = slot_float_feasigns[info.slot_value_idx]; slot_fea.clear(); for (int j = 0; j < num; ++j) { float feasign = strtof(endptr, &endptr); if (fabs(feasign) < 1e-6 && !used_slots_info_[info.used_idx].dense) { continue; } slot_fea.push_back(feasign); ++float_total_slot_num; } } else if (info.type[0] == 'u') { // uint64 auto& slot_fea = slot_uint64_feasigns[info.slot_value_idx]; slot_fea.clear(); for (int j = 0; j < num; ++j) { uint64_t feasign = static_cast(strtoull(endptr, &endptr, 10)); slot_fea.push_back(feasign); ++uint64_total_slot_num; } } pos = endptr - str; } else { for (int j = 0; j <= num; ++j) { // pos = line.find_first_of(' ', pos + 1); while (line[pos + 1] != ' ') { pos++; } } } } rec->slot_float_feasigns_.add_slot_feasigns(slot_float_feasigns, float_total_slot_num); rec->slot_uint64_feasigns_.add_slot_feasigns(slot_uint64_feasigns, uint64_total_slot_num); return (uint64_total_slot_num > 0); } void SlotRecordInMemoryDataFeed::AssignFeedVar(const Scope& scope) { CheckInit(); for (int i = 0; i < use_slot_size_; ++i) { feed_vec_[i] = scope.FindVar(used_slots_info_[i].slot)->GetMutable(); } } void SlotRecordInMemoryDataFeed::PutToFeedVec(const SlotRecord* ins_vec, int num) { #if defined(PADDLE_WITH_CUDA) && defined(PADDLE_WITH_HETERPS) paddle::platform::SetDeviceId(place_.GetDeviceId()); pack_->pack_instance(ins_vec, num); BuildSlotBatchGPU(pack_->ins_num()); #else for (int j = 0; j < use_slot_size_; ++j) { auto& feed = feed_vec_[j]; if (feed == nullptr) { continue; } auto& slot_offset = offset_[j]; slot_offset.clear(); slot_offset.reserve(num + 1); slot_offset.push_back(0); int total_instance = 0; auto& info = used_slots_info_[j]; // fill slot value with default value 0 if (info.type[0] == 'f') { // float auto& batch_fea = batch_float_feasigns_[j]; batch_fea.clear(); for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; size_t fea_num = 0; float* slot_values = r->slot_float_feasigns_.get_values(info.slot_value_idx, &fea_num); batch_fea.resize(total_instance + fea_num); memcpy( &batch_fea[total_instance], slot_values, sizeof(float) * fea_num); total_instance += fea_num; slot_offset.push_back(total_instance); } float* feasign = batch_fea.data(); float* tensor_ptr = feed->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(float)); } else if (info.type[0] == 'u') { // uint64 auto& batch_fea = batch_uint64_feasigns_[j]; batch_fea.clear(); for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; size_t fea_num = 0; uint64_t* slot_values = r->slot_uint64_feasigns_.get_values(info.slot_value_idx, &fea_num); if (fea_num > 0) { batch_fea.resize(total_instance + fea_num); memcpy(&batch_fea[total_instance], slot_values, sizeof(uint64_t) * fea_num); total_instance += fea_num; } if (fea_num == 0) { batch_fea.resize(total_instance + fea_num); batch_fea[total_instance] = 0; total_instance += 1; } slot_offset.push_back(total_instance); } // no uint64_t type in paddlepaddle uint64_t* feasign = batch_fea.data(); int64_t* tensor_ptr = feed->mutable_data({total_instance, 1}, this->place_); CopyToFeedTensor(tensor_ptr, feasign, total_instance * sizeof(int64_t)); } if (info.dense) { if (info.inductive_shape_index != -1) { info.local_shape[info.inductive_shape_index] = total_instance / info.total_dims_without_inductive; } feed->Resize(phi::make_ddim(info.local_shape)); } else { LoD data_lod{slot_offset}; feed_vec_[j]->set_lod(data_lod); } } #endif } void SlotRecordInMemoryDataFeed::ExpandSlotRecord(SlotRecord* rec) { SlotRecord& ins = (*rec); if (ins->slot_float_feasigns_.slot_offsets.empty()) { return; } size_t total_value_size = ins->slot_float_feasigns_.slot_values.size(); if (float_total_dims_size_ == total_value_size) { return; } int float_slot_num = static_cast(float_total_dims_without_inductives_.size()); CHECK(float_slot_num == float_use_slot_size_); std::vector old_values; std::vector old_offsets; old_values.swap(ins->slot_float_feasigns_.slot_values); old_offsets.swap(ins->slot_float_feasigns_.slot_offsets); ins->slot_float_feasigns_.slot_values.resize(float_total_dims_size_); ins->slot_float_feasigns_.slot_offsets.assign(float_slot_num + 1, 0); auto& slot_offsets = ins->slot_float_feasigns_.slot_offsets; auto& slot_values = ins->slot_float_feasigns_.slot_values; uint32_t offset = 0; int num = 0; uint32_t old_off = 0; int dim = 0; for (int i = 0; i < float_slot_num; ++i) { dim = float_total_dims_without_inductives_[i]; old_off = old_offsets[i]; num = static_cast(old_offsets[i + 1] - old_off); if (num == 0) { // fill slot value with default value 0 for (int k = 0; k < dim; ++k) { slot_values[k + offset] = 0.0; } } else { if (num == dim) { memcpy(&slot_values[offset], &old_values[old_off], dim * sizeof(float)); } else { // position fea // record position index need fix values int pos_idx = static_cast(old_values[old_off]); for (int k = 0; k < dim; ++k) { if (k == pos_idx) { slot_values[k + offset] = 1.0; } else { slot_values[k + offset] = 0.0; } } } } slot_offsets[i] = offset; offset += dim; } slot_offsets[float_slot_num] = offset; CHECK(float_total_dims_size_ == static_cast(offset)); } bool SlotRecordInMemoryDataFeed::Start() { VLOG(4) << "entering SlotRecordInMemoryDataFeed::Start"; #ifdef _LINUX this->CheckSetFileList(); if (input_channel_->Size() != 0) { std::vector data; input_channel_->Read(data); } #endif if (batch_offsets_.size() > 0) { VLOG(3) << "batch_size offsets: " << batch_offsets_.size(); enable_heterps_ = true; this->offset_index_ = 0; } this->finish_start_ = true; #if defined(PADDLE_WITH_CUDA) && defined(PADDLE_WITH_HETERPS) CHECK(paddle::platform::is_gpu_place(this->place_)); pack_ = BatchGpuPackMgr().get(this->GetPlace(), used_slots_info_); #endif #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) gpu_graph_data_generator_.SetFeedVec(feed_vec_); #endif return true; } int SlotRecordInMemoryDataFeed::Next() { #ifdef _LINUX this->CheckStart(); if (!gpu_graph_mode_) { VLOG(3) << "enable heter next: " << offset_index_ << " batch_offsets: " << batch_offsets_.size(); if (offset_index_ >= batch_offsets_.size()) { VLOG(3) << "offset_index: " << offset_index_ << " batch_offsets: " << batch_offsets_.size(); return 0; } auto& batch = batch_offsets_[offset_index_++]; this->batch_size_ = batch.second; VLOG(3) << "batch_size_=" << this->batch_size_ << ", thread_id=" << thread_id_; if (this->batch_size_ != 0) { PutToFeedVec(&records_[batch.first], this->batch_size_); } else { VLOG(3) << "finish reading for heterps, batch size zero, thread_id=" << thread_id_; } VLOG(3) << "enable heter next: " << offset_index_ << " batch_offsets: " << batch_offsets_.size() << " baych_size: " << this->batch_size_; } else { VLOG(3) << "datafeed in gpu graph mode"; #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) this->batch_size_ = gpu_graph_data_generator_.GenerateBatch(); #endif } return this->batch_size_; #else return 0; #endif } #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) void SlotRecordInMemoryDataFeed::DoWalkandSage() { gpu_graph_data_generator_.DoWalkandSage(); } #endif void SlotRecordInMemoryDataFeed::DumpWalkPath(std::string dump_path, size_t dump_rate) { VLOG(3) << "INTO SlotRecordInMemoryDataFeed::DumpWalkPath"; #if defined(PADDLE_WITH_GPU_GRAPH) && defined(PADDLE_WITH_HETERPS) std::string path = string::format_string("%s/part-%03d", dump_path.c_str(), thread_id_); gpu_graph_data_generator_.DumpWalkPath(path, dump_rate); #endif } #if defined(PADDLE_WITH_CUDA) && defined(PADDLE_WITH_HETERPS) void SlotRecordInMemoryDataFeed::BuildSlotBatchGPU(const int ins_num) { int offset_cols_size = (ins_num + 1); size_t slot_total_num = (use_slot_size_ * offset_cols_size); pack_->resize_gpu_slot_offsets(slot_total_num * sizeof(size_t)); auto& value = pack_->value(); const UsedSlotGpuType* used_slot_gpu_types = static_cast(pack_->get_gpu_slots()); FillSlotValueOffset(ins_num, use_slot_size_, reinterpret_cast(pack_->gpu_slot_offsets()), value.d_uint64_offset.data(), uint64_use_slot_size_, value.d_float_offset.data(), float_use_slot_size_, used_slot_gpu_types); size_t* d_slot_offsets = reinterpret_cast(pack_->gpu_slot_offsets()); HostBuffer& offsets = pack_->offsets(); offsets.resize(slot_total_num); HostBuffer& h_tensor_ptrs = pack_->h_tensor_ptrs(); h_tensor_ptrs.resize(use_slot_size_); // alloc gpu memory pack_->resize_tensor(); phi::DenseTensor& float_tensor = pack_->float_tensor(); phi::DenseTensor& uint64_tensor = pack_->uint64_tensor(); int64_t float_offset = 0; int64_t uint64_offset = 0; // copy index CUDA_CHECK(cudaMemcpy(offsets.data(), d_slot_offsets, slot_total_num * sizeof(size_t), cudaMemcpyDeviceToHost)); for (int j = 0; j < use_slot_size_; ++j) { auto& feed = feed_vec_[j]; if (feed == nullptr) { h_tensor_ptrs[j] = nullptr; continue; } size_t* off_start_ptr = &offsets[j * offset_cols_size]; int total_instance = static_cast(off_start_ptr[offset_cols_size - 1]); CHECK(total_instance >= 0) << "slot idx:" << j << ", total instance:" << total_instance; auto& info = used_slots_info_[j]; // fill slot value with default value 0 if (info.type[0] == 'f') { // float if (total_instance > 0) { feed->ShareDataWith(float_tensor.Slice( static_cast(float_offset), static_cast(float_offset + total_instance))); feed->Resize({total_instance, 1}); float_offset += total_instance; h_tensor_ptrs[j] = feed->mutable_data(this->place_); } else { h_tensor_ptrs[j] = feed->mutable_data({total_instance, 1}, this->place_); } } else if (info.type[0] == 'u') { // uint64 if (total_instance > 0) { feed->ShareDataWith(uint64_tensor.Slice( static_cast(uint64_offset), static_cast(uint64_offset + total_instance))); feed->Resize({total_instance, 1}); uint64_offset += total_instance; h_tensor_ptrs[j] = feed->mutable_data(this->place_); } else { h_tensor_ptrs[j] = feed->mutable_data({total_instance, 1}, this->place_); } } if (info.dense) { if (info.inductive_shape_index != -1) { info.local_shape[info.inductive_shape_index] = total_instance / info.total_dims_without_inductive; } feed->Resize(phi::make_ddim(info.local_shape)); } else { LoD& lod = (*feed->mutable_lod()); lod.resize(1); lod[0].resize(offset_cols_size); paddle::framework::MixVector mixv_lod(&lod[0]); memcpy(mixv_lod.MutableData(platform::CPUPlace()), off_start_ptr, offset_cols_size * sizeof(size_t)); } } void** dest_gpu_p = reinterpret_cast(pack_->slot_buf_ptr()); CUDA_CHECK(cudaMemcpy(dest_gpu_p, h_tensor_ptrs.data(), use_slot_size_ * sizeof(void*), cudaMemcpyHostToDevice)); CopyForTensor(ins_num, use_slot_size_, dest_gpu_p, (const size_t*)pack_->gpu_slot_offsets(), (const uint64_t*)value.d_uint64_keys.data(), (const int*)value.d_uint64_offset.data(), (const int*)value.d_uint64_lens.data(), uint64_use_slot_size_, (const float*)value.d_float_keys.data(), (const int*)value.d_float_offset.data(), (const int*)value.d_float_lens.data(), float_use_slot_size_, used_slot_gpu_types); } MiniBatchGpuPack::MiniBatchGpuPack(const paddle::platform::Place& place, const std::vector& infos) { place_ = place; stream_ = dynamic_cast( platform::DeviceContextPool::Instance().Get(place)) ->stream(); ins_num_ = 0; pv_num_ = 0; used_float_num_ = 0; used_uint64_num_ = 0; used_slot_size_ = static_cast(infos.size()); for (int i = 0; i < used_slot_size_; ++i) { auto& info = infos[i]; if (info.type[0] == 'u') { gpu_used_slots_.push_back({1, info.slot_value_idx}); ++used_uint64_num_; } else { gpu_used_slots_.push_back({0, info.slot_value_idx}); ++used_float_num_; } } copy_host2device(&gpu_slots_, gpu_used_slots_.data(), gpu_used_slots_.size()); slot_buf_ptr_ = memory::AllocShared(place_, used_slot_size_ * sizeof(void*)); int device_id = place_.GetDeviceId(); VLOG(3) << "begin get batch pack device id: " << device_id; // sync CUDA_CHECK(cudaStreamSynchronize(stream_)); } MiniBatchGpuPack::~MiniBatchGpuPack() {} void MiniBatchGpuPack::reset(const paddle::platform::Place& place) { place_ = place; stream_ = dynamic_cast( platform::DeviceContextPool::Instance().Get(place)) ->stream(); ins_num_ = 0; pv_num_ = 0; } void MiniBatchGpuPack::pack_all_data(const SlotRecord* ins_vec, int num) { int uint64_total_num = 0; int float_total_num = 0; buf_.h_uint64_lens.resize(num + 1); buf_.h_uint64_lens[0] = 0; buf_.h_float_lens.resize(num + 1); buf_.h_float_lens[0] = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; uint64_total_num += r->slot_uint64_feasigns_.slot_values.size(); buf_.h_uint64_lens[i + 1] = uint64_total_num; float_total_num += r->slot_float_feasigns_.slot_values.size(); buf_.h_float_lens[i + 1] = float_total_num; } int uint64_cols = (used_uint64_num_ + 1); buf_.h_uint64_offset.resize(uint64_cols * num); buf_.h_uint64_keys.resize(uint64_total_num); int float_cols = (used_float_num_ + 1); buf_.h_float_offset.resize(float_cols * num); buf_.h_float_keys.resize(float_total_num); size_t fea_num = 0; uint64_total_num = 0; float_total_num = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; auto& uint64_feasigns = r->slot_uint64_feasigns_; fea_num = uint64_feasigns.slot_values.size(); if (fea_num > 0) { memcpy(&buf_.h_uint64_keys[uint64_total_num], uint64_feasigns.slot_values.data(), fea_num * sizeof(uint64_t)); } uint64_total_num += fea_num; // copy uint64 offset memcpy(&buf_.h_uint64_offset[i * uint64_cols], uint64_feasigns.slot_offsets.data(), sizeof(int) * uint64_cols); auto& float_feasigns = r->slot_float_feasigns_; fea_num = float_feasigns.slot_values.size(); memcpy(&buf_.h_float_keys[float_total_num], float_feasigns.slot_values.data(), fea_num * sizeof(float)); float_total_num += fea_num; // copy float offset memcpy(&buf_.h_float_offset[i * float_cols], float_feasigns.slot_offsets.data(), sizeof(int) * float_cols); } CHECK(uint64_total_num == static_cast(buf_.h_uint64_lens.back())) << "uint64 value length error"; CHECK(float_total_num == static_cast(buf_.h_float_lens.back())) << "float value length error"; } void MiniBatchGpuPack::pack_uint64_data(const SlotRecord* ins_vec, int num) { int uint64_total_num = 0; buf_.h_float_lens.clear(); buf_.h_float_keys.clear(); buf_.h_float_offset.clear(); buf_.h_uint64_lens.resize(num + 1); buf_.h_uint64_lens[0] = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; uint64_total_num += r->slot_uint64_feasigns_.slot_values.size(); buf_.h_uint64_lens[i + 1] = uint64_total_num; } int uint64_cols = (used_uint64_num_ + 1); buf_.h_uint64_offset.resize(uint64_cols * num); buf_.h_uint64_keys.resize(uint64_total_num); size_t fea_num = 0; uint64_total_num = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; auto& uint64_feasigns = r->slot_uint64_feasigns_; fea_num = uint64_feasigns.slot_values.size(); if (fea_num > 0) { memcpy(&buf_.h_uint64_keys[uint64_total_num], uint64_feasigns.slot_values.data(), fea_num * sizeof(uint64_t)); } uint64_total_num += fea_num; // copy uint64 offset memcpy(&buf_.h_uint64_offset[i * uint64_cols], uint64_feasigns.slot_offsets.data(), sizeof(int) * uint64_cols); } CHECK(uint64_total_num == static_cast(buf_.h_uint64_lens.back())) << "uint64 value length error"; } void MiniBatchGpuPack::pack_float_data(const SlotRecord* ins_vec, int num) { int float_total_num = 0; buf_.h_uint64_lens.clear(); buf_.h_uint64_offset.clear(); buf_.h_uint64_keys.clear(); buf_.h_float_lens.resize(num + 1); buf_.h_float_lens[0] = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; float_total_num += r->slot_float_feasigns_.slot_values.size(); buf_.h_float_lens[i + 1] = float_total_num; } int float_cols = (used_float_num_ + 1); buf_.h_float_offset.resize(float_cols * num); buf_.h_float_keys.resize(float_total_num); size_t fea_num = 0; float_total_num = 0; for (int i = 0; i < num; ++i) { auto r = ins_vec[i]; auto& float_feasigns = r->slot_float_feasigns_; fea_num = float_feasigns.slot_values.size(); memcpy(&buf_.h_float_keys[float_total_num], float_feasigns.slot_values.data(), fea_num * sizeof(float)); float_total_num += fea_num; // copy float offset memcpy(&buf_.h_float_offset[i * float_cols], float_feasigns.slot_offsets.data(), sizeof(int) * float_cols); } CHECK(float_total_num == static_cast(buf_.h_float_lens.back())) << "float value length error"; } void MiniBatchGpuPack::pack_instance(const SlotRecord* ins_vec, int num) { ins_num_ = num; batch_ins_ = ins_vec; CHECK(used_uint64_num_ > 0 || used_float_num_ > 0); // uint64 and float if (used_uint64_num_ > 0 && used_float_num_ > 0) { pack_all_data(ins_vec, num); } else if (used_uint64_num_ > 0) { // uint64 pack_uint64_data(ins_vec, num); } else { // only float pack_float_data(ins_vec, num); } // to gpu transfer_to_gpu(); } void MiniBatchGpuPack::transfer_to_gpu(void) { copy_host2device(&value_.d_uint64_lens, buf_.h_uint64_lens); copy_host2device(&value_.d_uint64_keys, buf_.h_uint64_keys); copy_host2device(&value_.d_uint64_offset, buf_.h_uint64_offset); copy_host2device(&value_.d_float_lens, buf_.h_float_lens); copy_host2device(&value_.d_float_keys, buf_.h_float_keys); copy_host2device(&value_.d_float_offset, buf_.h_float_offset); CUDA_CHECK(cudaStreamSynchronize(stream_)); } #endif } // namespace framework } // namespace paddle