<p>This is the complete list of members for <aclass="el"href="classop_1_1_datum_producer.html">op::DatumProducer< TDatumsNoPtr ></a>, including all inherited members.</p>
<tr><tdclass="entry"><aclass="el"href="classop_1_1_datum_producer.html#a627c16307864f02251eeb77f2320052f">DatumProducer</a>(const std::shared_ptr< Producer >&producerSharedPtr, const unsigned long long frameFirst=0, const unsigned long long frameLast=std::numeric_limits< unsigned long long >::max(), const std::shared_ptr< std::pair< std::atomic< bool >, std::atomic< int >>>&videoSeekSharedPtr=nullptr)</td><tdclass="entry"><aclass="el"href="classop_1_1_datum_producer.html">op::DatumProducer< TDatumsNoPtr ></a></td><tdclass="entry"><spanclass="mlabel">explicit</span></td></tr>
<tr><tdclass="entry"><aclass="el"href="classop_1_1_datum_producer.html#a308c34ca52e7d6db2d85e27a348a15bd">DatumProducer</a>(const std::shared_ptr< Producer >&producerSharedPtr, const unsigned long long frameFirst=0, const unsigned long long frameStep=1, const unsigned long long frameLast=std::numeric_limits< unsigned long long >::max(), const std::shared_ptr< std::pair< std::atomic< bool >, std::atomic< int >>>&videoSeekSharedPtr=nullptr)</td><tdclass="entry"><aclass="el"href="classop_1_1_datum_producer.html">op::DatumProducer< TDatumsNoPtr ></a></td><tdclass="entry"><spanclass="mlabel">explicit</span></td></tr>
<trclass="memitem:a627c16307864f02251eeb77f2320052f"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="classop_1_1_datum_producer.html#a627c16307864f02251eeb77f2320052f">DatumProducer</a> (const std::shared_ptr<<aclass="el"href="classop_1_1_producer.html">Producer</a>>&producerSharedPtr, const unsigned long long frameFirst=0, const unsigned long long frameLast=std::numeric_limits< unsigned long long >::max(), const std::shared_ptr< std::pair< std::atomic< bool >, std::atomic< int >>>&videoSeekSharedPtr=nullptr)</td></tr>
<trclass="memitem:a308c34ca52e7d6db2d85e27a348a15bd"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="classop_1_1_datum_producer.html#a308c34ca52e7d6db2d85e27a348a15bd">DatumProducer</a> (const std::shared_ptr<<aclass="el"href="classop_1_1_producer.html">Producer</a>>&producerSharedPtr, const unsigned long long frameFirst=0, const unsigned long long frameStep=1, const unsigned long long frameLast=std::numeric_limits< unsigned long long >::max(), const std::shared_ptr< std::pair< std::atomic< bool >, std::atomic< int >>>&videoSeekSharedPtr=nullptr)</td></tr>
<trclass="memitem:a1433eaf1c5eb42e406e76bc6f8e517c3"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="flags_8hpp.html#a1433eaf1c5eb42e406e76bc6f8e517c3">DEFINE_uint64</a> (frame_first, 0,"Start on desired frame number. Indexes are 0-based, i.e. the first frame has index 0.")</td></tr>
<trclass="memitem:a22572531e5a4896c510639ac57cf522c"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="flags_8hpp.html#a22572531e5a4896c510639ac57cf522c">DEFINE_uint64</a> (frame_step, 1,"Step or gap between processed frames. E.g., `--frame_step 5` would read and process frames"" 0, 5, 10, etc..")</td></tr>
<trclass="memitem:ac8fef8bb0234286e74b75214a750d674"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="flags_8hpp.html#ac8fef8bb0234286e74b75214a750d674">DEFINE_uint64</a> (frame_last,-1,"Finish on desired frame number. Select -1 to disable. Indexes are 0-based, e.g. if set to"" 10, it will process 11 frames (0-10).")</td></tr>
<trclass="memitem:a64c46584d79ef0b947ecd0a98b282839"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="flags_8hpp.html#a64c46584d79ef0b947ecd0a98b282839">DEFINE_bool</a> (frame_flip, false,"Flip/mirror each frame (e.g. for real time webcam demonstrations).")</td></tr>
['define_5fint32',['DEFINE_int32',['../flags_8hpp.html#a73ee51843ead02ee6358fe39dcbeffde',1,'DEFINE_int32(logging_level, 3,"The logging level. Integer in the range [0, 255]. 0 will output any log() message, while"" 255 will not output any. Current OpenPose library messages are in the range 0-4: 1 for"" low priority messages and 4 for important ones."): flags.hpp'],['../flags_8hpp.html#a8e9de971b409cfe7fdded7f0d47c502d',1,'DEFINE_int32(profile_speed, 1000,"If PROFILER_ENABLED was set in CMake or Makefile.config files, OpenPose will show some"" runtime statistics at this frame number."): flags.hpp'],['../flags_8hpp.html#a10efaeb1ea3a8478388cc7d0bfd4e59e',1,'DEFINE_int32(camera,-1,"The camera index for cv::VideoCapture. Integer in the range [0, 9]. Select a negative"" number (by default), to auto-detect and open the first available camera."): flags.hpp'],['../flags_8hpp.html#ab41c02abe3634f0db65123ecda964a31',1,'DEFINE_int32(flir_camera_index,-1,"Select -1 (default) to run on all detected flir cameras at once. Otherwise, select the flir"" camera index to run, where 0 corresponds to the detected flir camera with the lowest"" serial number, and `n` to the `n`-th lowest serial number camera."): flags.hpp'],['../flags_8hpp.html#a6c9c8a5843dd8b93e009bf29dc31cde2',1,'DEFINE_int32(frame_rotate, 0,"Rotate each frame, 4 possible values: 0, 90, 180, 270."): flags.hpp'],['../flags_8hpp.html#a284252d3d255ad5e5c35815d720fd67a',1,'DEFINE_int32(num_gpu,-1,"The number of GPU devices to use. If negative, it will use all the available GPUs in your"" machine."): flags.hpp'],['../flags_8hpp.html#a8511765700f652000f2c1c2b1a5df9f9',1,'DEFINE_int32(num_gpu_start, 0,"GPU device start number."): flags.hpp'],['../flags_8hpp.html#a18fc454ffeef53c7c73d69c67d1a73fc',1,'DEFINE_int32(keypoint_scale, 0,"Scaling of the (x,y) coordinates of the final pose data array, i.e. the scale of the (x,y)"" coordinates that will be saved with the `write_json` & `write_keypoint` flags."" Select `0` to scale it to the original source resolution; `1`to scale it to the net output"" size (set with `net_resolution`); `2` to scale it to the final output size (set with"" `resolution`); `3` to scale it in the range [0,1], where (0,0) would be the top-left"" corner of the image, and (1,1) the bottom-right one; and 4 for range [-1,1], where"" (-1,-1) would be the top-left corner of the image, and (1,1) the bottom-right one. Non"" related with `scale_number` and `scale_gap`."): flags.hpp'],['../flags_8hpp.html#aa5a1826a500d7131fefb480ccd1713fb',1,'DEFINE_int32(number_people_max,-1,"This parameter will limit the maximum number of people detected, by keeping the people with"" top scores. The score is based in person area over the image, body part score, as well as"" joint score (between each pair of connected body parts). Useful if you know the exact"" number of people in the scene, so it can remove false positives (if all the people have"" been detected. However, it might also include false negatives by removing very small or"" highly occluded people. -1 will keep them all."): flags.hpp'],['../flags_8hpp.html#a1edea5f45026b353b6e6cc3b196767a0',1,'DEFINE_int32(scale_number, 1,"Number of scales to average."): flags.hpp'],['../flags_8hpp.html#aa80e5f9914dd35c852941282aa229b21',1,'DEFINE_int32(heatmaps_scale, 2,"Set 0 to scale op::Datum::poseHeatMaps in the range [-1,1], 1 for [0,1]; 2 for integer"" rounded [0,255]; and 3 for no scaling."): flags.hpp'],['../flags_8hpp.html#a311e487137e17445e6939abb7cc0cf8f',1,'DEFINE_int32(hand_scale_number, 1,"Analogous to `scale_number` but applied to the hand keypoint detector. Our best results"" were found with `hand_scale_number` = 6 and `hand_scale_range` = 0.4."): flags.hpp'],['../flags_8hpp.html#aa797dd033c6f3c4d2654e7000939d270',1,'DEFINE_int32(3d_min_views,-1,"Minimum number of views required to reconstruct each keypoint. By default (-1), it will"" require all the cameras to see the keypoint in order to reconstruct it."): flags.hpp'],['../flags_8hpp.html#a20c481950df0272f0b7b0cde67d8e72a',1,'DEFINE_int32(3d_views, 1,"Complementary option to `--image_dir` or `--video`. OpenPose will read as many images per"" iteration, allowing tasks such as stereo camera processing (`--3d`). Note that"" `--camera_parameters_folder` must be set. OpenPose must find as many `xml` files in the"" parameter folder as this number indicates."): flags.hpp'],['../flags_8hpp.html#afdfef14901f7b5e324d6983845f5ab50',1,'DEFINE_int32(tracking,-1,"Experimental, not available yet. Whether to enable people tracking across frames. The"" value indicates the number of frames where tracking is run between each OpenPose keypoint"" detection. Select -1 (default) to disable it or 0 to run simultaneously OpenPose keypoint"" detector and tracking for potentially higher accurary than only OpenPose."): flags.hpp'],['../flags_8hpp.html#a072bd893f4003b48bc7c99735eeeed39',1,'DEFINE_int32(ik_threads, 0,"Experimental, not available yet. Whether to enable inverse kinematics (IK) from 3-D"" keypoints to obtain 3-D joint angles. By default (0 threads), it is disabled. Increasing"" the number of threads will increase the speed but also the global system latency."): flags.hpp'],['../flags_8hpp.html#a4cead735de5b43cfcae5c1139df3be1a',1,'DEFINE_int32(part_to_show, 0,"Prediction channel to visualize (default: 0). 0 for all the body parts, 1-18 for each body"" part heat map, 19 for the background heat map, 20 for all the body part heat maps"" together, 21 for all the PAFs, 22-40 for each body part pair PAF."): flags.hpp'],['../flags_8hpp.html#ab2af299b6380dcd6dc06a95cceb056d4',1,'DEFINE_int32(render_pose,-1,"Set to 0 for no rendering, 1 for CPU rendering (slightly faster), and 2 for GPU rendering"" (slower but greater functionality, e.g. `alpha_X` flags). If -1, it will pick CPU if"" CPU_ONLY is enabled, or GPU if CUDA is enabled. If rendering is enabled, it will render"" both `outputData` and `cvOutputData` with the original image and desired body part to be"" shown (i.e. keypoints, heat maps or PAFs)."): flags.hpp'],['../flags_8hpp.html#ad0269da28dc2033e23b8ea84b7e793a2',1,'DEFINE_int32(face_render,-1,"Analogous to `render_pose` but applied to the face. Extra option: -1 to use the same"" configuration that `render_pose` is using."): flags.hpp'],['../flags_8hpp.html#ada5b7fef5063818fd668359e9e0b0504',1,'DEFINE_int32(hand_render,-1,"Analogous to `render_pose` but applied to the hand. Extra option: -1 to use the same"" configuration that `render_pose` is using."): flags.hpp'],['../flags_8hpp.html#a9a34e10f75069cf3283e535a77006775',1,'DEFINE_int32(display,-1,"Display mode: -1 for automatic selection; 0 for no display (useful if there is no X server"" and/or to slightly speed up the processing if visual output is not required); 2 for 2-D"" display; 3 for 3-D display (if `--3d` enabled); and 1 for both 2-D and 3-D display."): flags.hpp']]],
['define_5fstring',['DEFINE_string',['../flags_8hpp.html#a9e7b1394aea185360a1fe19cf4e20a89',1,'DEFINE_string(camera_resolution,"-1x-1","Set the camera resolution (either `--camera` or `--flir_camera`). `-1x-1` will use the"" default 1280x720 for `--camera`, or the maximum flir camera resolution available for"" `--flir_camera`"): flags.hpp'],['../flags_8hpp.html#aab02d4078b5f7999a582d9c5f4248676',1,'DEFINE_string(video,"","Use a video file instead of the camera. Use `examples/media/video.avi` for our default"" example video."): flags.hpp'],['../flags_8hpp.html#a7773e867133822f5601899975dc06adb',1,'DEFINE_string(image_dir,"","Process a directory of images. Use `examples/media/` for our default example folder with 20"" images. Read all standard formats (jpg, png, bmp, etc.)."): flags.hpp'],['../flags_8hpp.html#a5db17b8bc6df4fe40b556af7157dcbf7',1,'DEFINE_string(ip_camera,"","String with the IP camera URL. It supports protocols like RTSP and HTTP."): flags.hpp'],['../flags_8hpp.html#aff61246512375ff5941dc4110e127ca3',1,'DEFINE_string(camera_parameter_folder,"models/cameraParameters/flir/","String with the folder where the camera parameters are located."): flags.hpp'],['../flags_8hpp.html#a9ab689ebe20a261b20587af79123e79a',1,'DEFINE_string(model_folder,"models/","Folder path (absolute or relative) where the models (pose, face, ...) are located."): flags.hpp'],['../flags_8hpp.html#a85be61a31eaa438a7e9c7d2baf51da47',1,'DEFINE_string(output_resolution,"-1x-1","The image resolution (display and output). Use \"-1x-1\" to force the program to use the"" input image resolution."): flags.hpp'],['../flags_8hpp.html#acfd124be44003a59f2591c7584fd3c75',1,'DEFINE_string(model_pose,"BODY_25","Model to be used. E.g. `COCO` (18 keypoints), `MPI` (15 keypoints, ~10% faster), ""`MPI_4_layers` (15 keypoints, even faster but less accurate)."): flags.hpp'],['../flags_8hpp.html#a830bcfa6645bf39a18f59d3b72f75edf',1,'DEFINE_string(net_resolution,"-1x368","Multiples of 16. If it is increased, the accuracy potentially increases. If it is"" decreased, the speed increases. For maximum speed-accuracy balance, it should keep the"" closest aspect ratio possible to the images or videos to be processed. Using `-1` in"" any of the dimensions, OP will choose the optimal aspect ratio depending on the user's"" input value. E.g. the default `-1x368` is equivalent to `656x368` in 16:9 resolutions,"" e.g. full HD (1980x1080) and HD (1280x720) resolutions."): flags.hpp'],['../flags_8hpp.html#af5ee5f61f0d36a03bb8647408f5e236b',1,'DEFINE_string(face_net_resolution,"368x368","Multiples of 16 and squared. Analogous to `net_resolution` but applied to the face keypoint"" detector. 320x320 usually works fine while giving a substantial speed up when multiple"" faces on the image."): flags.hpp'],['../flags_8hpp.html#afaf97bbf6a49576782d25147bc865bed',1,'DEFINE_string(hand_net_resolution,"368x368","Multiples of 16 and squared. Analogous to `net_resolution` but applied to the hand keypoint"" detector."): flags.hpp'],['../flags_8hpp.html#a7a3597e9216885470199ca1578eb7f69',1,'DEFINE_string(write_images,"","Directory to write rendered frames in `write_images_format` image format."): flags.hpp'],['../flags_8hpp.html#ab077893503ebb5fba8cb300bd5f93d62',1,'DEFINE_string(write_images_format,"png","File extension and format for `write_images`, e.g. png, jpg or bmp. Check the OpenCV"" function cv::imwrite for all compatible extensions."): flags.hpp'],['../flags_8hpp.html#a3051d21eb51cc39eed5f781d8eaed960',1,'DEFINE_string(write_video,"","Full file path to write rendered frames in motion JPEG video format. It might fail if the"" final path does not finish in `.avi`. It internally uses cv::VideoWriter. Flag"" `camera_fps` controls FPS."): flags.hpp'],['../flags_8hpp.html#a9aa48b2ab293842bc42b96df9e97c9b8',1,'DEFINE_string(write_json,"","Directory to write OpenPose output in JSON format. It includes body, hand, and face pose"" keypoints (2-D and 3-D), as well as pose candidates (if `--part_candidates` enabled)."): flags.hpp'],['../flags_8hpp.html#a26d1c7340fc87d4593dda754d54145a2',1,'DEFINE_string(write_coco_json,"","Full file path to write people pose data with JSON COCO validation format."): flags.hpp'],['../flags_8hpp.html#ac6c2099e630e05f867ee10b43f35dc65',1,'DEFINE_string(write_coco_foot_json,"","Full file path to write people foot pose data with JSON COCO validation format."): flags.hpp'],['../flags_8hpp.html#aa3fe7c4c07492e6553a6c2d25ebd76b4',1,'DEFINE_string(write_heatmaps,"","Directory to write body pose heatmaps in PNG format. At least 1 `add_heatmaps_X` flag"" must be enabled."): flags.hpp'],['../flags_8hpp.html#a242473077549869f06534e8a9ea1ddd6',1,'DEFINE_string(write_heatmaps_format,"png","File extension and format for `write_heatmaps`, analogous to `write_images_format`."" For lossless compression, recommended `png` for integer `heatmaps_scale` and `float` for"" floating values."): flags.hpp'],['../flags_8hpp.html#ad8c5173beb83c0f9996362b3a3fba820',1,'DEFINE_string(write_keypoint,"","(Deprecated, use `write_json`) Directory to write the people pose keypoint data. Set format"" with `write_keypoint_format`."): flags.hpp'],['../flags_8hpp.html#a5aaba99bdb163516d0297d2e09dd0c7d',1,'DEFINE_string(write_keypoint_format,"yml","(Deprecated, use `write_json`) File extension and format for `write_keypoint`: json, xml,"" yaml & yml. Json not available for OpenCV < 3.0, use `write_json` instead."): flags.hpp'],['../flags_8hpp.html#a6c5341914694863528d3d93b23b45f01',1,'DEFINE_string(write_video_adam,"","Experimental, not available yet. E.g.: `~/Desktop/adamResult.avi`. Flag `camera_fps`"" controls FPS."): flags.hpp'],['../flags_8hpp.html#ad0c09dd7ede747d69d36dc86c7ffa11c',1,'DEFINE_string(write_bvh,"","Experimental, not available yet. E.g.: `~/Desktop/mocapResult.bvh`."): flags.hpp'],['../flags_8hpp.html#a63936bd2bc53a453ba74066289ab9d29',1,'DEFINE_string(udp_host,"","Experimental, not available yet. IP for UDP communication. E.g., `192.168.0.1`."): flags.hpp'],['../flags_8hpp.html#aa7b1c758c6d6e35c3227994a9ced0236',1,'DEFINE_string(udp_port,"8051","Experimental, not available yet. Port number for UDP communication."): flags.hpp']]],
['define_5fuint64',['DEFINE_uint64',['../flags_8hpp.html#a1433eaf1c5eb42e406e76bc6f8e517c3',1,'DEFINE_uint64(frame_first, 0,"Start on desired frame number. Indexes are 0-based, i.e. the first frame has index 0."): flags.hpp'],['../flags_8hpp.html#ac8fef8bb0234286e74b75214a750d674',1,'DEFINE_uint64(frame_last,-1,"Finish on desired frame number. Select -1 to disable. Indexes are 0-based, e.g. if set to"" 10, it will process 11 frames (0-10)."): flags.hpp']]],
['define_5fuint64',['DEFINE_uint64',['../flags_8hpp.html#a1433eaf1c5eb42e406e76bc6f8e517c3',1,'DEFINE_uint64(frame_first, 0,"Start on desired frame number. Indexes are 0-based, i.e. the first frame has index 0."): flags.hpp'],['../flags_8hpp.html#a22572531e5a4896c510639ac57cf522c',1,'DEFINE_uint64(frame_step, 1,"Step or gap between processed frames. E.g., `--frame_step 5` would read and process frames"" 0, 5, 10, etc.."): flags.hpp'],['../flags_8hpp.html#ac8fef8bb0234286e74b75214a750d674',1,'DEFINE_uint64(frame_last,-1,"Finish on desired frame number. Select -1 to disable. Indexes are 0-based, e.g. if set to"" 10, it will process 11 frames (0-10)."): flags.hpp']]],
<trclass="even"><tdclass="entry"><aclass="el"href="structop_1_1_wrapper_struct_input.html#a67d604b247a6ff54682b009c42ad61e2">WrapperStructInput</a>(const std::shared_ptr< Producer > producerSharedPtr=nullptr, const unsigned long long frameFirst=0, const unsigned long long frameStep=1, const unsigned long long frameLast=std::numeric_limits< unsigned long long >::max(), const bool realTimeProcessing=false, const bool frameFlip=false, const int frameRotate=0, const bool framesRepeat=false)</td><tdclass="entry"><aclass="el"href="structop_1_1_wrapper_struct_input.html">op::WrapperStructInput</a></td><tdclass="entry"></td></tr>
<trclass="memitem:a06aaa39948ac1a83a234b99f270ee14b"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="structop_1_1_wrapper_struct_input.html#a06aaa39948ac1a83a234b99f270ee14b">WrapperStructInput</a> (const std::shared_ptr<<aclass="el"href="classop_1_1_producer.html">Producer</a>><aclass="el"href="structop_1_1_wrapper_struct_input.html#adb77e57fd11db9950d9754a2c8c8eaa5">producerSharedPtr</a>=nullptr, const unsigned long long <aclass="el"href="structop_1_1_wrapper_struct_input.html#acc72b8efe09ec3888823ed5680a19fe4">frameFirst</a>=0, const unsigned long long <aclass="el"href="structop_1_1_wrapper_struct_input.html#a5cffb282052bdd812217e54f0b2ec7d5">frameLast</a>=-1, const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a2eeea9ee711a1dcbec99c3dc871fbc47">realTimeProcessing</a>=false, const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a5ee9722814fe2b5a695511cabd12b613">frameFlip</a>=false, const int <aclass="el"href="structop_1_1_wrapper_struct_input.html#a86df98e50b680b30afe100d8b2b50685">frameRotate</a>=0, const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a7c80f6a3687696ba30d3ce0902ac162f">framesRepeat</a>=false)</td></tr>
<trclass="memitem:a67d604b247a6ff54682b009c42ad61e2"><tdclass="memItemLeft"align="right"valign="top"> </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="structop_1_1_wrapper_struct_input.html#a67d604b247a6ff54682b009c42ad61e2">WrapperStructInput</a> (const std::shared_ptr<<aclass="el"href="classop_1_1_producer.html">Producer</a>><aclass="el"href="structop_1_1_wrapper_struct_input.html#adb77e57fd11db9950d9754a2c8c8eaa5">producerSharedPtr</a>=nullptr, const unsigned long long <aclass="el"href="structop_1_1_wrapper_struct_input.html#acc72b8efe09ec3888823ed5680a19fe4">frameFirst</a>=0, const unsigned long long <aclass="el"href="structop_1_1_wrapper_struct_input.html#ac4349e123d359f436cc01d4068231dc2">frameStep</a>=1, const unsigned long long <aclass="el"href="structop_1_1_wrapper_struct_input.html#a5cffb282052bdd812217e54f0b2ec7d5">frameLast</a>=std::numeric_limits< unsigned long long >::max(), const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a2eeea9ee711a1dcbec99c3dc871fbc47">realTimeProcessing</a>=false, const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a5ee9722814fe2b5a695511cabd12b613">frameFlip</a>=false, const int <aclass="el"href="structop_1_1_wrapper_struct_input.html#a86df98e50b680b30afe100d8b2b50685">frameRotate</a>=0, const bool <aclass="el"href="structop_1_1_wrapper_struct_input.html#a7c80f6a3687696ba30d3ce0902ac162f">framesRepeat</a>=false)</td></tr>
<trclass="memitem:acc72b8efe09ec3888823ed5680a19fe4"><tdclass="memItemLeft"align="right"valign="top">unsigned long long </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="structop_1_1_wrapper_struct_input.html#acc72b8efe09ec3888823ed5680a19fe4">frameFirst</a></td></tr>
<trclass="memitem:ac4349e123d359f436cc01d4068231dc2"><tdclass="memItemLeft"align="right"valign="top">unsigned long long </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="structop_1_1_wrapper_struct_input.html#ac4349e123d359f436cc01d4068231dc2">frameStep</a></td></tr>
<trclass="memitem:a5cffb282052bdd812217e54f0b2ec7d5"><tdclass="memItemLeft"align="right"valign="top">unsigned long long </td><tdclass="memItemRight"valign="bottom"><aclass="el"href="structop_1_1_wrapper_struct_input.html#a5cffb282052bdd812217e54f0b2ec7d5">frameLast</a></td></tr>
<divclass="textblock"><p><aclass="el"href="structop_1_1_wrapper_struct_input.html">WrapperStructInput</a>: Input (images, video, webcam, etc.) configuration struct. <aclass="el"href="structop_1_1_wrapper_struct_input.html">WrapperStructInput</a> allows the user to set up the input frames generator. </p>
</div><h2class="groupheader">Constructor & Destructor Documentation</h2>
@@ -273,7 +294,7 @@ Public Attributes</h2></td></tr>
</tr>
</table>
</div><divclass="memdoc">
<p><aclass="el"href="classop_1_1_producer.html">Producer</a> which will generate the frames. Set to nullptr to disable the whole input, i.e. if the user is going to use his own frames generator. </p>
<p><aclass="el"href="classop_1_1_producer.html">Producer</a> which will generate the frames. Set to nullptr to disable the whole input, i.e., if the user is going to use his own frames generator. </p>
<divclass="ttc"id="structop_1_1_wrapper_struct_input_html_ac4349e123d359f436cc01d4068231dc2"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_input.html#ac4349e123d359f436cc01d4068231dc2">op::WrapperStructInput::frameStep</a></div><divclass="ttdeci">unsigned long long frameStep</div><divclass="ttdef"><b>Definition:</b> wrapperStructInput.hpp:33</div></div>
<divclass="ttc"id="structop_1_1_wrapper_struct_input_html_a5cffb282052bdd812217e54f0b2ec7d5"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_input.html#a5cffb282052bdd812217e54f0b2ec7d5">op::WrapperStructInput::frameLast</a></div><divclass="ttdeci">unsigned long long frameLast</div><divclass="ttdef"><b>Definition:</b> wrapperStructInput.hpp:31</div></div>
<divclass="ttc"id="structop_1_1_wrapper_struct_input_html_a5cffb282052bdd812217e54f0b2ec7d5"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_input.html#a5cffb282052bdd812217e54f0b2ec7d5">op::WrapperStructInput::frameLast</a></div><divclass="ttdeci">unsigned long long frameLast</div><divclass="ttdef"><b>Definition:</b> wrapperStructInput.hpp:39</div></div>
<divclass="ttc"id="structop_1_1_wrapper_struct_face_html_a9845712fd6ebb66fccb0c1647e3491a0"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_face.html#a9845712fd6ebb66fccb0c1647e3491a0">op::WrapperStructFace::netInputSize</a></div><divclass="ttdeci">Point< int > netInputSize</div><divclass="ttdef"><b>Definition:</b> wrapperStructFace.hpp:27</div></div>
<divclass="ttc"id="structop_1_1_wrapper_struct_input_html_acc72b8efe09ec3888823ed5680a19fe4"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_input.html#acc72b8efe09ec3888823ed5680a19fe4">op::WrapperStructInput::frameFirst</a></div><divclass="ttdeci">unsigned long long frameFirst</div><divclass="ttdef"><b>Definition:</b> wrapperStructInput.hpp:25</div></div>
<divclass="ttc"id="structop_1_1_wrapper_struct_input_html_acc72b8efe09ec3888823ed5680a19fe4"><divclass="ttname"><ahref="structop_1_1_wrapper_struct_input.html#acc72b8efe09ec3888823ed5680a19fe4">op::WrapperStructInput::frameFirst</a></div><divclass="ttdeci">unsigned long long frameFirst</div><divclass="ttdef"><b>Definition:</b> wrapperStructInput.hpp:26</div></div>