forked from olgaliak/active-learning-detect
-
Notifications
You must be signed in to change notification settings - Fork 9
Expand file tree
/
Copy pathconfig_oikonos_mega.ini
More file actions
70 lines (70 loc) · 3.33 KB
/
config_oikonos_mega.ini
File metadata and controls
70 lines (70 loc) · 3.33 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
# AZURE STORAGE ACCOUNT INFORMATION
AZURE_STORAGE_ACCOUNT=oikonos
AZURE_STORAGE_KEY="put key here"
image_container_name=oikonos-mega-images
label_container_name=oikonos-mega-labels
# IMAGE INFORMATION
user_folders=True
classes=cat,dog,rabbit,rat,shearwater,Unknown,zorzal,petrel,kestrel,pigeon,human,goat,mouse,coati,cow,horse,pig,owl
# Provide preferred distribution of images-review ratio.
# Last value corresponds to images were no object were detected.
# In the example below: 60% of images that use will be reviewing have at least one bbox with objct class1 (knot),
# 30% images that have bboxes for class (defects)
# and 10% of images get class "NULL" -- were neither knots nor defects were detected by the model
ideal_class_balance=0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
filetype=*.JPG
# TAGGING MACHINE
tagging_location=C:\Users\ConservationMetrics\Desktop\oikonos_tagging
pick_max=False
max_tags_per_pixel=1
#
# CUSTOM VISION
# Uncomment lines below if using Azure Custom Vision Service training_key=<CUSTOM VISION TRAINING KEY>
# prediction_key=<CUSTOM VISION PREDICTION KEY>
# project_id=<CUSTOM VISION PROJECT ID>
#
# TRAINING MACHINE
# Locations
python_file_directory=/home/cmi/repos/models/research/active-learning-detect/train
data_dir=/home/cmi/data/oikonosmega
train_dir=/home/cmi/data/oikonosmega/training
inference_output_dir=/home/cmi/data/oikonosmega/training/oikonosmega_inference_graphs
tf_models_location=/home/cmi/repos/models/research
download_location=/home/cmi/downloads
# Training
train_iterations=100
eval_iterations=100
min_confidence=.1
test_percentage=.15
model_name=megadetector_v3
optional_pipeline_url=https://lilablobssc.blob.core.windows.net/models/camera_traps/megadetector/megadetector_v3.config
#Init Predictions
init_model_name=faster_rcnn_resnet101_coco_2018_01_28
# Config File Details
old_label_path=/megadetectorv3/PyCharm/CameraTraps/detection/experiments/megadetector_v3/label_map.pbtxt
old_train_path=/disk/megadetectorv3_tfrecords/???????~train-?????-of-?????
old_val_path=/disk/megadetectorv3_tfrecords/???????~val__-?????-of-?????
old_checkpoint_path=PATH_TO_BE_CONFIGURED/model.ckpt
num_examples_marker=num_examples:
num_steps_marker=num_steps:
num_classes_marker=num_classes:
# Calculated
num_classes="$(awk -F ',' '{print NF}' <<< ${classes})"
image_dir=${data_dir}/AllImages
untagged_output=${data_dir}/untagged.csv
tagged_output=${data_dir}/tagged.csv
tagged_predictions=${data_dir}/tagged_preds.csv
test_output=${data_dir}/test.csv
validation_output=${data_dir}/val.csv
tf_location=${tf_models_location}/object_detection
tf_location_legacy=${tf_models_location}/object_detection/legacy
PYTHONPATH=$PYTHONPATH:${tf_models_location}:${tf_models_location}/slim/
label_map_path=${data_dir}/pascal_label_map.pbtxt
tf_record_location=${data_dir}/stamps.record
tf_train_record=${tf_record_location%.*}_train.${tf_record_location##*.}
tf_val_record=${tf_record_location%.*}_val.${tf_record_location##*.}
tf_url=https://lilablobssc.blob.core.windows.net/models/camera_traps/megadetector/${model_name}_checkpoint.zip
pipeline_file=${download_location}/${model_name}/pipeline.config
fine_tune_checkpoint=${download_location}/${model_name}/${model_name}_checkpoint/model.ckpt
init_pred_tf_url=http://download.tensorflow.org/models/object_detection/${init_model_name}.tar.gz
init_model_graph=${download_location}/${init_model_name}/frozen_inference_graph.pb