-
Notifications
You must be signed in to change notification settings - Fork 5
/
d42_sd_sync.py
executable file
·1238 lines (1036 loc) · 61.2 KB
/
d42_sd_sync.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
__author__ = 'Roman Nyschuk'
import os
import sys
import logging
import json
import argparse
import datetime
from device42 import Device42
from freshservice import FreshService, FreshServiceDuplicateValueError
import xml.etree.ElementTree as eTree
from xmljson import badgerfish as bf
import time
import math
logger = logging.getLogger('log')
logger.setLevel(logging.INFO)
ch = logging.StreamHandler(sys.stdout)
ch.setFormatter(logging.Formatter('%(asctime)-15s\t%(levelname)s\t %(message)s'))
logger.addHandler(ch)
CUR_DIR = os.path.dirname(os.path.abspath(__file__))
RELATIONSHIP_BATCH_SIZE = 20
# With v1 of the API, we were able to create about 4 relationships per second.
# So we will assume that we will be able to create them at the same rate with
# the asynchronous background jobs.
RELATIONSHIPS_CREATED_PER_SECOND = 4
# The number of seconds to wait before we check the status of create relationships jobs.
RELATIONSHIPS_JOB_WAIT_SECONDS = int(math.ceil(RELATIONSHIP_BATCH_SIZE / float(RELATIONSHIPS_CREATED_PER_SECOND)))
ASSET_TYPE_BUSINESS_SERVICE = "Business Service"
ASSET_TYPE_SERVER = "Server"
ASSET_TYPE_UNIX_SERVER = "Unix Server"
ASSET_TYPE_WINDOWS_SERVER = "Windows Server"
ASSET_TYPE_HOST = "Host"
parser = argparse.ArgumentParser(description="freshservice")
parser.add_argument('-d', '--debug', action='store_true', help='Enable debug output')
parser.add_argument('-q', '--quiet', action='store_true', help='Quiet mode - outputs only errors')
parser.add_argument('-c', '--config', help='Config file', default='mapping.xml')
parser.add_argument('-l', '--logfolder', help='log folder path', default='.')
freshservice = None
default_approver = None
fs_cache = dict()
class JSONEncoder(json.JSONEncoder):
def default(self, o):
if isinstance(o, datetime):
return o.strftime("%Y %m %d %H:%M:%S")
return json.JSONEncoder.default(self, o)
def escape_value(name):
if name:
name = name.replace('<', '[')
name = name.replace('>', ']')
# Replace Unicode no-break spaces with normal spaces.
name = name.replace(u'\xa0', ' ')
name = name.strip()
return name
def find_object_by_name(assets, name):
for asset in assets:
if asset["name"].lower() == escape_value(name).lower():
return asset
return None
def find_object_in_map(objects_map, name):
if name:
return objects_map.get(escape_value(name).lower())
return None
def find_object_id_in_map(objects_map, name):
obj = find_object_in_map(objects_map, name)
if obj:
return obj["id"]
return None
def get_asset_type_field(asset_type_fields, map_info):
for section in asset_type_fields:
if section["field_header"] == map_info["@target-header"]:
for field in section["fields"]:
name = map_info["@target"]
if "@target-field" in map_info:
name = map_info["@target-field"]
if field["asset_type_id"] is not None:
name += "_" + str(field["asset_type_id"])
if field["name"] == name:
return field
return None
def get_map_value_from_device42(source, map_info, b_add=False, asset_type_id=None):
d42_value = source[map_info["@resource"]]
if d42_value is None and "@resource-secondary" in map_info:
d42_value = source[map_info["@resource-secondary"]]
if "@is-array" in map_info and map_info["@is-array"]:
d42_vals = d42_value
d42_value = None
for d42_val in d42_vals:
if map_info["@sub-key"] in d42_val:
d42_value = d42_val[map_info["@sub-key"]]
break
else:
if "value-mapping" in map_info:
cache_key = "%s-%s" % (map_info["@resource"], map_info["@target"])
if cache_key not in fs_cache:
if isinstance(map_info["value-mapping"]["item"], list):
items = map_info["value-mapping"]["item"]
else:
items = [map_info["value-mapping"]["item"]]
fs_cache[cache_key] = {item["@key"]: item["@value"] for item in items}
d42_val = fs_cache[cache_key].get(d42_value)
if d42_val is None and "@default" in map_info["value-mapping"]:
default_value = map_info["value-mapping"]["@default"]
# If we send a software status of "", we get the following error from the API:
# Error 400: {"description":"Validation failed","errors":[{"field":"status",
# "message":"It should be one of these values: 'blacklisted,ignored,managed'","code":"invalid_value"}]}
# So if we have a value in D42 that does not map to Freshservice (we don't currently have a value that
# does not map), instead of clearing the value in Freshservice by sending a "", it will try to set that
# value and this is not one of the available options. However, if we set the software status to None,
# the value in Freshservice will get cleared.
if default_value == "null":
d42_val = None
else:
d42_val = default_value
d42_value = d42_val
else:
pass
if "@target-foregin-key" in map_info:
target_foregin = map_info["@target-foregin"]
if target_foregin not in fs_cache:
fs_cache[target_foregin] = freshservice.get_objects_map("api/v2/%s" % target_foregin, target_foregin, map_info["@target-foregin-key"])
value = find_object_id_in_map(fs_cache[target_foregin], d42_value)
if b_add and value is None and "@not-null" in map_info and map_info["@not-null"]: # and "@required" in map_info and map_info["@required"]
if d42_value is not None:
if "@max-length" in map_info and len(d42_value) > map_info["@max-length"]:
name = d42_value[0:map_info["@max-length"] - 3] + "..."
else:
name = d42_value
if target_foregin in ["vendors", "groups", "agents"]:
new_item = freshservice.insert_and_get_by_name(target_foregin, name, None, map_info["@target-foregin-key"])
else:
new_item = freshservice.insert_and_get_by_name(target_foregin, name, asset_type_id, map_info["@target-foregin-key"])
fs_cache[target_foregin][new_item[map_info["@target-foregin-key"]].lower()] = new_item
d42_value = new_item["id"]
else:
d42_value = None
else:
# If value is None, that means we could not find a match for the D42 value in Freshservice.
# We will return the same D42 value since for product we will call this function again with
# the required asset_type_id which is needed to create the value in Freshservice.
if value is not None or "@not-null" not in map_info:
d42_value = value
if "@escape" in map_info and map_info["@escape"]:
d42_value = escape_value(d42_value)
return d42_value
def get_asset_type_field_from_map(asset_type_fields_map, asset_type_id, asset_type_fields, map_info):
if asset_type_id not in asset_type_fields_map:
asset_type_fields_map[asset_type_id] = dict()
target_header = map_info["@target-header"] if "@target-header" in map_info else ""
key = map_info["@resource"] + "-" + target_header
if key in asset_type_fields_map[asset_type_id]:
asset_type_field = asset_type_fields_map[asset_type_id][key]
else:
asset_type_field = get_asset_type_field(asset_type_fields, map_info)
asset_type_fields_map[asset_type_id][key] = asset_type_field
return asset_type_field
def submit_relationship_create_job(relationships_to_create):
logger.info("adding relationship create job")
# Creating relationships using the v2 API is now an asynchronous operation and is
# performed using background jobs. We will get back the job ID which can then be
# used to query the status of the job.
job_id = freshservice.insert_relationships({"relationships": relationships_to_create})
logger.info("added new relationship create job %s" % job_id)
return {
"job_id": job_id,
"relationships_to_create_count": len(relationships_to_create)
}
def update_objects_from_server(sources, _target, mapping):
global freshservice
# This method gets called for both devices and business apps. Since it gets called first for devices,
# that is when the assets from Freshservice will get added to the cache. When this method gets called
# for business apps, we can get the objects out of the cache.
if "assets" in fs_cache:
logger.info("Getting all existing assets in FS from cache.")
existing_objects_map = fs_cache["assets"]
logger.info("finished getting all existing assets in FS from cache.")
else:
logger.info("Getting all existing assets in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing assets in FS.")
fs_cache["assets"] = existing_objects_map
if "asset_types" not in fs_cache:
asset_types_map = freshservice.get_objects_map("api/v2/asset_types", "asset_types")
fs_cache["asset_types"] = asset_types_map
else:
asset_types_map = fs_cache["asset_types"]
if "asset_type_fields" not in fs_cache:
fs_cache["asset_type_fields"] = {}
asset_type_fields_map = dict()
server_asset_type_id = find_object_id_in_map(asset_types_map, ASSET_TYPE_SERVER)
unix_server_asset_type_id = find_object_id_in_map(asset_types_map, ASSET_TYPE_UNIX_SERVER)
windows_server_asset_type_id = find_object_id_in_map(asset_types_map, ASSET_TYPE_WINDOWS_SERVER)
host_asset_type_id = find_object_id_in_map(asset_types_map, ASSET_TYPE_HOST)
for source in sources:
error_skip = False
while True:
try:
existing_object = find_object_in_map(existing_objects_map, source["name"])
source_asset_type_id = find_object_id_in_map(asset_types_map, source["asset_type"])
# If we have an existing asset with an asset type of Windows Server or Unix Server and
# we determined that the asset type should be Host, we will update it to Host since previously
# we were bringing in host devices as Windows Server or Unix Server asset types instead of the
# Host asset type.
if existing_object is None or existing_object["asset_type_id"] == server_asset_type_id or \
(existing_object["asset_type_id"] in [unix_server_asset_type_id, windows_server_asset_type_id] and \
source_asset_type_id == host_asset_type_id):
asset_type_id = source_asset_type_id
else:
asset_type_id = existing_object["asset_type_id"]
if asset_type_id in fs_cache["asset_type_fields"]:
asset_type_fields = fs_cache["asset_type_fields"][asset_type_id]
else:
asset_type_fields = freshservice.get_asset_type_fields(asset_type_id)
fs_cache["asset_type_fields"][asset_type_id] = asset_type_fields
data = dict()
data['asset_type_id'] = asset_type_id
data["type_fields"] = dict()
# if there is only one field in the mapping, it will be dict.
if isinstance(mapping["field"], dict):
mapping["field"] = [mapping["field"]]
# validation
for map_info in mapping["field"]:
if error_skip and "@error-skip" in map_info and map_info["@error-skip"]:
continue
asset_type_field = get_asset_type_field_from_map(asset_type_fields_map, asset_type_id, asset_type_fields, map_info)
if asset_type_field is None:
continue
value = get_map_value_from_device42(source, map_info)
if asset_type_field["asset_type_id"] is not None:
data["type_fields"][asset_type_field["name"]] = value
else:
data[map_info["@target"]] = value
is_valid = True
if value is not None and "@min-length" in map_info and len(value) < map_info["@min-length"]:
is_valid = False
if value == "" and "@set-space" in map_info and map_info["@set-space"]:
is_valid = True
value = " " * map_info["@min-length"]
# value might have been translated to an associated ID in Freshservice by get_map_value_from_device42
# which is why we need to check that value is a string using isinstance.
if value is not None and "@max-length" in map_info and isinstance(value, str) and len(value) > map_info["@max-length"]:
value = value[0:map_info["@max-length"] - 3] + "..."
if value is None and "@not-null" in map_info and map_info["@not-null"]:
if map_info["@target"] == "asset_tag":
is_valid = False
else:
# There is an issue with the Freshservice API where sending a null value for
# a field will result in the API returning an error like "Has 0 characters,
# it should have minimum of 1 characters and can have maximum of 255 characters".
# This prevents us from being able to clear these field values in Freshservice (even though
# the Freshservice UI allows you to clear these fields). To get around this, we will send
# a single space for string values and a 0 for integer and float values when the value
# coming from D42 is null.
if "@target-type" in map_info:
target_type = map_info["@target-type"]
if target_type == "integer" or target_type == "float":
value = 0
elif target_type == "date":
value = None
is_valid = False
else:
value = " "
else:
value = " "
if "@target-foregin-key" in map_info:
value = get_map_value_from_device42(source, map_info, True, data["asset_type_id"])
is_valid = value is not None
if "@target-type" in map_info and value is not None:
target_type = map_info["@target-type"]
if target_type == "integer":
try:
value = int(value)
except Exception as e:
logger.exception(str(e))
is_valid = False
elif target_type == "dropdown":
try:
option = None
for choice in asset_type_field["choices"]:
d42_value = value.lower()
choice_value = choice[0].lower()
if d42_value in choice_value or choice_value in d42_value:
option = choice[0]
break
if option is None:
is_valid = False
else:
value = option
except Exception as e:
logger.exception(str(e))
is_valid = False
if not is_valid:
logger.debug("argument '%s' is invalid." % map_info["@target"])
if asset_type_field["asset_type_id"] is not None:
data["type_fields"].pop(asset_type_field["name"], None)
else:
data.pop(map_info["@target"], None)
if is_valid:
if asset_type_field["asset_type_id"] is not None:
data["type_fields"][asset_type_field["name"]] = value
else:
data[map_info["@target"]] = value
if existing_object is None:
logger.info("adding asset %s" % source["name"])
new_asset = freshservice.insert_asset(data)
logger.info("added new asset %d" % new_asset["id"])
# We added a new object to Freshservice. Add it to the map of objects that we know exist
# in Freshservice.
existing_objects_map[new_asset["name"].lower()] = new_asset
else:
logger.info("updating asset %s" % source["name"])
# This is a workaround for an issue with the Freshservice API where if a business service
# asset has the Managed By field filled in and we don't send an agent_id to update this
# field (we don't map any D42 data to this field and shouldn't need to because
# the API will only update the fields that we send), it will result in a validation
# error with the message:
# Assigned agent isn't a member of the group.
# So, if the business service asset has an agent_id already populated, we will send that
# same value over and that will avoid this error.
if source["asset_type"] == ASSET_TYPE_BUSINESS_SERVICE and "agent_id" in existing_object and existing_object["agent_id"]:
data["agent_id"] = existing_object["agent_id"]
updated_asset_id = freshservice.update_asset(data, existing_object["display_id"])
logger.info("updated existing asset %d" % updated_asset_id)
# If the asset type changed for this asset, update it in the cache.
if existing_object["asset_type_id"] != asset_type_id:
existing_object["asset_type_id"] = asset_type_id
break
except FreshServiceDuplicateValueError:
if not error_skip:
error_skip = True
continue
break
except Exception as e:
log = "Error (%s) updating device %s" % (str(e), source["name"])
logger.exception(log)
break
def delete_objects_from_server(sources, _target, mapping):
global freshservice
logger.info("Getting all existing devices in FS.")
existing_objects = freshservice.request(_target["@path"] + "?include=type_fields", "GET", _target["@model"])
logger.info("finished getting all existing devices in FS.")
for existing_object in existing_objects:
exist = False
for source in sources:
if source[mapping["@key"]] == existing_object[mapping["@key"]]:
exist = True
break
if not exist:
try:
logger.info("deleting device %s" % existing_object["name"])
freshservice.delete_asset(existing_object["display_id"])
logger.info("deleted asset %s" % existing_object["name"])
except Exception as e:
log = "Error (%s) deleting device %s" % (str(e), existing_object["name"])
logger.exception(log)
def update_softwares_from_server(sources, _target, mapping):
global freshservice
logger.info("Getting all existing softwares in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing softwares in FS.")
fs_cache["softwares"] = existing_objects_map
for source in sources:
try:
existing_object = find_object_in_map(existing_objects_map, source["name"])
data = dict()
for map_info in mapping["field"]:
value = get_map_value_from_device42(source, map_info)
# value might have been translated to an associated ID in Freshservice by get_map_value_from_device42
# which is why we need to check that value is a string using isinstance.
if value is not None and "@max-length" in map_info and isinstance(value, str) and len(value) > map_info["@max-length"]:
value = value[0:map_info["@max-length"] - 3] + "..."
data[map_info["@target"]] = value
if existing_object is None:
logger.info("adding software %s" % source["name"])
new_software = freshservice.insert_software(data)
logger.info("added new software %d" % new_software["id"])
# We added a new object to Freshservice. Add it to the map of objects that we know exist
# in Freshservice.
existing_objects_map[new_software["name"].lower()] = new_software
else:
logger.info("updating software %s" % source["name"])
updated_software_id = freshservice.update_software(data, existing_object["id"])
logger.info("updated existing software %d" % updated_software_id)
except Exception as e:
log = "Error (%s) updating software %s" % (str(e), source["name"])
logger.exception(log)
def delete_softwares_from_server(sources, _target, mapping):
global freshservice
logger.info("Getting all existing softwares in FS.")
existing_objects = freshservice.request(_target["@path"] + "?include=type_fields", "GET", _target["@model"])
logger.info("finished getting all existing devices in FS.")
for existing_object in existing_objects:
exist = False
for source in sources:
if source[mapping["@key"]] == existing_object[mapping["@key"]]:
exist = True
break
if not exist:
try:
logger.info("deleting software %s" % existing_object["name"])
freshservice.delete_software(existing_object["id"])
logger.info("deleted software %s" % existing_object["name"])
except Exception as e:
log = "Error (%s) deleting software %s" % (str(e), existing_object["name"])
logger.exception(log)
def update_products_from_server(sources, _target, mapping):
global freshservice
logger.info("Getting all existing products in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing products in FS.")
asset_types_map = freshservice.get_objects_map("api/v2/asset_types", "asset_types")
fs_cache["asset_types"] = asset_types_map
asset_type_id = find_object_id_in_map(asset_types_map, _target["@asset-type"])
for source in sources:
try:
existing_object = find_object_in_map(existing_objects_map, source["name"])
data = dict()
for map_info in mapping["field"]:
value = get_map_value_from_device42(source, map_info)
# value might have been translated to an associated ID in Freshservice by get_map_value_from_device42
# which is why we need to check that value is a string using isinstance.
if value is not None and "@max-length" in map_info and isinstance(value, str) and len(value) > map_info["@max-length"]:
value = value[0:map_info["@max-length"] - 3] + "..."
data[map_info["@target"]] = value
if existing_object is None:
logger.info("adding product %s" % source["name"])
data['asset_type_id'] = asset_type_id
new_product = freshservice.insert_product(data)
logger.info("added new product %d" % new_product["id"])
# We added a new object to Freshservice. Add it to the map of objects that we know exist
# in Freshservice.
existing_objects_map[new_product["name"].lower()] = new_product
else:
logger.info("updating product %s" % source["name"])
updated_product_id = freshservice.update_product(data, existing_object["id"])
logger.info("updated existing product %d" % updated_product_id)
except Exception as e:
log = "Error (%s) updating product %s" % (str(e), source["name"])
logger.exception(log)
def create_installation_from_software_in_use(sources, _target, mapping):
global freshservice
if "assets" in fs_cache:
logger.info("Getting all existing assets in FS from cache.")
existing_objects_map = fs_cache["assets"]
logger.info("finished getting all existing assets in FS from cache.")
else:
logger.info("Getting all existing assets in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing assets in FS.")
fs_cache["assets"] = existing_objects_map
if "softwares" in fs_cache:
logger.info("Getting all existing softwares in FS from cache.")
existing_softwares_map = fs_cache["softwares"]
logger.info("finished getting all existing softwares in FS from cache.")
else:
logger.info("Getting all existing softwares in FS.")
existing_softwares_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing softwares in FS.")
fs_cache["softwares"] = existing_softwares_map
software_to_assets_map = dict()
for source in sources:
try:
logger.info("Processing %s - %s." % (source[mapping["@device-name"]], source[mapping["@software-name"]]))
asset = find_object_in_map(existing_objects_map, source[mapping["@device-name"]])
software = find_object_in_map(existing_softwares_map, source[mapping["@software-name"]])
if asset is None:
log = "There is no asset(%s) in FS." % source[mapping["@device-name"]]
logger.exception(log)
continue
if software is None:
log = "There is no software(%s) in FS." % source[mapping["@software-name"]]
logger.exception(log)
continue
if software["id"] not in software_to_assets_map:
installations = freshservice.get_installations_by_id(software["id"])
software_to_assets_map[software["id"]] = {i["installation_machine_id"] for i in installations}
exist = asset["display_id"] in software_to_assets_map[software["id"]]
if exist:
logger.info("There is already installation in FS.")
continue
data = dict()
data["installation_machine_id"] = asset["display_id"]
data["version"] = source[mapping["@version"]]
data["installation_date"] = source[mapping["@install-date"]]
logger.info("adding installation %s-%s" % (source[mapping["@device-name"]], source[mapping["@software-name"]]))
freshservice.insert_installation(software["id"], data)
# We added a new installation to Freshservice. Add it to the map of installations that we know exist
# in Freshservice.
software_to_assets_map[software["id"]].add(asset["display_id"])
logger.info("added installation %s-%s" % (source[mapping["@device-name"]], source[mapping["@software-name"]]))
except Exception as e:
log = "Error (%s) creating installation %s" % (str(e), source[mapping["@device-name"]])
logger.exception(log)
def create_relationships_from_affinity_group(sources, _target, mapping):
global freshservice
if "assets" in fs_cache:
logger.info("Getting all existing assets in FS from cache.")
existing_objects_map = fs_cache["assets"]
logger.info("finished getting all existing assets in FS from cache.")
else:
logger.info("Getting all existing assets in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing assets in FS.")
fs_cache["assets"] = existing_objects_map
logger.info("Getting relationship type in FS.")
relationship_type = freshservice.get_relationship_type_by_content(mapping["@downstream-relationship"],
mapping["@upstream-relationship"])
logger.info("finished getting relationship type in FS.")
if relationship_type is None:
log = "There is no relationship type in FS. (%s - %s)" % (
mapping["@downstream-relationship"], mapping["@upstream-relationship"])
logger.info(log)
return
# The key will be the display_id of the primary asset and the value will be a set of
# the secondary asset display_id's that the primary asset is related to.
relationships_map = dict()
relationships_to_create = list()
source_count = len(sources)
submitted_jobs = list()
is_virtualized_by_rel_type = False
if relationship_type["downstream_relation"] == "Virtualized by" and relationship_type["upstream_relation"] == "Virtualizes":
is_virtualized_by_rel_type = True
# This will be used for deleting the incorrect Virtualized by/Virtualizes relationships.
assets_by_display_id = None
for idx, source in enumerate(sources):
try:
logger.info("Processing %s - %s." % (source[mapping["@key"]], source[mapping["@target-key"]]))
primary_asset = find_object_in_map(existing_objects_map, source[mapping["@key"]])
if primary_asset is None:
log = "There is no dependent asset(%s) in FS." % source[mapping["@key"]]
logger.exception(log)
continue
secondary_asset = find_object_in_map(existing_objects_map, source[mapping["@target-key"]])
if secondary_asset is None:
log = "There is no dependency asset(%s) in FS." % source[mapping["@target-key"]]
logger.exception(log)
continue
primary_asset_display_id = primary_asset["display_id"]
if primary_asset_display_id not in relationships_map:
relationships_map[primary_asset_display_id] = set()
relationships = freshservice.get_relationships_by_id(primary_asset_display_id)
for relationship in relationships:
if relationship["relationship_type_id"] == relationship_type["id"]:
deleted_relationship = False
# We used to created Virtualized by/Virtualizes relationships incorrectly. It would read
# as the host is virtualized by the VM and the VM virtualizes the host. This was wrong.
# It should read as the VM is virtualized by the host and the host virtualizes the VM.
# We will delete these incorrect relationships and the correct ones will be created if
# they do not exist.
if is_virtualized_by_rel_type:
if assets_by_display_id is None:
# Get both active assets (i.e. those assets that are not in the trash) and trashed
# assets (i.e. those assets that have been deleted and moved to the trash).
# We get the trashed assets because there could be incorrect relationships
# where the host has been deleted and moved to the trash, but the VM is active.
# The key for this dictionary will be the display id since this is what the
# relationships returned from the Freshservice API use (i.e. the display id for
# the primary and secondary assets in the relationship).
assets_by_display_id = {v["display_id"]: v for k, v in existing_objects_map.items()}
# For the key of the dictionary that gets returned, we will use the display id. We cannot
# use the asset name as the key because multiple assets in the trash could have the same
# name and this would result in only one of those assets with the same name being in
# the dictionary since they would each be using the same key. Even though these assets
# would have the same name, they would have different display ids. In addition, an
# asset in the trash could have the same name as an active asset.
trashed_assets = freshservice.get_objects_map("api/v2/assets?trashed=true", "assets", "display_id")
assets_by_display_id.update(trashed_assets)
# For the incorrect relationships, a host asset would have been the primary asset in the
# relationship.
if relationship["primary_id"] in assets_by_display_id:
pri = assets_by_display_id[relationship["primary_id"]]
if pri["asset_type_id"] in [fs_cache["asset_types"]["host"]["id"], fs_cache["asset_types"]["vmware vcenter host"]["id"]]:
freshservice.detach_relationship(relationship["id"])
logger.info("deleted incorrect Virtualized by/Virtualizes relationship %d" % relationship["id"])
deleted_relationship = True
if not deleted_relationship:
relationships_map[primary_asset_display_id].add(relationship["secondary_id"])
exist = secondary_asset["display_id"] in relationships_map[primary_asset_display_id]
if exist:
logger.info("There is already relationship in FS.")
continue
relationships_to_create.append({
"relationship_type_id": relationship_type["id"],
"primary_id": primary_asset_display_id,
"primary_type": "asset",
"secondary_id": secondary_asset["display_id"],
"secondary_type": "asset"
})
relationships_to_create_count = len(relationships_to_create)
relationships_map[primary_asset_display_id].add(secondary_asset["display_id"])
# Create a new job if we reached our batch size or we are on the last item (which
# means this is the last batch we will be submitting).
if relationships_to_create_count >= RELATIONSHIP_BATCH_SIZE or idx == source_count - 1:
submitted_jobs.append(submit_relationship_create_job(relationships_to_create))
# Clear the list for the next batch of relationships we are going to send.
del relationships_to_create[:]
except Exception as e:
log = "Error (%s) creating relationship %s" % (str(e), source[mapping["@key"]])
logger.exception(log)
# We may not have submitted the last batch of relationships to create if the last item in
# sources did not result in a relationship needing to be created (e.g. one of the assets
# in the relationship did not exist in Freshservice, the relationship already existed in
# Freshservice, etc.). So if we have any relationships that we need to create that have
# not been submitted, submit them now.
if relationships_to_create:
submitted_jobs.append(submit_relationship_create_job(relationships_to_create))
del relationships_to_create[:]
if submitted_jobs:
jobs_to_check = list(submitted_jobs)
next_jobs_to_check = list()
# We will make attempts to check the status of the jobs and see if they have
# completed. The max time we will wait is the number of jobs we submitted
# times the amount of time it takes to create a full batch of relationships.
# This total wait time will be broken into chunks based on how long it would
# take a single batch of relationships to be created. For example, if we
# submitted 3 jobs and each job had a batch of 20 relationships to create,
# then it should take 5 seconds to create the 20 relationships based on being
# able to create them at a rate of 4 per second. We will wait 5 seconds, then
# check the status of all jobs. If there are any jobs still waiting to complete,
# then we will wait another 5 seconds and check the status of the jobs that were
# previously waiting to complete.
# Added 20% padding to wait a little bit longer for the jobs to complete
# if needed.
for i in range(int(math.ceil(len(submitted_jobs) * 1.2))):
time.sleep(RELATIONSHIPS_JOB_WAIT_SECONDS)
for job_to_check in jobs_to_check:
try:
job = freshservice.get_job(job_to_check["job_id"])
status = job["status"]
if status == "success":
# All relationships were created.
logger.info("Job %s created all %d relationships successfully." % (job_to_check["job_id"], job_to_check["relationships_to_create_count"]))
elif status in ["failed", "partial"]:
# No relationships were created (failed status) or some relationships
# were created and some were not (partial status).
for relationship in job["relationships"]:
if not relationship["success"]:
log = "Job %s failed to create relationship: %s" % (job_to_check["job_id"], relationship)
logger.exception(log)
elif status in ["queued", "in progress"]:
# The job has not completed yet.
next_jobs_to_check.append(job_to_check)
log = "Job %s has not completed yet. The job status is %s." % (job_to_check["job_id"], status)
logger.info(log)
else:
raise Exception("Received unknown job status of %s." % status)
except Exception as e:
log = "Error (%s) checking job %s" % (str(e), job_to_check["job_id"])
logger.exception(log)
# Clear the list.
del jobs_to_check[:]
if next_jobs_to_check:
# We still have jobs we need to check.
jobs_to_check.extend(next_jobs_to_check)
# Clear the list so that we can add the next set of jobs that are
# still waiting to complete.
del next_jobs_to_check[:]
else:
# There are no more jobs that we need to check, so we can stop
# checking.
break
if jobs_to_check:
submitted_jobs_count = len(submitted_jobs)
jobs_not_completed_count = len(jobs_to_check)
logger.info("%d of %d relationship create jobs did not complete." % (jobs_not_completed_count, submitted_jobs_count))
def delete_relationships_from_affinity_group(sources, _target, mapping):
global freshservice
logger.info("Getting all existing devices in FS.")
existing_objects = freshservice.request("api/v2/assets" + "?include=type_fields", "GET", _target["@model"])
logger.info("finished getting all existing devices in FS.")
logger.info("Getting relationship type in FS.")
relationship_type = freshservice.get_relationship_type_by_content(mapping["@downstream-relationship"],
mapping["@upstream-relationship"])
logger.info("finished getting relationship type in FS.")
if relationship_type is None:
log = "There is no relationship type in FS. (%s - %s)" % (
mapping["@downstream-relationship"], mapping["@upstream-relationship"])
logger.info(log)
return
for source in sources:
try:
logger.info("Processing %s - %s." % (source[mapping["@key"]], source[mapping["@target-key"]]))
primary_asset = find_object_by_name(existing_objects, source[mapping["@key"]])
secondary_asset = find_object_by_name(existing_objects, source[mapping["@target-key"]])
if primary_asset is None:
logger.info("There is no dependent asset(%s) in FS." % source[mapping["@key"]])
continue
if secondary_asset is None:
logger.info("There is no dependency asset(%s) in FS." % source[mapping["@target-key"]])
continue
relationships = freshservice.get_relationships_by_id(primary_asset["display_id"])
remove_relationship = None
for relationship in relationships:
if relationship["relationship_type_id"] == relationship_type["id"]:
if relationship["secondary_id"] == secondary_asset["display_id"]:
remove_relationship = relationship
break
if remove_relationship is None:
logger.info("There is no relationship in FS.")
continue
freshservice.detach_relationship(remove_relationship["id"])
logger.info("detached relationship %d" % remove_relationship["id"])
except Exception as e:
log = "Error (%s) deleting relationship %s" % (str(e), source[mapping["@key"]])
logger.exception(log)
def create_relationships_from_business_app(sources, _target, mapping):
create_relationships_from_affinity_group(sources, _target, mapping)
def delete_relationships_from_business_app(sources, _target, mapping):
global freshservice
logger.info("Getting all existing devices in FS.")
existing_objects = freshservice.request("api/v2/assets" + "?include=type_fields", "GET", _target["@model"])
logger.info("finished getting all existing devices in FS.")
logger.info("Getting relationship type in FS.")
relationship_type = freshservice.get_relationship_type_by_content(mapping["@downstream-relationship"],
mapping["@upstream-relationship"])
logger.info("finished getting relationship type in FS.")
if relationship_type is None:
log = "There is no relationship type in FS. (%s - %s)" % (
mapping["@downstream-relationship"], mapping["@upstream-relationship"])
logger.info(log)
return
for existing_object in existing_objects:
try:
logger.info("Checking relationship of asset(%s)." % existing_object["name"])
relationships = freshservice.get_relationships_by_id(existing_object["display_id"])
for relationship in relationships:
if relationship["relationship_type_id"] == relationship_type["id"] and \
relationship["primary_id"] == existing_object["display_id"]:
remove_relationship = relationship
target_display_id = relationship["secondary_id"]
for source in sources:
if source[mapping["@key"]] == existing_object["name"]:
secondary_asset = find_object_by_name(existing_objects, source[mapping["@target-key"]])
if target_display_id == secondary_asset["display_id"]:
remove_relationship = None
break
if remove_relationship is None:
continue
freshservice.detach_relationship(remove_relationship["id"])
logger.info("detached relationship %d" % remove_relationship["id"])
except Exception as e:
log = "Error (%s) deleting relationship %s" % (str(e), existing_object[mapping["@key"]])
logger.exception(log)
def update_contracts_from_server(sources, _target, mapping):
global freshservice
logger.info("Getting all existing contracts in FS.")
existing_objects_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing contracts in FS.")
fs_cache["contracts"] = existing_objects_map
if "softwares" in fs_cache:
logger.info("Getting all existing softwares in FS from cache.")
existing_softwares_map = fs_cache["softwares"]
logger.info("finished getting all existing softwares in FS from cache.")
else:
logger.info("Getting all existing softwares in FS.")
existing_softwares_map = freshservice.get_objects_map(_target["@path"], _target["@model"])
logger.info("finished getting all existing softwares in FS.")
fs_cache["softwares"] = existing_softwares_map
for source in sources:
error_skip = False
while True:
try:
existing_object = find_object_in_map(existing_objects_map, source["name"])
data = dict()
if not default_approver:
raise Exception("The approver is required.")
data['approver_id'] = int(default_approver)
# validation
for map_info in mapping["field"]:
if error_skip and "@error-skip" in map_info and map_info["@error-skip"]:
continue
if "@target-foregin-key" in map_info and map_info["@target-foregin"] == "applications":
existing_software = find_object_in_map(existing_softwares_map, source[map_info["@resource"]])
value = existing_software["id"]
else:
value = get_map_value_from_device42(source, map_info)
if "@target-sub-key" not in map_info:
data[map_info["@target"]] = value
else: # Item cost detail attributes
if map_info["@target"] not in data:
data[map_info["@target"]] = [{
map_info["@target-sub-key"]: value
}]
else:
data[map_info["@target"]][0][map_info["@target-sub-key"]] = value
is_valid = True
if value is not None and "@min-length" in map_info and len(value) < map_info["@min-length"]:
is_valid = False
if value == "" and "@set-space" in map_info and map_info["@set-space"]:
is_valid = True
value = " " * map_info["@min-length"]
# value might have been translated to an associated ID in Freshservice by get_map_value_from_device42
# which is why we need to check that value is a string using isinstance.
if value is not None and "@max-length" in map_info and isinstance(value, str) and len(value) > map_info["@max-length"]:
value = value[0:map_info["@max-length"] - 3] + "..."
if value is None and "@not-null" in map_info and map_info["@not-null"]:
# There is an issue with the Freshservice API where sending a null value for
# a field will result in the API returning an error like "Has 0 characters,
# it should have minimum of 1 characters and can have maximum of 255 characters".
# This prevents us from being able to clear these field values in Freshservice (even though
# the Freshservice UI allows you to clear these fields). To get around this, we will send
# a single space for string values and a 0 for integer and float values when the value
# coming from D42 is null.
if "@target-type" in map_info:
target_type = map_info["@target-type"]
if target_type == "integer" or target_type == "float":
value = 0
elif target_type == "date":
value = None
is_valid = False
else:
value = " "
else:
value = " "
if value == 0 and "@not-zero" in map_info and map_info["@not-zero"]:
# Some fields in Freshservice do not allow a 0 value.
# D42 does allow a 0 for the value being synced over,
# so when we try to sync that data to Freshservice,
# the API returns an error like "It should be a Positive Number
# less than or equal to 99999999.99" when we send a 0 value to a field
# that does not accept 0.
# To get around this, we will send 1 for integer values and a 0.01
# for float values when the value coming from D42 is 0.
if "@target-type" in map_info:
target_type = map_info["@target-type"]
if target_type == "integer":
value = 1
elif target_type == "float":
value = 0.01
else:
value = None
is_valid = False
else:
value = 0.01
if "@target-foregin-key" in map_info and value is not None and isinstance(value, str):
value = get_map_value_from_device42(source, map_info, True)
is_valid = value is not None
if "@target-type" in map_info and value is not None:
target_type = map_info["@target-type"]
if target_type == "integer":
try:
value = int(value)
except Exception as e:
logger.exception(str(e))
is_valid = False