%PDF- %PDF-
Direktori : /lib/python3/dist-packages/samba/kcc/ |
Current File : //lib/python3/dist-packages/samba/kcc/kcc_utils.py |
# KCC topology utilities # # Copyright (C) Dave Craft 2011 # Copyright (C) Jelmer Vernooij 2011 # Copyright (C) Andrew Bartlett 2015 # # Andrew Bartlett's alleged work performed by his underlings Douglas # Bagnall and Garming Sam. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import sys import ldb import uuid from samba import dsdb from samba.dcerpc import ( drsblobs, drsuapi, misc, ) from samba.samdb import dsdb_Dn from samba.ndr import ndr_unpack, ndr_pack from collections import Counter class KCCError(Exception): pass class NCType(object): (unknown, schema, domain, config, application) = range(0, 5) # map the NCType enum to strings for debugging nctype_lut = dict((v, k) for k, v in NCType.__dict__.items() if k[:2] != '__') class NamingContext(object): """Base class for a naming context. Holds the DN, GUID, SID (if available) and type of the DN. Subclasses may inherit from this and specialize """ def __init__(self, nc_dnstr): """Instantiate a NamingContext :param nc_dnstr: NC dn string """ self.nc_dnstr = nc_dnstr self.nc_guid = None self.nc_sid = None self.nc_type = NCType.unknown def __str__(self): '''Debug dump string output of class''' text = "%s:" % (self.__class__.__name__,) +\ "\n\tnc_dnstr=%s" % self.nc_dnstr +\ "\n\tnc_guid=%s" % str(self.nc_guid) if self.nc_sid is None: text = text + "\n\tnc_sid=<absent>" else: text = text + "\n\tnc_sid=<present>" text = text + "\n\tnc_type=%s (%s)" % (nctype_lut[self.nc_type], self.nc_type) return text def load_nc(self, samdb): attrs = ["objectGUID", "objectSid"] try: res = samdb.search(base=self.nc_dnstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e: (enum, estr) = e.args raise KCCError("Unable to find naming context (%s) - (%s)" % (self.nc_dnstr, estr)) msg = res[0] if "objectGUID" in msg: self.nc_guid = misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) if "objectSid" in msg: self.nc_sid = msg["objectSid"][0] assert self.nc_guid is not None def is_config(self): '''Return True if NC is config''' assert self.nc_type != NCType.unknown return self.nc_type == NCType.config def identify_by_basedn(self, samdb): """Given an NC object, identify what type is is thru the samdb basedn strings and NC sid value """ # Invoke loader to initialize guid and more # importantly sid value (sid is used to identify # domain NCs) if self.nc_guid is None: self.load_nc(samdb) # We check against schema and config because they # will be the same for all nTDSDSAs in the forest. # That leaves the domain NCs which can be identified # by sid and application NCs as the last identified if self.nc_dnstr == str(samdb.get_schema_basedn()): self.nc_type = NCType.schema elif self.nc_dnstr == str(samdb.get_config_basedn()): self.nc_type = NCType.config elif self.nc_sid is not None: self.nc_type = NCType.domain else: self.nc_type = NCType.application def identify_by_dsa_attr(self, samdb, attr): """Given an NC which has been discovered thru the nTDSDSA database object, determine what type of NC it is (i.e. schema, config, domain, application) via the use of the schema attribute under which the NC was found. :param attr: attr of nTDSDSA object where NC DN appears """ # If the NC is listed under msDS-HasDomainNCs then # this can only be a domain NC and it is our default # domain for this dsa if attr == "msDS-HasDomainNCs": self.nc_type = NCType.domain # If the NC is listed under hasPartialReplicaNCs # this is only a domain NC elif attr == "hasPartialReplicaNCs": self.nc_type = NCType.domain # NCs listed under hasMasterNCs are either # default domain, schema, or config. We # utilize the identify_by_basedn() to # identify those elif attr == "hasMasterNCs": self.identify_by_basedn(samdb) # Still unknown (unlikely) but for completeness # and for finally identifying application NCs if self.nc_type == NCType.unknown: self.identify_by_basedn(samdb) class NCReplica(NamingContext): """Naming context replica that is relative to a specific DSA. This is a more specific form of NamingContext class (inheriting from that class) and it identifies unique attributes of the DSA's replica for a NC. """ def __init__(self, dsa, nc_dnstr): """Instantiate a Naming Context Replica :param dsa_guid: GUID of DSA where replica appears :param nc_dnstr: NC dn string """ self.rep_dsa_dnstr = dsa.dsa_dnstr self.rep_dsa_guid = dsa.dsa_guid self.rep_default = False # replica for DSA's default domain self.rep_partial = False self.rep_ro = False self.rep_instantiated_flags = 0 self.rep_fsmo_role_owner = None # RepsFromTo tuples self.rep_repsFrom = [] # RepsFromTo tuples self.rep_repsTo = [] # The (is present) test is a combination of being # enumerated in (hasMasterNCs or msDS-hasFullReplicaNCs or # hasPartialReplicaNCs) as well as its replica flags found # thru the msDS-HasInstantiatedNCs. If the NC replica meets # the first enumeration test then this flag is set true self.rep_present_criteria_one = False # Call my super class we inherited from NamingContext.__init__(self, nc_dnstr) def __str__(self): '''Debug dump string output of class''' text = "%s:" % self.__class__.__name__ +\ "\n\tdsa_dnstr=%s" % self.rep_dsa_dnstr +\ "\n\tdsa_guid=%s" % self.rep_dsa_guid +\ "\n\tdefault=%s" % self.rep_default +\ "\n\tro=%s" % self.rep_ro +\ "\n\tpartial=%s" % self.rep_partial +\ "\n\tpresent=%s" % self.is_present() +\ "\n\tfsmo_role_owner=%s" % self.rep_fsmo_role_owner +\ "".join("\n%s" % rep for rep in self.rep_repsFrom) +\ "".join("\n%s" % rep for rep in self.rep_repsTo) return "%s\n%s" % (NamingContext.__str__(self), text) def set_instantiated_flags(self, flags=0): '''Set or clear NC replica instantiated flags''' self.rep_instantiated_flags = flags def identify_by_dsa_attr(self, samdb, attr): """Given an NC which has been discovered thru the nTDSDSA database object, determine what type of NC replica it is (i.e. partial, read only, default) :param attr: attr of nTDSDSA object where NC DN appears """ # If the NC was found under hasPartialReplicaNCs # then a partial replica at this dsa if attr == "hasPartialReplicaNCs": self.rep_partial = True self.rep_present_criteria_one = True # If the NC is listed under msDS-HasDomainNCs then # this can only be a domain NC and it is the DSA's # default domain NC elif attr == "msDS-HasDomainNCs": self.rep_default = True # NCs listed under hasMasterNCs are either # default domain, schema, or config. We check # against schema and config because they will be # the same for all nTDSDSAs in the forest. That # leaves the default domain NC remaining which # may be different for each nTDSDSAs (and thus # we don't compare against this samdb's default # basedn elif attr == "hasMasterNCs": self.rep_present_criteria_one = True if self.nc_dnstr != str(samdb.get_schema_basedn()) and \ self.nc_dnstr != str(samdb.get_config_basedn()): self.rep_default = True # RODC only elif attr == "msDS-hasFullReplicaNCs": self.rep_present_criteria_one = True self.rep_ro = True # Not RODC elif attr == "msDS-hasMasterNCs": self.rep_present_criteria_one = True self.rep_ro = False # Now use this DSA attribute to identify the naming # context type by calling the super class method # of the same name NamingContext.identify_by_dsa_attr(self, samdb, attr) def is_default(self): """Whether this is a default domain for the dsa that this NC appears on """ return self.rep_default def is_ro(self): '''Return True if NC replica is read only''' return self.rep_ro def is_partial(self): '''Return True if NC replica is partial''' return self.rep_partial def is_present(self): """Given an NC replica which has been discovered thru the nTDSDSA database object and populated with replica flags from the msDS-HasInstantiatedNCs; return whether the NC replica is present (true) or if the IT_NC_GOING flag is set then the NC replica is not present (false) """ if self.rep_present_criteria_one and \ self.rep_instantiated_flags & dsdb.INSTANCE_TYPE_NC_GOING == 0: return True return False def load_repsFrom(self, samdb): """Given an NC replica which has been discovered thru the nTDSDSA database object, load the repsFrom attribute for the local replica. held by my dsa. The repsFrom attribute is not replicated so this attribute is relative only to the local DSA that the samdb exists on """ try: res = samdb.search(base=self.nc_dnstr, scope=ldb.SCOPE_BASE, attrs=["repsFrom"]) except ldb.LdbError as e1: (enum, estr) = e1.args raise KCCError("Unable to find NC for (%s) - (%s)" % (self.nc_dnstr, estr)) msg = res[0] # Possibly no repsFrom if this is a singleton DC if "repsFrom" in msg: for value in msg["repsFrom"]: try: unpacked = ndr_unpack(drsblobs.repsFromToBlob, value) except RuntimeError as e: print("bad repsFrom NDR: %r" % (value), file=sys.stderr) continue rep = RepsFromTo(self.nc_dnstr, unpacked) self.rep_repsFrom.append(rep) def commit_repsFrom(self, samdb, ro=False): """Commit repsFrom to the database""" # XXX - This is not truly correct according to the MS-TECH # docs. To commit a repsFrom we should be using RPCs # IDL_DRSReplicaAdd, IDL_DRSReplicaModify, and # IDL_DRSReplicaDel to affect a repsFrom change. # # Those RPCs are missing in samba, so I'll have to # implement them to get this to more accurately # reflect the reference docs. As of right now this # commit to the database will work as its what the # older KCC also did modify = False newreps = [] delreps = [] for repsFrom in self.rep_repsFrom: # Leave out any to be deleted from # replacement list. Build a list # of to be deleted reps which we will # remove from rep_repsFrom list below if repsFrom.to_be_deleted: delreps.append(repsFrom) modify = True continue if repsFrom.is_modified(): repsFrom.set_unmodified() modify = True # current (unmodified) elements also get # appended here but no changes will occur # unless something is "to be modified" or # "to be deleted" newreps.append(ndr_pack(repsFrom.ndr_blob)) # Now delete these from our list of rep_repsFrom for repsFrom in delreps: self.rep_repsFrom.remove(repsFrom) delreps = [] # Nothing to do if no reps have been modified or # need to be deleted or input option has informed # us to be "readonly" (ro). Leave database # record "as is" if not modify or ro: return m = ldb.Message() m.dn = ldb.Dn(samdb, self.nc_dnstr) m["repsFrom"] = \ ldb.MessageElement(newreps, ldb.FLAG_MOD_REPLACE, "repsFrom") try: samdb.modify(m) except ldb.LdbError as estr: raise KCCError("Could not set repsFrom for (%s) - (%s)" % (self.nc_dnstr, estr)) def load_replUpToDateVector(self, samdb): """Given an NC replica which has been discovered thru the nTDSDSA database object, load the replUpToDateVector attribute for the local replica. held by my dsa. The replUpToDateVector attribute is not replicated so this attribute is relative only to the local DSA that the samdb exists on """ try: res = samdb.search(base=self.nc_dnstr, scope=ldb.SCOPE_BASE, attrs=["replUpToDateVector"]) except ldb.LdbError as e2: (enum, estr) = e2.args raise KCCError("Unable to find NC for (%s) - (%s)" % (self.nc_dnstr, estr)) msg = res[0] # Possibly no replUpToDateVector if this is a singleton DC if "replUpToDateVector" in msg: value = msg["replUpToDateVector"][0] blob = ndr_unpack(drsblobs.replUpToDateVectorBlob, value) if blob.version != 2: # Samba only generates version 2, and this runs locally raise AttributeError("Unexpected replUpToDateVector version %d" % blob.version) self.rep_replUpToDateVector_cursors = blob.ctr.cursors else: self.rep_replUpToDateVector_cursors = [] def dumpstr_to_be_deleted(self): return '\n'.join(str(x) for x in self.rep_repsFrom if x.to_be_deleted) def dumpstr_to_be_modified(self): return '\n'.join(str(x) for x in self.rep_repsFrom if x.is_modified()) def load_fsmo_roles(self, samdb): """Given an NC replica which has been discovered thru the nTDSDSA database object, load the fSMORoleOwner attribute. """ try: res = samdb.search(base=self.nc_dnstr, scope=ldb.SCOPE_BASE, attrs=["fSMORoleOwner"]) except ldb.LdbError as e3: (enum, estr) = e3.args raise KCCError("Unable to find NC for (%s) - (%s)" % (self.nc_dnstr, estr)) msg = res[0] # Possibly no fSMORoleOwner if "fSMORoleOwner" in msg: self.rep_fsmo_role_owner = msg["fSMORoleOwner"] def is_fsmo_role_owner(self, dsa_dnstr): if self.rep_fsmo_role_owner is not None and \ self.rep_fsmo_role_owner == dsa_dnstr: return True return False def load_repsTo(self, samdb): """Given an NC replica which has been discovered thru the nTDSDSA database object, load the repsTo attribute for the local replica. held by my dsa. The repsTo attribute is not replicated so this attribute is relative only to the local DSA that the samdb exists on This is responsible for push replication, not scheduled pull replication. Not to be confused for repsFrom. """ try: res = samdb.search(base=self.nc_dnstr, scope=ldb.SCOPE_BASE, attrs=["repsTo"]) except ldb.LdbError as e4: (enum, estr) = e4.args raise KCCError("Unable to find NC for (%s) - (%s)" % (self.nc_dnstr, estr)) msg = res[0] # Possibly no repsTo if this is a singleton DC if "repsTo" in msg: for value in msg["repsTo"]: try: unpacked = ndr_unpack(drsblobs.repsFromToBlob, value) except RuntimeError as e: print("bad repsTo NDR: %r" % (value), file=sys.stderr) continue rep = RepsFromTo(self.nc_dnstr, unpacked) self.rep_repsTo.append(rep) def commit_repsTo(self, samdb, ro=False): """Commit repsTo to the database""" # XXX - This is not truly correct according to the MS-TECH # docs. To commit a repsTo we should be using RPCs # IDL_DRSReplicaAdd, IDL_DRSReplicaModify, and # IDL_DRSReplicaDel to affect a repsTo change. # # Those RPCs are missing in samba, so I'll have to # implement them to get this to more accurately # reflect the reference docs. As of right now this # commit to the database will work as its what the # older KCC also did modify = False newreps = [] delreps = [] for repsTo in self.rep_repsTo: # Leave out any to be deleted from # replacement list. Build a list # of to be deleted reps which we will # remove from rep_repsTo list below if repsTo.to_be_deleted: delreps.append(repsTo) modify = True continue if repsTo.is_modified(): repsTo.set_unmodified() modify = True # current (unmodified) elements also get # appended here but no changes will occur # unless something is "to be modified" or # "to be deleted" newreps.append(ndr_pack(repsTo.ndr_blob)) # Now delete these from our list of rep_repsTo for repsTo in delreps: self.rep_repsTo.remove(repsTo) delreps = [] # Nothing to do if no reps have been modified or # need to be deleted or input option has informed # us to be "readonly" (ro). Leave database # record "as is" if not modify or ro: return m = ldb.Message() m.dn = ldb.Dn(samdb, self.nc_dnstr) m["repsTo"] = \ ldb.MessageElement(newreps, ldb.FLAG_MOD_REPLACE, "repsTo") try: samdb.modify(m) except ldb.LdbError as estr: raise KCCError("Could not set repsTo for (%s) - (%s)" % (self.nc_dnstr, estr)) class DirectoryServiceAgent(object): def __init__(self, dsa_dnstr): """Initialize DSA class. Class is subsequently fully populated by calling the load_dsa() method :param dsa_dnstr: DN of the nTDSDSA """ self.dsa_dnstr = dsa_dnstr self.dsa_guid = None self.dsa_ivid = None self.dsa_is_ro = False self.dsa_is_istg = False self.options = 0 self.dsa_behavior = 0 self.default_dnstr = None # default domain dn string for dsa # NCReplicas for this dsa that are "present" # Indexed by DN string of naming context self.current_rep_table = {} # NCReplicas for this dsa that "should be present" # Indexed by DN string of naming context self.needed_rep_table = {} # NTDSConnections for this dsa. These are current # valid connections that are committed or pending a commit # in the database. Indexed by DN string of connection self.connect_table = {} def __str__(self): '''Debug dump string output of class''' text = "%s:" % self.__class__.__name__ if self.dsa_dnstr is not None: text = text + "\n\tdsa_dnstr=%s" % self.dsa_dnstr if self.dsa_guid is not None: text = text + "\n\tdsa_guid=%s" % str(self.dsa_guid) if self.dsa_ivid is not None: text = text + "\n\tdsa_ivid=%s" % str(self.dsa_ivid) text += "\n\tro=%s" % self.is_ro() +\ "\n\tgc=%s" % self.is_gc() +\ "\n\tistg=%s" % self.is_istg() +\ "\ncurrent_replica_table:" +\ "\n%s" % self.dumpstr_current_replica_table() +\ "\nneeded_replica_table:" +\ "\n%s" % self.dumpstr_needed_replica_table() +\ "\nconnect_table:" +\ "\n%s" % self.dumpstr_connect_table() return text def get_current_replica(self, nc_dnstr): return self.current_rep_table.get(nc_dnstr) def is_istg(self): '''Returns True if dsa is intersite topology generator for it's site''' # The KCC on an RODC always acts as an ISTG for itself return self.dsa_is_istg or self.dsa_is_ro def is_ro(self): '''Returns True if dsa a read only domain controller''' return self.dsa_is_ro def is_gc(self): '''Returns True if dsa hosts a global catalog''' if (self.options & dsdb.DS_NTDSDSA_OPT_IS_GC) != 0: return True return False def is_minimum_behavior(self, version): """Is dsa at minimum windows level greater than or equal to (version) :param version: Windows version to test against (e.g. DS_DOMAIN_FUNCTION_2008) """ if self.dsa_behavior >= version: return True return False def is_translate_ntdsconn_disabled(self): """Whether this allows NTDSConnection translation in its options.""" if (self.options & dsdb.DS_NTDSDSA_OPT_DISABLE_NTDSCONN_XLATE) != 0: return True return False def get_rep_tables(self): """Return DSA current and needed replica tables """ return self.current_rep_table, self.needed_rep_table def get_parent_dnstr(self): """Get the parent DN string of this object.""" head, sep, tail = self.dsa_dnstr.partition(',') return tail def load_dsa(self, samdb): """Load a DSA from the samdb. Prior initialization has given us the DN of the DSA that we are to load. This method initializes all other attributes, including loading the NC replica table for this DSA. """ attrs = ["objectGUID", "invocationID", "options", "msDS-isRODC", "msDS-Behavior-Version"] try: res = samdb.search(base=self.dsa_dnstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e5: (enum, estr) = e5.args raise KCCError("Unable to find nTDSDSA for (%s) - (%s)" % (self.dsa_dnstr, estr)) msg = res[0] self.dsa_guid = misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) # RODCs don't originate changes and thus have no invocationId, # therefore we must check for existence first if "invocationId" in msg: self.dsa_ivid = misc.GUID(samdb.schema_format_value("objectGUID", msg["invocationId"][0])) if "options" in msg: self.options = int(msg["options"][0]) if "msDS-isRODC" in msg and str(msg["msDS-isRODC"][0]) == "TRUE": self.dsa_is_ro = True else: self.dsa_is_ro = False if "msDS-Behavior-Version" in msg: self.dsa_behavior = int(msg['msDS-Behavior-Version'][0]) # Load the NC replicas that are enumerated on this dsa self.load_current_replica_table(samdb) # Load the nTDSConnection that are enumerated on this dsa self.load_connection_table(samdb) def load_current_replica_table(self, samdb): """Method to load the NC replica's listed for DSA object. This method queries the samdb for (hasMasterNCs, msDS-hasMasterNCs, hasPartialReplicaNCs, msDS-HasDomainNCs, msDS-hasFullReplicaNCs, and msDS-HasInstantiatedNCs) to determine complete list of NC replicas that are enumerated for the DSA. Once a NC replica is loaded it is identified (schema, config, etc) and the other replica attributes (partial, ro, etc) are determined. :param samdb: database to query for DSA replica list """ ncattrs = [ # not RODC - default, config, schema (old style) "hasMasterNCs", # not RODC - default, config, schema, app NCs "msDS-hasMasterNCs", # domain NC partial replicas "hasPartialReplicaNCs", # default domain NC "msDS-HasDomainNCs", # RODC only - default, config, schema, app NCs "msDS-hasFullReplicaNCs", # Identifies if replica is coming, going, or stable "msDS-HasInstantiatedNCs" ] try: res = samdb.search(base=self.dsa_dnstr, scope=ldb.SCOPE_BASE, attrs=ncattrs) except ldb.LdbError as e6: (enum, estr) = e6.args raise KCCError("Unable to find nTDSDSA NCs for (%s) - (%s)" % (self.dsa_dnstr, estr)) # The table of NCs for the dsa we are searching tmp_table = {} # We should get one response to our query here for # the ntds that we requested if len(res[0]) > 0: # Our response will contain a number of elements including # the dn of the dsa as well as elements for each # attribute (e.g. hasMasterNCs). Each of these elements # is a dictionary list which we retrieve the keys for and # then iterate over them for k in res[0].keys(): if k == "dn": continue # For each attribute type there will be one or more DNs # listed. For instance DCs normally have 3 hasMasterNCs # listed. for value in res[0][k]: # Turn dn into a dsdb_Dn so we can use # its methods to parse a binary DN dsdn = dsdb_Dn(samdb, value.decode('utf8')) flags = dsdn.get_binary_integer() dnstr = str(dsdn.dn) if dnstr not in tmp_table: rep = NCReplica(self, dnstr) tmp_table[dnstr] = rep else: rep = tmp_table[dnstr] if k == "msDS-HasInstantiatedNCs": rep.set_instantiated_flags(flags) continue rep.identify_by_dsa_attr(samdb, k) # if we've identified the default domain NC # then save its DN string if rep.is_default(): self.default_dnstr = dnstr else: raise KCCError("No nTDSDSA NCs for (%s)" % self.dsa_dnstr) # Assign our newly built NC replica table to this dsa self.current_rep_table = tmp_table def add_needed_replica(self, rep): """Method to add a NC replica that "should be present" to the needed_rep_table. """ self.needed_rep_table[rep.nc_dnstr] = rep def load_connection_table(self, samdb): """Method to load the nTDSConnections listed for DSA object. :param samdb: database to query for DSA connection list """ try: res = samdb.search(base=self.dsa_dnstr, scope=ldb.SCOPE_SUBTREE, expression="(objectClass=nTDSConnection)") except ldb.LdbError as e7: (enum, estr) = e7.args raise KCCError("Unable to find nTDSConnection for (%s) - (%s)" % (self.dsa_dnstr, estr)) for msg in res: dnstr = str(msg.dn) # already loaded if dnstr in self.connect_table: continue connect = NTDSConnection(dnstr) connect.load_connection(samdb) self.connect_table[dnstr] = connect def commit_connections(self, samdb, ro=False): """Method to commit any uncommitted nTDSConnections modifications that are in our table. These would be identified connections that are marked to be added or deleted :param samdb: database to commit DSA connection list to :param ro: if (true) then perform internal operations but do not write to the database (readonly) """ delconn = [] for dnstr, connect in self.connect_table.items(): if connect.to_be_added: connect.commit_added(samdb, ro) if connect.to_be_modified: connect.commit_modified(samdb, ro) if connect.to_be_deleted: connect.commit_deleted(samdb, ro) delconn.append(dnstr) # Now delete the connection from the table for dnstr in delconn: del self.connect_table[dnstr] def add_connection(self, dnstr, connect): assert dnstr not in self.connect_table self.connect_table[dnstr] = connect def get_connection_by_from_dnstr(self, from_dnstr): """Scan DSA nTDSConnection table and return connection with a "fromServer" dn string equivalent to method input parameter. :param from_dnstr: search for this from server entry """ answer = [] for connect in self.connect_table.values(): if connect.get_from_dnstr() == from_dnstr: answer.append(connect) return answer def dumpstr_current_replica_table(self): '''Debug dump string output of current replica table''' return '\n'.join(str(x) for x in self.current_rep_table) def dumpstr_needed_replica_table(self): '''Debug dump string output of needed replica table''' return '\n'.join(str(x) for x in self.needed_rep_table) def dumpstr_connect_table(self): '''Debug dump string output of connect table''' return '\n'.join(str(x) for x in self.connect_table) def new_connection(self, options, system_flags, transport, from_dnstr, sched): """Set up a new connection for the DSA based on input parameters. Connection will be added to the DSA connect_table and will be marked as "to be added" pending a call to commit_connections() """ dnstr = "CN=%s," % str(uuid.uuid4()) + self.dsa_dnstr connect = NTDSConnection(dnstr) connect.to_be_added = True connect.enabled = True connect.from_dnstr = from_dnstr connect.options = options connect.system_flags = system_flags if transport is not None: connect.transport_dnstr = transport.dnstr connect.transport_guid = transport.guid if sched is not None: connect.schedule = sched else: # Create schedule. Attribute value set according to MS-TECH # intra-site connection creation document connect.schedule = new_connection_schedule() self.add_connection(dnstr, connect) return connect class NTDSConnection(object): """Class defines a nTDSConnection found under a DSA """ def __init__(self, dnstr): self.dnstr = dnstr self.guid = None self.enabled = False self.whenCreated = 0 self.to_be_added = False # new connection needs to be added self.to_be_deleted = False # old connection needs to be deleted self.to_be_modified = False self.options = 0 self.system_flags = 0 self.transport_dnstr = None self.transport_guid = None self.from_dnstr = None self.schedule = None def __str__(self): '''Debug dump string output of NTDSConnection object''' text = "%s:\n\tdn=%s" % (self.__class__.__name__, self.dnstr) +\ "\n\tenabled=%s" % self.enabled +\ "\n\tto_be_added=%s" % self.to_be_added +\ "\n\tto_be_deleted=%s" % self.to_be_deleted +\ "\n\tto_be_modified=%s" % self.to_be_modified +\ "\n\toptions=0x%08X" % self.options +\ "\n\tsystem_flags=0x%08X" % self.system_flags +\ "\n\twhenCreated=%d" % self.whenCreated +\ "\n\ttransport_dn=%s" % self.transport_dnstr if self.guid is not None: text += "\n\tguid=%s" % str(self.guid) if self.transport_guid is not None: text += "\n\ttransport_guid=%s" % str(self.transport_guid) text = text + "\n\tfrom_dn=%s" % self.from_dnstr if self.schedule is not None: text += "\n\tschedule.size=%s" % self.schedule.size +\ "\n\tschedule.bandwidth=%s" % self.schedule.bandwidth +\ ("\n\tschedule.numberOfSchedules=%s" % self.schedule.numberOfSchedules) for i, header in enumerate(self.schedule.headerArray): text += ("\n\tschedule.headerArray[%d].type=%d" % (i, header.type)) +\ ("\n\tschedule.headerArray[%d].offset=%d" % (i, header.offset)) +\ "\n\tschedule.dataArray[%d].slots[ " % i +\ "".join("0x%X " % slot for slot in self.schedule.dataArray[i].slots) +\ "]" return text def load_connection(self, samdb): """Given a NTDSConnection object with an prior initialization for the object's DN, search for the DN and load attributes from the samdb. """ attrs = ["options", "enabledConnection", "schedule", "whenCreated", "objectGUID", "transportType", "fromServer", "systemFlags"] try: res = samdb.search(base=self.dnstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e8: (enum, estr) = e8.args raise KCCError("Unable to find nTDSConnection for (%s) - (%s)" % (self.dnstr, estr)) msg = res[0] if "options" in msg: self.options = int(msg["options"][0]) if "enabledConnection" in msg: if str(msg["enabledConnection"][0]).upper().lstrip().rstrip() == "TRUE": self.enabled = True if "systemFlags" in msg: self.system_flags = int(msg["systemFlags"][0]) try: self.guid = \ misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) except KeyError: raise KCCError("Unable to find objectGUID in nTDSConnection " "for (%s)" % (self.dnstr)) if "transportType" in msg: dsdn = dsdb_Dn(samdb, msg["transportType"][0].decode('utf8')) self.load_connection_transport(samdb, str(dsdn.dn)) if "schedule" in msg: self.schedule = ndr_unpack(drsblobs.schedule, msg["schedule"][0]) if "whenCreated" in msg: self.whenCreated = ldb.string_to_time(str(msg["whenCreated"][0])) if "fromServer" in msg: dsdn = dsdb_Dn(samdb, msg["fromServer"][0].decode('utf8')) self.from_dnstr = str(dsdn.dn) assert self.from_dnstr is not None def load_connection_transport(self, samdb, tdnstr): """Given a NTDSConnection object which enumerates a transport DN, load the transport information for the connection object :param tdnstr: transport DN to load """ attrs = ["objectGUID"] try: res = samdb.search(base=tdnstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e9: (enum, estr) = e9.args raise KCCError("Unable to find transport (%s) - (%s)" % (tdnstr, estr)) if "objectGUID" in res[0]: msg = res[0] self.transport_dnstr = tdnstr self.transport_guid = \ misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) assert self.transport_dnstr is not None assert self.transport_guid is not None def commit_deleted(self, samdb, ro=False): """Local helper routine for commit_connections() which handles committed connections that are to be deleted from the database database """ assert self.to_be_deleted self.to_be_deleted = False # No database modification requested if ro: return try: samdb.delete(self.dnstr) except ldb.LdbError as e10: (enum, estr) = e10.args raise KCCError("Could not delete nTDSConnection for (%s) - (%s)" % (self.dnstr, estr)) def commit_added(self, samdb, ro=False): """Local helper routine for commit_connections() which handles committed connections that are to be added to the database """ assert self.to_be_added self.to_be_added = False # No database modification requested if ro: return # First verify we don't have this entry to ensure nothing # is programmatically amiss found = False try: msg = samdb.search(base=self.dnstr, scope=ldb.SCOPE_BASE) if len(msg) != 0: found = True except ldb.LdbError as e11: (enum, estr) = e11.args if enum != ldb.ERR_NO_SUCH_OBJECT: raise KCCError("Unable to search for (%s) - (%s)" % (self.dnstr, estr)) if found: raise KCCError("nTDSConnection for (%s) already exists!" % self.dnstr) if self.enabled: enablestr = "TRUE" else: enablestr = "FALSE" # Prepare a message for adding to the samdb m = ldb.Message() m.dn = ldb.Dn(samdb, self.dnstr) m["objectClass"] = \ ldb.MessageElement("nTDSConnection", ldb.FLAG_MOD_ADD, "objectClass") m["showInAdvancedViewOnly"] = \ ldb.MessageElement("TRUE", ldb.FLAG_MOD_ADD, "showInAdvancedViewOnly") m["enabledConnection"] = \ ldb.MessageElement(enablestr, ldb.FLAG_MOD_ADD, "enabledConnection") m["fromServer"] = \ ldb.MessageElement(self.from_dnstr, ldb.FLAG_MOD_ADD, "fromServer") m["options"] = \ ldb.MessageElement(str(self.options), ldb.FLAG_MOD_ADD, "options") m["systemFlags"] = \ ldb.MessageElement(str(self.system_flags), ldb.FLAG_MOD_ADD, "systemFlags") if self.transport_dnstr is not None: m["transportType"] = \ ldb.MessageElement(str(self.transport_dnstr), ldb.FLAG_MOD_ADD, "transportType") if self.schedule is not None: m["schedule"] = \ ldb.MessageElement(ndr_pack(self.schedule), ldb.FLAG_MOD_ADD, "schedule") try: samdb.add(m) except ldb.LdbError as e12: (enum, estr) = e12.args raise KCCError("Could not add nTDSConnection for (%s) - (%s)" % (self.dnstr, estr)) def commit_modified(self, samdb, ro=False): """Local helper routine for commit_connections() which handles committed connections that are to be modified to the database """ assert self.to_be_modified self.to_be_modified = False # No database modification requested if ro: return # First verify we have this entry to ensure nothing # is programmatically amiss try: # we don't use the search result, but it tests the status # of self.dnstr in the database. samdb.search(base=self.dnstr, scope=ldb.SCOPE_BASE) except ldb.LdbError as e13: (enum, estr) = e13.args if enum == ldb.ERR_NO_SUCH_OBJECT: raise KCCError("nTDSConnection for (%s) doesn't exist!" % self.dnstr) raise KCCError("Unable to search for (%s) - (%s)" % (self.dnstr, estr)) if self.enabled: enablestr = "TRUE" else: enablestr = "FALSE" # Prepare a message for modifying the samdb m = ldb.Message() m.dn = ldb.Dn(samdb, self.dnstr) m["enabledConnection"] = \ ldb.MessageElement(enablestr, ldb.FLAG_MOD_REPLACE, "enabledConnection") m["fromServer"] = \ ldb.MessageElement(self.from_dnstr, ldb.FLAG_MOD_REPLACE, "fromServer") m["options"] = \ ldb.MessageElement(str(self.options), ldb.FLAG_MOD_REPLACE, "options") m["systemFlags"] = \ ldb.MessageElement(str(self.system_flags), ldb.FLAG_MOD_REPLACE, "systemFlags") if self.transport_dnstr is not None: m["transportType"] = \ ldb.MessageElement(str(self.transport_dnstr), ldb.FLAG_MOD_REPLACE, "transportType") else: m["transportType"] = \ ldb.MessageElement([], ldb.FLAG_MOD_DELETE, "transportType") if self.schedule is not None: m["schedule"] = \ ldb.MessageElement(ndr_pack(self.schedule), ldb.FLAG_MOD_REPLACE, "schedule") else: m["schedule"] = \ ldb.MessageElement([], ldb.FLAG_MOD_DELETE, "schedule") try: samdb.modify(m) except ldb.LdbError as e14: (enum, estr) = e14.args raise KCCError("Could not modify nTDSConnection for (%s) - (%s)" % (self.dnstr, estr)) def set_modified(self, truefalse): self.to_be_modified = truefalse def is_schedule_minimum_once_per_week(self): """Returns True if our schedule includes at least one replication interval within the week. False otherwise """ # replinfo schedule is None means "always", while # NTDSConnection schedule is None means "never". if self.schedule is None or self.schedule.dataArray[0] is None: return False for slot in self.schedule.dataArray[0].slots: if (slot & 0x0F) != 0x0: return True return False def is_equivalent_schedule(self, sched): """Returns True if our schedule is equivalent to the input comparison schedule. :param shed: schedule to compare to """ # There are 4 cases, where either self.schedule or sched can be None # # | self. is None | self. is not None # --------------+-----------------+-------------------- # sched is None | True | False # --------------+-----------------+-------------------- # sched is not None | False | do calculations if self.schedule is None: return sched is None if sched is None: return False if ((self.schedule.size != sched.size or self.schedule.bandwidth != sched.bandwidth or self.schedule.numberOfSchedules != sched.numberOfSchedules)): return False for i, header in enumerate(self.schedule.headerArray): if self.schedule.headerArray[i].type != sched.headerArray[i].type: return False if self.schedule.headerArray[i].offset != \ sched.headerArray[i].offset: return False for a, b in zip(self.schedule.dataArray[i].slots, sched.dataArray[i].slots): if a != b: return False return True def is_rodc_topology(self): """Returns True if NTDS Connection specifies RODC topology only """ if self.options & dsdb.NTDSCONN_OPT_RODC_TOPOLOGY == 0: return False return True def is_generated(self): """Returns True if NTDS Connection was generated by the KCC topology algorithm as opposed to set by the administrator """ if self.options & dsdb.NTDSCONN_OPT_IS_GENERATED == 0: return False return True def is_override_notify_default(self): """Returns True if NTDS Connection should override notify default """ if self.options & dsdb.NTDSCONN_OPT_OVERRIDE_NOTIFY_DEFAULT == 0: return False return True def is_use_notify(self): """Returns True if NTDS Connection should use notify """ if self.options & dsdb.NTDSCONN_OPT_USE_NOTIFY == 0: return False return True def is_twoway_sync(self): """Returns True if NTDS Connection should use twoway sync """ if self.options & dsdb.NTDSCONN_OPT_TWOWAY_SYNC == 0: return False return True def is_intersite_compression_disabled(self): """Returns True if NTDS Connection intersite compression is disabled """ if self.options & dsdb.NTDSCONN_OPT_DISABLE_INTERSITE_COMPRESSION == 0: return False return True def is_user_owned_schedule(self): """Returns True if NTDS Connection has a user owned schedule """ if self.options & dsdb.NTDSCONN_OPT_USER_OWNED_SCHEDULE == 0: return False return True def is_enabled(self): """Returns True if NTDS Connection is enabled """ return self.enabled def get_from_dnstr(self): '''Return fromServer dn string attribute''' return self.from_dnstr class Partition(NamingContext): """A naming context discovered thru Partitions DN of the config schema. This is a more specific form of NamingContext class (inheriting from that class) and it identifies unique attributes enumerated in the Partitions such as which nTDSDSAs are cross referenced for replicas """ def __init__(self, partstr): self.partstr = partstr self.enabled = True self.system_flags = 0 self.rw_location_list = [] self.ro_location_list = [] # We don't have enough info to properly # fill in the naming context yet. We'll get that # fully set up with load_partition(). NamingContext.__init__(self, None) def load_partition(self, samdb): """Given a Partition class object that has been initialized with its partition dn string, load the partition from the sam database, identify the type of the partition (schema, domain, etc) and record the list of nTDSDSAs that appear in the cross reference attributes msDS-NC-Replica-Locations and msDS-NC-RO-Replica-Locations. :param samdb: sam database to load partition from """ attrs = ["nCName", "Enabled", "systemFlags", "msDS-NC-Replica-Locations", "msDS-NC-RO-Replica-Locations"] try: res = samdb.search(base=self.partstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e15: (enum, estr) = e15.args raise KCCError("Unable to find partition for (%s) - (%s)" % (self.partstr, estr)) msg = res[0] for k in msg.keys(): if k == "dn": continue if k == "Enabled": if str(msg[k][0]).upper().lstrip().rstrip() == "TRUE": self.enabled = True else: self.enabled = False continue if k == "systemFlags": self.system_flags = int(msg[k][0]) continue for value in msg[k]: dsdn = dsdb_Dn(samdb, value.decode('utf8')) dnstr = str(dsdn.dn) if k == "nCName": self.nc_dnstr = dnstr continue if k == "msDS-NC-Replica-Locations": self.rw_location_list.append(dnstr) continue if k == "msDS-NC-RO-Replica-Locations": self.ro_location_list.append(dnstr) continue # Now identify what type of NC this partition # enumerated self.identify_by_basedn(samdb) def is_enabled(self): """Returns True if partition is enabled """ return self.is_enabled def is_foreign(self): """Returns True if this is not an Active Directory NC in our forest but is instead something else (e.g. a foreign NC) """ if (self.system_flags & dsdb.SYSTEM_FLAG_CR_NTDS_NC) == 0: return True else: return False def should_be_present(self, target_dsa): """Tests whether this partition should have an NC replica on the target dsa. This method returns a tuple of needed=True/False, ro=True/False, partial=True/False :param target_dsa: should NC be present on target dsa """ ro = False partial = False # If this is the config, schema, or default # domain NC for the target dsa then it should # be present needed = (self.nc_type == NCType.config or self.nc_type == NCType.schema or (self.nc_type == NCType.domain and self.nc_dnstr == target_dsa.default_dnstr)) # A writable replica of an application NC should be present # if there a cross reference to the target DSA exists. Depending # on whether the DSA is ro we examine which type of cross reference # to look for (msDS-NC-Replica-Locations or # msDS-NC-RO-Replica-Locations if self.nc_type == NCType.application: if target_dsa.is_ro(): if target_dsa.dsa_dnstr in self.ro_location_list: needed = True else: if target_dsa.dsa_dnstr in self.rw_location_list: needed = True # If the target dsa is a gc then a partial replica of a # domain NC (other than the DSAs default domain) should exist # if there is also a cross reference for the DSA if (target_dsa.is_gc() and self.nc_type == NCType.domain and self.nc_dnstr != target_dsa.default_dnstr and (target_dsa.dsa_dnstr in self.ro_location_list or target_dsa.dsa_dnstr in self.rw_location_list)): needed = True partial = True # partial NCs are always readonly if needed and (target_dsa.is_ro() or partial): ro = True return needed, ro, partial def __str__(self): '''Debug dump string output of class''' text = "%s" % NamingContext.__str__(self) +\ "\n\tpartdn=%s" % self.partstr +\ "".join("\n\tmsDS-NC-Replica-Locations=%s" % k for k in self.rw_location_list) +\ "".join("\n\tmsDS-NC-RO-Replica-Locations=%s" % k for k in self.ro_location_list) return text class Site(object): """An individual site object discovered thru the configuration naming context. Contains all DSAs that exist within the site """ def __init__(self, site_dnstr, nt_now): self.site_dnstr = site_dnstr self.site_guid = None self.site_options = 0 self.site_topo_generator = None self.site_topo_failover = 0 # appears to be in minutes self.dsa_table = {} self.rw_dsa_table = {} self.nt_now = nt_now def load_site(self, samdb): """Loads the NTDS Site Settings options attribute for the site as well as querying and loading all DSAs that appear within the site. """ ssdn = "CN=NTDS Site Settings,%s" % self.site_dnstr attrs = ["options", "interSiteTopologyFailover", "interSiteTopologyGenerator"] try: res = samdb.search(base=ssdn, scope=ldb.SCOPE_BASE, attrs=attrs) self_res = samdb.search(base=self.site_dnstr, scope=ldb.SCOPE_BASE, attrs=['objectGUID']) except ldb.LdbError as e16: (enum, estr) = e16.args raise KCCError("Unable to find site settings for (%s) - (%s)" % (ssdn, estr)) msg = res[0] if "options" in msg: self.site_options = int(msg["options"][0]) if "interSiteTopologyGenerator" in msg: self.site_topo_generator = \ str(msg["interSiteTopologyGenerator"][0]) if "interSiteTopologyFailover" in msg: self.site_topo_failover = int(msg["interSiteTopologyFailover"][0]) msg = self_res[0] if "objectGUID" in msg: self.site_guid = misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) self.load_all_dsa(samdb) def load_all_dsa(self, samdb): """Discover all nTDSDSA thru the sites entry and instantiate and load the DSAs. Each dsa is inserted into the dsa_table by dn string. """ try: res = samdb.search(self.site_dnstr, scope=ldb.SCOPE_SUBTREE, expression="(objectClass=nTDSDSA)") except ldb.LdbError as e17: (enum, estr) = e17.args raise KCCError("Unable to find nTDSDSAs - (%s)" % estr) for msg in res: dnstr = str(msg.dn) # already loaded if dnstr in self.dsa_table: continue dsa = DirectoryServiceAgent(dnstr) dsa.load_dsa(samdb) # Assign this dsa to my dsa table # and index by dsa dn self.dsa_table[dnstr] = dsa if not dsa.is_ro(): self.rw_dsa_table[dnstr] = dsa def get_dsa(self, dnstr): """Return a previously loaded DSA object by consulting the sites dsa_table for the provided DSA dn string :return: None if DSA doesn't exist """ return self.dsa_table.get(dnstr) def select_istg(self, samdb, mydsa, ro): """Determine if my DC should be an intersite topology generator. If my DC is the istg and is both a writeable DC and the database is opened in write mode then we perform an originating update to set the interSiteTopologyGenerator attribute in the NTDS Site Settings object. An RODC always acts as an ISTG for itself. """ # The KCC on an RODC always acts as an ISTG for itself if mydsa.dsa_is_ro: mydsa.dsa_is_istg = True self.site_topo_generator = mydsa.dsa_dnstr return True c_rep = get_dsa_config_rep(mydsa) # Load repsFrom and replUpToDateVector if not already loaded # so we can get the current state of the config replica and # whether we are getting updates from the istg c_rep.load_repsFrom(samdb) c_rep.load_replUpToDateVector(samdb) # From MS-ADTS 6.2.2.3.1 ISTG selection: # First, the KCC on a writable DC determines whether it acts # as an ISTG for its site # # Let s be the object such that s!lDAPDisplayName = nTDSDSA # and classSchema in s!objectClass. # # Let D be the sequence of objects o in the site of the local # DC such that o!objectCategory = s. D is sorted in ascending # order by objectGUID. # # Which is a fancy way of saying "sort all the nTDSDSA objects # in the site by guid in ascending order". Place sorted list # in D_sort[] D_sort = sorted( self.rw_dsa_table.values(), key=lambda dsa: ndr_pack(dsa.dsa_guid)) # double word number of 100 nanosecond intervals since 1600s # Let f be the duration o!interSiteTopologyFailover seconds, or 2 hours # if o!interSiteTopologyFailover is 0 or has no value. # # Note: lastSuccess and ntnow are in 100 nanosecond intervals # so it appears we have to turn f into the same interval # # interSiteTopologyFailover (if set) appears to be in minutes # so we'll need to convert to seconds and then 100 nanosecond # intervals # XXX [MS-ADTS] 6.2.2.3.1 says it is seconds, not minutes. # # 10,000,000 is number of 100 nanosecond intervals in a second if self.site_topo_failover == 0: f = 2 * 60 * 60 * 10000000 else: f = self.site_topo_failover * 60 * 10000000 # Let o be the site settings object for the site of the local # DC, or NULL if no such o exists. d_dsa = self.dsa_table.get(self.site_topo_generator) # From MS-ADTS 6.2.2.3.1 ISTG selection: # If o != NULL and o!interSiteTopologyGenerator is not the # nTDSDSA object for the local DC and # o!interSiteTopologyGenerator is an element dj of sequence D: # if d_dsa is not None and d_dsa is not mydsa: # From MS-ADTS 6.2.2.3.1 ISTG Selection: # Let c be the cursor in the replUpToDateVector variable # associated with the NC replica of the config NC such # that c.uuidDsa = dj!invocationId. If no such c exists # (No evidence of replication from current ITSG): # Let i = j. # Let t = 0. # # Else if the current time < c.timeLastSyncSuccess - f # (Evidence of time sync problem on current ISTG): # Let i = 0. # Let t = 0. # # Else (Evidence of replication from current ITSG): # Let i = j. # Let t = c.timeLastSyncSuccess. # # last_success appears to be a double word containing # number of 100 nanosecond intervals since the 1600s j_idx = D_sort.index(d_dsa) found = False for cursor in c_rep.rep_replUpToDateVector_cursors: if d_dsa.dsa_ivid == cursor.source_dsa_invocation_id: found = True break if not found: i_idx = j_idx t_time = 0 # XXX doc says current time < c.timeLastSyncSuccess - f # which is true only if f is negative or clocks are wrong. # f is not negative in the default case (2 hours). elif self.nt_now - cursor.last_sync_success > f: i_idx = 0 t_time = 0 else: i_idx = j_idx t_time = cursor.last_sync_success # Otherwise (Nominate local DC as ISTG): # Let i be the integer such that di is the nTDSDSA # object for the local DC. # Let t = the current time. else: i_idx = D_sort.index(mydsa) t_time = self.nt_now # Compute a function that maintains the current ISTG if # it is alive, cycles through other candidates if not. # # Let k be the integer (i + ((current time - t) / # o!interSiteTopologyFailover)) MOD |D|. # # Note: We don't want to divide by zero here so they must # have meant "f" instead of "o!interSiteTopologyFailover" k_idx = (i_idx + ((self.nt_now - t_time) // f)) % len(D_sort) # The local writable DC acts as an ISTG for its site if and # only if dk is the nTDSDSA object for the local DC. If the # local DC does not act as an ISTG, the KCC skips the # remainder of this task. d_dsa = D_sort[k_idx] d_dsa.dsa_is_istg = True # Update if we are the ISTG, otherwise return if d_dsa is not mydsa: return False # Nothing to do if self.site_topo_generator == mydsa.dsa_dnstr: return True self.site_topo_generator = mydsa.dsa_dnstr # If readonly database then do not perform a # persistent update if ro: return True # Perform update to the samdb ssdn = "CN=NTDS Site Settings,%s" % self.site_dnstr m = ldb.Message() m.dn = ldb.Dn(samdb, ssdn) m["interSiteTopologyGenerator"] = \ ldb.MessageElement(mydsa.dsa_dnstr, ldb.FLAG_MOD_REPLACE, "interSiteTopologyGenerator") try: samdb.modify(m) except ldb.LdbError as estr: raise KCCError( "Could not set interSiteTopologyGenerator for (%s) - (%s)" % (ssdn, estr)) return True def is_intrasite_topology_disabled(self): '''Returns True if intra-site topology is disabled for site''' return (self.site_options & dsdb.DS_NTDSSETTINGS_OPT_IS_AUTO_TOPOLOGY_DISABLED) != 0 def is_intersite_topology_disabled(self): '''Returns True if inter-site topology is disabled for site''' return ((self.site_options & dsdb.DS_NTDSSETTINGS_OPT_IS_INTER_SITE_AUTO_TOPOLOGY_DISABLED) != 0) def is_random_bridgehead_disabled(self): '''Returns True if selection of random bridgehead is disabled''' return (self.site_options & dsdb.DS_NTDSSETTINGS_OPT_IS_RAND_BH_SELECTION_DISABLED) != 0 def is_detect_stale_disabled(self): '''Returns True if detect stale is disabled for site''' return (self.site_options & dsdb.DS_NTDSSETTINGS_OPT_IS_TOPL_DETECT_STALE_DISABLED) != 0 def is_cleanup_ntdsconn_disabled(self): '''Returns True if NTDS Connection cleanup is disabled for site''' return (self.site_options & dsdb.DS_NTDSSETTINGS_OPT_IS_TOPL_CLEANUP_DISABLED) != 0 def same_site(self, dsa): '''Return True if dsa is in this site''' if self.get_dsa(dsa.dsa_dnstr): return True return False def is_rodc_site(self): if len(self.dsa_table) > 0 and len(self.rw_dsa_table) == 0: return True return False def __str__(self): '''Debug dump string output of class''' text = "%s:" % self.__class__.__name__ +\ "\n\tdn=%s" % self.site_dnstr +\ "\n\toptions=0x%X" % self.site_options +\ "\n\ttopo_generator=%s" % self.site_topo_generator +\ "\n\ttopo_failover=%d" % self.site_topo_failover for key, dsa in self.dsa_table.items(): text = text + "\n%s" % dsa return text class GraphNode(object): """A graph node describing a set of edges that should be directed to it. Each edge is a connection for a particular naming context replica directed from another node in the forest to this node. """ def __init__(self, dsa_dnstr, max_node_edges): """Instantiate the graph node according to a DSA dn string :param max_node_edges: maximum number of edges that should ever be directed to the node """ self.max_edges = max_node_edges self.dsa_dnstr = dsa_dnstr self.edge_from = [] def __str__(self): text = "%s:" % self.__class__.__name__ +\ "\n\tdsa_dnstr=%s" % self.dsa_dnstr +\ "\n\tmax_edges=%d" % self.max_edges for i, edge in enumerate(self.edge_from): if isinstance(edge, str): text += "\n\tedge_from[%d]=%s" % (i, edge) return text def add_edge_from(self, from_dsa_dnstr): """Add an edge from the dsa to our graph nodes edge from list :param from_dsa_dnstr: the dsa that the edge emanates from """ assert isinstance(from_dsa_dnstr, str) # No edges from myself to myself if from_dsa_dnstr == self.dsa_dnstr: return False # Only one edge from a particular node if from_dsa_dnstr in self.edge_from: return False # Not too many edges if len(self.edge_from) >= self.max_edges: return False self.edge_from.append(from_dsa_dnstr) return True def add_edges_from_connections(self, dsa): """For each nTDSConnection object associated with a particular DSA, we test if it implies an edge to this graph node (i.e. the "fromServer" attribute). If it does then we add an edge from the server unless we are over the max edges for this graph node :param dsa: dsa with a dnstr equivalent to his graph node """ for connect in dsa.connect_table.values(): self.add_edge_from(connect.from_dnstr) def add_connections_from_edges(self, dsa, transport): """For each edge directed to this graph node, ensure there is a corresponding nTDSConnection object in the dsa. """ for edge_dnstr in self.edge_from: connections = dsa.get_connection_by_from_dnstr(edge_dnstr) # For each edge directed to the NC replica that # "should be present" on the local DC, the KCC determines # whether an object c exists such that: # # c is a child of the DC's nTDSDSA object. # c.objectCategory = nTDSConnection # # Given the NC replica ri from which the edge is directed, # c.fromServer is the dsname of the nTDSDSA object of # the DC on which ri "is present". # # c.options does not contain NTDSCONN_OPT_RODC_TOPOLOGY found_valid = False for connect in connections: if connect.is_rodc_topology(): continue found_valid = True if found_valid: continue # if no such object exists then the KCC adds an object # c with the following attributes # Generate a new dnstr for this nTDSConnection opt = dsdb.NTDSCONN_OPT_IS_GENERATED flags = (dsdb.SYSTEM_FLAG_CONFIG_ALLOW_RENAME | dsdb.SYSTEM_FLAG_CONFIG_ALLOW_MOVE) dsa.new_connection(opt, flags, transport, edge_dnstr, None) def has_sufficient_edges(self): '''Return True if we have met the maximum "from edges" criteria''' if len(self.edge_from) >= self.max_edges: return True return False class Transport(object): """Class defines a Inter-site transport found under Sites """ def __init__(self, dnstr): self.dnstr = dnstr self.options = 0 self.guid = None self.name = None self.address_attr = None self.bridgehead_list = [] def __str__(self): '''Debug dump string output of Transport object''' text = "%s:\n\tdn=%s" % (self.__class__.__name__, self.dnstr) +\ "\n\tguid=%s" % str(self.guid) +\ "\n\toptions=%d" % self.options +\ "\n\taddress_attr=%s" % self.address_attr +\ "\n\tname=%s" % self.name +\ "".join("\n\tbridgehead_list=%s" % dnstr for dnstr in self.bridgehead_list) return text def load_transport(self, samdb): """Given a Transport object with an prior initialization for the object's DN, search for the DN and load attributes from the samdb. """ attrs = ["objectGUID", "options", "name", "bridgeheadServerListBL", "transportAddressAttribute"] try: res = samdb.search(base=self.dnstr, scope=ldb.SCOPE_BASE, attrs=attrs) except ldb.LdbError as e18: (enum, estr) = e18.args raise KCCError("Unable to find Transport for (%s) - (%s)" % (self.dnstr, estr)) msg = res[0] self.guid = misc.GUID(samdb.schema_format_value("objectGUID", msg["objectGUID"][0])) if "options" in msg: self.options = int(msg["options"][0]) if "transportAddressAttribute" in msg: self.address_attr = str(msg["transportAddressAttribute"][0]) if "name" in msg: self.name = str(msg["name"][0]) if "bridgeheadServerListBL" in msg: for value in msg["bridgeheadServerListBL"]: dsdn = dsdb_Dn(samdb, value.decode('utf8')) dnstr = str(dsdn.dn) if dnstr not in self.bridgehead_list: self.bridgehead_list.append(dnstr) class RepsFromTo(object): """Class encapsulation of the NDR repsFromToBlob. Removes the necessity of external code having to understand about other_info or manipulation of update flags. """ def __init__(self, nc_dnstr=None, ndr_blob=None): self.__dict__['to_be_deleted'] = False self.__dict__['nc_dnstr'] = nc_dnstr self.__dict__['update_flags'] = 0x0 # XXX the following sounds dubious and/or better solved # elsewhere, but lets leave it for now. In particular, there # seems to be no reason for all the non-ndr generated # attributes to be handled in the round about way (e.g. # self.__dict__['to_be_deleted'] = False above). On the other # hand, it all seems to work. Hooray! Hands off!. # # WARNING: # # There is a very subtle bug here with python # and our NDR code. If you assign directly to # a NDR produced struct (e.g. t_repsFrom.ctr.other_info) # then a proper python GC reference count is not # maintained. # # To work around this we maintain an internal # reference to "dns_name(x)" and "other_info" elements # of repsFromToBlob. This internal reference # is hidden within this class but it is why you # see statements like this below: # # self.__dict__['ndr_blob'].ctr.other_info = \ # self.__dict__['other_info'] = drsblobs.repsFromTo1OtherInfo() # # That would appear to be a redundant assignment but # it is necessary to hold a proper python GC reference # count. if ndr_blob is None: self.__dict__['ndr_blob'] = drsblobs.repsFromToBlob() self.__dict__['ndr_blob'].version = 0x1 self.__dict__['dns_name1'] = None self.__dict__['dns_name2'] = None self.__dict__['ndr_blob'].ctr.other_info = \ self.__dict__['other_info'] = drsblobs.repsFromTo1OtherInfo() else: self.__dict__['ndr_blob'] = ndr_blob self.__dict__['other_info'] = ndr_blob.ctr.other_info if ndr_blob.version == 0x1: self.__dict__['dns_name1'] = ndr_blob.ctr.other_info.dns_name self.__dict__['dns_name2'] = None else: self.__dict__['dns_name1'] = ndr_blob.ctr.other_info.dns_name1 self.__dict__['dns_name2'] = ndr_blob.ctr.other_info.dns_name2 def __str__(self): '''Debug dump string output of class''' text = "%s:" % self.__class__.__name__ +\ "\n\tdnstr=%s" % self.nc_dnstr +\ "\n\tupdate_flags=0x%X" % self.update_flags +\ "\n\tversion=%d" % self.version +\ "\n\tsource_dsa_obj_guid=%s" % self.source_dsa_obj_guid +\ ("\n\tsource_dsa_invocation_id=%s" % self.source_dsa_invocation_id) +\ "\n\ttransport_guid=%s" % self.transport_guid +\ "\n\treplica_flags=0x%X" % self.replica_flags +\ ("\n\tconsecutive_sync_failures=%d" % self.consecutive_sync_failures) +\ "\n\tlast_success=%s" % self.last_success +\ "\n\tlast_attempt=%s" % self.last_attempt +\ "\n\tdns_name1=%s" % self.dns_name1 +\ "\n\tdns_name2=%s" % self.dns_name2 +\ "\n\tschedule[ " +\ "".join("0x%X " % slot for slot in self.schedule) +\ "]" return text def __setattr__(self, item, value): """Set an attribute and chyange update flag. Be aware that setting any RepsFromTo attribute will set the drsuapi.DRSUAPI_DRS_UPDATE_ADDRESS update flag. """ if item in ['schedule', 'replica_flags', 'transport_guid', 'source_dsa_obj_guid', 'source_dsa_invocation_id', 'consecutive_sync_failures', 'last_success', 'last_attempt']: if item in ['replica_flags']: self.__dict__['update_flags'] |= drsuapi.DRSUAPI_DRS_UPDATE_FLAGS elif item in ['schedule']: self.__dict__['update_flags'] |= drsuapi.DRSUAPI_DRS_UPDATE_SCHEDULE setattr(self.__dict__['ndr_blob'].ctr, item, value) elif item in ['dns_name1']: self.__dict__['dns_name1'] = value if self.__dict__['ndr_blob'].version == 0x1: self.__dict__['ndr_blob'].ctr.other_info.dns_name = \ self.__dict__['dns_name1'] else: self.__dict__['ndr_blob'].ctr.other_info.dns_name1 = \ self.__dict__['dns_name1'] elif item in ['dns_name2']: self.__dict__['dns_name2'] = value if self.__dict__['ndr_blob'].version == 0x1: raise AttributeError(item) else: self.__dict__['ndr_blob'].ctr.other_info.dns_name2 = \ self.__dict__['dns_name2'] elif item in ['nc_dnstr']: self.__dict__['nc_dnstr'] = value elif item in ['to_be_deleted']: self.__dict__['to_be_deleted'] = value elif item in ['version']: raise AttributeError("Attempt to set readonly attribute %s" % item) else: raise AttributeError("Unknown attribute %s" % item) self.__dict__['update_flags'] |= drsuapi.DRSUAPI_DRS_UPDATE_ADDRESS def __getattr__(self, item): """Overload of RepsFromTo attribute retrieval. Allows external code to ignore substructures within the blob """ if item in ['schedule', 'replica_flags', 'transport_guid', 'source_dsa_obj_guid', 'source_dsa_invocation_id', 'consecutive_sync_failures', 'last_success', 'last_attempt']: return getattr(self.__dict__['ndr_blob'].ctr, item) elif item in ['version']: return self.__dict__['ndr_blob'].version elif item in ['dns_name1']: if self.__dict__['ndr_blob'].version == 0x1: return self.__dict__['ndr_blob'].ctr.other_info.dns_name else: return self.__dict__['ndr_blob'].ctr.other_info.dns_name1 elif item in ['dns_name2']: if self.__dict__['ndr_blob'].version == 0x1: raise AttributeError(item) else: return self.__dict__['ndr_blob'].ctr.other_info.dns_name2 elif item in ['to_be_deleted']: return self.__dict__['to_be_deleted'] elif item in ['nc_dnstr']: return self.__dict__['nc_dnstr'] elif item in ['update_flags']: return self.__dict__['update_flags'] raise AttributeError("Unknown attribute %s" % item) def is_modified(self): return (self.update_flags != 0x0) def set_unmodified(self): self.__dict__['update_flags'] = 0x0 class SiteLink(object): """Class defines a site link found under sites """ def __init__(self, dnstr): self.dnstr = dnstr self.options = 0 self.system_flags = 0 self.cost = 0 self.schedule = None self.interval = None self.site_list = [] def __str__(self): '''Debug dump string output of Transport object''' text = "%s:\n\tdn=%s" % (self.__class__.__name__, self.dnstr) +\ "\n\toptions=%d" % self.options +\ "\n\tsystem_flags=%d" % self.system_flags +\ "\n\tcost=%d" % self.cost +\ "\n\tinterval=%s" % self.interval if self.schedule is not None: text += "\n\tschedule.size=%s" % self.schedule.size +\ "\n\tschedule.bandwidth=%s" % self.schedule.bandwidth +\ ("\n\tschedule.numberOfSchedules=%s" % self.schedule.numberOfSchedules) for i, header in enumerate(self.schedule.headerArray): text += ("\n\tschedule.headerArray[%d].type=%d" % (i, header.type)) +\ ("\n\tschedule.headerArray[%d].offset=%d" % (i, header.offset)) +\ "\n\tschedule.dataArray[%d].slots[ " % i +\ "".join("0x%X " % slot for slot in self.schedule.dataArray[i].slots) +\ "]" for guid, dn in self.site_list: text = text + "\n\tsite_list=%s (%s)" % (guid, dn) return text def load_sitelink(self, samdb): """Given a siteLink object with an prior initialization for the object's DN, search for the DN and load attributes from the samdb. """ attrs = ["options", "systemFlags", "cost", "schedule", "replInterval", "siteList"] try: res = samdb.search(base=self.dnstr, scope=ldb.SCOPE_BASE, attrs=attrs, controls=['extended_dn:0']) except ldb.LdbError as e19: (enum, estr) = e19.args raise KCCError("Unable to find SiteLink for (%s) - (%s)" % (self.dnstr, estr)) msg = res[0] if "options" in msg: self.options = int(msg["options"][0]) if "systemFlags" in msg: self.system_flags = int(msg["systemFlags"][0]) if "cost" in msg: self.cost = int(msg["cost"][0]) if "replInterval" in msg: self.interval = int(msg["replInterval"][0]) if "siteList" in msg: for value in msg["siteList"]: dsdn = dsdb_Dn(samdb, value.decode('utf8')) guid = misc.GUID(dsdn.dn.get_extended_component('GUID')) dnstr = str(dsdn.dn) if (guid, dnstr) not in self.site_list: self.site_list.append((guid, dnstr)) if "schedule" in msg: self.schedule = ndr_unpack(drsblobs.schedule, value) else: self.schedule = new_connection_schedule() class KCCFailedObject(object): def __init__(self, uuid, failure_count, time_first_failure, last_result, dns_name): self.uuid = uuid self.failure_count = failure_count self.time_first_failure = time_first_failure self.last_result = last_result self.dns_name = dns_name ################################################## # Global Functions and Variables ################################################## def get_dsa_config_rep(dsa): # Find configuration NC replica for the DSA for c_rep in dsa.current_rep_table.values(): if c_rep.is_config(): return c_rep raise KCCError("Unable to find config NC replica for (%s)" % dsa.dsa_dnstr) def new_connection_schedule(): """Create a default schedule for an NTDSConnection or Sitelink. This is packed differently from the repltimes schedule used elsewhere in KCC (where the 168 nibbles are packed into 84 bytes). """ # 168 byte instances of the 0x01 value. The low order 4 bits # of the byte equate to 15 minute intervals within a single hour. # There are 168 bytes because there are 168 hours in a full week # Effectively we are saying to perform replication at the end of # each hour of the week schedule = drsblobs.schedule() schedule.size = 188 schedule.bandwidth = 0 schedule.numberOfSchedules = 1 header = drsblobs.scheduleHeader() header.type = 0 header.offset = 20 schedule.headerArray = [header] data = drsblobs.scheduleSlots() data.slots = [0x01] * 168 schedule.dataArray = [data] return schedule ################################################## # DNS related calls ################################################## def uncovered_sites_to_cover(samdb, site_name): """ Discover which sites have no DCs and whose lowest single-hop cost distance for any link attached to that site is linked to the site supplied. We compare the lowest cost of your single-hop link to this site to all of those available (if it exists). This means that a lower ranked siteLink with only the uncovered site can trump any available links (but this can only be done with specific, poorly enacted user configuration). If the site is connected to more than one other site with the same siteLink, only the largest site (failing that sorted alphabetically) creates the DNS records. :param samdb database :param site_name origin site (with a DC) :return a list of sites this site should be covering (for DNS) """ sites_to_cover = [] server_res = samdb.search(base=samdb.get_config_basedn(), scope=ldb.SCOPE_SUBTREE, expression="(&(objectClass=server)" "(serverReference=*))") site_res = samdb.search(base=samdb.get_config_basedn(), scope=ldb.SCOPE_SUBTREE, expression="(objectClass=site)") sites_in_use = Counter() dc_count = 0 # Assume server is of form DC,Servers,Site-ABCD because of schema for msg in server_res: site_dn = msg.dn.parent().parent() sites_in_use[site_dn.canonical_str()] += 1 if site_dn.get_rdn_value().lower() == site_name.lower(): dc_count += 1 if len(sites_in_use) != len(site_res): # There is a possible uncovered site sites_uncovered = [] for msg in site_res: if msg.dn.canonical_str() not in sites_in_use: sites_uncovered.append(msg) own_site_dn = "CN={},CN=Sites,{}".format( ldb.binary_encode(site_name), ldb.binary_encode(str(samdb.get_config_basedn())) ) for site in sites_uncovered: encoded_dn = ldb.binary_encode(str(site.dn)) # Get a sorted list of all siteLinks featuring the uncovered site link_res1 = samdb.search(base=samdb.get_config_basedn(), scope=ldb.SCOPE_SUBTREE, attrs=["cost"], expression="(&(objectClass=siteLink)" "(siteList={}))".format(encoded_dn), controls=["server_sort:1:0:cost"]) # Get a sorted list of all siteLinks connecting this an the # uncovered site link_res2 = samdb.search(base=samdb.get_config_basedn(), scope=ldb.SCOPE_SUBTREE, attrs=["cost", "siteList"], expression="(&(objectClass=siteLink)" "(siteList={})(siteList={}))".format( own_site_dn, encoded_dn), controls=["server_sort:1:0:cost"]) # Add to list if your link is equal in cost to lowest cost link if len(link_res1) > 0 and len(link_res2) > 0: cost1 = int(link_res1[0]['cost'][0]) cost2 = int(link_res2[0]['cost'][0]) # Own siteLink must match the lowest cost link if cost1 != cost2: continue # In a siteLink with more than 2 sites attached, only pick the # largest site, and if there are multiple, the earliest # alphabetically. to_cover = True for site_val in link_res2[0]['siteList']: site_dn = ldb.Dn(samdb, str(site_val)) site_dn_str = site_dn.canonical_str() site_rdn = site_dn.get_rdn_value().lower() if sites_in_use[site_dn_str] > dc_count: to_cover = False break elif (sites_in_use[site_dn_str] == dc_count and site_rdn < site_name.lower()): to_cover = False break if to_cover: site_cover_rdn = site.dn.get_rdn_value() sites_to_cover.append(site_cover_rdn.lower()) return sites_to_cover