コード例 #1
0
def test_example_user_agent_policy():
    url = "https://bing.com"
    redirect_policy = RedirectPolicy()

    # [START user_agent_policy]
    from azure.core.pipeline.policies import UserAgentPolicy

    user_agent_policy = UserAgentPolicy()

    # The user-agent policy allows you to append a custom value to the header.
    user_agent_policy.add_user_agent("CustomValue")

    # You can also pass in a custom value per operation to append to the end of the user-agent.
    # This can be used together with the policy configuration to append multiple values.
    policies = [
        redirect_policy,
        user_agent_policy,
    ]
    client = PipelineClient(base_url=url, policies=policies)
    request = client.get(url)
    pipeline_response = client._pipeline.run(request,
                                             user_agent="AnotherValue")
    # [END user_agent_policy]

    response = pipeline_response.http_response
    assert response.status_code == 200
コード例 #2
0
    def start(self):
        """
        Build the create cluster request.
        """
        credentials, subscription_id, managed_identity_id = self._get_credentials(
        )

        # Fetch metadata about the instance
        metadata = get_instance_metadata()

        # Resource group
        resource_group = self.config.get('resourceGroup', None)
        dss_host_resource_group = metadata["compute"]["resourceGroupName"]
        if _is_none_or_blank(resource_group):
            resource_group = dss_host_resource_group
            logging.info(
                "Using same resource group as DSS: {}".format(resource_group))

        # Location
        location = self.config.get('location', None)
        if _is_none_or_blank(location):
            location = metadata["compute"]["location"]
            logging.info("Using same location as DSS: {}".format(location))

        # Consistency checks
        if _is_none_or_blank(resource_group):
            raise Exception(
                "A resource group to put the cluster in is required")
        if _is_none_or_blank(location):
            raise Exception("A location to put the cluster in is required")

        # AKS Client
        clusters_client = None

        # Credit the cluster to DATAIKU
        if os.environ.get("DISABLE_AZURE_USAGE_ATTRIBUTION", "0") == "1":
            logging.info("Azure usage attribution is disabled")
            clusters_client = ContainerServiceClient(credentials,
                                                     subscription_id)
        else:
            policy = UserAgentPolicy()
            policy.add_user_agent('pid-fd3813c7-273c-5eec-9221-77323f62a148')
            clusters_client = ContainerServiceClient(credentials,
                                                     subscription_id,
                                                     user_agent_policy=policy)

        # check that the cluster doesn't exist yet, otherwise azure will try to update it
        # and will almost always fail
        try:
            existing = clusters_client.managed_clusters.get(
                resource_group, self.cluster_name)
            if existing is not None:
                raise Exception(
                    "A cluster with name %s in resource group %s already exists"
                    % (self.cluster_name, resource_group))
        except CloudError as e:
            logging.info("Cluster doesn't seem to exist yet")
        except ResourceNotFoundError as e:
            logging.info("Cluster doesn't seem to exist yet")

        cluster_builder = ClusterBuilder(clusters_client)
        cluster_builder.with_name(self.cluster_name)
        cluster_builder.with_dns_prefix("{}-dns".format(self.cluster_name))
        cluster_builder.with_resource_group(resource_group)
        cluster_builder.with_location(location)
        cluster_builder.add_tags(self.config.get("tags", None))
        cluster_builder.with_linux_profile()  # default is None
        cluster_builder.with_network_profile(
            service_cidr=self.config.get("serviceCIDR", None),
            dns_service_ip=self.config.get("dnsServiceIP", None),
            load_balancer_sku=self.config.get("loadBalancerSku", None),
            outbound_type=self.config.get("outboundType", None),
            network_plugin=self.config.get("networkPlugin"),
            docker_bridge_cidr=self.config.get("dockerBridgeCidr"))

        if self.config.get("useCustomNodeResourceGroup", False):
            cluster_builder.with_node_resource_group(
                self.config.get("nodeResourceGroup"))

        # Cluster identity
        connection_info = self.config.get("connectionInfo", None)
        cluster_idendity_legacy_use_distinct_sp = self.config.get(
            "useDistinctSPForCluster", False)
        cluster_idendity_legacy_sp = self.config.get("clusterServicePrincipal",
                                                     None)
        cluster_identity_type = None
        cluster_identity = None
        if not _is_none_or_blank(
                connection_info) or cluster_idendity_legacy_use_distinct_sp:
            logging.warn(
                "Using legacy options to configure cluster identity. Clear them to use the new ones."
            )
            if not cluster_idendity_legacy_use_distinct_sp and not _is_none_or_blank(
                    connection_info):
                cluster_sp = connection_info
            elif cluster_idendity_legacy_use_distinct_sp and not _is_none_or_blank(
                    cluster_idendity_legacy_sp):
                cluster_sp = self.config.get("clusterServicePrincipal")
            else:
                raise Exception(
                    "Legacy options are not complete enough to determine cluster identity settings"
                )
            cluster_builder.with_cluster_sp_legacy(
                cluster_service_principal_connection_info=cluster_sp)
        else:
            cluster_identity = self.config.get(
                "clusterIdentity", {"identityType": "managed-identity"})
            cluster_identity_type = cluster_identity.get(
                "identityType", "managed-identity")
            if cluster_identity_type == "managed-identity":
                if cluster_identity.get("inheritDSSIdentity", True):
                    logging.info(
                        "Need to inspect Managed Identity infos from Azure")
                    if metadata is None:
                        metadata = get_instance_metadata()
                    vm_resource_group = metadata["compute"][
                        "resourceGroupName"]
                    vm_name = metadata["compute"]["name"]
                    compute_client = ComputeManagementClient(
                        credentials, subscription_id)
                    vm = compute_client.virtual_machines.get(
                        vm_resource_group, vm_name)
                    # No choice here but to use the first one
                    if managed_identity_id is None:
                        managed_identity_id = next(
                            iter(vm.identity.user_assigned_identities.keys()))
                    for managed_identity_resource_id, managed_identity_properties in vm.identity.user_assigned_identities.items(
                    ):
                        if managed_identity_id == managed_identity_resource_id or managed_identity_id == managed_identity_properties.client_id:
                            break
                    logging.info("Found managed identity id {}".format(
                        managed_identity_resource_id))
                    cluster_builder.with_managed_identity(
                        managed_identity_resource_id)
                    cluster_builder.with_kubelet_identity(
                        managed_identity_resource_id,
                        managed_identity_properties.client_id,
                        managed_identity_properties.principal_id)
                else:
                    control_plane_mi = None if cluster_identity.get(
                        "useAKSManagedIdentity", True
                    ) else cluster_identity["controlPlaneUserAssignedIdentity"]
                    cluster_builder.with_managed_identity(control_plane_mi)
                    if control_plane_mi is None:
                        logging.info(
                            "Configure cluster with system managed identity.")
                    else:
                        logging.info(
                            "Configure cluster with user assigned identity: {}"
                            .format(control_plane_mi))
                    if not cluster_identity.get("useAKSManagedKubeletIdentity",
                                                True):
                        kubelet_mi = cluster_identity[
                            "kubeletUserAssignedIdentity"]
                        _, _, mi_subscription_id, _, mi_resource_group, _, _, _, mi_name = kubelet_mi.split(
                            "/")
                        msiclient = ManagedServiceIdentityClient(
                            AzureIdentityCredentialAdapter(credentials),
                            mi_subscription_id)
                        mi = msiclient.user_assigned_identities.get(
                            mi_resource_group, mi_name)
                        cluster_builder.with_kubelet_identity(
                            kubelet_mi, mi.client_id, mi.principal_id)
                        logging.info(
                            "Configure kubelet identity with user assigned identity resourceId=\"{}\", clientId=\"{}\", objectId=\"{}\""
                            .format(kubelet_mi, mi.client_id, mi.principal_id))
            elif cluster_identity_type == "service-principal":
                cluster_builder.with_cluster_sp(cluster_identity["clientId"],
                                                cluster_identity["password"])
                logging.info("Configure cluster with service principal")
            else:
                raise Exception(
                    "Cluster identity type \"{}\" is unknown".format(
                        cluster_identity_type))

        # Fail fast for non existing ACRs to avoid drama in case of failure AFTER cluster is created
        acr_role_id = None
        authorization_client = None
        if cluster_identity_type is not None and cluster_identity is not None:
            if cluster_identity_type == "managed-identity" and cluster_identity.get(
                    "useAKSManagedKubeletIdentity",
                    True) and not cluster_identity.get("inheritDSSIdentity",
                                                       True):
                acr_name = cluster_identity.get("attachToACRName", None)
                if not _is_none_or_blank(acr_name):
                    # build acr scope
                    acr_identifier_splitted = acr_name.split('/')
                    acr_subscription_id = subscription_id
                    acr_resource_group = resource_group
                    if 9 == len(acr_identifier_splitted):
                        _, _, acr_subscription_id, _, acr_resource_group, _, _, _, acr_name = acr_identifier_splitted
                    elif 2 == len(acr_identifier_splitted):
                        acr_resource_group, acr_name = acr_identifier_splitted

                    authorization_client = AuthorizationManagementClient(
                        credentials, acr_subscription_id)
                    acr_scope = "/subscriptions/{acr_subscription_id}/resourceGroups/{acr_resource_group}/providers/Microsoft.ContainerRegistry/registries/{acr_name}".format(
                        **locals())
                    try:
                        acr_roles = list(
                            authorization_client.role_definitions.list(
                                acr_scope, "roleName eq 'AcrPull'"))
                    except ResourceNotFoundError as e:
                        raise Exception(
                            "ACR {} not found. Check it exists and you are Owner of it."
                            .format(acr_scope))
                    if 0 == len(acr_roles):
                        raise Exception(
                            "Could not find the AcrPull role on the ACR {}. Check you are Owner of it."
                            .format(acr_scope))
                    else:
                        acr_role_id = acr_roles[0].id
                        logging.info("ACR pull role id: %s", acr_role_id)

                    # Try to run a fake role assignment. Depending on the failure type we know if we are Owner or not
                    try:
                        fake_role_assignment = authorization_client.role_assignments.create(
                            scope=acr_scope,
                            role_assignment_name=str(uuid.uuid4()),
                            parameters={
                                "properties": {
                                    "role_definition_id":
                                    acr_role_id,
                                    "principal_id":
                                    "00000000-0000-0000-0000-000000000000",
                                },
                            },
                        )
                    except HttpResponseError as e:
                        if e.reason == "Forbidden" and "AuthorizationFailed" in str(
                                e.error):
                            raise Exception(
                                "Cannot create role assignments on ACR {}. Check that your are Owner of it or provide an existing Kubelet identity."
                                .format(acr_scope))
                        elif e.reason == "Bad Request" and "PrincipalNotFound" in str(
                                e.error):
                            logging.info(
                                "Fake role assignment on ACR looks ok. Identity should be allowed to assign roles in further steps."
                            )
                        else:
                            raise (e)
                    except Exception as e:
                        raise (e)

        # Sanity check for node pools
        node_pool_vnets = set()
        for idx, node_pool_conf in enumerate(self.config.get("nodePools", [])):
            node_pool_builder = cluster_builder.get_node_pool_builder()
            nodepool_vnet = node_pool_conf.get("vnet", None)
            nodepool_subnet = node_pool_conf.get("subnet", None)
            vnet, _ = node_pool_builder.resolve_network(
                inherit_from_host=node_pool_conf.get(
                    "useSameNetworkAsDSSHost"),
                cluster_vnet=nodepool_vnet,
                cluster_subnet=nodepool_subnet,
                connection_info=connection_info,
                credentials=credentials,
                resource_group=resource_group,
                dss_host_resource_group=dss_host_resource_group)
            node_pool_vnets.add(vnet)

        if 1 < len(node_pool_vnets):
            raise Exception(
                "Node pools must all share the same vnet. Current node pools configuration yields vnets {}."
                .format(",".join(node_pool_vnets)))
        elif 0 == len(node_pool_vnets):
            raise Exception(
                "You cannot deploy a cluster without any node pool.")

        # Check role assignments for vnet like on ACR for fail fast if not doable
        vnet_id = node_pool_vnets.pop()
        if not vnet_id.startswith("/"):
            vnet_name = vnet_id
            vnet_id = "/subscriptions/{subscription_id}/resourceGroups/{resource_group}/providers/Microsoft.Network/virtualNetworks/{vnet_name}".format(
                **locals())
        vnet_role_id = None
        if cluster_identity_type is not None and cluster_identity is not None:
            if cluster_identity_type == "managed-identity" and cluster_identity.get(
                    "useAKSManagedIdentity",
                    True) and not cluster_identity.get("inheritDSSIdentity",
                                                       True):
                authorization_client = AuthorizationManagementClient(
                    credentials, subscription_id)
                try:
                    vnet_roles = list(
                        authorization_client.role_definitions.list(
                            vnet_id, "roleName eq 'Contributor'"))
                except ResourceNotFoundError as e:
                    raise Exception(
                        "Vnet {} not found. Check it exists and you are Owner of it."
                        .format(vnet_id))
                if 0 == len(acr_roles):
                    raise Exception(
                        "Could not find the Contributor role on the vnet {}. Check you are Owner of it."
                        .format(vnet_id))
                else:
                    vnet_role_id = vnet_roles[0].id
                    logging.info("Vnet contributor role id: %s", acr_role_id)
                    # Try to run a fake role assignment. Depending on the failure type we know if we are Owner or not
                    try:
                        fake_role_assignment = authorization_client.role_assignments.create(
                            scope=vnet_id,
                            role_assignment_name=str(uuid.uuid4()),
                            parameters={
                                "properties": {
                                    "role_definition_id":
                                    vnet_role_id,
                                    "principal_id":
                                    "00000000-0000-0000-0000-000000000000",
                                },
                            },
                        )
                    except HttpResponseError as e:
                        if e.reason == "Forbidden" and "AuthorizationFailed" in str(
                                e.error):
                            raise Exception(
                                "Cannot create role assignments on Vnet {}. Check that your are Owner of it or provide an existing Controle Plane identity."
                                .format(vnet_id))
                        elif e.reason == "Bad Request" and "PrincipalNotFound" in str(
                                e.error):
                            logging.info(
                                "Fake role assignment on Vnet looks ok. Identity should be allowed to assign roles in further steps."
                            )
                        else:
                            raise (e)
                    except Exception as e:
                        raise (e)

        # Access level
        if self.config.get("privateAccess"):
            cluster_builder.with_private_access(
                self.config.get("privateAccess"))

        cluster_builder.with_cluster_version(
            self.config.get("clusterVersion", None))

        # Node pools
        for idx, node_pool_conf in enumerate(self.config.get("nodePools", [])):
            node_pool_builder = cluster_builder.get_node_pool_builder()
            node_pool_builder.with_idx(idx)
            node_pool_builder.with_vm_size(node_pool_conf.get("vmSize", None))
            vnet = node_pool_conf.get("vnet", None)
            subnet = node_pool_conf.get("subnet", None)
            node_pool_builder.with_network(
                inherit_from_host=node_pool_conf.get(
                    "useSameNetworkAsDSSHost"),
                cluster_vnet=vnet,
                cluster_subnet=subnet,
                connection_info=connection_info,
                credentials=credentials,
                resource_group=resource_group,
                dss_host_resource_group=dss_host_resource_group)

            node_pool_builder.with_availability_zones(
                use_availability_zones=node_pool_conf.get(
                    "useAvailabilityZones", True))

            node_pool_builder.with_node_count(
                enable_autoscaling=node_pool_conf.get("autoScaling", False),
                num_nodes=node_pool_conf.get("numNodes", None),
                min_num_nodes=node_pool_conf.get("minNumNodes", None),
                max_num_nodes=node_pool_conf.get("maxNumNodes", None))

            node_pool_builder.with_mode(
                mode=node_pool_conf.get("mode", "Automatic"),
                system_pods_only=node_pool_conf.get("systemPodsOnly", True))

            node_pool_builder.with_disk_size_gb(
                disk_size_gb=node_pool_conf.get("osDiskSizeGb", 0))
            node_pool_builder.with_node_labels(
                node_pool_conf.get("labels", None))
            node_pool_builder.with_node_taints(
                node_pool_conf.get("taints", None))
            node_pool_builder.add_tags(self.config.get("tags", None))
            node_pool_builder.add_tags(node_pool_conf.get("tags", None))
            node_pool_builder.build()
            cluster_builder.with_node_pool(
                node_pool=node_pool_builder.agent_pool_profile)

        # Run creation
        logging.info("Start creation of cluster")

        def do_creation():
            cluster_create_op = cluster_builder.build()
            return cluster_create_op.result()

        create_result = run_and_process_cloud_error(do_creation)
        logging.info("Cluster creation finished")

        # Attach to ACR
        acr_attachment = {}
        if cluster_identity_type is not None and cluster_identity is not None:
            if cluster_identity_type == "managed-identity" and cluster_identity.get(
                    "useAKSManagedKubeletIdentity",
                    True) and not cluster_identity.get("inheritDSSIdentity",
                                                       True):
                kubelet_mi_object_id = create_result.identity_profile.get(
                    "kubeletidentity").object_id
                logging.info("Kubelet Managed Identity object id: %s",
                             kubelet_mi_object_id)
                if not _is_none_or_blank(acr_role_id):
                    logging.info("Assign ACR pull role id %s to %s",
                                 acr_role_id, kubelet_mi_object_id)
                    role_assignment = authorization_client.role_assignments.create(
                        scope=acr_scope,
                        role_assignment_name=str(uuid.uuid4()),
                        parameters={
                            "properties": {
                                "role_definition_id": acr_role_id,
                                "principal_id": kubelet_mi_object_id,
                            },
                        },
                    )
                    acr_attachment.update({
                        "name":
                        acr_name,
                        "resource_group":
                        acr_resource_group,
                        "subscription_id":
                        acr_subscription_id,
                        "resource_id":
                        acr_scope,
                        "role_assignment":
                        role_assignment.as_dict(),
                    })

        # Attach to VNET to allow LoadBalancers creation
        vnet_attachment = {}
        if cluster_identity_type is not None and cluster_identity is not None:
            if cluster_identity_type == "managed-identity" and cluster_identity.get(
                    "useAKSManagedIdentity",
                    True) and not cluster_identity.get("inheritDSSIdentity",
                                                       True):
                # And here we are blocked because we cant get the principal id of a System Assigned Managed Id easily
                control_plane_object_id = create_result.identity.principal_id
                logging.info("Controle Plane Managed Identity object id: %s",
                             control_plane_object_id)
                if not _is_none_or_blank(vnet_role_id):
                    logging.info("Assign Vnet contributolr role id %s to %s",
                                 vnet_role_id, control_plane_object_id)
                    vnet_role_assignment = authorization_client.role_assignments.create(
                        scope=vnet_id,
                        role_assignment_name=str(uuid.uuid4()),
                        parameters={
                            "properties": {
                                "role_definition_id": vnet_role_id,
                                "principal_id": control_plane_object_id,
                            },
                        },
                    )
                    vnet_attachment.update({
                        "subscription_id":
                        subscription_id,
                        "resource_id":
                        vnet_id,
                        "role_assignment":
                        vnet_role_assignment.as_dict(),
                    })

        logging.info("Fetching kubeconfig for cluster {} in {}...".format(
            self.cluster_name, resource_group))

        def do_fetch():
            return clusters_client.managed_clusters.list_cluster_admin_credentials(
                resource_group, self.cluster_name)

        get_credentials_result = run_and_process_cloud_error(do_fetch)
        kube_config_content = get_credentials_result.kubeconfigs[
            0].value.decode("utf8")
        logging.info("Writing kubeconfig file...")
        kube_config_path = os.path.join(os.getcwd(), "kube_config")
        with open(kube_config_path, 'w') as f:
            f.write(kube_config_content)

        overrides = make_overrides(
            self.config,
            yaml.safe_load(kube_config_content),
            kube_config_path,
            acr_name=None
            if _is_none_or_blank(acr_attachment) else acr_attachment["name"],
        )

        return [
            overrides, {
                "kube_config_path": kube_config_path,
                "cluster": create_result.as_dict(),
                "acr_attachment": acr_attachment,
                "vnet_attachment": vnet_attachment
            }
        ]