Share via

Title: Azure Managed Instance for Apache Cassandra Deployment Failing with InternalServerError

Vikas Suhag | MAQ Software 0 Reputation points
2026-04-06T13:45:10.7033333+00:00

Hello,

I am trying to deploy Azure Managed Instance for Apache Cassandra using the Azure Portal / ARM template, but I am running into an InternalServerError and ResourceDeployement Failure.

Here is my setup and what I have tried:

  1. Created a Virtual Network and associated subnet.
  2. Assigned Network Contributor role to the subnet/resource group for Azure Cosmos DB.
  3. Attempted deployment of Managed Cassandra Cluster in multiple regions:
  • East US
  • East US 2
  • West US
  • Central India
  1. Renamed the cluster multiple times to rule out name conflicts.
  2. This issue started today, and earlier deployments were working without errors.

**Error Details from Azure CLI:
**"message": "The resource write operation failed to complete successfully, because it reached terminal provisioning state 'Failed'.",

"code": "ResourceDeploymentFailure"

What I’ve checked:

  • Microsoft.DocumentDB provider is registered.
  • CassandraClusters resource type is available in the subscription.
  • I have tried multiple regions to rule out regional capacity issues.

Question:

  1. Could this be due to a subscription-level restriction or backend quota issue?
  2. Are there any known issues today with Azure Managed Cassandra deployment?
  3. Any additional network or configuration requirements that could cause this?

Thanks in advance for any guidance!

Azure Cosmos DB
Azure Cosmos DB

An Azure NoSQL database service for app development.


1 answer

Sort by: Most helpful
  1. Q&A Assist
    2026-04-06T13:45:31.8733333+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    The error ResourceDeploymentFailure with terminal state Failed is a generic deployment failure. From the provided information, the most likely causes in this scenario are missing/incorrect network permissions for the Azure Cosmos DB service principal on the virtual network, or general Azure deployment issues (including quota or regional capacity), rather than a simple name conflict.

    Key checks and actions:

    1. Validate required VNet permissions for Managed Instance for Apache Cassandra

    Azure Managed Instance for Apache Cassandra requires that the Azure Cosmos DB service principal has a role that allows Microsoft.Network/virtualNetworks/subnets/join/action on the virtual network (or subnet scope). If this is missing or mis-scoped, deployments can fail with generic InternalServerError / ResourceDeploymentFailure.

    Required role assignment (CLI example):

    az role assignment create \
      --assignee a232010e-820c-4083-83bb-3ace5fc29d0b \
      --role 4d97b98b-1d4f-4787-a291-c67834d212e7 \
      --scope /subscriptions/<subscriptionID>/resourceGroups/<resourceGroupName>/providers/Microsoft.Network/virtualNetworks/<vnetName>
    

    Important points:

    • assignee must be a232010e-820c-4083-83bb-3ace5fc29d0b (Azure Cosmos DB service principal).
    • role must be 4d97b98b-1d4f-4787-a291-c67834d212e7 (Network Contributor built‑in role ID used in the docs).
    • Scope should be the VNet (or at least the VNet that contains the subnet used for the cluster), not just the resource group.
    • If the CLI command fails (for example, due to permission issues), the role can be assigned via the Azure portal:
      • Go to the VNet → Access control (IAM)Add role assignment.
      • Select Network Contributor.
      • Under Members, select the Azure Cosmos DB service principal.

    If the role was assigned only at the subnet or only at the resource group, re‑check that the scope matches the guidance above. Mis-scoped or missing permissions are a common cause of deployment failures.

    1. Confirm outbound network access requirements

    Deployment of Azure Managed Instance for Apache Cassandra requires internet access from the VNet/subnet. If outbound access is restricted (firewalls, NSGs, custom routes, or forced tunneling), deployment can fail with generic internal errors.

    Ensure the VNet/subnet used for the cluster is not blocking outbound access to these services:

    • Azure Storage
    • Azure Key Vault
    • Azure Virtual Machine Scale Sets
    • Azure Monitor
    • Microsoft Entra ID
    • Microsoft Defender for Cloud

    If using custom NSGs or route tables, verify that outbound rules allow required traffic and that there is no forced tunneling that blocks these services.

    1. Re-validate subscription and provider state

    From the context, common deployment errors include authorization and provider registration issues. Even though Microsoft.DocumentDB is registered and CassandraClusters is visible, confirm:

    • The deploying identity has sufficient permissions at the subscription/resource group level (no AuthorizationFailed in the detailed deployment error).
    • No other ARM error codes (such as AllocationFailed, SubscriptionNotRegistered, or quota-related errors) are present in the deployment operation details.

    Use the deployment’s Operation details in the portal or CLI to check for nested error codes beyond ResourceDeploymentFailure.

    1. Consider quota/region issues and retry

    The context shows that some deployment failures are region-specific or quota-related and may not clearly state the region issue in the top-level error. Since deployments previously worked and started failing “today,” a transient platform or regional capacity issue is possible.

    Actions:

    • Retry deployment after some time in one of the previously working regions.
    • If the failure persists across multiple regions and subscriptions are otherwise healthy, open an Azure support request and include the full deployment error details and correlation IDs.
    1. When to contact support

    If all of the following are true:

    • Azure Cosmos DB service principal has the correct Network Contributor role on the VNet (as shown above).
    • Outbound network access to the required Azure services is confirmed.
    • No obvious RBAC or provider registration errors appear in deployment operation details.
    • The issue persists across multiple regions and cluster names.

    Then the failure is likely due to a backend/platform issue or a subscription-specific condition that requires direct Azure support investigation.


    References:

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.