Long latency and dropped messages to the Front-end

Mikhail Soloviev 0 Reputation points
2025-10-22T15:28:27.9133333+00:00

we're experiencing delays on the front-end side with latency and messages dropped. This is a major disruption of our app function. This behaviour looks very similar to the one we had when we hit Cosmos DB quotas. Now however, there are no 429 Throttling messages and increasing the units does not help either.
MQTT messages from the IoT devices come normally. Using Azure explorer we can see that they arrive as they should.

Azure IoT Hub
Azure IoT Hub
An Azure service that enables bidirectional communication between internet of things (IoT) devices and applications.
{count} votes

1 answer

Sort by: Most helpful
  1. Vinodh247 39,291 Reputation points MVP Volunteer Moderator
    2025-10-22T16:32:06.4133333+00:00

    Hi ,

    Thanks for reaching out to Microsoft Q&A.

    The application is experiencing high latency and message drops on the front end, severely affecting functionality. MQTT messages from IoT devices are confirmed to arrive normally, as verified through Azure Explorer, indicating that the issue is not with device connectivity or message ingestion.

    Previously, similar symptoms occurred when CosmosDB throughput limits were reached, resulting in 429 throttling errors. However, this time, no throttling events are observed, and increasing the request units (RUs) does not improve performance, ruling out direct quota saturation.

    The issue likely lies between the ingestion and front-end delivery layers, potentially in the message processing, event routing (eventhub, stream analytics, or service bus), or API response path. Bottlenecks in query latency, partition key design, or hot partitions within cosmosDB could also contribute without triggering throttling.

    Next steps to check:

    Check end-to-end telemetry (App Insights, Azure Monitor) for latency spikes.

    Review Cosmos DB diagnostics for request latency, partition load, and RU consumption patterns.

    Verify if the API or streaming layer has backlog or scaling issues.

    Enable detailed logging to identify where message loss or delay begins.

    This is likely a downstream processing or read-path performance issue rather than ingestion failure.

    Please 'Upvote'(Thumbs-up) and 'Accept' as answer if the reply was helpful. This will be benefitting other community members who face the same issue.

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.