Skip to content

Commit c5059c2

Browse files
authored
Merge pull request #47837 from sidramadoss/patch-10
Max delay while scanning blob input.md
2 parents ab26530 + 195e725 commit c5059c2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

articles/stream-analytics/stream-analytics-define-inputs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -117,7 +117,7 @@ For scenarios with large quantities of unstructured data to store in the cloud,
117117

118118
Log processing is a commonly used scenario for using Blob storage inputs with Stream Analytics. In this scenario, telemetry data files have been captured from a system and need to be parsed and processed to extract meaningful data.
119119

120-
The default timestamp of Blob storage events in Stream Analytics is the timestamp that the blob was last modified, which is `BlobLastModifiedUtcTime`. To process the data as a stream using a timestamp in the event payload, you must use the [TIMESTAMP BY](https://msdn.microsoft.com/library/azure/dn834998.aspx) keyword.
120+
The default timestamp of Blob storage events in Stream Analytics is the timestamp that the blob was last modified, which is `BlobLastModifiedUtcTime`. To process the data as a stream using a timestamp in the event payload, you must use the [TIMESTAMP BY](https://msdn.microsoft.com/library/azure/dn834998.aspx) keyword. A Stream Analytics job pulls data from Azure Blob storage input every second if the blob file is available. If the blob file is unavailable, there is an exponential backoff with a maximum time delay of 90 seconds.
121121

122122
CSV-formatted inputs *require* a header row to define fields for the data set, and all header row fields must be unique.
123123

0 commit comments

Comments
 (0)