Skip to content

Commit bd2de1f

Browse files
Learn Build Service GitHub AppLearn Build Service GitHub App
authored andcommitted
Merging changes synced from https://github.com/MicrosoftDocs/dataexplorer-docs-pr (branch live)
2 parents be5c74c + 89e22ae commit bd2de1f

File tree

2 files changed

+2
-1
lines changed

2 files changed

+2
-1
lines changed

data-explorer/ingestion-supported-formats.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,7 @@ Data ingestion is the process by which data is added to a table and is made avai
2929
|---------|------------|-----------|
3030
|ApacheAvro|`.avro` |An [AVRO](https://avro.apache.org/docs/current/) format with support for [logical types](https://avro.apache.org/docs/++version++/specification/#Logical+Types). The following compression codecs are supported: `null`, `deflate`, and `snappy`. Reader implementation of the `apacheavro` format is based on the official [Apache Avro library](https://github.com/apache/avro). For information about ingesting Event Hubs Capture Avro files, see [Ingesting Event Hubs Capture Avro files](ingest-data-event-hub-overview.md#schema-mapping-for-event-hubs-capture-avro-files). |
3131
|Avro |`.avro` |A legacy implementation for [AVRO](https://avro.apache.org/docs/current/) format based on [.NET library](https://www.nuget.org/packages/Microsoft.Hadoop.Avro). The following compression codecs are supported: `null`, `deflate` (for `snappy` - use `ApacheAvro` data format). |
32+
|AzMonStream |N/A |Azure Monitor [exports data in this format](/azure/azure-monitor/platform/stream-monitoring-data-event-hubs#data-formats) to Azure Event Hubs. This format is supported only by Azure Event Hubs. |
3233
|CSV |`.csv` |A text file with comma-separated values (`,`). See [RFC 4180: _Common Format and MIME Type for Comma-Separated Values (CSV) Files_](https://www.ietf.org/rfc/rfc4180.txt).|
3334
|JSON |`.json` |A text file with JSON objects delimited by `\n` or `\r\n`. See [JSON Lines (JSONL)](http://jsonlines.org/).|
3435
|MultiJSON|`.multijson`|A text file with a JSON array of property bags (each representing a record), or any number of property bags delimited by whitespace, `\n`, or `\r\n`. Each property bag can be spread on multiple lines.|

data-explorer/kusto/management/data-ingestion/ingest-from-storage-queued.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ You must have at least [Table Ingestor](../../access-control/role-based-access-c
3838
|*SkipBatching*| `boolean` | | If set to `true`, the blobs are ingested individually rather than batched together with other blobs. The default value is `false`.|
3939
|*CompressionFactor*| `real` | |The compression factor (ratio) between the original size and the compressed size of blobs. Compression factor is used to estimate the original size of the data for batching purposes, when blobs are provided in a compressed format.|
4040
|*IngestionPropertyName*, *IngestionPropertyValue* | `string` | |Optional ingestion properties. For more information about ingestion properties, see [Data ingestion properties](../../ingestion-properties.md).|
41-
|*IngestionSource* | `string` | :heavy_check_mark: | The ingestion source. The source can be a URL, source file, or a list of blobs returned using the [.list blobs](list-blobs.md) command. |
41+
|*IngestionSource* | `string` | :heavy_check_mark: | The ingestion source. The source can be a URL or a list of blobs returned using [.list blobs](list-blobs.md) command. |
4242

4343
> [!NOTE]
4444
> The `.list blobs` command can be used with the `.ingest-from-storage-queued` command to return the blobs you want to ingest. For detailed information about the command and a full list of its parameters, see [.list blobs command](list-blobs.md).

0 commit comments

Comments
 (0)