@@ -110,25 +110,13 @@ service RerunCloudService {
110
110
// * Latest-at, range and dataframe queries.
111
111
// * Arbitrary Lance filters.
112
112
//
113
- // To fetch the actual chunks themselves, see `GetChunks `.
113
+ // To fetch the actual chunks themselves, see `FetchChunks `.
114
114
//
115
115
// Passing chunk IDs to this method effectively acts as a IF_EXIST filter.
116
116
//
117
117
// This endpoint requires the standard dataset headers.
118
118
rpc QueryDataset (QueryDatasetRequest ) returns (stream QueryDatasetResponse ) {}
119
119
120
- // Perform Rerun-native queries on a dataset, returning the underlying chunks.
121
- //
122
- // These Rerun-native queries include:
123
- // * Filtering by specific partition and chunk IDs.
124
- // * Latest-at, range and dataframe queries.
125
- // * Arbitrary Lance filters.
126
- //
127
- // To fetch only the actual chunk IDs rather than the chunks themselves, see `QueryDataset`.
128
- //
129
- // This endpoint requires the standard dataset headers.
130
- rpc GetChunks (GetChunksRequest ) returns (stream GetChunksResponse ) {}
131
-
132
120
// Fetch specific chunks from Rerun Cloud. In a 2-step query process, result of 1st phase,
133
121
// that is, the result of `QueryDataset` should include all the necessary information to send
134
122
// the actual chunk requests, which is the 2nd step of the query process.
@@ -530,63 +518,6 @@ message QueryRange {
530
518
reserved "fuzzy_descriptors" ;
531
519
}
532
520
533
- message GetChunksRequest {
534
- // Client can specify from which partitions to get chunks. If left unspecified (empty list),
535
- // data from all partition (that match other query parameters) will be included.
536
- repeated rerun.common.v1alpha1.PartitionId partition_ids = 2 ;
537
-
538
- // Client can specify chunk ids to include. If left unspecified (empty list),
539
- // all chunks (that match other query parameters) will be included.
540
- repeated rerun.common.v1alpha1.Tuid chunk_ids = 3 ;
541
-
542
- // Which entity paths are we interested in? Leave empty, and set `select_all_entity_paths`,
543
- // in order to query all of them.
544
- repeated rerun.common.v1alpha1.EntityPath entity_paths = 4 ;
545
-
546
- // If set, the query will cover all existing entity paths.
547
- //
548
- // `entity_paths` must be empty, otherwise an error will be raised.
549
- //
550
- // Truth table:
551
- // ```text
552
- // select_all_entity_paths | entity_paths | result
553
- // ------------------------+----------------+--------
554
- // false | [] | valid query, empty results (no entity paths selected)
555
- // false | ['foo', 'bar'] | valid query, 'foo' & 'bar' selected
556
- // true | [] | valid query, all entity paths selected
557
- // true | ['foo', 'bar'] | invalid query, error
558
- // ```
559
- bool select_all_entity_paths = 6 ;
560
-
561
- // Which components are we interested in?
562
- //
563
- // If left unspecified, all existing components are considered of interest.
564
- //
565
- // This will perform a basic fuzzy match on the available columns' descriptors.
566
- // The fuzzy logic is a simple case-sensitive `contains()` query.
567
- // For example, given a `log_tick__SeriesLines:width` index, all of the following
568
- // would match: `SeriesLines:width`, `Width`, `SeriesLines`, etc.
569
- repeated string fuzzy_descriptors = 9 ;
570
-
571
- // If set, static data will be excluded from the results.
572
- bool exclude_static_data = 7 ;
573
-
574
- // If set, temporal data will be excluded from the results.
575
- bool exclude_temporal_data = 8 ;
576
-
577
- // Query details
578
- Query query = 5 ;
579
-
580
- reserved 1 ;
581
- reserved "dataset_id" ;
582
- }
583
-
584
- message GetChunksResponse {
585
- // Every gRPC response, even within the confines of a stream, involves HTTP2 overhead, which isn't
586
- // cheap by any means, which is why we're returning a batch of `ArrowMsg` rather than a single one.
587
- repeated rerun.log_msg.v1alpha1.ArrowMsg chunks = 1 ;
588
- }
589
-
590
521
message FetchChunksRequest {
591
522
// Information about the chunks to fetch. These dataframes have to include the following columns:
592
523
// * `chunk_id` - Chunk unique identifier
0 commit comments