@@ -432,7 +432,7 @@ func (s *Source) pageChunker(
432
432
433
433
for objIdx , obj := range metadata .page .Contents {
434
434
if obj == nil {
435
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "nil_object" )
435
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "nil_object" , 0 )
436
436
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
437
437
ctx .Logger ().Error (err , "could not update progress for nil object" )
438
438
}
@@ -448,7 +448,7 @@ func (s *Source) pageChunker(
448
448
// Skip GLACIER and GLACIER_IR objects.
449
449
if obj .StorageClass == nil || strings .Contains (* obj .StorageClass , "GLACIER" ) {
450
450
ctx .Logger ().V (5 ).Info ("Skipping object in storage class" , "storage_class" , * obj .StorageClass )
451
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "storage_class" )
451
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "storage_class" , float64 ( * obj . Size ) )
452
452
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
453
453
ctx .Logger ().Error (err , "could not update progress for glacier object" )
454
454
}
@@ -458,7 +458,7 @@ func (s *Source) pageChunker(
458
458
// Ignore large files.
459
459
if * obj .Size > s .maxObjectSize {
460
460
ctx .Logger ().V (5 ).Info ("Skipping %d byte file (over maxObjectSize limit)" )
461
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "size_limit" )
461
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "size_limit" , float64 ( * obj . Size ) )
462
462
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
463
463
ctx .Logger ().Error (err , "could not update progress for large file" )
464
464
}
@@ -468,7 +468,7 @@ func (s *Source) pageChunker(
468
468
// File empty file.
469
469
if * obj .Size == 0 {
470
470
ctx .Logger ().V (5 ).Info ("Skipping empty file" )
471
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "empty_file" )
471
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "empty_file" , 0 )
472
472
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
473
473
ctx .Logger ().Error (err , "could not update progress for empty file" )
474
474
}
@@ -478,7 +478,7 @@ func (s *Source) pageChunker(
478
478
// Skip incompatible extensions.
479
479
if common .SkipFile (* obj .Key ) {
480
480
ctx .Logger ().V (5 ).Info ("Skipping file with incompatible extension" )
481
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "incompatible_extension" )
481
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "incompatible_extension" , float64 ( * obj . Size ) )
482
482
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
483
483
ctx .Logger ().Error (err , "could not update progress for incompatible file" )
484
484
}
@@ -493,7 +493,7 @@ func (s *Source) pageChunker(
493
493
494
494
if strings .HasSuffix (* obj .Key , "/" ) {
495
495
ctx .Logger ().V (5 ).Info ("Skipping directory" )
496
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "directory" )
496
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "directory" , float64 ( * obj . Size ) )
497
497
return nil
498
498
}
499
499
@@ -521,7 +521,7 @@ func (s *Source) pageChunker(
521
521
if err != nil {
522
522
if strings .Contains (err .Error (), "AccessDenied" ) {
523
523
ctx .Logger ().Error (err , "could not get S3 object; access denied" )
524
- s .metricsCollector .RecordObjectSkipped (metadata .bucket , "access_denied" )
524
+ s .metricsCollector .RecordObjectSkipped (metadata .bucket , "access_denied" , float64 ( * obj . Size ) )
525
525
} else {
526
526
ctx .Logger ().Error (err , "could not get S3 object" )
527
527
s .metricsCollector .RecordObjectError (metadata .bucket )
@@ -596,7 +596,7 @@ func (s *Source) pageChunker(
596
596
if err := s .checkpointer .UpdateObjectCompletion (ctx , objIdx , metadata .bucket , metadata .page .Contents ); err != nil {
597
597
ctx .Logger ().Error (err , "could not update progress for scanned object" )
598
598
}
599
- s .metricsCollector .RecordObjectScanned (metadata .bucket )
599
+ s .metricsCollector .RecordObjectScanned (metadata .bucket , float64 ( * obj . Size ) )
600
600
601
601
return nil
602
602
})
0 commit comments