@@ -12,7 +12,6 @@ import (
1212 "fmt"
1313 "io"
1414 "io/ioutil"
15- "log/slog"
1615 "net"
1716 "os"
1817 "path"
@@ -37,6 +36,7 @@ import (
3736 "github.com/klauspost/compress/zstd"
3837 "github.com/peterebden/go-cli-init/v4/logging"
3938 "github.com/prometheus/client_golang/prometheus"
39+ "github.com/sirupsen/logrus"
4040 "gocloud.dev/blob"
4141 "gocloud.dev/gcerrors"
4242 "golang.org/x/exp/slices"
@@ -65,6 +65,7 @@ const (
6565const DefaultMaxSize int64 = 200 * 1012 // 200 Kelly-Bootle standard units
6666
6767var log = logging .MustGetLogger ()
68+ var logr = logrus .New ()
6869
6970// emptyHash is the sha256 hash of the empty file.
7071var emptyHash = digest .Empty .Hash
@@ -294,7 +295,9 @@ func (s *server) UpdateActionResult(ctx context.Context, req *pb.UpdateActionRes
294295 InstanceName : req .InstanceName ,
295296 ActionDigest : req .ActionDigest ,
296297 }); err == nil {
297- slog .Debug ("Returning existing action result for UpdateActionResult request" , "hash" , req .ActionDigest .Hash )
298+ logr .WithFields (logrus.Fields {
299+ "hash" : req .ActionDigest .Hash ,
300+ }).Debug ("Returning existing action result for UpdateActionResult request" )
298301 return ar , nil
299302 }
300303 }
@@ -404,14 +407,18 @@ func (s *server) BatchUpdateBlobs(ctx context.Context, req *pb.BatchUpdateBlobsR
404407 rr .Status .Message = fmt .Sprintf ("Blob sizes do not match (%d / %d)" , len (r .Data ), r .Digest .SizeBytes )
405408 blobSizeMismatches .Inc ()
406409 } else if s .blobExists (ctx , "cas" , r .Digest , compressed , true ) {
407- slog .Debug ("Blob already exists remotely" , "hash" , r .Digest .Hash )
410+ logr .WithFields (logrus.Fields {
411+ "hash" : r .Digest .Hash ,
412+ }).Debug ("Blob already exists remotely" )
408413 } else if err := s .writeAll (ctx , r .Digest , r .Data , compressed ); err != nil {
409414 log .Errorf ("Error writing blob %s: %s" , r .Digest , err )
410415 rr .Status .Code = int32 (status .Code (err ))
411416 rr .Status .Message = err .Error ()
412417 blobsReceived .WithLabelValues (batchLabel (true , false ), compressorLabel (compressed )).Inc ()
413418 } else {
414- slog .Debug ("Stored blob" , "hash" , r .Digest .Hash )
419+ logr .WithFields (logrus.Fields {
420+ "hash" : r .Digest .Hash ,
421+ }).Debug ("Stored blob" )
415422 }
416423 wg .Done ()
417424 bytesReceived .WithLabelValues (batchLabel (true , false ), compressorLabel (compressed )).Add (float64 (r .Digest .SizeBytes ))
@@ -497,12 +504,17 @@ func (s *server) Read(req *bs.ReadRequest, srv bs.ByteStream_ReadServer) error {
497504 if err != nil {
498505 return err
499506 }
500- slog .Debug ("Received ByteStream.Read request" , "hash" , digest .Hash )
507+ logr .WithFields (logrus.Fields {
508+ "hash" : digest .Hash ,
509+ }).Debug ("Received ByteStream.Read request" )
501510 if req .ReadOffset < 0 || req .ReadOffset > digest .SizeBytes {
502511 return status .Errorf (codes .OutOfRange , "Invalid Read() request; offset %d is outside the range of blob %s which is %d bytes long" , req .ReadOffset , digest .Hash , digest .SizeBytes )
503512 } else if req .ReadOffset == digest .SizeBytes {
504513 // We know there is nothing left to read, just return immediately.
505- slog .Debug ("Completed ByteStream.Read request immediately at final byte" , "bytes" , digest .SizeBytes , "hash" , digest .Hash )
514+ logr .WithFields (logrus.Fields {
515+ "hash" : digest .Hash ,
516+ "bytes" : digest .SizeBytes ,
517+ }).Debug ("Completed ByteStream.Read request immediately at final byte" )
506518 return nil
507519 } else if req .ReadLimit == 0 || req .ReadOffset + req .ReadLimit >= digest .SizeBytes {
508520 req .ReadLimit = - 1
@@ -525,7 +537,12 @@ func (s *server) Read(req *bs.ReadRequest, srv bs.ByteStream_ReadServer) error {
525537 return err
526538 }
527539 bytesServed .WithLabelValues (batchLabel (false , true ), compressorLabel (compressed )).Add (float64 (n ))
528- slog .Debug ("Completed ByteStream.Read request of bytes_number bytes (starting at read_offset)" , "bytes_number" , n , "read_offset" , req .ReadOffset , "hash" , digest .Hash , "duration" , time .Since (start ))
540+ logr .WithFields (logrus.Fields {
541+ "hash" : digest .Hash ,
542+ "bytes" : n ,
543+ "readOffset" : req .ReadOffset ,
544+ "duration" : time .Since (start ),
545+ }).Debug ("Completed ByteStream.Read request (starting at readOffset)" )
529546 return nil
530547}
531548
@@ -588,13 +605,17 @@ func (s *server) Write(srv bs.ByteStream_WriteServer) error {
588605 if err != nil {
589606 return err
590607 }
591- slog .Debug ("Received ByteStream.Write request" , "hash" , digest .Hash )
608+ logr .WithFields (logrus.Fields {
609+ "hash" : digest .Hash ,
610+ }).Debug ("Received ByteStream.Write request" )
592611 r := & bytestreamReader {stream : srv , buf : req .Data }
593612 if err := s .writeBlob (ctx , "cas" , digest , bufio .NewReaderSize (r , 65536 ), compressed ); err != nil {
594613 return err
595614 }
596615 bytesReceived .WithLabelValues (batchLabel (false , true ), compressorLabel (compressed )).Add (float64 (r .TotalSize ))
597- slog .Debug ("Stored blob with hash" , "hash" , digest .Hash )
616+ logr .WithFields (logrus.Fields {
617+ "hash" : digest .Hash ,
618+ }).Debug ("Stored blob with hash" )
598619 blobsReceived .WithLabelValues (batchLabel (false , true ), compressorLabel (compressed )).Inc ()
599620 return srv .SendAndClose (& bs.WriteResponse {
600621 CommittedSize : r .TotalSize ,
0 commit comments