mirror of
https://github.com/trufflesecurity/trufflehog.git
synced 2024-11-14 00:47:21 +00:00
d317ddb51a
* [chore] Remove logrus from circleci, filesystem, gitlab, and s3 sources * Address comments
182 lines
4.3 KiB
Go
182 lines
4.3 KiB
Go
package filesystem
|
|
|
|
import (
|
|
"fmt"
|
|
"io"
|
|
"io/fs"
|
|
"os"
|
|
"path/filepath"
|
|
|
|
diskbufferreader "github.com/bill-rich/disk-buffer-reader"
|
|
"github.com/go-errors/errors"
|
|
"github.com/go-logr/logr"
|
|
"google.golang.org/protobuf/proto"
|
|
"google.golang.org/protobuf/types/known/anypb"
|
|
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/common"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/context"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/handlers"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/pb/source_metadatapb"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/pb/sourcespb"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/sanitizer"
|
|
"github.com/trufflesecurity/trufflehog/v3/pkg/sources"
|
|
)
|
|
|
|
const (
|
|
// These buffer sizes are mainly driven by our largest credential size, which is GCP @ ~2.25KB.
|
|
// Having a peek size larger than that ensures that we have complete credential coverage in our chunks.
|
|
BufferSize = 10 * 1024 // 10KB
|
|
PeekSize = 3 * 1024 // 3KB
|
|
)
|
|
|
|
type Source struct {
|
|
name string
|
|
sourceId int64
|
|
jobId int64
|
|
verify bool
|
|
paths []string
|
|
log logr.Logger
|
|
filter *common.Filter
|
|
sources.Progress
|
|
}
|
|
|
|
// Ensure the Source satisfies the interface at compile time
|
|
var _ sources.Source = (*Source)(nil)
|
|
|
|
// Type returns the type of source.
|
|
// It is used for matching source types in configuration and job input.
|
|
func (s *Source) Type() sourcespb.SourceType {
|
|
return sourcespb.SourceType_SOURCE_TYPE_FILESYSTEM
|
|
}
|
|
|
|
func (s *Source) SourceID() int64 {
|
|
return s.sourceId
|
|
}
|
|
|
|
func (s *Source) JobID() int64 {
|
|
return s.jobId
|
|
}
|
|
|
|
// Init returns an initialized Filesystem source.
|
|
func (s *Source) Init(aCtx context.Context, name string, jobId, sourceId int64, verify bool, connection *anypb.Any, _ int) error {
|
|
s.log = aCtx.Logger()
|
|
|
|
s.name = name
|
|
s.sourceId = sourceId
|
|
s.jobId = jobId
|
|
s.verify = verify
|
|
|
|
var conn sourcespb.Filesystem
|
|
if err := anypb.UnmarshalTo(connection, &conn, proto.UnmarshalOptions{}); err != nil {
|
|
return errors.WrapPrefix(err, "error unmarshalling connection", 0)
|
|
}
|
|
|
|
s.paths = conn.Directories
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Source) WithFilter(filter *common.Filter) {
|
|
s.filter = filter
|
|
}
|
|
|
|
// Chunks emits chunks of bytes over a channel.
|
|
func (s *Source) Chunks(ctx context.Context, chunksChan chan *sources.Chunk) error {
|
|
for i, path := range s.paths {
|
|
s.SetProgressComplete(i, len(s.paths), fmt.Sprintf("Path: %s", path), "")
|
|
|
|
cleanPath := filepath.Clean(path)
|
|
done := false
|
|
go func() {
|
|
<-ctx.Done()
|
|
done = true
|
|
}()
|
|
|
|
err := fs.WalkDir(os.DirFS(cleanPath), ".", func(relativePath string, d fs.DirEntry, err error) error {
|
|
if err != nil {
|
|
return nil
|
|
}
|
|
|
|
path := filepath.Join(cleanPath, relativePath)
|
|
logger := ctx.Logger().WithValues("file", path)
|
|
|
|
fileStat, err := os.Stat(path)
|
|
if err != nil {
|
|
logger.Error(err, "unable to stat file")
|
|
return nil
|
|
}
|
|
if !fileStat.Mode().IsRegular() {
|
|
return nil
|
|
}
|
|
|
|
if s.filter != nil && !s.filter.Pass(path) {
|
|
return nil
|
|
}
|
|
|
|
inputFile, err := os.Open(path)
|
|
if err != nil {
|
|
logger.V(2).Info("error opening file", "error", err)
|
|
return nil
|
|
}
|
|
defer inputFile.Close()
|
|
logger.V(5).Info("scanning file")
|
|
|
|
reReader, err := diskbufferreader.New(inputFile)
|
|
if err != nil {
|
|
logger.Error(err, "Could not create re-readable reader.")
|
|
}
|
|
defer reReader.Close()
|
|
|
|
chunkSkel := &sources.Chunk{
|
|
SourceType: s.Type(),
|
|
SourceName: s.name,
|
|
SourceID: s.SourceID(),
|
|
SourceMetadata: &source_metadatapb.MetaData{
|
|
Data: &source_metadatapb.MetaData_Filesystem{
|
|
Filesystem: &source_metadatapb.Filesystem{
|
|
File: sanitizer.UTF8(path),
|
|
},
|
|
},
|
|
},
|
|
Verify: s.verify,
|
|
}
|
|
if handlers.HandleFile(ctx, reReader, chunkSkel, chunksChan) {
|
|
return nil
|
|
}
|
|
|
|
if err := reReader.Reset(); err != nil {
|
|
return err
|
|
}
|
|
reReader.Stop()
|
|
data, err := io.ReadAll(reReader)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
chunksChan <- &sources.Chunk{
|
|
SourceType: s.Type(),
|
|
SourceName: s.name,
|
|
SourceID: s.SourceID(),
|
|
Data: data,
|
|
SourceMetadata: &source_metadatapb.MetaData{
|
|
Data: &source_metadatapb.MetaData_Filesystem{
|
|
Filesystem: &source_metadatapb.Filesystem{
|
|
File: sanitizer.UTF8(path),
|
|
},
|
|
},
|
|
},
|
|
Verify: s.verify,
|
|
}
|
|
return nil
|
|
})
|
|
|
|
if err != nil && err != io.EOF {
|
|
return errors.New(err)
|
|
}
|
|
|
|
if done {
|
|
return nil
|
|
}
|
|
|
|
}
|
|
return nil
|
|
}
|