/* Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments Copyright (C) ITsysCOM GmbH This program is free software: you can redistribute it and/or modify it under the terms of the GNU Affero General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more details. You should have received a copy of the GNU Affero General Public License along with this program. If not, see */ package ees import ( "bytes" "crypto/tls" "fmt" "net/http" "sync" "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/credentials" "github.com/aws/aws-sdk-go/aws/session" "github.com/aws/aws-sdk-go/service/s3/s3manager" "github.com/cgrates/cgrates/config" "github.com/cgrates/cgrates/utils" ) // NewS3EE creates a s3 poster func NewS3EE(cfg *config.EventExporterCfg, em *utils.ExporterMetrics) *S3EE { pstr := &S3EE{ cfg: cfg, em: em, reqs: newConcReq(cfg.ConcurrentRequests), } pstr.parseOpts(cfg.Opts) return pstr } // S3EE is a s3 poster type S3EE struct { awsRegion string awsID string awsKey string awsToken string bucket string folderPath string forcePathStyle bool skipTlsVerify bool session *session.Session up *s3manager.Uploader cfg *config.EventExporterCfg em *utils.ExporterMetrics reqs *concReq sync.RWMutex // protect connection bytePreparing } func (pstr *S3EE) parseOpts(opts *config.EventExporterOpts) { pstr.bucket = utils.DefaultQueueID if s3Opts := opts.AWS; s3Opts != nil { if s3Opts.S3BucketID != nil { pstr.bucket = *s3Opts.S3BucketID } if s3Opts.S3FolderPath != nil { pstr.folderPath = *s3Opts.S3FolderPath } if s3Opts.Region != nil { pstr.awsRegion = *s3Opts.Region } if s3Opts.Key != nil { pstr.awsID = *s3Opts.Key } if s3Opts.Secret != nil { pstr.awsKey = *s3Opts.Secret } if s3Opts.Token != nil { pstr.awsToken = *s3Opts.Token } if s3Opts.S3ForcePathStyle != nil { pstr.forcePathStyle = *s3Opts.S3ForcePathStyle } if s3Opts.S3SkipTlsVerify != nil { pstr.skipTlsVerify = *s3Opts.S3SkipTlsVerify } } } func (pstr *S3EE) Cfg() *config.EventExporterCfg { return pstr.cfg } func (pstr *S3EE) Connect() (err error) { pstr.Lock() defer pstr.Unlock() if pstr.session == nil { cfg := aws.Config{Endpoint: aws.String(pstr.Cfg().ExportPath)} if len(pstr.awsRegion) != 0 { cfg.Region = aws.String(pstr.awsRegion) } if len(pstr.awsID) != 0 && len(pstr.awsKey) != 0 { cfg.Credentials = credentials.NewStaticCredentials(pstr.awsID, pstr.awsKey, pstr.awsToken) } if pstr.forcePathStyle { cfg.S3ForcePathStyle = aws.Bool(true) // Required for custom S3-compatible endpoints } if pstr.skipTlsVerify { cfg.HTTPClient = &http.Client{ Transport: &http.Transport{ TLSClientConfig: &tls.Config{ InsecureSkipVerify: true, // Equivalent to verify=False for self-signed certificates }, }, } } pstr.session, err = session.NewSessionWithOptions( session.Options{ Config: cfg, }, ) if err != nil { return } } if pstr.up == nil { pstr.up, err = s3manager.NewUploader(pstr.session), nil } return } func (pstr *S3EE) ExportEvent(message any, key string) (err error) { pstr.reqs.get() pstr.RLock() _, err = pstr.up.Upload(&s3manager.UploadInput{ Bucket: aws.String(pstr.bucket), // Can also use the `filepath` standard library package to modify the // filename as need for an S3 object key. Such as turning absolute path // to a relative path. Key: aws.String(fmt.Sprintf("%s/%s.json", pstr.folderPath, key)), // The file to be uploaded. io.ReadSeeker is preferred as the Uploader // will be able to optimize memory when uploading large content. io.Reader // is supported, but will require buffering of the reader's bytes for // each part. Body: bytes.NewReader(message.([]byte)), }) pstr.RUnlock() pstr.reqs.done() return } func (pstr *S3EE) Close() (_ error) { return } func (pstr *S3EE) GetMetrics() *utils.ExporterMetrics { return pstr.em }