mirror of
https://github.com/cgrates/cgrates.git
synced 2026-02-11 18:16:24 +05:00
Added S3 Event Reader
This commit is contained in:
committed by
Dan Christian Bogos
parent
edcfa792e6
commit
49e423d21a
@@ -81,11 +81,13 @@ func (erS *ERService) ListenAndServe(cfgRldChan chan struct{}) (err error) {
|
||||
for {
|
||||
select {
|
||||
case err = <-erS.rdrErr: // got application error
|
||||
erS.closeAllRdrs()
|
||||
utils.Logger.Crit(
|
||||
fmt.Sprintf("<%s> running reader got error: <%s>",
|
||||
utils.ERs, err.Error()))
|
||||
return
|
||||
case <-erS.stopChan:
|
||||
erS.closeAllRdrs()
|
||||
return
|
||||
case erEv := <-erS.rdrEvents:
|
||||
if err := erS.processEvent(erEv.cgrEvent, erEv.rdrCfg, erEv.opts); err != nil {
|
||||
@@ -302,3 +304,9 @@ func (erS *ERService) processEvent(cgrEv *utils.CGREvent,
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func (erS *ERService) closeAllRdrs() {
|
||||
for _, stopL := range erS.stopLsn {
|
||||
close(stopL)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -57,6 +57,8 @@ func NewEventReader(cfg *config.CGRConfig, cfgIdx int,
|
||||
return NewJSONFileER(cfg, cfgIdx, rdrEvents, rdrErr, fltrS, rdrExit)
|
||||
case utils.MetaAMQPjsonMap:
|
||||
return NewAMQPER(cfg, cfgIdx, rdrEvents, rdrErr, fltrS, rdrExit)
|
||||
case utils.MetaS3jsonMap:
|
||||
return NewS3ER(cfg, cfgIdx, rdrEvents, rdrErr, fltrS, rdrExit)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
244
ers/s3.go
Normal file
244
ers/s3.go
Normal file
@@ -0,0 +1,244 @@
|
||||
/*
|
||||
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
|
||||
Copyright (C) ITsysCOM GmbH
|
||||
|
||||
This program is free software: you can redistribute it and/or modify
|
||||
it under the terms of the GNU General Public License as published by
|
||||
the Free Software Foundation, either version 3 of the License, or
|
||||
(at your option) any later version.
|
||||
|
||||
This program is distributed in the hope that it will be useful,
|
||||
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
GNU General Public License for more details.
|
||||
|
||||
You should have received a copy of the GNU General Public License
|
||||
along with this program. If not, see <http://www.gnu.org/licenses/>
|
||||
*/
|
||||
|
||||
package ers
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"io/ioutil"
|
||||
"time"
|
||||
|
||||
"github.com/aws/aws-sdk-go/aws"
|
||||
"github.com/aws/aws-sdk-go/aws/credentials"
|
||||
"github.com/aws/aws-sdk-go/aws/session"
|
||||
"github.com/aws/aws-sdk-go/service/s3"
|
||||
"github.com/cgrates/cgrates/agents"
|
||||
"github.com/cgrates/cgrates/config"
|
||||
"github.com/cgrates/cgrates/engine"
|
||||
"github.com/cgrates/cgrates/utils"
|
||||
)
|
||||
|
||||
// NewS3ER return a new s3 event reader
|
||||
func NewS3ER(cfg *config.CGRConfig, cfgIdx int,
|
||||
rdrEvents chan *erEvent, rdrErr chan error,
|
||||
fltrS *engine.FilterS, rdrExit chan struct{}) (er EventReader, err error) {
|
||||
|
||||
rdr := &S3ER{
|
||||
cgrCfg: cfg,
|
||||
cfgIdx: cfgIdx,
|
||||
fltrS: fltrS,
|
||||
rdrEvents: rdrEvents,
|
||||
rdrExit: rdrExit,
|
||||
rdrErr: rdrErr,
|
||||
}
|
||||
if concReq := rdr.Config().ConcurrentReqs; concReq != -1 {
|
||||
rdr.cap = make(chan struct{}, concReq)
|
||||
for i := 0; i < concReq; i++ {
|
||||
rdr.cap <- struct{}{}
|
||||
}
|
||||
}
|
||||
rdr.parseOpts(rdr.Config().Opts)
|
||||
return rdr, nil
|
||||
}
|
||||
|
||||
// S3ER implements EventReader interface for kafka message
|
||||
type S3ER struct {
|
||||
// sync.RWMutex
|
||||
cgrCfg *config.CGRConfig
|
||||
cfgIdx int // index of config instance within ERsCfg.Readers
|
||||
fltrS *engine.FilterS
|
||||
|
||||
rdrEvents chan *erEvent // channel to dispatch the events created to
|
||||
rdrExit chan struct{}
|
||||
rdrErr chan error
|
||||
cap chan struct{}
|
||||
|
||||
awsRegion string
|
||||
awsID string
|
||||
awsKey string
|
||||
awsToken string
|
||||
queueID string
|
||||
session *session.Session
|
||||
|
||||
poster engine.Poster
|
||||
}
|
||||
|
||||
// Config returns the curent configuration
|
||||
func (rdr *S3ER) Config() *config.EventReaderCfg {
|
||||
return rdr.cgrCfg.ERsCfg().Readers[rdr.cfgIdx]
|
||||
}
|
||||
|
||||
// Serve will start the gorutines needed to watch the kafka topic
|
||||
func (rdr *S3ER) Serve() (err error) {
|
||||
var sess *session.Session
|
||||
cfg := aws.Config{Endpoint: aws.String(rdr.Config().SourcePath)}
|
||||
if len(rdr.awsRegion) != 0 {
|
||||
cfg.Region = aws.String(rdr.awsRegion)
|
||||
}
|
||||
if len(rdr.awsID) != 0 &&
|
||||
len(rdr.awsKey) != 0 {
|
||||
cfg.Credentials = credentials.NewStaticCredentials(rdr.awsID, rdr.awsKey, rdr.awsToken)
|
||||
}
|
||||
if sess, err = session.NewSessionWithOptions(session.Options{Config: cfg}); err != nil {
|
||||
return
|
||||
}
|
||||
rdr.session = sess
|
||||
|
||||
if rdr.Config().RunDelay == time.Duration(0) { // 0 disables the automatic read, maybe done per API
|
||||
return
|
||||
}
|
||||
|
||||
go rdr.readLoop() // read until the connection is closed
|
||||
return
|
||||
}
|
||||
|
||||
func (rdr *S3ER) processMessage(body []byte) (err error) {
|
||||
var decodedMessage map[string]interface{}
|
||||
if err = json.Unmarshal(body, &decodedMessage); err != nil {
|
||||
return
|
||||
}
|
||||
|
||||
agReq := agents.NewAgentRequest(
|
||||
utils.MapStorage(decodedMessage), nil,
|
||||
nil, nil, nil, rdr.Config().Tenant,
|
||||
rdr.cgrCfg.GeneralCfg().DefaultTenant,
|
||||
utils.FirstNonEmpty(rdr.Config().Timezone,
|
||||
rdr.cgrCfg.GeneralCfg().DefaultTimezone),
|
||||
rdr.fltrS, nil, nil) // create an AgentRequest
|
||||
var pass bool
|
||||
if pass, err = rdr.fltrS.Pass(agReq.Tenant, rdr.Config().Filters,
|
||||
agReq); err != nil || !pass {
|
||||
return
|
||||
}
|
||||
if err = agReq.SetFields(rdr.Config().Fields); err != nil {
|
||||
return
|
||||
}
|
||||
rdr.rdrEvents <- &erEvent{
|
||||
cgrEvent: config.NMAsCGREvent(agReq.CGRRequest, agReq.Tenant, utils.NestingSep),
|
||||
rdrCfg: rdr.Config(),
|
||||
opts: config.NMAsMapInterface(agReq.Opts, utils.NestingSep),
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (rdr *S3ER) parseOpts(opts map[string]interface{}) {
|
||||
rdr.queueID = utils.DefaultQueueID
|
||||
if val, has := opts[utils.QueueID]; has {
|
||||
rdr.queueID = utils.IfaceAsString(val)
|
||||
}
|
||||
if val, has := opts[utils.AWSRegion]; has {
|
||||
rdr.awsRegion = utils.IfaceAsString(val)
|
||||
}
|
||||
if val, has := opts[utils.AWSKey]; has {
|
||||
rdr.awsID = utils.IfaceAsString(val)
|
||||
}
|
||||
if val, has := opts[utils.AWSSecret]; has {
|
||||
rdr.awsKey = utils.IfaceAsString(val)
|
||||
}
|
||||
if val, has := opts[utils.AWSToken]; has {
|
||||
rdr.awsToken = utils.IfaceAsString(val)
|
||||
}
|
||||
}
|
||||
|
||||
func (rdr *S3ER) readLoop() (err error) {
|
||||
scv := s3.New(rdr.session)
|
||||
var keys []string
|
||||
if err = scv.ListObjectsV2Pages(&s3.ListObjectsV2Input{Bucket: aws.String(rdr.queueID)},
|
||||
func(lovo *s3.ListObjectsV2Output, b bool) bool {
|
||||
for _, objMeta := range lovo.Contents {
|
||||
if objMeta.Key != nil {
|
||||
keys = append(keys, *objMeta.Key)
|
||||
}
|
||||
}
|
||||
return !rdr.isClosed()
|
||||
}); err != nil {
|
||||
rdr.rdrErr <- err
|
||||
return
|
||||
}
|
||||
if rdr.isClosed() {
|
||||
return
|
||||
}
|
||||
for _, key := range keys {
|
||||
go rdr.readMsg(scv, key)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (rdr *S3ER) createPoster() {
|
||||
processedOpt := getProcessOptions(rdr.Config().Opts)
|
||||
if len(processedOpt) == 0 &&
|
||||
len(rdr.Config().ProcessedPath) == 0 {
|
||||
return
|
||||
}
|
||||
rdr.poster = engine.NewKafkaPoster(utils.FirstNonEmpty(rdr.Config().ProcessedPath, rdr.Config().SourcePath),
|
||||
rdr.cgrCfg.GeneralCfg().PosterAttempts, processedOpt)
|
||||
}
|
||||
|
||||
func (rdr *S3ER) isClosed() bool {
|
||||
select {
|
||||
case <-rdr.rdrExit:
|
||||
return true
|
||||
default:
|
||||
return false
|
||||
}
|
||||
}
|
||||
|
||||
func (rdr *S3ER) readMsg(scv *s3.S3, key string) (err error) {
|
||||
if rdr.Config().ConcurrentReqs != -1 {
|
||||
<-rdr.cap // do not try to read if the limit is reached
|
||||
defer func() { rdr.cap <- struct{}{} }()
|
||||
}
|
||||
if rdr.isClosed() {
|
||||
return
|
||||
}
|
||||
|
||||
obj, err := scv.GetObject(&s3.GetObjectInput{Bucket: &rdr.queueID, Key: &key})
|
||||
if err != nil {
|
||||
rdr.rdrErr <- err
|
||||
return
|
||||
}
|
||||
var msg []byte
|
||||
if msg, err = ioutil.ReadAll(obj.Body); err != nil {
|
||||
utils.Logger.Warning(
|
||||
fmt.Sprintf("<%s> decoding message %s error: %s",
|
||||
utils.ERs, key, err.Error()))
|
||||
return
|
||||
}
|
||||
obj.Body.Close()
|
||||
if err = rdr.processMessage(msg); err != nil {
|
||||
utils.Logger.Warning(
|
||||
fmt.Sprintf("<%s> processing message %s error: %s",
|
||||
utils.ERs, key, err.Error()))
|
||||
return
|
||||
}
|
||||
if _, err = scv.DeleteObject(&s3.DeleteObjectInput{Bucket: &rdr.queueID, Key: &key}); err != nil {
|
||||
rdr.rdrErr <- err
|
||||
return
|
||||
}
|
||||
|
||||
if rdr.poster != nil { // post it
|
||||
if err = rdr.poster.Post(msg, key); err != nil {
|
||||
utils.Logger.Warning(
|
||||
fmt.Sprintf("<%s> writing message %s error: %s",
|
||||
utils.ERs, key, err.Error()))
|
||||
return
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
132
ers/s3_it_test.go
Normal file
132
ers/s3_it_test.go
Normal file
@@ -0,0 +1,132 @@
|
||||
// +build integration
|
||||
|
||||
/*
|
||||
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
|
||||
Copyright (C) ITsysCOM GmbH
|
||||
|
||||
This program is free software: you can redistribute it and/or modify
|
||||
it under the terms of the GNU General Public License as published by
|
||||
the Free Software Foundation, either version 3 of the License, or
|
||||
(at your option) any later version.
|
||||
|
||||
This program is distributed in the hope that it will be useful,
|
||||
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
GNU General Public License for more details.
|
||||
|
||||
You should have received a copy of the GNU General Public License
|
||||
along with this program. If not, see <http://www.gnu.org/licenses/>
|
||||
*/
|
||||
|
||||
package ers
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"flag"
|
||||
"fmt"
|
||||
"reflect"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/aws/aws-sdk-go/aws"
|
||||
"github.com/aws/aws-sdk-go/aws/credentials"
|
||||
"github.com/aws/aws-sdk-go/aws/session"
|
||||
"github.com/aws/aws-sdk-go/service/s3/s3manager"
|
||||
"github.com/cgrates/cgrates/config"
|
||||
"github.com/cgrates/cgrates/engine"
|
||||
"github.com/cgrates/cgrates/utils"
|
||||
)
|
||||
|
||||
var (
|
||||
itTestS3 = flag.Bool("s4", false, "Run the test for S3Reader")
|
||||
)
|
||||
|
||||
func TestS3ER(t *testing.T) {
|
||||
if !*itTestS3 {
|
||||
t.SkipNow()
|
||||
}
|
||||
cfg, err := config.NewCGRConfigFromJSONStringWithDefaults(`{
|
||||
"ers": { // EventReaderService
|
||||
"enabled": true, // starts the EventReader service: <true|false>
|
||||
"readers": [
|
||||
{
|
||||
"id": "s3", // identifier of the EventReader profile
|
||||
"type": "*s3_json_map", // reader type <*file_csv>
|
||||
"run_delay": "-1", // sleep interval in seconds between consecutive runs, -1 to use automation via inotify or 0 to disable running all together
|
||||
"concurrent_requests": 1024, // maximum simultaneous requests/files to process, 0 for unlimited
|
||||
"source_path": "s3.us-east-2.amazonaws.com", // read data from this path
|
||||
// "processed_path": "/var/spool/cgrates/ers/out", // move processed data here
|
||||
"tenant": "cgrates.org", // tenant used by import
|
||||
"filters": [], // limit parsing based on the filters
|
||||
"flags": [], // flags to influence the event processing
|
||||
"opts": {
|
||||
"queueID": "cgrates-cdrs",
|
||||
"awsRegion": "us-east-2",
|
||||
"awsKey": "AWSAccessKeyId",
|
||||
"awsSecret": "AWSSecretKey",
|
||||
// "awsToken": "".
|
||||
},
|
||||
"fields":[ // import fields template, tag will match internally CDR field, in case of .csv value will be represented by index of the field value
|
||||
{"tag": "CGRID", "type": "*composed", "value": "~*req.CGRID", "path": "*cgreq.CGRID"},
|
||||
],
|
||||
},
|
||||
],
|
||||
},
|
||||
}`)
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
rdrEvents = make(chan *erEvent, 1)
|
||||
rdrErr = make(chan error, 1)
|
||||
rdrExit = make(chan struct{}, 1)
|
||||
|
||||
if rdr, err = NewS3ER(cfg, 1, rdrEvents,
|
||||
rdrErr, new(engine.FilterS), rdrExit); err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
s3Rdr := rdr.(*S3ER)
|
||||
var sess *session.Session
|
||||
awsCfg := aws.Config{Endpoint: aws.String(rdr.Config().SourcePath)}
|
||||
awsCfg.Region = aws.String(s3Rdr.awsRegion)
|
||||
awsCfg.Credentials = credentials.NewStaticCredentials(s3Rdr.awsID, s3Rdr.awsKey, s3Rdr.awsToken)
|
||||
|
||||
if sess, err = session.NewSessionWithOptions(session.Options{Config: awsCfg}); err != nil {
|
||||
return
|
||||
}
|
||||
scv := s3manager.NewUploader(sess)
|
||||
|
||||
randomCGRID := utils.UUIDSha1Prefix()
|
||||
scv.Upload(&s3manager.UploadInput{
|
||||
Bucket: aws.String(s3Rdr.queueID),
|
||||
Key: aws.String("home/test.json"),
|
||||
Body: bytes.NewReader([]byte(fmt.Sprintf(`{"CGRID": "%s"}`, randomCGRID))),
|
||||
})
|
||||
|
||||
if err = rdr.Serve(); err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
select {
|
||||
case err = <-rdrErr:
|
||||
t.Error(err)
|
||||
case ev := <-rdrEvents:
|
||||
if ev.rdrCfg.ID != "s3" {
|
||||
t.Errorf("Expected 's3' received `%s`", ev.rdrCfg.ID)
|
||||
}
|
||||
expected := &utils.CGREvent{
|
||||
Tenant: "cgrates.org",
|
||||
ID: ev.cgrEvent.ID,
|
||||
Time: ev.cgrEvent.Time,
|
||||
Event: map[string]interface{}{
|
||||
"CGRID": randomCGRID,
|
||||
},
|
||||
}
|
||||
if !reflect.DeepEqual(ev.cgrEvent, expected) {
|
||||
t.Errorf("Expected %s ,received %s", utils.ToJSON(expected), utils.ToJSON(ev.cgrEvent))
|
||||
}
|
||||
case <-time.After(10 * time.Second):
|
||||
t.Fatal("Timeout")
|
||||
}
|
||||
close(rdrExit)
|
||||
}
|
||||
@@ -112,6 +112,7 @@ cgrates (0.11.0~dev) UNRELEASED; urgency=medium
|
||||
* [AttributeS] Add support for *prefix and *suffix type
|
||||
* [ConfigS] Add "redis_" prefix to "dataDB" option for redis
|
||||
* [DataDB] Add support for redis with TLS connection ( + integration test )
|
||||
* [ERs] Added support for *s3_json_map type
|
||||
|
||||
-- DanB <danb@cgrates.org> Wed, 19 Feb 2020 13:25:52 +0200
|
||||
|
||||
|
||||
Reference in New Issue
Block a user