/* Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments Copyright (C) ITsysCOM GmbH This program is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program. If not, see */ package engine import ( "bytes" "compress/zlib" "fmt" "io/ioutil" "strings" "time" "github.com/cgrates/cgrates/config" "github.com/cgrates/cgrates/guardian" "github.com/cgrates/cgrates/utils" "github.com/cgrates/ltcache" "github.com/cgrates/mgo" "github.com/cgrates/mgo/bson" ) const ( colDst = "destinations" colRds = "reverse_destinations" colAct = "actions" colApl = "action_plans" colAAp = "account_action_plans" colTsk = "tasks" colAtr = "action_triggers" colRpl = "rating_plans" colRpf = "rating_profiles" colAcc = "accounts" colShg = "shared_groups" colLcr = "lcr_rules" colDcs = "derived_chargers" colAls = "aliases" colRCfgs = "reverse_aliases" colStq = "stat_qeues" colPbs = "pubsub" colUsr = "users" colCrs = "cdr_stats" colLht = "load_history" colVer = "versions" colRsP = "resource_profiles" colRFI = "request_filter_indexes" colTmg = "timings" colRes = "resources" colSqs = "statqueues" colSqp = "statqueue_profiles" colTps = "threshold_profiles" colThs = "thresholds" colFlt = "filters" colSpp = "supplier_profiles" colAttr = "attribute_profiles" ColCDRs = "cdrs" colCpp = "charger_profiles" ) var ( CGRIDLow = strings.ToLower(utils.CGRID) RunIDLow = strings.ToLower(utils.RunID) OrderIDLow = strings.ToLower(utils.OrderID) OriginHostLow = strings.ToLower(utils.OriginHost) OriginIDLow = strings.ToLower(utils.OriginID) ToRLow = strings.ToLower(utils.ToR) CDRHostLow = strings.ToLower(utils.OriginHost) CDRSourceLow = strings.ToLower(utils.Source) RequestTypeLow = strings.ToLower(utils.RequestType) DirectionLow = strings.ToLower(utils.Direction) TenantLow = strings.ToLower(utils.Tenant) CategoryLow = strings.ToLower(utils.Category) AccountLow = strings.ToLower(utils.Account) SubjectLow = strings.ToLower(utils.Subject) SupplierLow = strings.ToLower(utils.SUPPLIER) DisconnectCauseLow = strings.ToLower(utils.DISCONNECT_CAUSE) SetupTimeLow = strings.ToLower(utils.SetupTime) AnswerTimeLow = strings.ToLower(utils.AnswerTime) CreatedAtLow = strings.ToLower(utils.CreatedAt) UpdatedAtLow = strings.ToLower(utils.UpdatedAt) UsageLow = strings.ToLower(utils.Usage) PDDLow = strings.ToLower(utils.PDD) CostDetailsLow = strings.ToLower(utils.CostDetails) DestinationLow = strings.ToLower(utils.Destination) CostLow = strings.ToLower(utils.COST) ) func NewMongoStorage(host, port, db, user, pass, storageType string, cdrsIndexes []string, cacheCfg config.CacheConfig) (ms *MongoStorage, err error) { url := host if port != "" { url += ":" + port } if user != "" && pass != "" { url = fmt.Sprintf("%s:%s@%s", user, pass, url) } var dbName string if db != "" { url += "/" + db dbName = strings.Split(db, "?")[0] // remove extra info after ? } session, err := mgo.Dial(url) if err != nil { return nil, err } session.SetMode(mgo.Strong, true) ms = &MongoStorage{db: dbName, session: session, storageType: storageType, ms: NewCodecMsgpackMarshaler(), cacheCfg: cacheCfg, cdrsIndexes: cdrsIndexes} if cNames, err := session.DB(ms.db).CollectionNames(); err != nil { return nil, err } else if len(cNames) == 0 { // create indexes only if database is empty if err = ms.EnsureIndexes(); err != nil { return nil, err } } ms.cnter = utils.NewCounter(time.Now().UnixNano(), 0) return } type MongoStorage struct { session *mgo.Session db string storageType string // datadb, stordb ms Marshaler cacheCfg config.CacheConfig cdrsIndexes []string cnter *utils.Counter } func (ms *MongoStorage) conn(col string) (*mgo.Session, *mgo.Collection) { sessionCopy := ms.session.Copy() return sessionCopy, sessionCopy.DB(ms.db).C(col) } // EnsureIndexes creates db indexes func (ms *MongoStorage) EnsureIndexes() (err error) { dbSession := ms.session.Copy() defer dbSession.Close() db := dbSession.DB(ms.db) if ms.storageType == utils.DataDB { idx := mgo.Index{ Key: []string{"key"}, Unique: true, // Prevent two documents from having the same index key DropDups: false, // Drop documents with the same index key as a previously indexed one Background: false, // Build index in background and return immediately Sparse: false, // Only index documents containing the Key fields } for _, col := range []string{colAct, colApl, colAAp, colAtr, colDcs, colRpl, colLcr, colDst, colRds, colAls, colUsr, colLht} { if err = db.C(col).EnsureIndex(idx); err != nil { return } } idx = mgo.Index{ Key: []string{"tenant", "id"}, Unique: true, DropDups: false, Background: false, Sparse: false, } for _, col := range []string{colRsP, colRes, colSqs, colSqp, colTps, colThs, colSpp, colAttr, colFlt, colCpp} { if err = db.C(col).EnsureIndex(idx); err != nil { return } } idx = mgo.Index{ Key: []string{"id"}, Unique: true, DropDups: false, Background: false, Sparse: false, } for _, col := range []string{colRpf, colShg, colCrs, colAcc} { if err = db.C(col).EnsureIndex(idx); err != nil { return } } } if ms.storageType == utils.StorDB { idx := mgo.Index{ Key: []string{"tpid", "id"}, Unique: true, DropDups: false, Background: false, Sparse: false, } for _, col := range []string{utils.TBLTPTimings, utils.TBLTPDestinations, utils.TBLTPDestinationRates, utils.TBLTPRatingPlans, utils.TBLTPSharedGroups, utils.TBLTPCdrStats, utils.TBLTPActions, utils.TBLTPActionPlans, utils.TBLTPActionTriggers, utils.TBLTPStats, utils.TBLTPResources} { if err = db.C(col).EnsureIndex(idx); err != nil { return } } idx = mgo.Index{ Key: []string{"tpid", "direction", "tenant", "category", "subject", "loadid"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPRateProfiles).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{"tpid", "direction", "tenant", "category", "account", "subject"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPLcrs).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{"tpid", "tenant", "username"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPUsers).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{"tpid", "direction", "tenant", "category", "account", "subject", "context"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPLcrs).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{"tpid", "tenant", "category", "subject", "account", "loadid"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPDerivedChargers).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{"tpid", "direction", "tenant", "account", "loadid"}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.TBLTPDerivedChargers).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{CGRIDLow, RunIDLow, OriginIDLow}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.CDRsTBL).EnsureIndex(idx); err != nil { return } for _, idxKey := range ms.cdrsIndexes { idx = mgo.Index{ Key: []string{idxKey}, Unique: false, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.CDRsTBL).EnsureIndex(idx); err != nil { return } } idx = mgo.Index{ Key: []string{CGRIDLow, RunIDLow}, Unique: true, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.SessionsCostsTBL).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{OriginHostLow, OriginIDLow}, Unique: false, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.SessionsCostsTBL).EnsureIndex(idx); err != nil { return } idx = mgo.Index{ Key: []string{RunIDLow, OriginIDLow}, Unique: false, DropDups: false, Background: false, Sparse: false, } if err = db.C(utils.SessionsCostsTBL).EnsureIndex(idx); err != nil { return } } return } func (ms *MongoStorage) getColNameForPrefix(prefix string) (name string, ok bool) { colMap := map[string]string{ utils.DESTINATION_PREFIX: colDst, utils.REVERSE_DESTINATION_PREFIX: colRds, utils.ACTION_PREFIX: colAct, utils.ACTION_PLAN_PREFIX: colApl, utils.AccountActionPlansPrefix: colAAp, utils.TASKS_KEY: colTsk, utils.ACTION_TRIGGER_PREFIX: colAtr, utils.RATING_PLAN_PREFIX: colRpl, utils.RATING_PROFILE_PREFIX: colRpf, utils.ACCOUNT_PREFIX: colAcc, utils.SHARED_GROUP_PREFIX: colShg, utils.LCR_PREFIX: colLcr, utils.DERIVEDCHARGERS_PREFIX: colDcs, utils.ALIASES_PREFIX: colAls, utils.REVERSE_ALIASES_PREFIX: colRCfgs, utils.PUBSUB_SUBSCRIBERS_PREFIX: colPbs, utils.USERS_PREFIX: colUsr, utils.CDR_STATS_PREFIX: colCrs, utils.LOADINST_KEY: colLht, utils.VERSION_PREFIX: colVer, //utils.CDR_STATS_QUEUE_PREFIX: colStq, utils.TimingsPrefix: colTmg, utils.ResourcesPrefix: colRes, utils.ResourceProfilesPrefix: colRsP, utils.ThresholdProfilePrefix: colTps, utils.StatQueueProfilePrefix: colSqp, utils.ThresholdPrefix: colThs, utils.FilterPrefix: colFlt, utils.SupplierProfilePrefix: colSpp, utils.AttributeProfilePrefix: colAttr, } name, ok = colMap[prefix] return } func (ms *MongoStorage) Close() { ms.session.Close() } func (ms *MongoStorage) Flush(ignore string) (err error) { dbSession := ms.session.Copy() defer dbSession.Close() return dbSession.DB(ms.db).DropDatabase() } func (ms *MongoStorage) Marshaler() Marshaler { return ms.ms } // DB returnes a database object with cloned session inside func (ms *MongoStorage) DB() *mgo.Database { return ms.session.Copy().DB(ms.db) } func (ms *MongoStorage) SelectDatabase(dbName string) (err error) { ms.db = dbName return } func (ms *MongoStorage) RebuildReverseForPrefix(prefix string) (err error) { if !utils.IsSliceMember([]string{utils.REVERSE_DESTINATION_PREFIX, utils.REVERSE_ALIASES_PREFIX, utils.AccountActionPlansPrefix}, prefix) { return utils.ErrInvalidKey } colName, ok := ms.getColNameForPrefix(prefix) if !ok { return utils.ErrInvalidKey } session, col := ms.conn(colName) defer session.Close() if _, err := col.RemoveAll(bson.M{}); err != nil { return err } var keys []string switch prefix { case utils.REVERSE_DESTINATION_PREFIX: if keys, err = ms.GetKeysForPrefix(utils.DESTINATION_PREFIX); err != nil { return } for _, key := range keys { dest, err := ms.GetDestination(key[len(utils.DESTINATION_PREFIX):], true, utils.NonTransactional) if err != nil { return err } if err = ms.SetReverseDestination(dest, utils.NonTransactional); err != nil { return err } } case utils.REVERSE_ALIASES_PREFIX: if keys, err = ms.GetKeysForPrefix(utils.ALIASES_PREFIX); err != nil { return } for _, key := range keys { al, err := ms.GetAlias(key[len(utils.ALIASES_PREFIX):], true, utils.NonTransactional) if err != nil { return err } if err = ms.SetReverseAlias(al, utils.NonTransactional); err != nil { return err } } case utils.AccountActionPlansPrefix: if keys, err = ms.GetKeysForPrefix(utils.ACTION_PLAN_PREFIX); err != nil { return } for _, key := range keys { apl, err := ms.GetActionPlan(key[len(utils.ACTION_PLAN_PREFIX):], true, utils.NonTransactional) if err != nil { return err } for acntID := range apl.AccountIDs { if err = ms.SetAccountActionPlans(acntID, []string{apl.Id}, false); err != nil { return err } } } } return nil } func (ms *MongoStorage) RemoveReverseForPrefix(prefix string) (err error) { if !utils.IsSliceMember([]string{utils.REVERSE_DESTINATION_PREFIX, utils.REVERSE_ALIASES_PREFIX, utils.AccountActionPlansPrefix}, prefix) { return utils.ErrInvalidKey } colName, ok := ms.getColNameForPrefix(prefix) if !ok { return utils.ErrInvalidKey } session, col := ms.conn(colName) defer session.Close() if _, err := col.RemoveAll(bson.M{}); err != nil { return err } var keys []string switch prefix { case utils.REVERSE_DESTINATION_PREFIX: if keys, err = ms.GetKeysForPrefix(utils.DESTINATION_PREFIX); err != nil { return } for _, key := range keys { dest, err := ms.GetDestination(key[len(utils.DESTINATION_PREFIX):], true, utils.NonTransactional) if err != nil { return err } if err := ms.RemoveDestination(dest.Id, utils.NonTransactional); err != nil { return err } } case utils.REVERSE_ALIASES_PREFIX: if keys, err = ms.GetKeysForPrefix(utils.ALIASES_PREFIX); err != nil { return } for _, key := range keys { al, err := ms.GetAlias(key[len(utils.ALIASES_PREFIX):], true, utils.NonTransactional) if err != nil { return err } if err := ms.RemoveAlias(al.GetId(), utils.NonTransactional); err != nil { return err } } case utils.AccountActionPlansPrefix: if keys, err = ms.GetKeysForPrefix(utils.ACTION_PLAN_PREFIX); err != nil { return } for _, key := range keys { apl, err := ms.GetActionPlan(key[len(utils.ACTION_PLAN_PREFIX):], true, utils.NonTransactional) if err != nil { return err } for acntID := range apl.AccountIDs { if err = ms.RemAccountActionPlans(acntID, []string{apl.Id}); err != nil { return err } } } } return nil } func (ms *MongoStorage) IsDBEmpty() (resp bool, err error) { session := ms.session.Copy() defer session.Close() db := session.DB(ms.db) cols, err := db.CollectionNames() if err != nil { return } return len(cols) == 0 || cols[0] == "cdrs", nil } func (ms *MongoStorage) GetKeysForPrefix(prefix string) (result []string, err error) { var category, subject string keyLen := len(utils.DESTINATION_PREFIX) if len(prefix) < keyLen { return nil, fmt.Errorf("unsupported prefix in GetKeysForPrefix: %s", prefix) } category = prefix[:keyLen] // prefix length tntID := utils.NewTenantID(prefix[keyLen:]) subject = fmt.Sprintf("^%s", prefix[keyLen:]) // old way, no tenant support session := ms.session.Copy() defer session.Close() db := session.DB(ms.db) keyResult := struct{ Key string }{} idResult := struct{ Tenant, Id string }{} switch category { case utils.DESTINATION_PREFIX: iter := db.C(colDst).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.DESTINATION_PREFIX+keyResult.Key) } case utils.REVERSE_DESTINATION_PREFIX: iter := db.C(colRds).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.REVERSE_DESTINATION_PREFIX+keyResult.Key) } case utils.RATING_PLAN_PREFIX: iter := db.C(colRpl).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.RATING_PLAN_PREFIX+keyResult.Key) } case utils.RATING_PROFILE_PREFIX: iter := db.C(colRpf).Find(bson.M{"id": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.RATING_PROFILE_PREFIX+idResult.Id) } case utils.ACTION_PREFIX: iter := db.C(colAct).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.ACTION_PREFIX+keyResult.Key) } case utils.ACTION_PLAN_PREFIX: iter := db.C(colApl).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.ACTION_PLAN_PREFIX+keyResult.Key) } case utils.ACTION_TRIGGER_PREFIX: iter := db.C(colAtr).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.ACTION_TRIGGER_PREFIX+keyResult.Key) } case utils.SHARED_GROUP_PREFIX: iter := db.C(colShg).Find(bson.M{"id": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.SHARED_GROUP_PREFIX+idResult.Id) } case utils.DERIVEDCHARGERS_PREFIX: iter := db.C(colDcs).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.DERIVEDCHARGERS_PREFIX+keyResult.Key) } case utils.LCR_PREFIX: iter := db.C(colLcr).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.LCR_PREFIX+keyResult.Key) } case utils.ACCOUNT_PREFIX: iter := db.C(colAcc).Find(bson.M{"id": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ACCOUNT_PREFIX+idResult.Id) } case utils.ALIASES_PREFIX: iter := db.C(colAls).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.ALIASES_PREFIX+keyResult.Key) } case utils.REVERSE_ALIASES_PREFIX: iter := db.C(colRCfgs).Find(bson.M{"key": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"key": 1}).Iter() for iter.Next(&keyResult) { result = append(result, utils.REVERSE_ALIASES_PREFIX+keyResult.Key) } case utils.ResourceProfilesPrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colRsP).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ResourceProfilesPrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.ResourcesPrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colRes).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ResourcesPrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.StatQueuePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colSqs).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.StatQueuePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.StatQueueProfilePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colSqp).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.StatQueueProfilePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.AccountActionPlansPrefix: iter := db.C(colAAp).Find(bson.M{"id": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.AccountActionPlansPrefix+idResult.Id) } case utils.TimingsPrefix: iter := db.C(colTmg).Find(bson.M{"id": bson.M{"$regex": bson.RegEx{Pattern: subject}}}).Select(bson.M{"id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.TimingsPrefix+idResult.Id) } case utils.FilterPrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colFlt).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.FilterPrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.ThresholdPrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colThs).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ThresholdPrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.ThresholdProfilePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colTps).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ThresholdProfilePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.SupplierProfilePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colSpp).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.SupplierProfilePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.AttributeProfilePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colAttr).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.AttributeProfilePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } case utils.ChargerProfilePrefix: qry := bson.M{} if tntID.Tenant != "" { qry["tenant"] = tntID.Tenant } if tntID.ID != "" { qry["id"] = bson.M{"$regex": bson.RegEx{Pattern: subject}} } iter := db.C(colCpp).Find(qry).Select(bson.M{"tenant": 1, "id": 1}).Iter() for iter.Next(&idResult) { result = append(result, utils.ChargerProfilePrefix+utils.ConcatenatedKey(idResult.Tenant, idResult.Id)) } default: err = fmt.Errorf("unsupported prefix in GetKeysForPrefix: %s", prefix) } return } func (ms *MongoStorage) HasDataDrv(category, subject, tenant string) (has bool, err error) { session := ms.session.Copy() defer session.Close() db := session.DB(ms.db) var count int switch category { case utils.DESTINATION_PREFIX: count, err = db.C(colDst).Find(bson.M{"key": subject}).Count() has = count > 0 case utils.RATING_PLAN_PREFIX: count, err = db.C(colRpl).Find(bson.M{"key": subject}).Count() has = count > 0 case utils.RATING_PROFILE_PREFIX: count, err = db.C(colRpf).Find(bson.M{"id": subject}).Count() has = count > 0 case utils.ACTION_PREFIX: count, err = db.C(colAct).Find(bson.M{"key": subject}).Count() has = count > 0 case utils.ACTION_PLAN_PREFIX: count, err = db.C(colApl).Find(bson.M{"key": subject}).Count() has = count > 0 case utils.ACCOUNT_PREFIX: count, err = db.C(colAcc).Find(bson.M{"id": subject}).Count() has = count > 0 case utils.ResourcesPrefix: count, err = db.C(colRes).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.ResourceProfilesPrefix: count, err = db.C(colRsP).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.StatQueuePrefix: count, err = db.C(colSqs).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.StatQueueProfilePrefix: count, err = db.C(colSqp).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.ThresholdPrefix: count, err = db.C(colTps).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.FilterPrefix: count, err = db.C(colFlt).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.SupplierProfilePrefix: count, err = db.C(colSpp).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.AttributeProfilePrefix: count, err = db.C(colAttr).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 case utils.ChargerProfilePrefix: count, err = db.C(colCpp).Find(bson.M{"tenant": tenant, "id": subject}).Count() has = count > 0 default: err = fmt.Errorf("unsupported category in HasData: %s", category) } return } func (ms *MongoStorage) GetRatingPlanDrv(key string) (rp *RatingPlan, err error) { var kv struct { Key string Value []byte } session, col := ms.conn(colRpl) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } b := bytes.NewBuffer(kv.Value) r, err := zlib.NewReader(b) if err != nil { return nil, err } out, err := ioutil.ReadAll(r) if err != nil { return nil, err } r.Close() if err = ms.ms.Unmarshal(out, &rp); err != nil { return nil, err } return } func (ms *MongoStorage) SetRatingPlanDrv(rp *RatingPlan) error { result, err := ms.ms.Marshal(rp) if err != nil { return err } var b bytes.Buffer w := zlib.NewWriter(&b) w.Write(result) w.Close() session, col := ms.conn(colRpl) defer session.Close() _, err = col.Upsert(bson.M{"key": rp.Id}, &struct { Key string Value []byte }{Key: rp.Id, Value: b.Bytes()}) return err } func (ms *MongoStorage) RemoveRatingPlanDrv(key string) error { session, col := ms.conn(colRpl) defer session.Close() var kv struct { Key string Value []byte } var rp RatingPlan iter := col.Find(bson.M{"key": key}).Iter() for iter.Next(&kv) { if err := col.Remove(bson.M{"key": kv.Key}); err != nil { return err } b := bytes.NewBuffer(kv.Value) r, err := zlib.NewReader(b) if err != nil { return err } out, err := ioutil.ReadAll(r) if err != nil { return err } if err = ms.ms.Unmarshal(out, &rp); err != nil { return err } } return iter.Close() } func (ms *MongoStorage) GetRatingProfileDrv(key string) (rp *RatingProfile, err error) { session, col := ms.conn(colRpf) defer session.Close() if err = col.Find(bson.M{"id": key}).One(&rp); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err // Make sure we don't return new object on error } return } func (ms *MongoStorage) SetRatingProfileDrv(rp *RatingProfile) (err error) { session, col := ms.conn(colRpf) defer session.Close() if _, err = col.Upsert(bson.M{"id": rp.Id}, rp); err != nil { return } return } func (ms *MongoStorage) RemoveRatingProfileDrv(key string) error { session, col := ms.conn(colRpf) defer session.Close() iter := col.Find(bson.M{"id": key}).Select(bson.M{"id": 1}).Iter() var result struct{ Id string } for iter.Next(&result) { if err := col.Remove(bson.M{"id": result.Id}); err != nil { return err } } return iter.Close() } func (ms *MongoStorage) GetLCRDrv(key string) (lcr *LCR, err error) { var result struct { Key string Value *LCR } session, col := ms.conn(colLcr) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&result); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } lcr = result.Value return } func (ms *MongoStorage) SetLCRDrv(lcr *LCR) (err error) { session, col := ms.conn(colLcr) defer session.Close() if _, err = col.Upsert(bson.M{"key": lcr.GetId()}, &struct { Key string Value *LCR }{lcr.GetId(), lcr}); err != nil { return } return } func (ms *MongoStorage) RemoveLCRDrv(id, transactionID string) (err error) { session, col := ms.conn(colLcr) defer session.Close() err = col.Remove(bson.M{"key": id}) return err } func (ms *MongoStorage) GetDestination(key string, skipCache bool, transactionID string) (result *Destination, err error) { if !skipCache { if x, ok := Cache.Get(utils.CacheDestinations, key); ok { if x == nil { return nil, utils.ErrNotFound } return x.(*Destination), nil } } var kv struct { Key string Value []byte } session, col := ms.conn(colDst) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheDestinations, key, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return } b := bytes.NewBuffer(kv.Value) r, err := zlib.NewReader(b) if err != nil { return nil, err } out, err := ioutil.ReadAll(r) if err != nil { return nil, err } r.Close() err = ms.ms.Unmarshal(out, &result) if err != nil { return nil, err } Cache.Set(utils.CacheDestinations, key, result, nil, cacheCommit(transactionID), transactionID) return } func (ms *MongoStorage) SetDestination(dest *Destination, transactionID string) (err error) { result, err := ms.ms.Marshal(dest) if err != nil { return err } var b bytes.Buffer w := zlib.NewWriter(&b) w.Write(result) w.Close() session, col := ms.conn(colDst) defer session.Close() if _, err = col.Upsert(bson.M{"key": dest.Id}, &struct { Key string Value []byte }{Key: dest.Id, Value: b.Bytes()}); err != nil { return } return } func (ms *MongoStorage) GetReverseDestination(prefix string, skipCache bool, transactionID string) (ids []string, err error) { if !skipCache { if x, ok := Cache.Get(utils.CacheReverseDestinations, prefix); ok { if x == nil { return nil, utils.ErrNotFound } return x.([]string), nil } } var result struct { Key string Value []string } session, col := ms.conn(colRds) defer session.Close() if err = col.Find(bson.M{"key": prefix}).One(&result); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheReverseDestinations, prefix, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return } ids = result.Value Cache.Set(utils.CacheReverseDestinations, prefix, ids, nil, cacheCommit(transactionID), transactionID) return } func (ms *MongoStorage) SetReverseDestination(dest *Destination, transactionID string) (err error) { session, col := ms.conn(colRds) defer session.Close() for _, p := range dest.Prefixes { if _, err = col.Upsert(bson.M{"key": p}, bson.M{"$addToSet": bson.M{"value": dest.Id}}); err != nil { break } } return } func (ms *MongoStorage) RemoveDestination(destID string, transactionID string) (err error) { session, col := ms.conn(colDst) // get destination for prefix list d, err := ms.GetDestination(destID, false, transactionID) if err != nil { if err == mgo.ErrNotFound { err = nil } return } err = col.Remove(bson.M{"key": destID}) if err != nil { return err } Cache.Remove(utils.CacheDestinations, destID, cacheCommit(transactionID), transactionID) session.Close() session, col = ms.conn(colRds) defer session.Close() for _, prefix := range d.Prefixes { err = col.Update(bson.M{"key": prefix}, bson.M{"$pull": bson.M{"value": destID}}) if err != nil { return err } ms.GetReverseDestination(prefix, true, transactionID) // it will recache the destination } return } func (ms *MongoStorage) UpdateReverseDestination(oldDest, newDest *Destination, transactionID string) error { session, col := ms.conn(colRds) defer session.Close() //log.Printf("Old: %+v, New: %+v", oldDest, newDest) var obsoletePrefixes []string var addedPrefixes []string var found bool if oldDest == nil { oldDest = new(Destination) // so we can process prefixes } for _, oldPrefix := range oldDest.Prefixes { found = false for _, newPrefix := range newDest.Prefixes { if oldPrefix == newPrefix { found = true break } } if !found { obsoletePrefixes = append(obsoletePrefixes, oldPrefix) } } for _, newPrefix := range newDest.Prefixes { found = false for _, oldPrefix := range oldDest.Prefixes { if newPrefix == oldPrefix { found = true break } } if !found { addedPrefixes = append(addedPrefixes, newPrefix) } } //log.Print("Obsolete prefixes: ", obsoletePrefixes) //log.Print("Added prefixes: ", addedPrefixes) // remove id for all obsolete prefixes cCommit := cacheCommit(transactionID) var err error for _, obsoletePrefix := range obsoletePrefixes { err = col.Update(bson.M{"key": obsoletePrefix}, bson.M{"$pull": bson.M{"value": oldDest.Id}}) if err != nil { return err } Cache.Remove(utils.CacheReverseDestinations, obsoletePrefix, cCommit, transactionID) } // add the id to all new prefixes for _, addedPrefix := range addedPrefixes { _, err = col.Upsert(bson.M{"key": addedPrefix}, bson.M{"$addToSet": bson.M{"value": newDest.Id}}) if err != nil { return err } } return nil } func (ms *MongoStorage) GetActionsDrv(key string) (as Actions, err error) { var result struct { Key string Value Actions } session, col := ms.conn(colAct) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&result); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } as = result.Value return } func (ms *MongoStorage) SetActionsDrv(key string, as Actions) error { session, col := ms.conn(colAct) defer session.Close() _, err := col.Upsert(bson.M{"key": key}, &struct { Key string Value Actions }{Key: key, Value: as}) return err } func (ms *MongoStorage) RemoveActionsDrv(key string) error { session, col := ms.conn(colAct) defer session.Close() err := col.Remove(bson.M{"key": key}) return err } func (ms *MongoStorage) GetSharedGroupDrv(key string) (sg *SharedGroup, err error) { session, col := ms.conn(colShg) defer session.Close() if err = col.Find(bson.M{"id": key}).One(&sg); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetSharedGroupDrv(sg *SharedGroup) (err error) { session, col := ms.conn(colShg) defer session.Close() if _, err = col.Upsert(bson.M{"id": sg.Id}, sg); err != nil { return } return } func (ms *MongoStorage) RemoveSharedGroupDrv(id, transactionID string) (err error) { session, col := ms.conn(colShg) defer session.Close() err = col.Remove(bson.M{"id": id}) return err } func (ms *MongoStorage) GetAccount(key string) (result *Account, err error) { result = new(Account) session, col := ms.conn(colAcc) defer session.Close() err = col.Find(bson.M{"id": key}).One(result) if err == mgo.ErrNotFound { err = utils.ErrNotFound result = nil } return } func (ms *MongoStorage) SetAccount(acc *Account) error { // never override existing account with an empty one // UPDATE: if all balances expired and were cleaned it makes // sense to write empty balance map if len(acc.BalanceMap) == 0 { if ac, err := ms.GetAccount(acc.ID); err == nil && !ac.allBalancesExpired() { ac.ActionTriggers = acc.ActionTriggers ac.UnitCounters = acc.UnitCounters ac.AllowNegative = acc.AllowNegative ac.Disabled = acc.Disabled acc = ac } } session, col := ms.conn(colAcc) defer session.Close() _, err := col.Upsert(bson.M{"id": acc.ID}, acc) return err } func (ms *MongoStorage) RemoveAccount(key string) (err error) { session, col := ms.conn(colAcc) defer session.Close() if err = col.Remove(bson.M{"id": key}); err == mgo.ErrNotFound { err = utils.ErrNotFound } return } func (ms *MongoStorage) GetCdrStatsQueueDrv(key string) (sq *CDRStatsQueue, err error) { var result struct { Key string Value *CDRStatsQueue } session, col := ms.conn(colStq) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&result); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } sq = result.Value return } func (ms *MongoStorage) SetCdrStatsQueueDrv(sq *CDRStatsQueue) (err error) { session, col := ms.conn(colStq) defer session.Close() _, err = col.Upsert(bson.M{"key": sq.GetId()}, &struct { Key string Value *CDRStatsQueue }{Key: sq.GetId(), Value: sq}) return } func (ms *MongoStorage) RemoveCdrStatsQueueDrv(id string) (err error) { session, col := ms.conn(colStq) defer session.Close() if err = col.Remove(bson.M{"key": id}); err != nil && err != mgo.ErrNotFound { return } return nil } func (ms *MongoStorage) GetSubscribersDrv() (result map[string]*SubscriberData, err error) { session, col := ms.conn(colPbs) defer session.Close() iter := col.Find(nil).Iter() result = make(map[string]*SubscriberData) var kv struct { Key string Value *SubscriberData } for iter.Next(&kv) { result[kv.Key] = kv.Value } err = iter.Close() return } func (ms *MongoStorage) SetSubscriberDrv(key string, sub *SubscriberData) (err error) { session, col := ms.conn(colPbs) defer session.Close() _, err = col.Upsert(bson.M{"key": key}, &struct { Key string Value *SubscriberData }{Key: key, Value: sub}) return err } func (ms *MongoStorage) RemoveSubscriberDrv(key string) (err error) { session, col := ms.conn(colPbs) defer session.Close() return col.Remove(bson.M{"key": key}) } func (ms *MongoStorage) SetUserDrv(up *UserProfile) (err error) { session, col := ms.conn(colUsr) defer session.Close() _, err = col.Upsert(bson.M{"key": up.GetId()}, &struct { Key string Value *UserProfile }{Key: up.GetId(), Value: up}) return err } func (ms *MongoStorage) GetUserDrv(key string) (up *UserProfile, err error) { var kv struct { Key string Value *UserProfile } session, col := ms.conn(colUsr) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } up = kv.Value return } func (ms *MongoStorage) GetUsersDrv() (result []*UserProfile, err error) { session, col := ms.conn(colUsr) defer session.Close() iter := col.Find(nil).Iter() var kv struct { Key string Value *UserProfile } for iter.Next(&kv) { result = append(result, kv.Value) } err = iter.Close() return } func (ms *MongoStorage) RemoveUserDrv(key string) (err error) { session, col := ms.conn(colUsr) defer session.Close() return col.Remove(bson.M{"key": key}) } func (ms *MongoStorage) GetAlias(key string, skipCache bool, transactionID string) (al *Alias, err error) { if !skipCache { if x, ok := Cache.Get(utils.CacheAliases, key); ok { if x == nil { return nil, utils.ErrNotFound } al = x.(*Alias) return } } var kv struct { Key string Value AliasValues } session, col := ms.conn(colAls) defer session.Close() cCommit := cacheCommit(transactionID) if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheAliases, key, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return } al = &Alias{Values: kv.Value} al.SetId(key) Cache.Set(utils.CacheAliases, key, al, nil, cCommit, transactionID) return } func (ms *MongoStorage) SetAlias(al *Alias, transactionID string) (err error) { session, col := ms.conn(colAls) defer session.Close() if _, err = col.Upsert(bson.M{"key": al.GetId()}, &struct { Key string Value AliasValues }{Key: al.GetId(), Value: al.Values}); err != nil { return } return err } func (ms *MongoStorage) GetReverseAlias(reverseID string, skipCache bool, transactionID string) (ids []string, err error) { if !skipCache { if x, ok := Cache.Get(utils.CacheReverseAliases, reverseID); ok { if x == nil { return nil, utils.ErrNotFound } return x.([]string), nil } } var result struct { Key string Value []string } session, col := ms.conn(colRCfgs) defer session.Close() if err = col.Find(bson.M{"key": reverseID}).One(&result); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheReverseAliases, reverseID, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return } ids = result.Value Cache.Set(utils.CacheReverseAliases, reverseID, ids, nil, cacheCommit(transactionID), transactionID) return } func (ms *MongoStorage) SetReverseAlias(al *Alias, transactionID string) (err error) { session, col := ms.conn(colRCfgs) defer session.Close() for _, value := range al.Values { for target, pairs := range value.Pairs { for _, alias := range pairs { rKey := strings.Join([]string{alias, target, al.Context}, "") id := utils.ConcatenatedKey(al.GetId(), value.DestinationId) if _, err = col.Upsert(bson.M{"key": rKey}, bson.M{"$addToSet": bson.M{"value": id}}); err != nil { break } } } } return } func (ms *MongoStorage) RemoveAlias(key, transactionID string) (err error) { al := &Alias{} al.SetId(key) origKey := key key = utils.ALIASES_PREFIX + key var kv struct { Key string Value AliasValues } session, col := ms.conn(colAls) if err := col.Find(bson.M{"key": origKey}).One(&kv); err != nil { if err == mgo.ErrNotFound { err = nil } return err } al.Values = kv.Value if err = col.Remove(bson.M{"key": origKey}); err != nil { return } cCommit := cacheCommit(transactionID) Cache.Remove(utils.CacheAliases, key, cCommit, transactionID) session.Close() session, col = ms.conn(colRCfgs) defer session.Close() for _, value := range al.Values { tmpKey := utils.ConcatenatedKey(al.GetId(), value.DestinationId) for target, pairs := range value.Pairs { for _, alias := range pairs { rKey := alias + target + al.Context if err = col.Update(bson.M{"key": rKey}, bson.M{"$pull": bson.M{"value": tmpKey}}); err != nil { if err == mgo.ErrNotFound { err = nil // cancel the error not to be propagated with return bellow } else { return err } } Cache.Remove(utils.CacheReverseAliases, rKey, cCommit, transactionID) } } } return } // Limit will only retrieve the last n items out of history, newest first func (ms *MongoStorage) GetLoadHistory(limit int, skipCache bool, transactionID string) (loadInsts []*utils.LoadInstance, err error) { if limit == 0 { return nil, nil } if !skipCache { if x, ok := Cache.Get(utils.LOADINST_KEY, ""); ok { if x != nil { items := x.([]*utils.LoadInstance) if len(items) < limit || limit == -1 { return items, nil } return items[:limit], nil } return nil, utils.ErrNotFound } } var kv struct { Key string Value []*utils.LoadInstance } session, col := ms.conn(colLht) defer session.Close() err = col.Find(bson.M{"key": utils.LOADINST_KEY}).One(&kv) cCommit := cacheCommit(transactionID) if err == nil { loadInsts = kv.Value Cache.Remove(utils.LOADINST_KEY, "", cCommit, transactionID) Cache.Set(utils.LOADINST_KEY, "", loadInsts, nil, cCommit, transactionID) } if len(loadInsts) < limit || limit == -1 { return loadInsts, nil } return loadInsts[:limit], nil } // Adds a single load instance to load history func (ms *MongoStorage) AddLoadHistory(ldInst *utils.LoadInstance, loadHistSize int, transactionID string) error { if loadHistSize == 0 { // Load history disabled return nil } // get existing load history var existingLoadHistory []*utils.LoadInstance var kv struct { Key string Value []*utils.LoadInstance } session, col := ms.conn(colLht) defer session.Close() err := col.Find(bson.M{"key": utils.LOADINST_KEY}).One(&kv) if err != nil && err != mgo.ErrNotFound { return err } else { if kv.Value != nil { existingLoadHistory = kv.Value } } _, err = guardian.Guardian.Guard(func() (interface{}, error) { // Make sure we do it locked since other instance can modify history while we read it // insert on first position existingLoadHistory = append(existingLoadHistory, nil) copy(existingLoadHistory[1:], existingLoadHistory[0:]) existingLoadHistory[0] = ldInst //check length histLen := len(existingLoadHistory) if histLen >= loadHistSize { // Have hit maximum history allowed, remove oldest element in order to add new one existingLoadHistory = existingLoadHistory[:loadHistSize] } session, col := ms.conn(colLht) defer session.Close() _, err = col.Upsert(bson.M{"key": utils.LOADINST_KEY}, &struct { Key string Value []*utils.LoadInstance }{Key: utils.LOADINST_KEY, Value: existingLoadHistory}) return nil, err }, 0, utils.LOADINST_KEY) Cache.Remove(utils.LOADINST_KEY, "", cacheCommit(transactionID), transactionID) return err } func (ms *MongoStorage) GetActionTriggersDrv(key string) (atrs ActionTriggers, err error) { var kv struct { Key string Value ActionTriggers } session, col := ms.conn(colAtr) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return } atrs = kv.Value return } func (ms *MongoStorage) SetActionTriggersDrv(key string, atrs ActionTriggers) (err error) { session, col := ms.conn(colAtr) defer session.Close() if len(atrs) == 0 { err = col.Remove(bson.M{"key": key}) if err == mgo.ErrNotFound { // Overwrite not found since it is not really mandatory here to be returned err = nil } return } _, err = col.Upsert(bson.M{"key": key}, &struct { Key string Value ActionTriggers }{Key: key, Value: atrs}) return err } func (ms *MongoStorage) RemoveActionTriggersDrv(key string) error { session, col := ms.conn(colAtr) defer session.Close() err := col.Remove(bson.M{"key": key}) return err } func (ms *MongoStorage) GetActionPlan(key string, skipCache bool, transactionID string) (ats *ActionPlan, err error) { if !skipCache { if x, err := Cache.GetCloned(utils.CacheActionPlans, key); err != nil { if err != ltcache.ErrNotFound { // Only consider cache if item was found return nil, err } } else if x == nil { // item was placed nil in cache return nil, utils.ErrNotFound } else { return x.(*ActionPlan), nil } } var kv struct { Key string Value []byte } session, col := ms.conn(colApl) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheActionPlans, key, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return nil, err } b := bytes.NewBuffer(kv.Value) r, err := zlib.NewReader(b) if err != nil { return nil, err } out, err := ioutil.ReadAll(r) if err != nil { return nil, err } r.Close() if err = ms.ms.Unmarshal(out, &ats); err != nil { return nil, err } Cache.Set(utils.CacheActionPlans, key, ats, nil, cacheCommit(transactionID), transactionID) return } func (ms *MongoStorage) SetActionPlan(key string, ats *ActionPlan, overwrite bool, transactionID string) (err error) { session, col := ms.conn(colApl) defer session.Close() // clean dots from account ids map cCommit := cacheCommit(transactionID) if len(ats.ActionTimings) == 0 { if err = col.Remove(bson.M{"key": key}); err != nil && err == mgo.ErrNotFound { err = nil // NotFound is good } Cache.Remove(utils.CacheActionPlans, key, cCommit, transactionID) return } if !overwrite { // get existing action plan to merge the account ids if existingAts, _ := ms.GetActionPlan(key, true, transactionID); existingAts != nil { if ats.AccountIDs == nil && len(existingAts.AccountIDs) > 0 { ats.AccountIDs = make(utils.StringMap) } for accID := range existingAts.AccountIDs { ats.AccountIDs[accID] = true } } } result, err := ms.ms.Marshal(ats) if err != nil { return err } var b bytes.Buffer w := zlib.NewWriter(&b) w.Write(result) w.Close() if _, err = col.Upsert(bson.M{"key": key}, &struct { Key string Value []byte }{Key: key, Value: b.Bytes()}); err != nil { return } return err } func (ms *MongoStorage) RemoveActionPlan(key string, transactionID string) error { session, col := ms.conn(colApl) defer session.Close() cCommit := cacheCommit(transactionID) Cache.Remove(utils.CacheActionPlans, key, cCommit, transactionID) err := col.Remove(bson.M{"key": key}) if err != nil && err == mgo.ErrNotFound { err = nil } return err } func (ms *MongoStorage) GetAllActionPlans() (ats map[string]*ActionPlan, err error) { keys, err := ms.GetKeysForPrefix(utils.ACTION_PLAN_PREFIX) if err != nil { return nil, err } ats = make(map[string]*ActionPlan, len(keys)) for _, key := range keys { ap, err := ms.GetActionPlan(key[len(utils.ACTION_PLAN_PREFIX):], false, utils.NonTransactional) if err != nil { return nil, err } ats[key[len(utils.ACTION_PLAN_PREFIX):]] = ap } return } func (ms *MongoStorage) GetAccountActionPlans(acntID string, skipCache bool, transactionID string) (aPlIDs []string, err error) { if !skipCache { if x, ok := Cache.Get(utils.CacheAccountActionPlans, acntID); ok { if x == nil { return nil, utils.ErrNotFound } return x.([]string), nil } } session, col := ms.conn(colAAp) defer session.Close() var kv struct { Key string Value []string } if err = col.Find(bson.M{"key": acntID}).One(&kv); err != nil { if err == mgo.ErrNotFound { Cache.Set(utils.CacheAccountActionPlans, acntID, nil, nil, cacheCommit(transactionID), transactionID) err = utils.ErrNotFound } return } aPlIDs = kv.Value Cache.Set(utils.CacheAccountActionPlans, acntID, aPlIDs, nil, cacheCommit(transactionID), transactionID) return } func (ms *MongoStorage) SetAccountActionPlans(acntID string, aPlIDs []string, overwrite bool) (err error) { session, col := ms.conn(colAAp) defer session.Close() if !overwrite { if oldaPlIDs, err := ms.GetAccountActionPlans(acntID, false, utils.NonTransactional); err != nil && err != utils.ErrNotFound { return err } else { for _, oldAPid := range oldaPlIDs { if !utils.IsSliceMember(aPlIDs, oldAPid) { aPlIDs = append(aPlIDs, oldAPid) } } } } _, err = col.Upsert(bson.M{"key": acntID}, &struct { Key string Value []string }{Key: acntID, Value: aPlIDs}) return } func (ms *MongoStorage) RemAccountActionPlans(acntID string, aPlIDs []string) (err error) { session, col := ms.conn(colAAp) defer session.Close() if len(aPlIDs) == 0 { err = col.Remove(bson.M{"key": acntID}) if err == mgo.ErrNotFound { err = utils.ErrNotFound } } oldAPlIDs, err := ms.GetAccountActionPlans(acntID, true, utils.NonTransactional) if err != nil { return err } for i := 0; i < len(oldAPlIDs); { if utils.IsSliceMember(aPlIDs, oldAPlIDs[i]) { oldAPlIDs = append(oldAPlIDs[:i], oldAPlIDs[i+1:]...) continue // if we have stripped, don't increase index so we can check next element by next run } i++ } if len(oldAPlIDs) == 0 { // no more elements, remove the reference return col.Remove(bson.M{"key": acntID}) } _, err = col.Upsert(bson.M{"key": acntID}, &struct { Key string Value []string }{Key: acntID, Value: oldAPlIDs}) return } func (ms *MongoStorage) PushTask(t *Task) error { session, col := ms.conn(colTsk) defer session.Close() return col.Insert(bson.M{"_id": bson.NewObjectId(), "task": t}) } func (ms *MongoStorage) PopTask() (t *Task, err error) { v := struct { ID bson.ObjectId `bson:"_id"` Task *Task }{} session, col := ms.conn(colTsk) defer session.Close() if err = col.Find(nil).One(&v); err == nil { if remErr := col.Remove(bson.M{"_id": v.ID}); remErr != nil { return nil, remErr } t = v.Task } return } func (ms *MongoStorage) GetDerivedChargersDrv(key string) (dcs *utils.DerivedChargers, err error) { var kv struct { Key string Value *utils.DerivedChargers } session, col := ms.conn(colDcs) defer session.Close() if err = col.Find(bson.M{"key": key}).One(&kv); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return } dcs = kv.Value return } func (ms *MongoStorage) SetDerivedChargers(key string, dcs *utils.DerivedChargers, transactionID string) (err error) { cCommit := cacheCommit(transactionID) if dcs == nil || len(dcs.Chargers) == 0 { session, col := ms.conn(colDcs) defer session.Close() if err = col.Remove(bson.M{"key": key}); err != nil && err != mgo.ErrNotFound { return } Cache.Remove(utils.CacheDerivedChargers, key, cCommit, transactionID) return nil } session, col := ms.conn(colDcs) defer session.Close() if _, err = col.Upsert(bson.M{"key": key}, &struct { Key string Value *utils.DerivedChargers }{Key: key, Value: dcs}); err != nil { return } return err } func (ms *MongoStorage) RemoveDerivedChargersDrv(id, transactionID string) (err error) { cCommit := cacheCommit(transactionID) session, col := ms.conn(colDcs) defer session.Close() if err = col.Remove(bson.M{"key": id}); err != nil && err != mgo.ErrNotFound { return } Cache.Remove(utils.CacheDerivedChargers, id, cCommit, transactionID) return nil } func (ms *MongoStorage) SetCdrStatsDrv(cs *CdrStats) error { session, col := ms.conn(colCrs) defer session.Close() _, err := col.Upsert(bson.M{"id": cs.Id}, cs) return err } func (ms *MongoStorage) GetCdrStatsDrv(key string) (cs *CdrStats, err error) { cs = new(CdrStats) session, col := ms.conn(colCrs) defer session.Close() if err = col.Find(bson.M{"id": key}).One(cs); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) GetAllCdrStatsDrv() (css []*CdrStats, err error) { session, col := ms.conn(colCrs) defer session.Close() iter := col.Find(nil).Iter() var cs CdrStats for iter.Next(&cs) { clone := cs // avoid using the same pointer in append css = append(css, &clone) } err = iter.Close() return } func (ms *MongoStorage) GetResourceProfileDrv(tenant, id string) (rp *ResourceProfile, err error) { session, col := ms.conn(colRsP) defer session.Close() rp = new(ResourceProfile) if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(rp); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetResourceProfileDrv(rp *ResourceProfile) (err error) { session, col := ms.conn(colRsP) defer session.Close() _, err = col.Upsert(bson.M{"tenant": rp.Tenant, "id": rp.ID}, rp) return } func (ms *MongoStorage) RemoveResourceProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colRsP) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetResourceDrv(tenant, id string) (r *Resource, err error) { session, col := ms.conn(colRes) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetResourceDrv(r *Resource) (err error) { session, col := ms.conn(colRes) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveResourceDrv(tenant, id string) (err error) { session, col := ms.conn(colRes) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetTimingDrv(id string) (t *utils.TPTiming, err error) { session, col := ms.conn(colTmg) defer session.Close() if err = col.Find(bson.M{"id": id}).One(&t); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetTimingDrv(t *utils.TPTiming) (err error) { session, col := ms.conn(colTmg) defer session.Close() _, err = col.Upsert(bson.M{"id": t.ID}, t) return } func (ms *MongoStorage) RemoveTimingDrv(id string) (err error) { session, col := ms.conn(colTmg) defer session.Close() if err = col.Remove(bson.M{"id": id}); err != nil { return } return nil } // GetFilterIndexesDrv retrieves Indexes from dataDB //filterType is used togheter with fieldName:Val func (ms *MongoStorage) GetFilterIndexesDrv(cacheID, itemIDPrefix, filterType string, fldNameVal map[string]string) (indexes map[string]utils.StringMap, err error) { session, col := ms.conn(colRFI) defer session.Close() type result struct { Key string Value []string } var results []result dbKey := utils.CacheInstanceToPrefix[cacheID] + itemIDPrefix if len(fldNameVal) != 0 { for fldName, fldValue := range fldNameVal { if err = col.Find(bson.M{"key": utils.ConcatenatedKey(dbKey, filterType, fldName, fldValue)}).All(&results); err != nil { if err == mgo.ErrNotFound { return nil, utils.ErrNotFound } } } } else { for _, character := range []string{".", "*"} { dbKey = strings.Replace(dbKey, character, `\`+character, strings.Count(dbKey, character)) } //inside bson.RegEx add carrot to match the prefix (optimization) findWithoutIndexes := bson.M{"key": bson.M{"$regex": bson.RegEx{"^" + dbKey, ""}}} if err = col.Find(findWithoutIndexes).All(&results); err != nil { if err == mgo.ErrNotFound { return nil, utils.ErrNotFound } } } indexes = make(map[string]utils.StringMap) for _, res := range results { if len(res.Value) == 0 { continue } keys := strings.Split(res.Key, ":") indexKey := utils.ConcatenatedKey(keys[1], keys[2], keys[3]) //check here if itemIDPrefix has context if len(strings.Split(itemIDPrefix, ":")) == 2 { indexKey = utils.ConcatenatedKey(keys[2], keys[3], keys[4]) } if _, hasIt := indexes[indexKey]; !hasIt { indexes[indexKey] = make(utils.StringMap) } indexes[indexKey] = utils.StringMapFromSlice(res.Value) } if len(indexes) == 0 { return nil, utils.ErrNotFound } return indexes, nil } // SetFilterIndexesDrv stores Indexes into DataDB func (ms *MongoStorage) SetFilterIndexesDrv(cacheID, itemIDPrefix string, indexes map[string]utils.StringMap, commit bool, transactionID string) (err error) { session, col := ms.conn(colRFI) defer session.Close() originKey := utils.CacheInstanceToPrefix[cacheID] + itemIDPrefix dbKey := originKey if transactionID != "" { dbKey = "tmp_" + utils.ConcatenatedKey(originKey, transactionID) } if commit && transactionID != "" { regexKey := originKey for _, character := range []string{".", "*"} { regexKey = strings.Replace(regexKey, character, `\`+character, strings.Count(regexKey, character)) } //inside bson.RegEx add carrot to match the prefix (optimization) _, err = col.RemoveAll(bson.M{"key": bson.M{"$regex": bson.RegEx{"^" + regexKey, ""}}}) if err != nil && err != mgo.ErrNotFound { return } pairs := []interface{}{} for key, itmMp := range indexes { pairs = append(pairs, bson.M{"key": utils.ConcatenatedKey(originKey, key)}) pairs = append(pairs, bson.M{"$set": bson.M{"key": utils.ConcatenatedKey(originKey, key), "value": itmMp.Slice()}}) } if len(pairs) != 0 { bulk := col.Bulk() bulk.Unordered() bulk.Upsert(pairs...) _, err = bulk.Run() } oldKey := "tmp_" + utils.ConcatenatedKey(originKey, transactionID) for _, character := range []string{".", "*"} { oldKey = strings.Replace(oldKey, character, `\`+character, strings.Count(oldKey, character)) } //inside bson.RegEx add carrot to match the prefix (optimization) _, err = col.RemoveAll(bson.M{"key": bson.M{"$regex": bson.RegEx{"^" + oldKey, ""}}}) return err } else { pairs := []interface{}{} for key, itmMp := range indexes { pairs = append(pairs, bson.M{"key": utils.ConcatenatedKey(dbKey, key)}) if len(itmMp) == 0 { pairs = append(pairs, bson.M{"$unset": bson.M{"value": 1}}) } else { pairs = append(pairs, bson.M{"$set": bson.M{"key": utils.ConcatenatedKey(dbKey, key), "value": itmMp.Slice()}}) } } if len(pairs) != 0 { bulk := col.Bulk() bulk.Unordered() bulk.Upsert(pairs...) _, err = bulk.Run() } } return } func (ms *MongoStorage) RemoveFilterIndexesDrv(cacheID, itemIDPrefix string) (err error) { session, col := ms.conn(colRFI) defer session.Close() regexKey := utils.CacheInstanceToPrefix[cacheID] + itemIDPrefix for _, character := range []string{".", "*"} { regexKey = strings.Replace(regexKey, character, `\`+character, strings.Count(regexKey, character)) } //inside bson.RegEx add carrot to match the prefix (optimization) _, err = col.RemoveAll(bson.M{"key": bson.M{"$regex": bson.RegEx{"^" + regexKey, ""}}}) // redis compatibility if err == mgo.ErrNotFound { err = nil } return } func (ms *MongoStorage) MatchFilterIndexDrv(cacheID, itemIDPrefix, filterType, fldName, fldVal string) (itemIDs utils.StringMap, err error) { session, col := ms.conn(colRFI) defer session.Close() var result struct { Key string Value []string } dbKey := utils.CacheInstanceToPrefix[cacheID] + itemIDPrefix if err = col.Find( bson.M{"key": utils.ConcatenatedKey(dbKey, filterType, fldName, fldVal)}).One(&result); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } itemIDs = utils.StringMapFromSlice(result.Value) return } // GetStatQueueProfileDrv retrieves a StatQueueProfile from dataDB func (ms *MongoStorage) GetStatQueueProfileDrv(tenant string, id string) (sq *StatQueueProfile, err error) { session, col := ms.conn(colSqp) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&sq); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } // SetStatQueueProfileDrv stores a StatsQueue into DataDB func (ms *MongoStorage) SetStatQueueProfileDrv(sq *StatQueueProfile) (err error) { session, col := ms.conn(colSqp) defer session.Close() _, err = col.UpsertId(bson.M{"tenant": sq.Tenant, "id": sq.ID}, sq) return } // RemStatQueueProfileDrv removes a StatsQueue from dataDB func (ms *MongoStorage) RemStatQueueProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colSqp) err = col.Remove(bson.M{"tenant": tenant, "id": id}) if err != nil { return err } session.Close() return } // GetStoredStatQueueDrv retrieves a StoredStatQueue func (ms *MongoStorage) GetStoredStatQueueDrv(tenant, id string) (sq *StoredStatQueue, err error) { session, col := ms.conn(colSqs) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&sq); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } // SetStoredStatQueueDrv stores the metrics for a StoredStatQueue func (ms *MongoStorage) SetStoredStatQueueDrv(sq *StoredStatQueue) (err error) { session, col := ms.conn(colSqs) defer session.Close() _, err = col.Upsert(bson.M{"tenant": sq.Tenant, "id": sq.ID}, sq) return } // RemStoredStatQueueDrv removes stored metrics for a StoredStatQueue func (ms *MongoStorage) RemStoredStatQueueDrv(tenant, id string) (err error) { session, col := ms.conn(colSqs) defer session.Close() err = col.Remove(bson.M{"tenant": tenant, "id": id}) if err == mgo.ErrNotFound { err = utils.ErrNotFound } return err } // GetThresholdProfileDrv retrieves a ThresholdProfile from dataDB func (ms *MongoStorage) GetThresholdProfileDrv(tenant, ID string) (tp *ThresholdProfile, err error) { session, col := ms.conn(colTps) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": ID}).One(&tp); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } // SetThresholdProfileDrv stores a ThresholdProfile into DataDB func (ms *MongoStorage) SetThresholdProfileDrv(tp *ThresholdProfile) (err error) { session, col := ms.conn(colTps) defer session.Close() _, err = col.UpsertId(bson.M{"tenant": tp.Tenant, "id": tp.ID}, tp) return } // RemoveThresholdProfile removes a ThresholdProfile from dataDB/cache func (ms *MongoStorage) RemThresholdProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colTps) defer session.Close() err = col.Remove(bson.M{"tenant": tenant, "id": id}) if err != nil { return err } return } func (ms *MongoStorage) GetThresholdDrv(tenant, id string) (r *Threshold, err error) { session, col := ms.conn(colThs) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetThresholdDrv(r *Threshold) (err error) { session, col := ms.conn(colThs) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveThresholdDrv(tenant, id string) (err error) { session, col := ms.conn(colThs) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetFilterDrv(tenant, id string) (r *Filter, err error) { session, col := ms.conn(colFlt) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } for _, fltr := range r.Rules { if err = fltr.CompileValues(); err != nil { return } } return } func (ms *MongoStorage) SetFilterDrv(r *Filter) (err error) { session, col := ms.conn(colFlt) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveFilterDrv(tenant, id string) (err error) { session, col := ms.conn(colFlt) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetSupplierProfileDrv(tenant, id string) (r *SupplierProfile, err error) { session, col := ms.conn(colSpp) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetSupplierProfileDrv(r *SupplierProfile) (err error) { session, col := ms.conn(colSpp) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveSupplierProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colSpp) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetAttributeProfileDrv(tenant, id string) (r *AttributeProfile, err error) { session, col := ms.conn(colAttr) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetAttributeProfileDrv(r *AttributeProfile) (err error) { session, col := ms.conn(colAttr) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveAttributeProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colAttr) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil } func (ms *MongoStorage) GetChargerProfileDrv(tenant, id string) (r *ChargerProfile, err error) { session, col := ms.conn(colCpp) defer session.Close() if err = col.Find(bson.M{"tenant": tenant, "id": id}).One(&r); err != nil { if err == mgo.ErrNotFound { err = utils.ErrNotFound } return nil, err } return } func (ms *MongoStorage) SetChargerProfileDrv(r *ChargerProfile) (err error) { session, col := ms.conn(colCpp) defer session.Close() _, err = col.Upsert(bson.M{"tenant": r.Tenant, "id": r.ID}, r) return } func (ms *MongoStorage) RemoveChargerProfileDrv(tenant, id string) (err error) { session, col := ms.conn(colCpp) defer session.Close() if err = col.Remove(bson.M{"tenant": tenant, "id": id}); err != nil { return } return nil }