This commit is contained in:
edwardro22
2017-09-13 09:16:11 +00:00
15 changed files with 645 additions and 74 deletions

View File

@@ -47,9 +47,10 @@ func (self *ApierV1) GetTPRate(attrs AttrGetTPRate, reply *utils.TPRate) error {
return utils.NewErrMandatoryIeMissing(missing...)
}
if rs, err := self.StorDb.GetTPRates(attrs.TPid, attrs.ID); err != nil {
return utils.NewErrServerError(err)
} else if len(rs) == 0 {
return utils.ErrNotFound
if err.Error() != utils.ErrNotFound.Error() {
err = utils.NewErrServerError(err)
}
return err
} else {
*reply = *rs[0]
}

240
apier/v1/tprates_it_test.go Normal file
View File

@@ -0,0 +1,240 @@
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
Copyright (C) ITsysCOM GmbH
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>
*/
package v1
import (
"github.com/cgrates/cgrates/config"
"github.com/cgrates/cgrates/engine"
"github.com/cgrates/cgrates/utils"
"net/rpc"
"net/rpc/jsonrpc"
"path"
"reflect"
"testing"
)
var (
tpRateCfgPath string
tpRateCfg *config.CGRConfig
tpRateRPC *rpc.Client
tpRateDataDir = "/usr/share/cgrates"
tpRate *utils.TPRate
tpRateDelay int
tpRateConfigDIR string //run tests for specific configuration
)
var sTestsTPRates = []func(t *testing.T){
testTPRatesInitCfg,
testTPRatesResetStorDb,
testTPRatesStartEngine,
testTPRatesRpcConn,
testTPRatesGetTPRateforeSet,
testTPRatesSetTPRate,
testTPRatesGetTPRateAfterSet,
testTPRatesGetTPRateIds,
testTPRatesUpdateTPRate,
testTPRatesGetTPRateAfterUpdate,
testTPRatesRemTPRate,
testTPRatesGetTPRateAfterRemove,
testTPRatesKillEngine,
}
//Test start here
func TestTPRatesITMySql(t *testing.T) {
tpRateConfigDIR = "tutmysql"
for _, stest := range sTestsTPRates {
t.Run(tpRateConfigDIR, stest)
}
}
func TestTPRatesITMongo(t *testing.T) {
tpRateConfigDIR = "tutmongo"
for _, stest := range sTestsTPRates {
t.Run(tpRateConfigDIR, stest)
}
}
func TestTPRatesITPG(t *testing.T) {
tpRateConfigDIR = "tutpostgres"
for _, stest := range sTestsTPRates {
t.Run(tpRateConfigDIR, stest)
}
}
func testTPRatesInitCfg(t *testing.T) {
var err error
tpRateCfgPath = path.Join(tpRateDataDir, "conf", "samples", tpRateConfigDIR)
tpRateCfg, err = config.NewCGRConfigFromFolder(tpRateCfgPath)
if err != nil {
t.Error(err)
}
tpRateCfg.DataFolderPath = tpRateDataDir // Share DataFolderPath through config towards StoreDb for Flush()
config.SetCgrConfig(tpRateCfg)
switch tpRateConfigDIR {
case "tutmongo": // Mongo needs more time to reset db, need to investigate
tpRateDelay = 2000
default:
tpRateDelay = 1000
}
}
// Wipe out the cdr database
func testTPRatesResetStorDb(t *testing.T) {
if err := engine.InitStorDb(tpRateCfg); err != nil {
t.Fatal(err)
}
}
// Start CGR Engine
func testTPRatesStartEngine(t *testing.T) {
if _, err := engine.StopStartEngine(tpRateCfgPath, tpRateDelay); err != nil {
t.Fatal(err)
}
}
// Connect rpc client to rater
func testTPRatesRpcConn(t *testing.T) {
var err error
tpRateRPC, err = jsonrpc.Dial("tcp", tpRateCfg.RPCJSONListen) // We connect over JSON so we can also troubleshoot if needed
if err != nil {
t.Fatal(err)
}
}
func testTPRatesGetTPRateforeSet(t *testing.T) {
var reply *utils.TPRate
if err := tpRateRPC.Call("ApierV1.GetTPRate", &AttrGetTPRate{TPid: "TPidTpRate", ID: "RT_FS_USERS"}, &reply); err == nil || err.Error() != utils.ErrNotFound.Error() {
t.Error(err)
}
}
func testTPRatesSetTPRate(t *testing.T) {
tpRate = &utils.TPRate{
TPid: "TPidTpRate",
ID: "RT_FS_USERS",
RateSlots: []*utils.RateSlot{
&utils.RateSlot{
ConnectFee: 12,
Rate: 3,
RateUnit: "6s",
RateIncrement: "6s",
GroupIntervalStart: "0s",
},
&utils.RateSlot{
ConnectFee: 12,
Rate: 3,
RateUnit: "4s",
RateIncrement: "6s",
GroupIntervalStart: "1s",
},
},
}
var result string
if err := tpRateRPC.Call("ApierV1.SetTPRate", tpRate, &result); err != nil {
t.Error(err)
} else if result != utils.OK {
t.Error("Unexpected reply returned", result)
}
}
func testTPRatesGetTPRateAfterSet(t *testing.T) {
var reply *utils.TPRate
if err := tpRateRPC.Call("ApierV1.GetTPRate", &AttrGetTPRate{TPid: "TPidTpRate", ID: tpRate.ID}, &reply); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(tpRate, reply) {
t.Errorf("Expecting : %+v, received: %+v", tpRate, reply)
}
}
func testTPRatesGetTPRateIds(t *testing.T) {
var result []string
expectedTPID := []string{"RT_FS_USERS"}
if err := tpRateRPC.Call("ApierV1.GetTPRateIds", &AttrGetTPRateIds{TPid: "TPidTpRate"}, &result); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(expectedTPID, result) {
t.Errorf("Expecting: %+v, received: %+v", expectedTPID, result)
}
}
func testTPRatesUpdateTPRate(t *testing.T) {
var result string
tpRate.RateSlots = []*utils.RateSlot{
&utils.RateSlot{
ConnectFee: 12,
Rate: 3,
RateUnit: "6s",
RateIncrement: "6s",
GroupIntervalStart: "0s",
},
&utils.RateSlot{
ConnectFee: 12,
Rate: 10,
RateUnit: "4s",
RateIncrement: "6s",
GroupIntervalStart: "1s",
},
&utils.RateSlot{
ConnectFee: 5,
Rate: 10,
RateUnit: "4s",
RateIncrement: "6s",
GroupIntervalStart: "3s",
},
}
if err := tpRateRPC.Call("ApierV1.SetTPRate", tpRate, &result); err != nil {
t.Error(err)
} else if result != utils.OK {
t.Error("Unexpected reply returned", result)
}
}
func testTPRatesGetTPRateAfterUpdate(t *testing.T) {
var reply *utils.TPRate
if err := tpRateRPC.Call("ApierV1.GetTPRate", &AttrGetTPRate{TPid: "TPidTpRate", ID: tpRate.ID}, &reply); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(tpRate, reply) {
t.Errorf("Expecting : %+v, received: %+v", tpRate, reply)
}
}
func testTPRatesRemTPRate(t *testing.T) {
var resp string
if err := tpRateRPC.Call("ApierV1.RemTPRate", &AttrGetTPRate{TPid: "TPidTpRate", ID: "RT_FS_USERS"}, &resp); err != nil {
t.Error(err)
} else if resp != utils.OK {
t.Error("Unexpected reply returned", resp)
}
}
func testTPRatesGetTPRateAfterRemove(t *testing.T) {
var reply *utils.TPRate
if err := tpRateRPC.Call("ApierV1.GetTPRate", &AttrGetTPRate{TPid: "TPidTpRate", ID: "RT_FS_USERS"}, &reply); err == nil || err.Error() != utils.ErrNotFound.Error() {
t.Error(err)
}
}
func testTPRatesKillEngine(t *testing.T) {
if err := engine.KillEngine(tpRateDelay); err != nil {
t.Error(err)
}
}

View File

@@ -46,7 +46,7 @@ type AttrGetTPRatingProfileByLoadId struct {
func (self *ApierV1) GetTPRatingProfilesByLoadId(attrs utils.TPRatingProfile, reply *[]*utils.TPRatingProfile) error {
mndtryFlds := []string{"TPid", "LoadId"}
if len(attrs.Subject) != 0 { // If Subject provided as filter, make all related fields mandatory
mndtryFlds = append(mndtryFlds, "Tenant", "TOR", "Direction", "Subject")
mndtryFlds = append(mndtryFlds, "Tenant", "Category", "Direction", "Subject")
}
if missing := utils.MissingStructFields(&attrs, mndtryFlds); len(missing) != 0 { //Params missing
return utils.NewErrMandatoryIeMissing(missing...)
@@ -68,7 +68,7 @@ func (self *ApierV1) GetTPRatingProfileLoadIds(attrs utils.AttrTPRatingProfileId
}
if ids, err := self.StorDb.GetTpTableIds(attrs.TPid, utils.TBLTPRateProfiles, utils.TPDistinctIds{"loadid"}, map[string]string{
"tenant": attrs.Tenant,
"tor": attrs.Category,
"category": attrs.Category,
"direction": attrs.Direction,
"subject": attrs.Subject,
}, new(utils.Paginator)); err != nil {
@@ -96,9 +96,10 @@ func (self *ApierV1) GetTPRatingProfile(attrs AttrGetTPRatingProfile, reply *uti
return err
}
if rpfs, err := self.StorDb.GetTPRatingProfiles(tmpRpf); err != nil {
return utils.NewErrServerError(err)
} else if len(rpfs) == 0 {
return utils.ErrNotFound
if err.Error() != utils.ErrNotFound.Error() {
err = utils.NewErrServerError(err)
}
return err
} else {
*reply = *rpfs[0]
}

View File

@@ -0,0 +1,262 @@
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
Copyright (C) ITsysCOM GmbH
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>
*/
package v1
import (
"github.com/cgrates/cgrates/config"
"github.com/cgrates/cgrates/engine"
"github.com/cgrates/cgrates/utils"
"net/rpc"
"net/rpc/jsonrpc"
"path"
"reflect"
"testing"
)
var (
tpRatingProfileCfgPath string
tpRatingProfileCfg *config.CGRConfig
tpRatingProfileRPC *rpc.Client
tpRatingProfileDataDir = "/usr/share/cgrates"
tpRatingProfile *utils.TPRatingProfile
tpRatingProfileDelay int
tpRatingProfileConfigDIR string //run tests for specific configuration
tpRatingProfileID = "RPrf:*out:Tenant1:Category:Subject"
)
var sTestsTPRatingProfiles = []func(t *testing.T){
testTPRatingProfilesInitCfg,
testTPRatingProfilesResetStorDb,
testTPRatingProfilesStartEngine,
testTPRatingProfilesRpcConn,
testTPRatingProfilesGetTPRatingProfileBeforeSet,
testTPRatingProfilesSetTPRatingProfile,
testTPRatingProfilesGetTPRatingProfileAfterSet,
testTPRatingProfilesGetTPRatingProfileLoadIds,
testTPRatingProfilesUpdateTPRatingProfile,
testTPRatingProfilesGetTPRatingProfileAfterUpdate,
testTPRatingProfilesRemTPRatingProfile,
testTPRatingProfilesGetTPRatingProfileAfterRemove,
testTPRatingProfilesKillEngine,
}
//Test start here
func TestTPRatingProfilesITMySql(t *testing.T) {
tpRatingProfileConfigDIR = "tutmysql"
for _, stest := range sTestsTPRatingProfiles {
t.Run(tpRatingProfileConfigDIR, stest)
}
}
func TestTPRatingProfilesITMongo(t *testing.T) {
tpRatingProfileConfigDIR = "tutmongo"
for _, stest := range sTestsTPRatingProfiles {
t.Run(tpRatingProfileConfigDIR, stest)
}
}
func TestTPRatingProfilesITPG(t *testing.T) {
tpRatingProfileConfigDIR = "tutpostgres"
for _, stest := range sTestsTPRatingProfiles {
t.Run(tpRatingProfileConfigDIR, stest)
}
}
func testTPRatingProfilesInitCfg(t *testing.T) {
var err error
tpRatingProfileCfgPath = path.Join(tpRatingProfileDataDir, "conf", "samples", tpRatingProfileConfigDIR)
tpRatingProfileCfg, err = config.NewCGRConfigFromFolder(tpRatingProfileCfgPath)
if err != nil {
t.Error(err)
}
tpRatingProfileCfg.DataFolderPath = tpRatingProfileDataDir // Share DataFolderPath through config towards StoreDb for Flush()
config.SetCgrConfig(tpRatingProfileCfg)
switch tpRatingProfileConfigDIR {
case "tutmongo": // Mongo needs more time to reset db, need to investigate
tpRatingProfileDelay = 2000
default:
tpRatingProfileDelay = 1000
}
}
// Wipe out the cdr database
func testTPRatingProfilesResetStorDb(t *testing.T) {
if err := engine.InitStorDb(tpRatingProfileCfg); err != nil {
t.Fatal(err)
}
}
// Start CGR Engine
func testTPRatingProfilesStartEngine(t *testing.T) {
if _, err := engine.StopStartEngine(tpRatingProfileCfgPath, tpRatingProfileDelay); err != nil {
t.Fatal(err)
}
}
// Connect rpc client to rater
func testTPRatingProfilesRpcConn(t *testing.T) {
var err error
tpRatingProfileRPC, err = jsonrpc.Dial("tcp", tpRatingProfileCfg.RPCJSONListen) // We connect over JSON so we can also troubleshoot if needed
if err != nil {
t.Fatal(err)
}
}
func testTPRatingProfilesGetTPRatingProfileBeforeSet(t *testing.T) {
var reply *utils.TPRatingProfile
if err := tpRatingProfileRPC.Call("ApierV1.GetTPRatingProfile", &AttrGetTPRatingProfile{TPid: "TPRProf1", RatingProfileId: tpRatingProfileID}, &reply); err == nil || err.Error() != utils.ErrNotFound.Error() {
t.Error(err)
}
}
func testTPRatingProfilesSetTPRatingProfile(t *testing.T) {
tpRatingProfile = &utils.TPRatingProfile{
TPid: "TPRProf1",
LoadId: "RPrf",
Direction: "*out",
Tenant: "Tenant1",
Category: "Category",
Subject: "Subject",
RatingPlanActivations: []*utils.TPRatingActivation{
&utils.TPRatingActivation{
ActivationTime: "2014-07-29T15:00:00Z",
RatingPlanId: "PlanOne",
FallbackSubjects: "FallBack",
CdrStatQueueIds: "RandomId",
},
&utils.TPRatingActivation{
ActivationTime: "2015-07-29T10:00:00Z",
RatingPlanId: "PlanTwo",
FallbackSubjects: "FallOut",
CdrStatQueueIds: "RandomIdTwo",
},
},
}
var result string
if err := tpRatingProfileRPC.Call("ApierV1.SetTPRatingProfile", tpRatingProfile, &result); err != nil {
t.Error(err)
} else if result != utils.OK {
t.Error("Unexpected reply returned", result)
}
}
func testTPRatingProfilesGetTPRatingProfileAfterSet(t *testing.T) {
var respond *utils.TPRatingProfile
if err := tpRatingProfileRPC.Call("ApierV1.GetTPRatingProfile", &AttrGetTPRatingProfile{TPid: "TPRProf1", RatingProfileId: tpRatingProfileID}, &respond); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(tpRatingProfile.TPid, respond.TPid) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.TPid, respond.TPid)
} else if !reflect.DeepEqual(tpRatingProfile.LoadId, respond.LoadId) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.LoadId, respond.LoadId)
} else if !reflect.DeepEqual(tpRatingProfile.Direction, respond.Direction) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Direction, respond.Direction)
} else if !reflect.DeepEqual(tpRatingProfile.Tenant, respond.Tenant) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Tenant, respond.Tenant)
} else if !reflect.DeepEqual(tpRatingProfile.Category, respond.Category) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Category, respond.Category)
} else if !reflect.DeepEqual(tpRatingProfile.Subject, respond.Subject) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Subject, respond.Subject)
} else if !reflect.DeepEqual(len(tpRatingProfile.RatingPlanActivations), len(respond.RatingPlanActivations)) {
t.Errorf("Expecting : %+v, received: %+v", len(tpRatingProfile.RatingPlanActivations), len(respond.RatingPlanActivations))
}
}
func testTPRatingProfilesGetTPRatingProfileLoadIds(t *testing.T) {
var result []string
expected := []string{"RPrf"}
if err := tpRatingProfileRPC.Call("ApierV1.GetTPRatingProfileLoadIds", &utils.AttrTPRatingProfileIds{TPid: tpRatingProfile.TPid, Tenant: tpRatingProfile.Tenant, Category: tpRatingProfile.Category, Direction: tpRatingProfile.Direction, Subject: tpRatingProfile.Subject}, &result); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(expected, result) {
t.Errorf("Expecting: %+v, received: %+v", expected, result)
}
}
func testTPRatingProfilesUpdateTPRatingProfile(t *testing.T) {
var result string
tpRatingProfile.RatingPlanActivations = []*utils.TPRatingActivation{
&utils.TPRatingActivation{
ActivationTime: "2014-07-29T15:00:00Z",
RatingPlanId: "PlanOne",
FallbackSubjects: "FallBack",
CdrStatQueueIds: "RandomId",
},
&utils.TPRatingActivation{
ActivationTime: "2015-07-29T10:00:00Z",
RatingPlanId: "PlanTwo",
FallbackSubjects: "FallOut",
CdrStatQueueIds: "RandomIdTwo",
},
&utils.TPRatingActivation{
ActivationTime: "2017-07-29T10:00:00Z",
RatingPlanId: "BackupPlan",
FallbackSubjects: "Retreat",
CdrStatQueueIds: "DefenseID",
},
}
if err := tpRatingProfileRPC.Call("ApierV1.SetTPRatingProfile", tpRatingProfile, &result); err != nil {
t.Error(err)
} else if result != utils.OK {
t.Error("Unexpected reply returned", result)
}
}
func testTPRatingProfilesGetTPRatingProfileAfterUpdate(t *testing.T) {
var respond *utils.TPRatingProfile
if err := tpRatingProfileRPC.Call("ApierV1.GetTPRatingProfile", &AttrGetTPRatingProfile{TPid: "TPRProf1", RatingProfileId: tpRatingProfileID}, &respond); err != nil {
t.Error(err)
} else if !reflect.DeepEqual(tpRatingProfile.TPid, respond.TPid) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.TPid, respond.TPid)
} else if !reflect.DeepEqual(tpRatingProfile.LoadId, respond.LoadId) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.LoadId, respond.LoadId)
} else if !reflect.DeepEqual(tpRatingProfile.Direction, respond.Direction) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Direction, respond.Direction)
} else if !reflect.DeepEqual(tpRatingProfile.Tenant, respond.Tenant) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Tenant, respond.Tenant)
} else if !reflect.DeepEqual(tpRatingProfile.Category, respond.Category) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Category, respond.Category)
} else if !reflect.DeepEqual(tpRatingProfile.Subject, respond.Subject) {
t.Errorf("Expecting : %+v, received: %+v", tpRatingProfile.Subject, respond.Subject)
} else if !reflect.DeepEqual(len(tpRatingProfile.RatingPlanActivations), len(respond.RatingPlanActivations)) {
t.Errorf("Expecting : %+v, received: %+v", len(tpRatingProfile.RatingPlanActivations), len(respond.RatingPlanActivations))
}
}
func testTPRatingProfilesRemTPRatingProfile(t *testing.T) {
var resp string
if err := tpRatingProfileRPC.Call("ApierV1.RemTPRatingProfile", &AttrGetTPRatingProfile{TPid: "TPRProf1", RatingProfileId: tpRatingProfile.GetRatingProfilesId()}, &resp); err != nil {
t.Error(err)
} else if resp != utils.OK {
t.Error("Unexpected reply returned", resp)
}
}
func testTPRatingProfilesGetTPRatingProfileAfterRemove(t *testing.T) {
var respond *utils.TPRatingProfile
if err := tpRatingProfileRPC.Call("ApierV1.GetTPRatingProfile", &AttrGetTPRatingProfile{TPid: "TPRProf1", RatingProfileId: tpRatingProfileID}, &respond); err == nil || err.Error() != utils.ErrNotFound.Error() {
t.Error(err)
}
}
func testTPRatingProfilesKillEngine(t *testing.T) {
if err := engine.KillEngine(tpRatingProfileDelay); err != nil {
t.Error(err)
}
}

View File

@@ -1,4 +1,4 @@
// +build integration
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
@@ -89,7 +89,7 @@ func testTPResInitCfg(t *testing.T) {
config.SetCgrConfig(tpResCfg)
switch tpResConfigDIR {
case "tutmongo": // Mongo needs more time to reset db, need to investigate
tpResDelay = 4000
tpResDelay = 2000
default:
tpResDelay = 1000
}

View File

@@ -1,4 +1,4 @@
// +build integration
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
@@ -90,9 +90,9 @@ func testTPSharedGroupsInitCfg(t *testing.T) {
config.SetCgrConfig(tpSharedGroupCfg)
switch tpSharedGroupConfigDIR {
case "tutmongo": // Mongo needs more time to reset db
tpSharedGroupDelay = 4000
default:
tpSharedGroupDelay = 2000
default:
tpSharedGroupDelay = 1000
}
}

View File

@@ -1,4 +1,4 @@
// +build integration
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
@@ -89,7 +89,7 @@ func testTPStatsInitCfg(t *testing.T) {
config.SetCgrConfig(tpStatCfg)
switch tpStatConfigDIR {
case "tutmongo": // Mongo needs more time to reset db
tpStatDelay = 4000
tpStatDelay = 2000
default:
tpStatDelay = 1000
}

View File

@@ -1,4 +1,4 @@
// +build integration
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
@@ -90,9 +90,9 @@ func testTPTimingsInitCfg(t *testing.T) {
config.SetCgrConfig(tpTimingCfg)
switch tpTimingConfigDIR {
case "tutmongo": // Mongo needs more time to reset db
tpTimingDelay = 4000
default:
tpTimingDelay = 2000
default:
tpTimingDelay = 1000
}
}

View File

@@ -1,4 +1,4 @@
// +build integration
// +build offline_TP
/*
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
@@ -90,9 +90,9 @@ func testTPUsersInitCfg(t *testing.T) {
config.SetCgrConfig(tpUserCfg)
switch tpUserConfigDIR {
case "tutmongo": // Mongo needs more time to reset db
tpUserDelay = 4000
default:
tpUserDelay = 2000
default:
tpUserDelay = 1000
}
}

View File

@@ -30,6 +30,11 @@ type DataManager struct {
dataDB DataDB
}
// DataDB exports access to dataDB
func (dm *DataManager) DataDB() DataDB {
return dm.dataDB
}
// GetStatQueue retrieves a StatQueue from dataDB
// handles caching and deserialization of metrics
func (dm *DataManager) GetStatQueue(tenant, id string, skipCache bool, transactionID string) (sq *StatQueue, err error) {

View File

@@ -99,7 +99,7 @@ func testDMitCRUDStatQueue(t *testing.T) {
},
},
}
cacheKey := utils.StatQueuePrefix + sq.SqID()
cacheKey := utils.StatQueuePrefix + sq.TenantID()
if _, rcvErr := dm.GetStatQueue(sq.Tenant, sq.ID, false, ""); rcvErr != utils.ErrNotFound {
t.Error(rcvErr)
}

View File

@@ -151,7 +151,7 @@ type StatQueue struct {
}
// SqID will compose the unique identifier for the StatQueue out of Tenant and ID
func (sq *StatQueue) SqID() string {
func (sq *StatQueue) TenantID() string {
return utils.ConcatenatedKey(sq.Tenant, sq.ID)
}

View File

@@ -292,7 +292,7 @@ func (rS *ResourceService) storeResources() {
break // no more keys, backup completed
}
if rIf, ok := cache.Get(utils.ResourcesPrefix + rID); !ok || rIf == nil {
utils.Logger.Warning(fmt.Sprintf("<ResourceS> failed retrieving from cache resource with ID: %s"))
utils.Logger.Warning(fmt.Sprintf("<ResourceS> failed retrieving from cache resource with ID: %s", rID))
} else if err := rS.StoreResource(rIf.(*Resource)); err != nil {
failedRIDs = append(failedRIDs, rID) // record failure so we can schedule it for next backup
}

View File

@@ -19,83 +19,139 @@ along with this program. If not, see <http://www.gnu.org/licenses/>
package engine
import (
"fmt"
"math/rand"
"sync"
"time"
"github.com/cgrates/cgrates/cache"
"github.com/cgrates/cgrates/utils"
)
func init() {
rand.Seed(time.Now().UnixNano())
}
/*
// NewStatService initializes a StatService
func NewStatService(dataDB DataDB, ms Marshaler, storeInterval time.Duration) (ss *StatService, err error) {
ss = &StatService{dataDB: dataDB, ms: ms, storeInterval: storeInterval,
stopStoring: make(chan struct{})}
sqPrfxs, err := dataDB.GetKeysForPrefix(utils.StatsConfigPrefix)
if err != nil {
return nil, err
}
go ss.dumpStoredMetrics() // start dumpStoredMetrics loop
return
func NewStatService(dm *DataManager, storeInterval time.Duration) (ss *StatService, err error) {
return &StatService{dm: dm, storeInterval: storeInterval,
stopBackup: make(chan struct{})}, nil
}
// StatService builds stats for events
type StatService struct {
dataDB DataDB
ms Marshaler
storeInterval time.Duration
stopStoring chan struct{}
dm *DataManager
storeInterval time.Duration
stopBackup chan struct{}
storedStatQueues utils.StringMap // keep a record of stats which need saving, map[statsTenantID]bool
ssqMux sync.RWMutex // protects storedStatQueues
}
// ListenAndServe loops keeps the service alive
func (ss *StatService) ListenAndServe(exitChan chan bool) error {
func (sS *StatService) ListenAndServe(exitChan chan bool) error {
go sS.runBackup() // start backup loop
e := <-exitChan
exitChan <- e // put back for the others listening for shutdown request
return nil
}
// Called to shutdown the service
// ToDo: improve with context, ie following http implementation
func (ss *StatService) Shutdown() error {
// Shutdown is called to shutdown the service
func (sS *StatService) Shutdown() error {
utils.Logger.Info("<StatS> service shutdown initialized")
close(ss.stopStoring)
ss.storeMetrics()
close(sS.stopBackup)
sS.storeStats()
utils.Logger.Info("<StatS> service shutdown complete")
return nil
}
// runBackup will regularly store resources changed to dataDB
func (sS *StatService) runBackup() {
if sS.storeInterval <= 0 {
return
}
for {
select {
case <-sS.stopBackup:
return
}
sS.storeStats()
}
time.Sleep(sS.storeInterval)
}
// storeResources represents one task of complete backup
func (sS *StatService) storeStats() {
var failedSqIDs []string
for { // don't stop untill we store all dirty statQueues
sS.ssqMux.Lock()
sID := sS.storedStatQueues.GetOne()
if sID != "" {
delete(sS.storedStatQueues, sID)
}
sS.ssqMux.Unlock()
if sID == "" {
break // no more keys, backup completed
}
if sqIf, ok := cache.Get(utils.StatQueuePrefix + sID); !ok || sqIf == nil {
utils.Logger.Warning(fmt.Sprintf("<StatS> failed retrieving from cache stat queue with ID: %s", sID))
} else if err := sS.StoreStatQueue(sqIf.(*StatQueue)); err != nil {
failedSqIDs = append(failedSqIDs, sID) // record failure so we can schedule it for next backup
}
// randomize the CPU load and give up thread control
time.Sleep(time.Duration(rand.Intn(1000)) * time.Nanosecond)
}
if len(failedSqIDs) != 0 { // there were errors on save, schedule the keys for next backup
sS.ssqMux.Lock()
for _, sqID := range failedSqIDs {
sS.storedStatQueues[sqID] = true
}
sS.ssqMux.Unlock()
}
}
// StoreStatQueue stores the statQueue in DB and corrects dirty flag
func (sS *StatService) StoreStatQueue(sq *StatQueue) (err error) {
if sq.dirty == nil || !*sq.dirty {
return
}
if err = sS.dm.SetStatQueue(sq); err != nil {
utils.Logger.Warning(
fmt.Sprintf("<StatS> failed saving StatQueue with ID: %s, error: %s",
sq.TenantID(), err.Error()))
} else {
*sq.dirty = false
}
return
}
/*
// setQueue adds or modifies a queue into cache
// sort will reorder the ss.queues
// sort will reorder the sS.queues
func (ss *StatService) loadQueue(qID string) (q *StatQueue, err error) {
sq, err := ss.dataDB.GetStatsConfig(qID)
sq, err := sS.dataDB.GetStatsConfig(qID)
if err != nil {
return nil, err
}
return NewStatQueue(ss.evCache, ss.ms, sq, sqSM)
return NewStatQueue(sS.evCache, sS.ms, sq, sqSM)
}
func (ss *StatService) setQueue(q *StatQueue) {
ss.queuesCache[q.cfg.ID] = q
ss.queues = append(ss.queues, q)
sS.queuesCache[q.cfg.ID] = q
sS.queues = append(sS.queues, q)
}
// remQueue will remove a queue based on it's ID
func (ss *StatService) remQueue(qID string) (si *StatQueue) {
si = ss.queuesCache[qID]
ss.queues.remWithID(qID)
delete(ss.queuesCache, qID)
si = sS.queuesCache[qID]
sS.queues.remWithID(qID)
delete(sS.queuesCache, qID)
return
}
// store stores the necessary storedMetrics to dataDB
func (ss *StatService) storeMetrics() {
for _, si := range ss.queues {
for _, si := range sS.queues {
if !si.cfg.Store || !si.dirty { // no need to save
continue
}
if siSM := si.GetStoredMetrics(); siSM != nil {
if err := ss.dataDB.SetSQStoredMetrics(siSM); err != nil {
if err := sS.dataDB.SetSQStoredMetrics(siSM); err != nil {
utils.Logger.Warning(
fmt.Sprintf("<StatService> failed saving StoredMetrics for QueueID: %s, error: %s",
si.cfg.ID, err.Error()))
@@ -111,11 +167,11 @@ func (ss *StatService) storeMetrics() {
func (ss *StatService) dumpStoredMetrics() {
for {
select {
case <-ss.stopStoring:
case <-sS.stopStoring:
return
}
ss.storeMetrics()
time.Sleep(ss.storeInterval)
sS.storeMetrics()
time.Sleep(sS.storeInterval)
}
}
@@ -125,7 +181,7 @@ func (ss *StatService) processEvent(ev StatsEvent) (err error) {
if evStatsID == "" { // ID is mandatory
return errors.New("missing ID field")
}
for _, stInst := range ss.queues {
for _, stInst := range sS.queues {
if err := stInst.ProcessEvent(ev); err != nil {
utils.Logger.Warning(
fmt.Sprintf("<StatService> QueueID: %s, ignoring event with ID: %s, error: %s",
@@ -140,7 +196,7 @@ func (ss *StatService) processEvent(ev StatsEvent) (err error) {
// V1ProcessEvent implements StatV1 method for processing an Event
func (ss *StatService) V1ProcessEvent(ev StatsEvent, reply *string) (err error) {
if err = ss.processEvent(ev); err == nil {
if err = sS.processEvent(ev); err == nil {
*reply = utils.OK
}
return
@@ -148,10 +204,10 @@ func (ss *StatService) V1ProcessEvent(ev StatsEvent, reply *string) (err error)
// V1GetQueueIDs returns list of queue IDs configured in the service
func (ss *StatService) V1GetQueueIDs(ignored struct{}, reply *[]string) (err error) {
if len(ss.queuesCache) == 0 {
if len(sS.queuesCache) == 0 {
return utils.ErrNotFound
}
for k := range ss.queuesCache {
for k := range sS.queuesCache {
*reply = append(*reply, k)
}
return
@@ -159,7 +215,7 @@ func (ss *StatService) V1GetQueueIDs(ignored struct{}, reply *[]string) (err err
// V1GetStringMetrics returns the metrics as string values
func (ss *StatService) V1GetStringMetrics(queueID string, reply *map[string]string) (err error) {
sq, has := ss.queuesCache[queueID]
sq, has := sS.queuesCache[queueID]
if !has {
return utils.ErrNotFound
}
@@ -173,7 +229,7 @@ func (ss *StatService) V1GetStringMetrics(queueID string, reply *map[string]stri
// V1GetFloatMetrics returns the metrics as float64 values
func (ss *StatService) V1GetFloatMetrics(queueID string, reply *map[string]float64) (err error) {
sq, has := ss.queuesCache[queueID]
sq, has := sS.queuesCache[queueID]
if !has {
return utils.ErrNotFound
}
@@ -195,7 +251,7 @@ type ArgsLoadQueues struct {
func (ss *StatService) V1LoadQueues(args ArgsLoadQueues, reply *string) (err error) {
qIDs := args.QueueIDs
if qIDs == nil {
sqPrfxs, err := ss.dataDB.GetKeysForPrefix(utils.StatsConfigPrefix)
sqPrfxs, err := sS.dataDB.GetKeysForPrefix(utils.StatsConfigPrefix)
if err != nil {
return err
}
@@ -212,10 +268,10 @@ func (ss *StatService) V1LoadQueues(args ArgsLoadQueues, reply *string) (err err
}
var sQs []*StatQueue // cache here so we lock only later when data available
for _, qID := range *qIDs {
if _, hasPrev := ss.queuesCache[qID]; hasPrev {
if _, hasPrev := sS.queuesCache[qID]; hasPrev {
continue // don't overwrite previous, could be extended in the future by carefully checking cached events
}
if q, err := ss.loadQueue(qID); err != nil {
if q, err := sS.loadQueue(qID); err != nil {
utils.Logger.Err(fmt.Sprintf("<StatS> failed loading quueue with id: <%s>, err: <%s>",
q.cfg.ID, err.Error()))
continue
@@ -223,12 +279,12 @@ func (ss *StatService) V1LoadQueues(args ArgsLoadQueues, reply *string) (err err
sQs = append(sQs, q)
}
}
ss.Lock()
sS.Lock()
for _, q := range sQs {
ss.setQueue(q)
sS.setQueue(q)
}
ss.queues.Sort()
ss.Unlock()
sS.queues.Sort()
sS.Unlock()
*reply = utils.OK
return
}

6
offline_tp_test.sh Executable file
View File

@@ -0,0 +1,6 @@
./test.sh
gen=$?
echo 'go test github.com/cgrates/cgrates/apier/v1 -tags=offline_TP'
go test github.com/cgrates/cgrates/apier/v1 -tags=offline_TP
exit $gen