mirror of
https://github.com/cgrates/cgrates.git
synced 2026-02-12 18:46:24 +05:00
324 lines
10 KiB
Go
324 lines
10 KiB
Go
//go:build integration
|
|
// +build integration
|
|
|
|
/*
|
|
Real-time Online/Offline Charging System (OCS) for Telecom & ISP environments
|
|
Copyright (C) ITsysCOM GmbH
|
|
|
|
This program is free software: you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation, either version 3 of the License, or
|
|
(at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program. If not, see <http://www.gnu.org/licenses/>
|
|
*/
|
|
|
|
package general_tests
|
|
|
|
// import (
|
|
// "bytes"
|
|
// "flag"
|
|
// "fmt"
|
|
// "net/rpc"
|
|
// "os"
|
|
// "os/exec"
|
|
// "path"
|
|
// "reflect"
|
|
// "testing"
|
|
// "time"
|
|
|
|
// "github.com/cgrates/cgrates/config"
|
|
// "github.com/cgrates/cgrates/engine"
|
|
// "github.com/cgrates/cgrates/utils"
|
|
// )
|
|
|
|
// /*
|
|
// * Documentation:
|
|
// * This code should work on redis 5 or later:
|
|
// * `redis-cli --cluster create 127.0.0.1:7001 127.0.0.1:7002 127.0.0.1:7003 127.0.0.1:7004 127.0.0.1:7005 127.0.0.1:7006 --cluster-replicas 1`
|
|
// * For redis 4 or before you need to create create the cluster manualy:
|
|
// * - install ruby
|
|
// * - install redis gem: `gem install redis`
|
|
// * - download the `redis-trib.rb` from the source code
|
|
// * - start the 6 nodes with the command `redis-server node1.conf`
|
|
// * - configure the cluster with the following command:
|
|
// * `./redis-trib.rb create --replicas 1 127.0.0.1:7001 127.0.0.1:7002 127.0.0.1:7003 127.0.0.1:7004 127.0.0.1:7005 127.0.0.1:7006`
|
|
// *
|
|
// * To run the tests you need to specify the `redisCluster` flag and have the redis stopped:
|
|
// * `go test github.com/cgrates/cgrates/general_tests -tags=integration -dbtype=*mysql -run=TestRedisCluster -redisCluster -v`
|
|
// *
|
|
// * The configuration of the cluster is the following:
|
|
// * - node1 127.0.0.1:7001 master
|
|
// * - node2 127.0.0.1:7002 master
|
|
// * - node3 127.0.0.1:7003 master
|
|
// * - node4 127.0.0.1:7004 replica
|
|
// * - node5 127.0.0.1:7005 replica
|
|
// * - node6 127.0.0.1:7006 replica
|
|
// * The replicas do not allways select the same master
|
|
// */
|
|
|
|
// var (
|
|
// clsrConfig *config.CGRConfig
|
|
// clsrRPC *rpc.Client
|
|
|
|
// clsrNodeCfgPath = path.Join(*dataDir, "redisCluster", "node%v.conf")
|
|
// clsrEngineCfgPath = path.Join(*dataDir, "conf", "samples", "redisCluster")
|
|
// clsrNodes = make(map[string]*exec.Cmd)
|
|
// clsrOutput = make(map[string]*bytes.Buffer) // in order to debug if something is not working
|
|
// clsrNoNodes = 6 // this is the minimum number of nodes for a cluster with 1 replica for each master
|
|
// clsrRedisFlag = flag.Bool("redisCluster", false, "Run tests for redis cluster")
|
|
// clsrTests = []func(t *testing.T){
|
|
// testClsrPrepare,
|
|
// testClsrStartNodes,
|
|
// testClsrCreateCluster,
|
|
// testClsrInitConfig,
|
|
// testClsrFlushDb,
|
|
// testClsrStartEngine,
|
|
// testClsrRPCConection,
|
|
// testClsrSetGetAttribute,
|
|
// testClsrStopMaster,
|
|
// testClsrSetGetAttribute2,
|
|
// testClsrReStartMaster,
|
|
// testClsrGetAttribute,
|
|
// testClsrStopNodes,
|
|
// testClsrKillEngine,
|
|
// testClsrDeleteFolder,
|
|
// // testClsrPrintOutput,
|
|
// }
|
|
|
|
// clsrRedisCliArgs = []string{
|
|
// "--cluster", "create",
|
|
// "127.0.0.1:7001",
|
|
// "127.0.0.1:7002",
|
|
// "127.0.0.1:7003",
|
|
// "127.0.0.1:7004",
|
|
// "127.0.0.1:7005",
|
|
// "127.0.0.1:7006",
|
|
// "--cluster-replicas", "1",
|
|
// }
|
|
// )
|
|
|
|
// const (
|
|
// clsrRedisCmd = "redis-server"
|
|
// clsrRedisCliCmd = "redis-cli"
|
|
// clsrDir = "/tmp/cluster/"
|
|
// )
|
|
|
|
// func TestRedisCluster(t *testing.T) {
|
|
// if !*clsrRedisFlag {
|
|
// t.SkipNow()
|
|
// }
|
|
// switch *dbType {
|
|
// case utils.MetaMySQL:
|
|
// case utils.MetaInternal,
|
|
// utils.MetaMongo,
|
|
// utils.MetaPostgres:
|
|
// t.SkipNow()
|
|
// default:
|
|
// t.Fatal("Unknown Database type")
|
|
// }
|
|
// for _, stest := range clsrTests {
|
|
// t.Run("TestRedisCluster", stest)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrPrepare(t *testing.T) {
|
|
// if err := os.MkdirAll(clsrDir, 0755); err != nil {
|
|
// t.Fatalf("Error creating folder<%s>:%s", clsrDir, err)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrStartNodes(t *testing.T) {
|
|
// for i := 1; i <= clsrNoNodes; i++ {
|
|
// path := fmt.Sprintf(clsrNodeCfgPath, i)
|
|
// clsrNodes[path] = exec.Command(clsrRedisCmd, path)
|
|
// clsrOutput[path] = bytes.NewBuffer(nil)
|
|
// clsrNodes[path].Stdout = clsrOutput[path]
|
|
// if err := clsrNodes[path].Start(); err != nil {
|
|
// t.Fatalf("Could not start node %v because %s", i, err)
|
|
// }
|
|
// }
|
|
// }
|
|
|
|
// func testClsrCreateCluster(t *testing.T) {
|
|
// cmd := exec.Command(clsrRedisCliCmd, clsrRedisCliArgs...)
|
|
// cmd.Stdin = bytes.NewBuffer([]byte("yes\n"))
|
|
// var stdOut bytes.Buffer
|
|
// cmd.Stdout = &stdOut
|
|
// if err := cmd.Run(); err != nil {
|
|
// t.Errorf("Could not create the cluster because %s", err)
|
|
// t.Logf("The output was:\n %s", stdOut.String()) // print the output to debug the error
|
|
// }
|
|
// time.Sleep(200 * time.Millisecond)
|
|
// }
|
|
|
|
// func testClsrInitConfig(t *testing.T) {
|
|
// var err error
|
|
// clsrConfig, err = config.NewCGRConfigFromPath(clsrEngineCfgPath)
|
|
// if err != nil {
|
|
// t.Error(err)
|
|
// }
|
|
// clsrConfig.DataFolderPath = *dataDir // Share DataFolderPath through config towards StoreDb for Flush()
|
|
// }
|
|
|
|
// func testClsrFlushDb(t *testing.T) {
|
|
// if err := engine.InitDataDB(clsrConfig); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrStartEngine(t *testing.T) {
|
|
// if _, err := engine.StopStartEngine(clsrEngineCfgPath, 200); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrRPCConection(t *testing.T) {
|
|
// var err error
|
|
// clsrRPC, err = newRPCClient(clsrConfig.ListenCfg()) // We connect over JSON so we can also troubleshoot if needed
|
|
// if err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrSetGetAttribute(t *testing.T) {
|
|
// alsPrf := &engine.AttributeProfile{
|
|
// Tenant: "cgrates.org",
|
|
// ID: "ClsrTest",
|
|
// FilterIDs: []string{"*string:~*req.Account:1001", "*string:~*opts.*context:*sessions|*cdrs"},
|
|
// Attributes: []*engine.Attribute{
|
|
// {
|
|
// Path: utils.MetaReq + utils.NestingSep + utils.Subject,
|
|
// Value: config.NewRSRParsersMustCompile("1001", utils.InfieldSep),
|
|
// },
|
|
// },
|
|
// Weight: 20,
|
|
// }
|
|
// alsPrf.Compile()
|
|
// var result string
|
|
// if err := clsrRPC.Call(utils.APIerSv1SetAttributeProfile, alsPrf, &result); err != nil {
|
|
// t.Error(err)
|
|
// } else if result != utils.OK {
|
|
// t.Error("Unexpected reply returned", result)
|
|
// }
|
|
// var reply *engine.AttributeProfile
|
|
// if err := clsrRPC.Call(utils.APIerSv1GetAttributeProfile,
|
|
// &utils.TenantID{Tenant: "cgrates.org", ID: "ClsrTest"}, &reply); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// reply.Compile()
|
|
// if !reflect.DeepEqual(alsPrf, reply) {
|
|
// t.Errorf("Expecting : %+v, received: %+v", alsPrf, reply)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrStopMaster(t *testing.T) {
|
|
// path := fmt.Sprintf(clsrNodeCfgPath, 3)
|
|
// if err = clsrNodes[path].Process.Kill(); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// time.Sleep(time.Second)
|
|
// }
|
|
|
|
// func testClsrSetGetAttribute2(t *testing.T) {
|
|
// alsPrf := &engine.AttributeProfile{
|
|
// Tenant: "cgrates.org",
|
|
// ID: "ClsrTest",
|
|
// FilterIDs: []string{"*string:~*req.Account:1001", "*string:~*opts.*context:*sessions|*cdrs"},
|
|
// Attributes: []*engine.Attribute{
|
|
// {
|
|
// Path: utils.MetaReq + utils.NestingSep + utils.Subject,
|
|
// Value: config.NewRSRParsersMustCompile("1001", utils.InfieldSep),
|
|
// },
|
|
// },
|
|
// Weight: 20,
|
|
// }
|
|
// alsPrf.Compile()
|
|
// var reply *engine.AttributeProfile
|
|
// if err := clsrRPC.Call(utils.APIerSv1GetAttributeProfile,
|
|
// &utils.TenantID{Tenant: "cgrates.org", ID: "ClsrTest"}, &reply); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// reply.Compile()
|
|
// if !reflect.DeepEqual(alsPrf, reply) {
|
|
// t.Errorf("Expecting : %+v, received: %+v", alsPrf, reply)
|
|
// }
|
|
// // add another attribute
|
|
// alsPrf.ID += "2"
|
|
// var result string
|
|
// if err := clsrRPC.Call(utils.APIerSv1SetAttributeProfile, alsPrf, &result); err != nil {
|
|
// t.Error(err)
|
|
// } else if result != utils.OK {
|
|
// t.Error("Unexpected reply returned", result)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrReStartMaster(t *testing.T) {
|
|
// path := fmt.Sprintf(clsrNodeCfgPath, 3)
|
|
// clsrNodes[path] = exec.Command(clsrRedisCmd, path)
|
|
// clsrOutput[path] = bytes.NewBuffer(nil)
|
|
// clsrNodes[path].Stdout = clsrOutput[path]
|
|
// if err := clsrNodes[path].Start(); err != nil {
|
|
// t.Fatalf("Could not start node %v because %s", 3, err)
|
|
// }
|
|
// time.Sleep(200 * time.Millisecond)
|
|
// }
|
|
|
|
// func testClsrGetAttribute(t *testing.T) {
|
|
// alsPrf := &engine.AttributeProfile{
|
|
// Tenant: "cgrates.org",
|
|
// ID: "ClsrTest2",
|
|
// FilterIDs: []string{"*string:~*req.Account:1001", "*string:~*opts.*context:*sessions|*cdrs"},
|
|
// Attributes: []*engine.Attribute{
|
|
// {
|
|
// Path: utils.MetaReq + utils.NestingSep + utils.Subject,
|
|
// Value: config.NewRSRParsersMustCompile("1001", utils.InfieldSep),
|
|
// },
|
|
// },
|
|
// Weight: 20,
|
|
// }
|
|
// alsPrf.Compile()
|
|
// var reply *engine.AttributeProfile
|
|
// if err := clsrRPC.Call(utils.APIerSv1GetAttributeProfile,
|
|
// &utils.TenantID{Tenant: "cgrates.org", ID: "ClsrTest2"}, &reply); err != nil {
|
|
// t.Fatal(err)
|
|
// }
|
|
// reply.Compile()
|
|
// if !reflect.DeepEqual(alsPrf, reply) {
|
|
// t.Errorf("Expecting : %+v, received: %+v", alsPrf, reply)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrStopNodes(t *testing.T) {
|
|
// for path, node := range clsrNodes {
|
|
// if err := node.Process.Kill(); err != nil {
|
|
// t.Fatalf("Could not stop node with path <%s> because %s", path, err)
|
|
// }
|
|
// }
|
|
// }
|
|
|
|
// func testClsrPrintOutput(t *testing.T) {
|
|
// for path, node := range clsrOutput {
|
|
// t.Logf("The output of the node <%s> is:\n%s", path, node.String())
|
|
// t.Logf("==========================================================")
|
|
// }
|
|
// }
|
|
|
|
// func testClsrKillEngine(t *testing.T) {
|
|
// if err := engine.KillEngine(200); err != nil {
|
|
// t.Error(err)
|
|
// }
|
|
// }
|
|
|
|
// func testClsrDeleteFolder(t *testing.T) {
|
|
// if err := os.RemoveAll(clsrDir); err != nil {
|
|
// t.Fatalf("Error removing folder<%s>: %s", clsrDir, err)
|
|
// }
|
|
// }
|