1
0
mirror of https://github.com/redis/go-redis.git synced 2025-04-17 20:17:02 +03:00

fix(tests): enable testing with Redis CE 8.0-M4 in CI (#3247)

* introduce github workflow for ci similar to the one in redis-py

use prerelease for 8.0-M4

* Enable osscluster tests in CI

* Add redis major version env

Enable filtering test per redis major version
Fix test for FT.SEARCH WITHSCORE, the default scorer
has changed.

fix Makefile syntax

remove filter from github action

fix makefile

use the container name in Makefile

* remove 1.20 from doctests

* self review, cleanup, add comments

* add comments, reorder prints, add default value for REDIS_MAJOR_VERSION
This commit is contained in:
Nedyalko Dyakov 2025-01-31 16:14:11 +02:00 committed by GitHub
parent 9f9fa221a8
commit 1139bc3aa9
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
16 changed files with 445 additions and 160 deletions

62
.github/actions/run-tests/action.yml vendored Normal file
View File

@ -0,0 +1,62 @@
name: 'Run go-redis tests'
description: 'Runs go-redis tests against different Redis versions and configurations'
inputs:
go-version:
description: 'Go version to use for running tests'
default: '1.23'
redis-version:
description: 'Redis version to test against'
required: true
runs:
using: "composite"
steps:
- name: Set up ${{ inputs.go-version }}
uses: actions/setup-go@v5
with:
go-version: ${{ inputs.go-version }}
- name: Setup Test environment
env:
REDIS_VERSION: ${{ inputs.redis-version }}
CLIENT_LIBS_TEST_IMAGE: "redislabs/client-libs-test:${{ inputs.redis-version }}"
run: |
set -e
redis_major_version=$(echo "$REDIS_VERSION" | grep -oP '^\d+')
if (( redis_major_version < 8 )); then
echo "Using redis-stack for module tests"
else
echo "Using redis CE for module tests"
fi
# Mapping of redis version to redis testing containers
declare -A redis_version_mapping=(
["8.0-M03"]="8.0-M04-pre"
["7.4.2"]="rs-7.4.0-v2"
["7.2.7"]="rs-7.2.0-v14"
)
if [[ -v redis_version_mapping[$REDIS_VERSION] ]]; then
echo "REDIS_MAJOR_VERSION=${redis_major_version}" >> $GITHUB_ENV
echo "REDIS_IMAGE=redis:${{ inputs.redis-version }}" >> $GITHUB_ENV
echo "CLIENT_LIBS_TEST_IMAGE=redislabs/client-libs-test:${redis_version_mapping[$REDIS_VERSION]}" >> $GITHUB_ENV
else
echo "Version not found in the mapping."
exit 1
fi
sleep 10 # time to settle
shell: bash
- name: Set up Docker Compose environment with redis ${{ inputs.redis-version }}
run: docker compose --profile all up -d
shell: bash
- name: Run tests
env:
RCE_DOCKER: "true"
RE_CLUSTER: "false"
run: |
go test \
--ginkgo.skip-file="ring_test.go" \
--ginkgo.skip-file="sentinel_test.go" \
--ginkgo.skip-file="pubsub_test.go" \
--ginkgo.skip-file="gears_commands_test.go" \
--ginkgo.label-filter="!NonRedisEnterprise"
shell: bash

View File

@ -16,15 +16,7 @@ jobs:
strategy: strategy:
fail-fast: false fail-fast: false
matrix: matrix:
go-version: [1.19.x, 1.20.x, 1.21.x] go-version: [1.21.x, 1.22.x, 1.23.x]
services:
redis:
image: redis/redis-stack-server:latest
options: >-
--health-cmd "redis-cli ping" --health-interval 10s --health-timeout 5s --health-retries 5
ports:
- 6379:6379
steps: steps:
- name: Set up ${{ matrix.go-version }} - name: Set up ${{ matrix.go-version }}
@ -50,39 +42,22 @@ jobs:
strategy: strategy:
fail-fast: false fail-fast: false
matrix: matrix:
redis_version: redis-version:
- "8.0-M01" - "8.0-M03" # 8.0 milestone 4
- "7.4.1" - "7.4.2" # should use redis stack 7.4
- "7.2.6" - "7.2.7" # should redis stack 7.2
- "6.2.16"
go-version: go-version:
- "1.19.x" - "1.22.x"
- "1.20.x" - "1.23.x"
- "1.21.x"
steps: steps:
- name: Set up ${{ matrix.go-version }}
uses: actions/setup-go@v5
with:
go-version: ${{ matrix.go-version }}
- name: Checkout code - name: Checkout code
uses: actions/checkout@v4 uses: actions/checkout@v4
# Set up Docker Compose environment
- name: Set up Docker Compose environment
run: |
docker compose --profile all up -d
- name: Run tests - name: Run tests
env: uses: ./.github/actions/run-tests
USE_CONTAINERIZED_REDIS: "true" with:
RE_CLUSTER: "true" go-version: ${{matrix.go-version}}
run: | redis-version: ${{ matrix.redis-version }}
go test \
--ginkgo.skip-file="ring_test.go" \
--ginkgo.skip-file="sentinel_test.go" \
--ginkgo.skip-file="osscluster_test.go" \
--ginkgo.skip-file="pubsub_test.go" \
--ginkgo.skip-file="gears_commands_test.go" \
--ginkgo.label-filter='!NonRedisEnterprise'

View File

@ -25,7 +25,7 @@ jobs:
strategy: strategy:
fail-fast: false fail-fast: false
matrix: matrix:
go-version: [ "1.18", "1.19", "1.20", "1.21" ] go-version: [ "1.21", "1.22", "1.23" ]
steps: steps:
- name: Set up ${{ matrix.go-version }} - name: Set up ${{ matrix.go-version }}

View File

@ -15,7 +15,7 @@ jobs:
strategy: strategy:
fail-fast: false fail-fast: false
matrix: matrix:
go-version: [1.21.x] go-version: [1.23.x]
re-build: ["7.4.2-54"] re-build: ["7.4.2-54"]
steps: steps:
@ -46,8 +46,8 @@ jobs:
- name: Test - name: Test
env: env:
RE_CLUSTER: "1" RE_CLUSTER: true
USE_CONTAINERIZED_REDIS: "1" REDIS_MAJOR_VERSION: 7
run: | run: |
go test \ go test \
--ginkgo.skip-file="ring_test.go" \ --ginkgo.skip-file="ring_test.go" \

1
.gitignore vendored
View File

@ -1,4 +1,3 @@
dockers/
*.rdb *.rdb
testdata/* testdata/*
.idea/ .idea/

View File

@ -1,6 +1,8 @@
GO_MOD_DIRS := $(shell find . -type f -name 'go.mod' -exec dirname {} \; | sort) GO_MOD_DIRS := $(shell find . -type f -name 'go.mod' -exec dirname {} \; | sort)
export REDIS_MAJOR_VERSION := 7
test: testdeps test: testdeps
docker start go-redis-redis-stack || docker run -d --name go-redis-redis-stack -p 6379:6379 -e REDIS_ARGS="--enable-debug-command yes --enable-module-command yes" redis/redis-stack-server:latest
$(eval GO_VERSION := $(shell go version | cut -d " " -f 3 | cut -d. -f2)) $(eval GO_VERSION := $(shell go version | cut -d " " -f 3 | cut -d. -f2))
set -e; for dir in $(GO_MOD_DIRS); do \ set -e; for dir in $(GO_MOD_DIRS); do \
if echo "$${dir}" | grep -q "./example" && [ "$(GO_VERSION)" = "19" ]; then \ if echo "$${dir}" | grep -q "./example" && [ "$(GO_VERSION)" = "19" ]; then \
@ -19,6 +21,7 @@ test: testdeps
done done
cd internal/customvet && go build . cd internal/customvet && go build .
go vet -vettool ./internal/customvet/customvet go vet -vettool ./internal/customvet/customvet
docker stop go-redis-redis-stack
testdeps: testdata/redis/src/redis-server testdeps: testdata/redis/src/redis-server
@ -32,7 +35,7 @@ build:
testdata/redis: testdata/redis:
mkdir -p $@ mkdir -p $@
wget -qO- https://download.redis.io/releases/redis-7.4-rc2.tar.gz | tar xvz --strip-components=1 -C $@ wget -qO- https://download.redis.io/releases/redis-7.4.2.tar.gz | tar xvz --strip-components=1 -C $@
testdata/redis/src/redis-server: testdata/redis testdata/redis/src/redis-server: testdata/redis
cd $< && make all cd $< && make all

View File

@ -277,7 +277,7 @@ func BenchmarkXRead(b *testing.B) {
func newClusterScenario() *clusterScenario { func newClusterScenario() *clusterScenario {
return &clusterScenario{ return &clusterScenario{
ports: []string{"8220", "8221", "8222", "8223", "8224", "8225"}, ports: []string{"16600", "16601", "16602", "16603", "16604", "16605"},
nodeIDs: make([]string, 6), nodeIDs: make([]string, 6),
processes: make(map[string]*redisProcess, 6), processes: make(map[string]*redisProcess, 6),
clients: make(map[string]*redis.Client, 6), clients: make(map[string]*redis.Client, 6),

View File

@ -441,7 +441,6 @@ var _ = Describe("Commands", func() {
It("should Command", Label("NonRedisEnterprise"), func() { It("should Command", Label("NonRedisEnterprise"), func() {
cmds, err := client.Command(ctx).Result() cmds, err := client.Command(ctx).Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
Expect(len(cmds)).To(BeNumerically("~", 240, 25))
cmd := cmds["mget"] cmd := cmds["mget"]
Expect(cmd.Name).To(Equal("mget")) Expect(cmd.Name).To(Equal("mget"))

View File

@ -1,21 +1,140 @@
--- ---
services: services:
redis:
redis-stanalone: image: ${CLIENT_LIBS_TEST_IMAGE:-redislabs/client-libs-test:7.4.1}
image: redislabs/client-libs-test:8.0-M02
container_name: redis-standalone container_name: redis-standalone
environment: environment:
- TLS_ENABLED=yes
- REDIS_CLUSTER=no - REDIS_CLUSTER=no
- PORT=6379 - PORT=6379
- TLS_PORT=6666 - TLS_PORT=6666
command: ${REDIS_EXTRA_ARGS:---enable-debug-command yes --enable-module-command yes --tls-auth-clients optional --save ""}
ports: ports:
- 6379:6379 - 6379:6379
- 6380:6379
- 6666:6666 # TLS port - 6666:6666 # TLS port
volumes: volumes:
- "./dockers/redis-standalone:/redis/work" - "./dockers/standalone:/redis/work"
profiles: profiles:
- standalone - standalone
- sentinel
- all-stack
- all
cluster:
image: ${CLIENT_LIBS_TEST_IMAGE:-redislabs/client-libs-test:7.4.1}
container_name: redis-cluster
environment:
- NODES=6
- PORT=16600
command: "--cluster-enabled yes"
ports:
- "16600-16605:16600-16605"
volumes:
- "./dockers/cluster:/redis/work"
profiles:
- cluster
- all-stack
- all
sentinel:
image: ${REDIS_IMAGE:-redis:7.4.1}
container_name: redis-sentinel
depends_on:
- redis
entrypoint: "redis-sentinel /redis.conf --port 26379"
ports:
- 26379:26379
volumes:
- "./dockers/sentinel.conf:/redis.conf"
profiles:
- sentinel
- all-stack
- all
sentinel2:
image: ${REDIS_IMAGE:-redis:7.4.1}
container_name: redis-sentinel2
depends_on:
- redis
entrypoint: "redis-sentinel /redis.conf --port 26380"
ports:
- 26380:26380
volumes:
- "./dockers/sentinel.conf:/redis.conf"
profiles:
- sentinel
- all-stack
- all
sentinel3:
image: ${REDIS_IMAGE:-redis:7.4.1}
container_name: redis-sentinel3
depends_on:
- redis
entrypoint: "redis-sentinel /redis.conf --port 26381"
ports:
- 26381:26381
volumes:
- "./dockers/sentinel.conf:/redis.conf"
profiles:
- sentinel
- all-stack
- all
redisRing1:
image: ${CLIENT_LIBS_TEST_IMAGE:-redislabs/client-libs-test:7.4.1}
container_name: redis-ring-1
environment:
- TLS_ENABLED=yes
- REDIS_CLUSTER=no
- PORT=6390
command: ${REDIS_EXTRA_ARGS:---enable-debug-command yes --enable-module-command yes --tls-auth-clients optional --save ""}
ports:
- 6390:6390
volumes:
- "./dockers/ring1:/redis/work"
profiles:
- ring
- cluster
- sentinel
- all-stack
- all
redisRing2:
image: ${CLIENT_LIBS_TEST_IMAGE:-redislabs/client-libs-test:7.4.1}
container_name: redis-ring-2
environment:
- TLS_ENABLED=yes
- REDIS_CLUSTER=no
- PORT=6391
command: ${REDIS_EXTRA_ARGS:---enable-debug-command yes --enable-module-command yes --tls-auth-clients optional --save ""}
ports:
- 6391:6391
volumes:
- "./dockers/ring2:/redis/work"
profiles:
- ring
- cluster
- sentinel
- all-stack
- all
redisRing3:
image: ${CLIENT_LIBS_TEST_IMAGE:-redislabs/client-libs-test:7.4.1}
container_name: redis-ring-3
environment:
- TLS_ENABLED=yes
- REDIS_CLUSTER=no
- PORT=6392
command: ${REDIS_EXTRA_ARGS:---enable-debug-command yes --enable-module-command yes --tls-auth-clients optional --save ""}
ports:
- 6392:6392
volumes:
- "./dockers/ring3:/redis/work"
profiles:
- ring
- cluster
- sentinel
- all-stack
- all - all

1
dockers/.gitignore vendored Normal file
View File

@ -0,0 +1 @@
*/

5
dockers/sentinel.conf Normal file
View File

@ -0,0 +1,5 @@
sentinel resolve-hostnames yes
sentinel monitor go-redis-test redis 6379 2
sentinel down-after-milliseconds go-redis-test 5000
sentinel failover-timeout go-redis-test 60000
sentinel parallel-syncs go-redis-test 1

View File

@ -13,7 +13,6 @@ import (
. "github.com/bsm/ginkgo/v2" . "github.com/bsm/ginkgo/v2"
. "github.com/bsm/gomega" . "github.com/bsm/gomega"
"github.com/redis/go-redis/v9" "github.com/redis/go-redis/v9"
) )
@ -28,7 +27,7 @@ const (
) )
const ( const (
sentinelName = "mymaster" sentinelName = "go-redis-test"
sentinelMasterPort = "9123" sentinelMasterPort = "9123"
sentinelSlave1Port = "9124" sentinelSlave1Port = "9124"
sentinelSlave2Port = "9125" sentinelSlave2Port = "9125"
@ -43,8 +42,8 @@ var (
) )
var ( var (
rediStackPort = "6379" redisStackPort = "6379"
rediStackAddr = ":" + rediStackPort redisStackAddr = ":" + redisStackPort
) )
var ( var (
@ -59,14 +58,22 @@ var (
) )
var cluster = &clusterScenario{ var cluster = &clusterScenario{
ports: []string{"8220", "8221", "8222", "8223", "8224", "8225"}, ports: []string{"16600", "16601", "16602", "16603", "16604", "16605"},
nodeIDs: make([]string, 6), nodeIDs: make([]string, 6),
processes: make(map[string]*redisProcess, 6), processes: make(map[string]*redisProcess, 6),
clients: make(map[string]*redis.Client, 6), clients: make(map[string]*redis.Client, 6),
} }
// Redis Software Cluster
var RECluster = false var RECluster = false
var USE_CONTAINERIZED_REDIS = false
// Redis Community Edition Docker
var RCEDocker = false
// Notes the major version of redis we are executing tests.
// This can be used before we change the bsm fork of ginkgo for one,
// which have support for label sets, so we can filter tests per redis major version.
var REDIS_MAJOR_VERSION = 7
func registerProcess(port string, p *redisProcess) { func registerProcess(port string, p *redisProcess) {
if processes == nil { if processes == nil {
@ -83,8 +90,19 @@ var _ = BeforeSuite(func() {
} }
var err error var err error
RECluster, _ = strconv.ParseBool(os.Getenv("RE_CLUSTER")) RECluster, _ = strconv.ParseBool(os.Getenv("RE_CLUSTER"))
USE_CONTAINERIZED_REDIS, _ = strconv.ParseBool(os.Getenv("USE_CONTAINERIZED_REDIS")) RCEDocker, _ = strconv.ParseBool(os.Getenv("RCE_DOCKER"))
if !RECluster || !USE_CONTAINERIZED_REDIS {
REDIS_MAJOR_VERSION, _ = strconv.Atoi(os.Getenv("REDIS_MAJOR_VERSION"))
if REDIS_MAJOR_VERSION == 0 {
REDIS_MAJOR_VERSION = 7
}
Expect(REDIS_MAJOR_VERSION).To(BeNumerically(">=", 6))
Expect(REDIS_MAJOR_VERSION).To(BeNumerically("<=", 8))
fmt.Printf("RECluster: %v\n", RECluster)
fmt.Printf("RCEDocker: %v\n", RCEDocker)
fmt.Printf("REDIS_MAJOR_VERSION: %v\n", REDIS_MAJOR_VERSION)
if !RECluster && !RCEDocker {
redisMain, err = startRedis(redisPort) redisMain, err = startRedis(redisPort)
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
@ -121,19 +139,25 @@ var _ = BeforeSuite(func() {
err = startCluster(ctx, cluster) err = startCluster(ctx, cluster)
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
} else { } else {
redisPort = rediStackPort redisPort = redisStackPort
redisAddr = rediStackAddr redisAddr = redisStackAddr
if !RECluster {
// populate cluster node information
Expect(configureClusterTopology(ctx, cluster)).NotTo(HaveOccurred())
}
} }
}) })
var _ = AfterSuite(func() { var _ = AfterSuite(func() {
if !RECluster { if !RECluster {
Expect(cluster.Close()).NotTo(HaveOccurred()) Expect(cluster.Close()).NotTo(HaveOccurred())
}
// NOOP if there are no processes registered
for _, p := range processes { for _, p := range processes {
Expect(p.Close()).NotTo(HaveOccurred()) Expect(p.Close()).NotTo(HaveOccurred())
} }
}
processes = nil processes = nil
}) })
@ -156,8 +180,8 @@ func redisOptions() *redis.Options {
ContextTimeoutEnabled: true, ContextTimeoutEnabled: true,
MaxRetries: -1, MaxRetries: -1,
PoolSize: 10,
PoolSize: 10,
PoolTimeout: 30 * time.Second, PoolTimeout: 30 * time.Second,
ConnMaxIdleTime: time.Minute, ConnMaxIdleTime: time.Minute,
} }

View File

@ -22,7 +22,7 @@ var _ = Describe("Monitor command", Label("monitor"), func() {
if os.Getenv("RUN_MONITOR_TEST") != "true" { if os.Getenv("RUN_MONITOR_TEST") != "true" {
Skip("Skipping Monitor command test. Set RUN_MONITOR_TEST=true to run it.") Skip("Skipping Monitor command test. Set RUN_MONITOR_TEST=true to run it.")
} }
client = redis.NewClient(&redis.Options{Addr: ":6379"}) client = redis.NewClient(&redis.Options{Addr: redisPort})
Expect(client.FlushDB(ctx).Err()).NotTo(HaveOccurred()) Expect(client.FlushDB(ctx).Err()).NotTo(HaveOccurred())
}) })
@ -33,7 +33,7 @@ var _ = Describe("Monitor command", Label("monitor"), func() {
It("should monitor", Label("monitor"), func() { It("should monitor", Label("monitor"), func() {
ress := make(chan string) ress := make(chan string)
client1 := redis.NewClient(&redis.Options{Addr: rediStackAddr}) client1 := redis.NewClient(&redis.Options{Addr: redisPort})
mn := client1.Monitor(ctx, ress) mn := client1.Monitor(ctx, ress)
mn.Start() mn.Start()
// Wait for the Redis server to be in monitoring mode. // Wait for the Redis server to be in monitoring mode.
@ -61,7 +61,7 @@ func TestMonitorCommand(t *testing.T) {
} }
ctx := context.TODO() ctx := context.TODO()
client := redis.NewClient(&redis.Options{Addr: ":6379"}) client := redis.NewClient(&redis.Options{Addr: redisPort})
if err := client.FlushDB(ctx).Err(); err != nil { if err := client.FlushDB(ctx).Err(); err != nil {
t.Fatalf("FlushDB failed: %v", err) t.Fatalf("FlushDB failed: %v", err)
} }
@ -73,7 +73,7 @@ func TestMonitorCommand(t *testing.T) {
}() }()
ress := make(chan string, 10) // Buffer to prevent blocking ress := make(chan string, 10) // Buffer to prevent blocking
client1 := redis.NewClient(&redis.Options{Addr: ":6379"}) // Adjust the Addr field as necessary client1 := redis.NewClient(&redis.Options{Addr: redisPort}) // Adjust the Addr field as necessary
mn := client1.Monitor(ctx, ress) mn := client1.Monitor(ctx, ress)
mn.Start() mn.Start()
// Wait for the Redis server to be in monitoring mode. // Wait for the Redis server to be in monitoring mode.

View File

@ -25,6 +25,10 @@ type clusterScenario struct {
clients map[string]*redis.Client clients map[string]*redis.Client
} }
func (s *clusterScenario) slots() []int {
return []int{0, 5461, 10923, 16384}
}
func (s *clusterScenario) masters() []*redis.Client { func (s *clusterScenario) masters() []*redis.Client {
result := make([]*redis.Client, 3) result := make([]*redis.Client, 3)
for pos, port := range s.ports[:3] { for pos, port := range s.ports[:3] {
@ -83,37 +87,39 @@ func (s *clusterScenario) newClusterClient(
} }
func (s *clusterScenario) Close() error { func (s *clusterScenario) Close() error {
ctx := context.TODO()
for _, master := range s.masters() {
err := master.FlushAll(ctx).Err()
if err != nil {
return err
}
// since 7.2 forget calls should be propagated, calling only master
// nodes should be sufficient.
for _, nID := range s.nodeIDs {
master.ClusterForget(ctx, nID)
}
}
for _, port := range s.ports { for _, port := range s.ports {
if process, ok := processes[port]; ok { if process, ok := processes[port]; ok {
if process != nil {
process.Close() process.Close()
}
delete(processes, port) delete(processes, port)
} }
} }
return nil return nil
} }
func startCluster(ctx context.Context, scenario *clusterScenario) error { func configureClusterTopology(ctx context.Context, scenario *clusterScenario) error {
// Start processes and collect node ids err := collectNodeInformation(ctx, scenario)
for pos, port := range scenario.ports {
process, err := startRedis(port, "--cluster-enabled", "yes")
if err != nil { if err != nil {
return err return err
} }
client := redis.NewClient(&redis.Options{
Addr: ":" + port,
})
info, err := client.ClusterNodes(ctx).Result()
if err != nil {
return err
}
scenario.processes[port] = process
scenario.clients[port] = client
scenario.nodeIDs[pos] = info[:40]
}
// Meet cluster nodes. // Meet cluster nodes.
for _, client := range scenario.clients { for _, client := range scenario.clients {
err := client.ClusterMeet(ctx, "127.0.0.1", scenario.ports[0]).Err() err := client.ClusterMeet(ctx, "127.0.0.1", scenario.ports[0]).Err()
@ -122,8 +128,7 @@ func startCluster(ctx context.Context, scenario *clusterScenario) error {
} }
} }
// Bootstrap masters. slots := scenario.slots()
slots := []int{0, 5000, 10000, 16384}
for pos, master := range scenario.masters() { for pos, master := range scenario.masters() {
err := master.ClusterAddSlotsRange(ctx, slots[pos], slots[pos+1]-1).Err() err := master.ClusterAddSlotsRange(ctx, slots[pos], slots[pos+1]-1).Err()
if err != nil { if err != nil {
@ -157,35 +162,36 @@ func startCluster(ctx context.Context, scenario *clusterScenario) error {
// Wait until all nodes have consistent info. // Wait until all nodes have consistent info.
wanted := []redis.ClusterSlot{{ wanted := []redis.ClusterSlot{{
Start: 0, Start: 0,
End: 4999, End: 5460,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8220", Addr: "127.0.0.1:16600",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8223", Addr: "127.0.0.1:16603",
}}, }},
}, { }, {
Start: 5000, Start: 5461,
End: 9999, End: 10922,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8221", Addr: "127.0.0.1:16601",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8224", Addr: "127.0.0.1:16604",
}}, }},
}, { }, {
Start: 10000, Start: 10923,
End: 16383, End: 16383,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8222", Addr: "127.0.0.1:16602",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8225", Addr: "127.0.0.1:16605",
}}, }},
}} }}
for _, client := range scenario.clients { for _, client := range scenario.clients {
err := eventually(func() error { err := eventually(func() error {
res, err := client.ClusterSlots(ctx).Result() res, err := client.ClusterSlots(ctx).Result()
@ -193,7 +199,7 @@ func startCluster(ctx context.Context, scenario *clusterScenario) error {
return err return err
} }
return assertSlotsEqual(res, wanted) return assertSlotsEqual(res, wanted)
}, 30*time.Second) }, 60*time.Second)
if err != nil { if err != nil {
return err return err
} }
@ -202,6 +208,37 @@ func startCluster(ctx context.Context, scenario *clusterScenario) error {
return nil return nil
} }
func collectNodeInformation(ctx context.Context, scenario *clusterScenario) error {
for pos, port := range scenario.ports {
client := redis.NewClient(&redis.Options{
Addr: ":" + port,
})
info, err := client.ClusterNodes(ctx).Result()
if err != nil {
return err
}
scenario.clients[port] = client
scenario.nodeIDs[pos] = info[:40]
}
return nil
}
// startCluster start a cluster
func startCluster(ctx context.Context, scenario *clusterScenario) error {
// Start processes and collect node ids
for _, port := range scenario.ports {
process, err := startRedis(port, "--cluster-enabled", "yes")
if err != nil {
return err
}
scenario.processes[port] = process
}
return configureClusterTopology(ctx, scenario)
}
func assertSlotsEqual(slots, wanted []redis.ClusterSlot) error { func assertSlotsEqual(slots, wanted []redis.ClusterSlot) error {
outerLoop: outerLoop:
for _, s2 := range wanted { for _, s2 := range wanted {
@ -301,17 +338,19 @@ var _ = Describe("ClusterClient", func() {
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
} }
client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error { err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
defer GinkgoRecover() defer GinkgoRecover()
Eventually(func() string { Eventually(func() string {
return master.Info(ctx, "keyspace").Val() return master.Info(ctx, "keyspace").Val()
}, 30*time.Second).Should(Or( }, 30*time.Second).Should(Or(
ContainSubstring("keys=31"), ContainSubstring("keys=32"),
ContainSubstring("keys=29"), ContainSubstring("keys=36"),
ContainSubstring("keys=40"), ContainSubstring("keys=32"),
)) ))
return nil return nil
}) })
Expect(err).NotTo(HaveOccurred())
}) })
It("distributes keys when using EVAL", func() { It("distributes keys when using EVAL", func() {
@ -327,17 +366,19 @@ var _ = Describe("ClusterClient", func() {
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())
} }
client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error { err := client.ForEachMaster(ctx, func(ctx context.Context, master *redis.Client) error {
defer GinkgoRecover() defer GinkgoRecover()
Eventually(func() string { Eventually(func() string {
return master.Info(ctx, "keyspace").Val() return master.Info(ctx, "keyspace").Val()
}, 30*time.Second).Should(Or( }, 30*time.Second).Should(Or(
ContainSubstring("keys=31"), ContainSubstring("keys=32"),
ContainSubstring("keys=29"), ContainSubstring("keys=36"),
ContainSubstring("keys=40"), ContainSubstring("keys=32"),
)) ))
return nil return nil
}) })
Expect(err).NotTo(HaveOccurred())
}) })
It("distributes scripts when using Script Load", func() { It("distributes scripts when using Script Load", func() {
@ -347,13 +388,14 @@ var _ = Describe("ClusterClient", func() {
script.Load(ctx, client) script.Load(ctx, client)
client.ForEachShard(ctx, func(ctx context.Context, shard *redis.Client) error { err := client.ForEachShard(ctx, func(ctx context.Context, shard *redis.Client) error {
defer GinkgoRecover() defer GinkgoRecover()
val, _ := script.Exists(ctx, shard).Result() val, _ := script.Exists(ctx, shard).Result()
Expect(val[0]).To(Equal(true)) Expect(val[0]).To(Equal(true))
return nil return nil
}) })
Expect(err).NotTo(HaveOccurred())
}) })
It("checks all shards when using Script Exists", func() { It("checks all shards when using Script Exists", func() {
@ -727,33 +769,33 @@ var _ = Describe("ClusterClient", func() {
wanted := []redis.ClusterSlot{{ wanted := []redis.ClusterSlot{{
Start: 0, Start: 0,
End: 4999, End: 5460,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8220", Addr: "127.0.0.1:16600",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8223", Addr: "127.0.0.1:16603",
}}, }},
}, { }, {
Start: 5000, Start: 5461,
End: 9999, End: 10922,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8221", Addr: "127.0.0.1:16601",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8224", Addr: "127.0.0.1:16604",
}}, }},
}, { }, {
Start: 10000, Start: 10923,
End: 16383, End: 16383,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
ID: "", ID: "",
Addr: "127.0.0.1:8222", Addr: "127.0.0.1:16602",
}, { }, {
ID: "", ID: "",
Addr: "127.0.0.1:8225", Addr: "127.0.0.1:16605",
}}, }},
}} }}
Expect(assertSlotsEqual(res, wanted)).NotTo(HaveOccurred()) Expect(assertSlotsEqual(res, wanted)).NotTo(HaveOccurred())
@ -1122,14 +1164,14 @@ var _ = Describe("ClusterClient", func() {
client, err := client.SlaveForKey(ctx, "test") client, err := client.SlaveForKey(ctx, "test")
Expect(err).ToNot(HaveOccurred()) Expect(err).ToNot(HaveOccurred())
info := client.Info(ctx, "server") info := client.Info(ctx, "server")
Expect(info.Val()).Should(ContainSubstring("tcp_port:8224")) Expect(info.Val()).Should(ContainSubstring("tcp_port:16604"))
}) })
It("should return correct master for key", func() { It("should return correct master for key", func() {
client, err := client.MasterForKey(ctx, "test") client, err := client.MasterForKey(ctx, "test")
Expect(err).ToNot(HaveOccurred()) Expect(err).ToNot(HaveOccurred())
info := client.Info(ctx, "server") info := client.Info(ctx, "server")
Expect(info.Val()).Should(ContainSubstring("tcp_port:8221")) Expect(info.Val()).Should(ContainSubstring("tcp_port:16601"))
}) })
assertClusterClient() assertClusterClient()
@ -1176,18 +1218,18 @@ var _ = Describe("ClusterClient", func() {
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) { opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
slots := []redis.ClusterSlot{{ slots := []redis.ClusterSlot{{
Start: 0, Start: 0,
End: 4999, End: 5460,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard1Port, Addr: ":" + ringShard1Port,
}}, }},
}, { }, {
Start: 5000, Start: 5461,
End: 9999, End: 10922,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard2Port, Addr: ":" + ringShard2Port,
}}, }},
}, { }, {
Start: 10000, Start: 10923,
End: 16383, End: 16383,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard3Port, Addr: ":" + ringShard3Port,
@ -1230,18 +1272,18 @@ var _ = Describe("ClusterClient", func() {
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) { opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
slots := []redis.ClusterSlot{{ slots := []redis.ClusterSlot{{
Start: 0, Start: 0,
End: 4999, End: 5460,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard1Port, Addr: ":" + ringShard1Port,
}}, }},
}, { }, {
Start: 5000, Start: 5461,
End: 9999, End: 10922,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard2Port, Addr: ":" + ringShard2Port,
}}, }},
}, { }, {
Start: 10000, Start: 10923,
End: 16383, End: 16383,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":" + ringShard3Port, Addr: ":" + ringShard3Port,
@ -1284,27 +1326,27 @@ var _ = Describe("ClusterClient", func() {
opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) { opt.ClusterSlots = func(ctx context.Context) ([]redis.ClusterSlot, error) {
slots := []redis.ClusterSlot{{ slots := []redis.ClusterSlot{{
Start: 0, Start: 0,
End: 4999, End: 5460,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":8220", Addr: ":16600",
}, { }, {
Addr: ":8223", Addr: ":16603",
}}, }},
}, { }, {
Start: 5000, Start: 5461,
End: 9999, End: 10922,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":8221", Addr: ":16601",
}, { }, {
Addr: ":8224", Addr: ":16604",
}}, }},
}, { }, {
Start: 10000, Start: 10923,
End: 16383, End: 16383,
Nodes: []redis.ClusterNode{{ Nodes: []redis.ClusterNode{{
Addr: ":8222", Addr: ":16602",
}, { }, {
Addr: ":8225", Addr: ":16605",
}}, }},
}} }}
return slots, nil return slots, nil

View File

@ -282,6 +282,9 @@ type FTSearchSortBy struct {
Desc bool Desc bool
} }
// FTSearchOptions hold options that can be passed to the FT.SEARCH command.
// More information about the options can be found
// in the documentation for FT.SEARCH https://redis.io/docs/latest/commands/ft.search/
type FTSearchOptions struct { type FTSearchOptions struct {
NoContent bool NoContent bool
Verbatim bool Verbatim bool
@ -299,6 +302,10 @@ type FTSearchOptions struct {
InOrder bool InOrder bool
Language string Language string
Expander string Expander string
// Scorer is used to set scoring function, if not set passed, a default will be used.
// The default scorer depends on the Redis version:
// - `BM25` for Redis >= 8
// - `TFIDF` for Redis < 8
Scorer string Scorer string
ExplainScore bool ExplainScore bool
Payload string Payload string

View File

@ -371,7 +371,56 @@ var _ = Describe("RediSearch commands Resp 2", Label("search"), func() {
Expect(names).To(ContainElement("John")) Expect(names).To(ContainElement("John"))
}) })
// up until redis 8 the default scorer was TFIDF, in redis 8 it is BM25
// this test expect redis major version >= 8
It("should FTSearch WithScores", Label("search", "ftsearch"), func() { It("should FTSearch WithScores", Label("search", "ftsearch"), func() {
if REDIS_MAJOR_VERSION < 8 {
Skip("(redis major version < 8) default scorer is not BM25")
}
text1 := &redis.FieldSchema{FieldName: "description", FieldType: redis.SearchFieldTypeText}
val, err := client.FTCreate(ctx, "idx1", &redis.FTCreateOptions{}, text1).Result()
Expect(err).NotTo(HaveOccurred())
Expect(val).To(BeEquivalentTo("OK"))
WaitForIndexing(client, "idx1")
client.HSet(ctx, "doc1", "description", "The quick brown fox jumps over the lazy dog")
client.HSet(ctx, "doc2", "description", "Quick alice was beginning to get very tired of sitting by her quick sister on the bank, and of having nothing to do.")
res, err := client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeNumerically("<=", 0.236))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "TFIDF"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeEquivalentTo(float64(1)))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "TFIDF.DOCNORM"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeEquivalentTo(0.14285714285714285))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "BM25"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeNumerically("<=", 0.22471909420069797))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "DISMAX"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeEquivalentTo(float64(2)))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "DOCSCORE"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeEquivalentTo(float64(1)))
res, err = client.FTSearchWithArgs(ctx, "idx1", "quick", &redis.FTSearchOptions{WithScores: true, Scorer: "HAMMING"}).Result()
Expect(err).NotTo(HaveOccurred())
Expect(*res.Docs[0].Score).To(BeEquivalentTo(float64(0)))
})
// up until redis 8 the default scorer was TFIDF, in redis 8 it is BM25
// this test expect redis major version <=7
It("should FTSearch WithScores", Label("search", "ftsearch"), func() {
if REDIS_MAJOR_VERSION > 7 {
Skip("(redis major version > 7) default scorer is not TFIDF")
}
text1 := &redis.FieldSchema{FieldName: "description", FieldType: redis.SearchFieldTypeText} text1 := &redis.FieldSchema{FieldName: "description", FieldType: redis.SearchFieldTypeText}
val, err := client.FTCreate(ctx, "idx1", &redis.FTCreateOptions{}, text1).Result() val, err := client.FTCreate(ctx, "idx1", &redis.FTCreateOptions{}, text1).Result()
Expect(err).NotTo(HaveOccurred()) Expect(err).NotTo(HaveOccurred())