@@ -20,9 +20,6 @@ import (
2020 "testing"
2121 "time"
2222
23- "github.com/stretchr/testify/require"
24-
25- "go.etcd.io/etcd/api/v3/version"
2623 "go.etcd.io/etcd/client/pkg/v3/fileutil"
2724 "go.etcd.io/etcd/server/v3/etcdserver"
2825 "go.etcd.io/etcd/tests/v3/framework/e2e"
@@ -50,16 +47,6 @@ var trafficProfiles = []TrafficProfile{
5047 Traffic : traffic .EtcdPutDeleteLease ,
5148 Profile : traffic .LowTraffic ,
5249 },
53- {
54- Name : "KubernetesHighTraffic" ,
55- Traffic : traffic .Kubernetes ,
56- Profile : traffic .HighTrafficProfile ,
57- },
58- {
59- Name : "KubernetesLowTraffic" ,
60- Traffic : traffic .Kubernetes ,
61- Profile : traffic .LowTraffic ,
62- },
6350}
6451
6552type TestScenario struct {
@@ -160,75 +147,72 @@ func Exploratory(_ *testing.T) []TestScenario {
160147}
161148
162149func Regression (t * testing.T ) []TestScenario {
163- v , err := e2e .GetVersionFromBinary (e2e .BinPath .Etcd )
164- require .NoErrorf (t , err , "Failed checking etcd version binary, binary: %q" , e2e .BinPath .Etcd )
165-
166- scenarios := []TestScenario {}
167- scenarios = append (scenarios , TestScenario {
168- Name : "Issue14370" ,
169- Failpoint : failpoint .RaftBeforeSavePanic ,
170- Profile : traffic .LowTraffic .WithoutWatchLoop (),
171- Traffic : traffic .EtcdPutDeleteLease ,
172- Cluster : * e2e .NewConfig (
173- e2e .WithClusterSize (1 ),
174- e2e .WithGoFailEnabled (true ),
175- ),
176- })
177- scenarios = append (scenarios , TestScenario {
178- Name : "Issue14685" ,
179- Failpoint : failpoint .DefragBeforeCopyPanic ,
180- Profile : traffic .LowTraffic .WithoutWatchLoop (),
181- Traffic : traffic .EtcdPutDeleteLease ,
182- Cluster : * e2e .NewConfig (
183- e2e .WithClusterSize (1 ),
184- e2e .WithGoFailEnabled (true ),
185- ),
186- })
187- scenarios = append (scenarios , TestScenario {
188- Name : "Issue13766" ,
189- Failpoint : failpoint .KillFailpoint ,
190- Profile : traffic .HighTrafficProfile .WithoutWatchLoop (),
191- Traffic : traffic .EtcdPut ,
192- Cluster : * e2e .NewConfig (
193- e2e .WithSnapshotCount (100 ),
194- ),
195- })
196- scenarios = append (scenarios , TestScenario {
197- Name : "Issue15220" ,
198- Watch : client.WatchConfig {
199- RequestProgress : true ,
200- },
201- Profile : traffic .LowTraffic ,
202- Traffic : traffic .EtcdPutDeleteLease ,
203- Failpoint : failpoint .KillFailpoint ,
204- Cluster : * e2e .NewConfig (
205- e2e .WithClusterSize (1 ),
206- ),
207- })
208- scenarios = append (scenarios , TestScenario {
209- Name : "Issue17529" ,
210- Profile : traffic .HighTrafficProfile ,
211- Traffic : traffic .Kubernetes ,
212- Failpoint : failpoint .SleepBeforeSendWatchResponse ,
213- Cluster : * e2e .NewConfig (
214- e2e .WithClusterSize (1 ),
215- e2e .WithGoFailEnabled (true ),
216- options .WithSnapshotCount (100 ),
217- ),
218- })
219-
220- scenarios = append (scenarios , TestScenario {
221- Name : "Issue17780" ,
222- Profile : traffic .LowTraffic .WithoutCompaction ().WithoutWatchLoop (),
223- Failpoint : failpoint .BatchCompactBeforeSetFinishedCompactPanic ,
224- Traffic : traffic .Kubernetes ,
225- Cluster : * e2e .NewConfig (
226- e2e .WithClusterSize (1 ),
227- e2e .WithCompactionBatchLimit (300 ),
228- e2e .WithSnapshotCount (1000 ),
229- e2e .WithGoFailEnabled (true ),
230- ),
231- })
150+ // scenarios := []TestScenario{}
151+ // scenarios = append(scenarios, TestScenario{
152+ // Name: "Issue14370",
153+ // Failpoint: failpoint.RaftBeforeSavePanic,
154+ // Profile: traffic.LowTraffic.WithoutWatchLoop(),
155+ // Traffic: traffic.EtcdPutDeleteLease,
156+ // Cluster: *e2e.NewConfig(
157+ // e2e.WithClusterSize(1),
158+ // e2e.WithGoFailEnabled(true),
159+ // ),
160+ // })
161+ // scenarios = append(scenarios, TestScenario{
162+ // Name: "Issue14685",
163+ // Failpoint: failpoint.DefragBeforeCopyPanic,
164+ // Profile: traffic.LowTraffic.WithoutWatchLoop(),
165+ // Traffic: traffic.EtcdPutDeleteLease,
166+ // Cluster: *e2e.NewConfig(
167+ // e2e.WithClusterSize(1),
168+ // e2e.WithGoFailEnabled(true),
169+ // ),
170+ // })
171+ // scenarios = append(scenarios, TestScenario{
172+ // Name: "Issue13766",
173+ // Failpoint: failpoint.KillFailpoint,
174+ // Profile: traffic.HighTrafficProfile.WithoutWatchLoop(),
175+ // Traffic: traffic.EtcdPut,
176+ // Cluster: *e2e.NewConfig(
177+ // e2e.WithSnapshotCount(100),
178+ // ),
179+ // })
180+ // scenarios = append(scenarios, TestScenario{
181+ // Name: "Issue15220",
182+ // Watch: client.WatchConfig{
183+ // RequestProgress: true,
184+ // },
185+ // Profile: traffic.LowTraffic,
186+ // Traffic: traffic.EtcdPutDeleteLease,
187+ // Failpoint: failpoint.KillFailpoint,
188+ // Cluster: *e2e.NewConfig(
189+ // e2e.WithClusterSize(1),
190+ // ),
191+ // })
192+ // scenarios = append(scenarios, TestScenario{
193+ // Name: "Issue17529",
194+ // Profile: traffic.HighTrafficProfile,
195+ // Traffic: traffic.Kubernetes,
196+ // Failpoint: failpoint.SleepBeforeSendWatchResponse,
197+ // Cluster: *e2e.NewConfig(
198+ // e2e.WithClusterSize(1),
199+ // e2e.WithGoFailEnabled(true),
200+ // options.WithSnapshotCount(100),
201+ // ),
202+ // })
203+
204+ // scenarios = append(scenarios, TestScenario{
205+ // Name: "Issue17780",
206+ // Profile: traffic.LowTraffic.WithoutCompaction().WithoutWatchLoop(),
207+ // Failpoint: failpoint.BatchCompactBeforeSetFinishedCompactPanic,
208+ // Traffic: traffic.Kubernetes,
209+ // Cluster: *e2e.NewConfig(
210+ // e2e.WithClusterSize(1),
211+ // e2e.WithCompactionBatchLimit(300),
212+ // e2e.WithSnapshotCount(1000),
213+ // e2e.WithGoFailEnabled(true),
214+ // ),
215+ // })
232216
233217 // NOTE:
234218 //
@@ -242,51 +226,6 @@ func Regression(t *testing.T) []TestScenario {
242226 // burstable value. A higher QPS can generate more new keys than
243227 // expected, making it difficult to determine an optimal compaction
244228 // batch limit within a larger key space.
245- scenarios = append (scenarios , TestScenario {
246- Name : "Issue19179" ,
247- Profile : traffic.Profile {
248- MinimalQPS : 50 ,
249- MaximalQPS : 100 ,
250- BurstableQPS : 100 ,
251- MemberClientCount : 6 ,
252- ClusterClientCount : 2 ,
253- MaxNonUniqueRequestConcurrency : 3 ,
254- }.WithoutCompaction (),
255- Failpoint : failpoint .BatchCompactBeforeSetFinishedCompactPanic ,
256- Traffic : traffic .KubernetesCreateDelete ,
257- Cluster : * e2e .NewConfig (
258- e2e .WithClusterSize (1 ),
259- e2e .WithCompactionBatchLimit (50 ),
260- e2e .WithSnapshotCount (1000 ),
261- e2e .WithGoFailEnabled (true ),
262- ),
263- })
264- scenarios = append (scenarios , TestScenario {
265- Name : "Issue18089" ,
266- Profile : traffic .LowTraffic .WithCompactionPeriod (100 * time .Millisecond ), // Use frequent compaction for high reproduce rate
267- Failpoint : failpoint .SleepBeforeSendWatchResponse ,
268- Traffic : traffic .EtcdDelete ,
269- Cluster : * e2e .NewConfig (
270- e2e .WithClusterSize (1 ),
271- e2e .WithGoFailEnabled (true ),
272- ),
273- })
274- if v .Compare (version .V3_5 ) >= 0 {
275- opts := []e2e.EPClusterOption {
276- e2e .WithSnapshotCount (100 ),
277- e2e .WithPeerProxy (true ),
278- e2e .WithIsPeerTLS (true ),
279- }
280- if e2e .CouldSetSnapshotCatchupEntries (e2e .BinPath .Etcd ) {
281- opts = append (opts , e2e .WithSnapshotCatchUpEntries (100 ))
282- }
283- scenarios = append (scenarios , TestScenario {
284- Name : "Issue15271" ,
285- Failpoint : failpoint .BlackholeUntilSnapshot ,
286- Profile : traffic .HighTrafficProfile ,
287- Traffic : traffic .EtcdPut ,
288- Cluster : * e2e .NewConfig (opts ... ),
289- })
290- }
291- return scenarios
229+ // scenarios = append(scenarios, TestScenario{
230+ return nil
292231}
0 commit comments