-
Notifications
You must be signed in to change notification settings - Fork 4
Issues: zeebe-io/zeebe-chaos
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Hypothesis: CPU stress on standalone gateway should not cause harm cluster wise
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: Medium
The issue is not so likely.
#28
opened Jun 11, 2020 by
ChrisKujawa
Hypothesis: High CPU consumption on one Node will not break cluster
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#6
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Large State does not cause leader changes
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#8
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: We can handle storage errors
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Low
The issue is really unlikely.
#9
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Operate receives data eventually
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: Low
The issue is really unlikely.
#10
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: We can handle corrupted events
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Low
The issue is really unlikely.
#11
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Cluster restart will be handled proberly
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Low
The issue is really unlikely.
#12
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Operate always shows consistent data
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: Low
The issue is really unlikely.
#13
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: We can recover after full disk due to failed exporting
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Low
The issue is really unlikely.
#14
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: With Long Polling I receive Job eventually
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Likelihood: Low
The issue is really unlikely.
#15
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Operate can handle connection Errors
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Low
The issue is really unlikely.
#16
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Multiple Workflows should not affect the system
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#18
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Clock reset will not fail exporting
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: Low
The issue is really unlikely.
#3
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: No difference in Standalone and embedded Gateway
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Performability
This issue will contribute to build up confidence in performability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#19
opened Mar 4, 2020 by
ChrisKujawa
Hypothesis: Partition one Broker with Gateway doesn't affect other partitions
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#29
opened Jun 26, 2020 by
ChrisKujawa
Hypothesis:I believe I can start a lot of timers without impacting the general throughput
Await Fix
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Performability
This issue will contribute to build up confidence in performability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Low
The issue has an low impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#31
opened Jul 9, 2020 by
ChrisKujawa
Hypothesis: We stop accepting commands on reaching disk space
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: High
The likelihood of this issue is really high!
Likelihood: Low
The issue is really unlikely.
#32
opened Jul 16, 2020 by
ChrisKujawa
Hypothesis: Reaching the maxMessageLimit in an workflow instance will not break the partition
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Performability
This issue will contribute to build up confidence in performability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: Medium
The issue has an medium impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#33
opened Jul 16, 2020 by
ChrisKujawa
Chaos: Inject latency to broker network
Chaos Experiment
This issue describes a chaos experiments, which should be created.
#37
opened Sep 28, 2020 by
ChrisKujawa
Chaos: Run cont. load during experiment
Chaos Experiment
This issue describes a chaos experiments, which should be created.
#40
opened Oct 13, 2020 by
ChrisKujawa
Hypothesis: Disconnecting Leader and one Follower should not make cluster disruptive
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Contribution: Reliability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: High
The likelihood of this issue is really high!
#45
opened Nov 17, 2020 by
ChrisKujawa
Hypothesis: If a leader has a write error, which is not recoverable, it will step down and another leader should take over.
Contribution: Availability
This issue will contribute to build up confidence in reliability.
Hypothesis
A thing which worries us and is ready for exploration.
Impact: High
The issue has an high impact on the system.
Likelihood: Medium
The issue is not so likely.
#52
opened Mar 31, 2021 by
ChrisKujawa
Previous Next
ProTip!
Follow long discussions with comments:>50.