A/A test shows significant results
Table of contents
How to recognize the error
Your A/A test shows a significant result for one or more metrics after a few days.
How to fix the error
An A/A test is used to check the reliability of the test setup. Two identical variants are tested against each other. Theoretically, both variants should achieve the same results. If an A/A test shows a significant result usually indicates that not on a real effect, but on Statistical or technical causes there.
1. importance of significance in the A/A test
Varify evaluates test results with a frequentist approach.
When the Coincidence largely ruled out the reciprocal of the p-value as Significance displayed.
If this value is above 95 %, Varify shows a significant result .
In an A/A test, however, this does not mean that one variant is actually „better“ - but that the data is randomly distributed in such a way that the probability of a difference appears greater than it actually is.
2. common causes for a „false“ significant result
a) Too short a term
With too little data, chance can still have a strong effect. A significant short-term deflection is normal and not a reliable signal.
b) Too many goals
Each additional target increases the probability of a so-called Alpha error. This means that the chance that a difference will be found somewhere by chance increases with the number of metrics.
c) Uneven traffic allocation
If visitors are not evenly distributed across the variants (e.g. through caching, bot traffic or incomplete playout), the result can be distorted.
d) Sample too small
Metrics fluctuate greatly with low conversion figures. A difference of just a few conversions can lead to a seemingly high significance.
3. best practice for A/A tests
To ensure that A/A tests provide reliable results, we recommend
- Running time: at least 10 days 
- Amount of data: at least 500 conversions per variant 
- Goals: maximum 3 Metrics, with a focus on the main KPIs 
- Ignore intermediate results: Significance values may fluctuate during the term. Only the Final result at the end of the test is meaningful. 
This ensures that the influence of chance remains low and the evaluation delivers realistic results.
4. conclusion
A significant result in an A/A test means in most cases No real signal, but is limited to Random or test configuration attributable.
Only when sufficient data has been collected over a longer period of time and chance can be statistically excluded is a result really reliable.
					 First steps 
							
			
			
		
						
				
					
					 Tracking & web analytics integrations 
							
			
			
		
						
				- Tracking with Varify
- Manual Google Tag Manager tracking integration
- Automatic GA4 tracking integration
- Shopify Custom Pixel Integration via Google Tag Manager
- Shopify Tracking
- BigQuery
- PostHog evaluations
- Matomo - Integration via Matomo Tag Manager
- etracker integration
- Piwik Pro Integration
- Consent - Tracking via Consent
- Advanced Settings
- Tracking with Varify
- Manual Google Tag Manager tracking integration
- Automatic GA4 tracking integration
- Shopify Custom Pixel Integration via Google Tag Manager
- Shopify Tracking
- BigQuery
- PostHog evaluations
- Matomo - Integration via Matomo Tag Manager
- etracker integration
- Piwik Pro Integration
- Consent - Tracking via Consent
- Advanced Settings
					 Create experiment 
							
			
			
		
						
				
					 Targeting 
							
			
			
		
						
				
					
					 Reporting & evaluation 
							
			
			
		
						
				- GA4 reporting in Varify.io
- BigQuery
- Segment and filter reports
- Share report
- Audience-based evaluation in GA4
- Segment-based evaluation in GA 4
- PostHog Tracking
- Exporting the experiment results from Varify
- Matomo - Results analysis
- etracker evaluation
- Calculate significance
- User-defined click events
- Evaluate custom events in explorative reports
- GA4 - Cross-Domain Tracking
- GA4 reporting in Varify.io
- BigQuery
- Segment and filter reports
- Share report
- Audience-based evaluation in GA4
- Segment-based evaluation in GA 4
- PostHog Tracking
- Exporting the experiment results from Varify
- Matomo - Results analysis
- etracker evaluation
- Calculate significance
- User-defined click events
- Evaluate custom events in explorative reports
- GA4 - Cross-Domain Tracking
					 Visual editor 
							
			
			
		
						
				- Campaign Booster: Arrow Up
- Campaign Booster: Exit Intent Layer
- Campaign Booster: Information Bar
- Campaign Booster: Notification
- Campaign Booster: USP Bar
- Add Link Target
- Browse Mode
- Custom Selector Picker
- Edit Content
- Edit Text
- Move elements
- Hide Element
- Keyword Insertion
- Redirect & Split URL Testing
- Remove Element
- Replace Image
- Responsive Device Switcher
- Style & Layout Changes
- Campaign Booster: Arrow Up
- Campaign Booster: Exit Intent Layer
- Campaign Booster: Information Bar
- Campaign Booster: Notification
- Campaign Booster: USP Bar
- Add Link Target
- Browse Mode
- Custom Selector Picker
- Edit Content
- Edit Text
- Move elements
- Hide Element
- Keyword Insertion
- Redirect & Split URL Testing
- Remove Element
- Replace Image
- Responsive Device Switcher
- Style & Layout Changes