Phase Two: Execution
Setting up your Delphi study inside Durvey.org ensures that everything is structured, traceable, and ready to produce reliable results. If you have never created a Delphi before, don't worry – every step is designed to guide you and prevent common mistakes.
What You'll Learn: Clear explanations about each execution step, why it matters, and how it will help you create a professional, high-quality Delphi survey.
Step 1: Create Your Project and Configure Survey Settings
Before you can start collecting expert opinions, you need to define the framework of your study. This foundational setup determines how your Delphi process will unfold.
Create Your First Project
Project Name and Description
Give your study a clear, descriptive name and provide a short explanation of its purpose.
Example: "Future of AI in Public Health 2035" – Exploring expert consensus on emerging technologies that will transform healthcare delivery.
Survey Mode Configuration
Real-Time Delphi
Participants see an updated group average as soon as more responses come in.
When to use:
- Need faster consensus-building
- Want dynamic, adaptive feedback
- Working with time-sensitive topics
Manual Rounds
Feedback is only shared after each round is closed by the moderator.
When to use:
- Want to reduce potential groupthink
- Need controlled feedback timing
- Conducting classical Delphi research
Important: This choice affects how opinions evolve over time. Real-Time Delphi can speed up consensus-building, while Manual Rounds can reduce groupthink by preventing early responses from overly influencing later ones.
Anonymous Participation Settings
Configure whether expert identities remain anonymous. Anonymity reduces bias from authority and encourages honest ratings without fear of judgment.
Full Anonymity
Experts cannot see who else is participating. Only aggregated group statistics are shown.
Visible Experts
Panel members can see who is participating, but individual responses remain anonymous.
Pre-Phase Configuration
Define how many participants you want to submit their first ratings before showing aggregated feedback. This prevents bias when the first person enters a rating (because otherwise there is no average yet).
Recommended Setting: Wait for at least 3-5 responses before showing group statistics to ensure meaningful averages and prevent anchoring bias from a single early response.
Automated Notifications
Enable reminders to inform early participants when they should return to complete a second evaluation once the group average is available.
Notification Types
- Round opening notifications: Alert experts when a new round begins
- Feedback available notifications: Inform early respondents when group data is ready
- Gentle reminders: Nudge non-respondents before deadlines
- Round closing alerts: Final reminder before round closes
Step 2: Design Your Delphi Questions and Rating Dimensions
The quality of your questions determines the clarity and usefulness of your results. If your statements are vague or your rating scales inconsistent, experts will struggle to provide meaningful input. That's why this step is essential.
Create, Duplicate, or Delete Pages
Organize your survey into logical sections or topics. This helps experts navigate complex studies and keeps related questions together.
Best Practices
- Group related theses into themed pages
- Use page duplications to maintain consistent formatting
- Avoid having too many theses on one page (max 10-15)
- Use clear page titles that describe the topic
State Your Theses or Questions
Each statement should be specific, relevant, and neutral in tone to avoid bias.
Good Example:
"Artificial intelligence will be the main diagnostic tool in public health by 2035."
Poor Example:
"Don't you think AI might somehow change healthcare?" (Too vague, leading, lacks timeframe)
Define Rating Dimensions
Decide how you want experts to rate each thesis. Common options include:
Likert Scales
1-5 or 1-7 agreement scales
Example: 1 = Strongly Disagree, 5 = Strongly Agree
Time Expectations
Year-based predictions
Example: "In which year do you expect this to happen?"
Probability Estimates
0-100% likelihood scales
Example: "What is the probability this will occur?"
Importance Ratings
1-10 priority scales
Example: "How important is this factor?"
Configure Comments and Additional Inputs
Enable rich expert interaction beyond numerical ratings. Durvey.org provides powerful commenting features that capture nuanced expert reasoning and facilitate engagement.
Pro/Contra Comments
Experts can add comments tagged with sentiment (pro or contra). This helps identify why experts support or oppose specific theses and provides context for ratings.
Like and Interact
Experts can "like" other comments to show agreement, creating a lightweight way to express support for particular viewpoints without writing duplicate comments.
Explanation Fields
Ask experts to justify their ratings. This qualitative data enriches your analysis and helps other participants understand different perspectives.
Research Benefit: These interactive features generate rich qualitative data alongside quantitative consensus metrics. Your final analysis will include both statistical agreement and deep expert insights.
Step 3: Define Metadata and Stakeholder Variables
Metadata enables you to analyze results by group and see patterns you would otherwise miss. For example, technology experts may rate a trend very differently than policymakers. Collecting this information enriches your analysis.
Identify Key Variables
Think about what categories are relevant. Examples:
Professional Background
- Researcher
- Practitioner
- Policymaker
- Industry expert
Experience Level
- Years in field
- Seniority level
- Publications
- Certifications
Geographic Context
- Region
- Country
- Urban/Rural
- Organization type
Explain Why You're Collecting Metadata
In your survey introduction, tell participants that this information will help compare perspectives across sectors and increase the quality of your findings.
Warning: If you skip this step, you lose the ability to see valuable differences between subgroups, which weakens your study's insights.
Step 4: Finalize and Test Your Survey
Before you invite experts to participate, you need to make sure everything works perfectly. Testing helps catch typos, technical issues, and confusing instructions. This reduces dropout rates and increases trust in your process.
Randomize Page Order (if needed)
Prevent order effects that could bias results by randomizing the sequence in which pages are presented to different participants.
Review Your Survey in Test Mode
Check the experience exactly as your participants will see it.
Testing Checklist
- All questions display correctly
- Rating scales work properly
- Navigation flows smoothly
- Instructions are clear
- Time to complete is reasonable (≤20 minutes)
- Mobile compatibility verified
Confirm Rating Scales and Instructions
Make sure all labels are clear and consistent across all questions.
Step 5: Create and Manage Participant Profiles
Your Delphi results are only as good as your expert panel. That's why it's critical to record clear participant profiles and manage them in a structured way.
Add Participants Manually
Add each expert manually, including:
- Name: Full name of the expert
- Email address: Valid email for invitations
- Professional background: Role, organization, expertise area
- Relevant metadata: Any stakeholder variables you defined
Bulk Import Expert Lists
If you already have a list of contacts in Excel, you can import them all at once to save time.
Tip: Ensure your spreadsheet includes all required fields (name, email, metadata) to avoid errors during import.
Step 6: Draft Professional Invitations, Reminders, and Thank-You Emails
A clear, motivating invitation increases the chance that experts will respond promptly and stay engaged throughout multiple rounds. Durvey.org provides comprehensive email management tools to streamline communication and maintain expert engagement.
Durvey's Email Management Features
- Custom templates: Create branded email templates with your institution's style
- Automated sending: Schedule invitations to go out at optimal times
- Smart reminders: Automated reminders sent only to non-respondents
- Personalization: Merge fields for expert names and custom details
- Tracking: See who opened emails and clicked links
Write an Invitation Email
Clearly explain:
- The purpose and relevance of your Delphi study
- Why the recipient was selected (acknowledge their expertise)
- How their input will make a difference (impact and value)
- What the expected time commitment is (be honest and specific)
Use or Customize Templates
Durvey.org provides ready-to-use templates you can adapt to your tone and audience.
Professional Tone
For academic or policy contexts requiring formal language
Conversational Tone
For industry or practitioner panels with collaborative culture
Compose Reminders and Thank-You Messages
Reminders
Schedule friendly reminders. This reduces dropouts and ensures high-quality longitudinal data.
Recommended: Send 3 days before deadline, then 1 day before
Thank-You Messages
At the end of each round, show appreciation and share a short update on next steps.
Step 7: Launch Your Delphi Study and Manage Invitations
Launching the survey officially starts data collection. You control when invitations go out and when rounds open and close.
Launch Your Delphi Study
Once you launch, participants can begin submitting their ratings and comments.
Ready to Launch? Double-check all settings, test the survey, and ensure your participant list is complete before hitting launch.
Send Targeted Invitations and Reminders
Send invitations and remind only those who haven't completed the round.
Smart Communication
- Send invitations in batches if you have a large panel
- Target reminders only to non-respondents
- Avoid over-messaging (max 2-3 reminders per round)
- Personalize messages when possible
Track Opened and Pending Invitations
See who has opened your email, started the survey, or not yet responded.
Tip: Regular communication shows professionalism and helps maintain engagement over several rounds.
Step 8: Monitor Participation Progress in Real Time
Ongoing monitoring ensures you can react early if engagement drops or certain stakeholder groups are underrepresented.
Review Completion Rates
See the percentage of invited experts who have finished each round.
Target Response Rates
- Excellent: ≥80% completion
- Good: 70-79% completion
- Acceptable: 60-69% completion
- Concerning: <60% completion (take action)
Check Participation by Segment
Analyze who is participating (e.g., by region, sector, experience) to ensure a balanced dataset.
Watch For: Systematic underrepresentation of certain stakeholder groups. If one segment has low participation, send targeted follow-ups or recruit additional experts from that group.
Step 9: Close the Delphi Study
Formally closing each round locks the data, allowing you to prepare feedback and start the next stage of analysis or iteration.
Decide When to Close the Round
You can close it based on numerous indicators:
- Time-based: Deadline has passed
- Response-based: Sufficient participation achieved
- Stability-based: No significant changes in recent responses
- Consensus-based: Target agreement levels reached
Notify Participants (Optional)
Send a closing message summarizing participation or thanking participants for their contributions.
Ready to Analyze Your Results?
Once data collection is complete, it's time to conduct thorough analysis and interpret your expert consensus findings.
Next: Analysis Phase →