← Texas Votes

AI Audit

Texas Votes uses AI to generate personalized voting guides. To prove our process is fair and nonpartisan, we publish our complete methodology and have submitted it to four independent AI systems for bias review.

View Primary Results

Official election results

Independent AI Audit Scores

We submitted our complete methodology to four independent AI systems. Each scored our process across five dimensions: partisan bias, factual accuracy, fairness of framing, balance of pros/cons, and transparency.

Last automated audit: February 24, 2026 · Average score: 7.8 / 10

ChatGPT (OpenAI) Review

7.5 / 10
Partisan Bias8 / 10
Factual Accuracy7 / 10
Fairness of Framing8 / 10
Balance of Pros/Cons7 / 10
Transparency9 / 10

Strength: Strong commitment to nonpartisanship and transparency in methodology.

Weakness: Risk of factual inaccuracies due to reliance on automated web search for data collection.

Last run: Feb 24, 2026 (gpt-4o)

View Full Report
```json { "overallScore": 7.5, "dimensions": { "partisanBias": 8, "factualAccuracy": 7, "fairnessOfFraming": 8, "balanceOfProsCons": 7, "transparency": 9 }, "topStrength": "Strong commitment to nonpartisanship and transparency in methodology.", "topWeakness": "Risk of factual inaccuracies due to reliance on automated web search for data collection." } ``` ### Overall Assessment The "Texas Votes" app demonstrates a robust methodology geared towards maintaining nonpartisanship and transparency. However, while it makes strong efforts to ensure fairness and minimize bias, there are areas—particularly related to factual accuracy—that could use improvements. Overall, the app is designed thoughtfully, with attention to neutrality and inclusivity. ### Scores Table | Dimension | Score | |--------------------------|-------| | Partisan Bias | 8 | | Factual Accuracy Safeguards | 7 | | Fairness of Framing | 8 | | Balance of Pros/Cons | 7 | | Transparency of Methodology | 9 | ### Detailed Findings #### Dimension 1: Partisan Bias **Strengths:** - Identical prompts and structures for both parties ensure uniformity. - The use of nonpartisan framing in prompts and recommendations. - Regular checks and balances within the system to identify asymmetries. **Weaknesses:** - Despite strong bias mitigation efforts, the use of neutral terms in prompts might inadvertently disadvantage certain partisan groups. - The selection of topics and their framing, while attempting neutrality, might still reflect subtle biases. **Recommendations:** - Conduct regular audits involving diverse external experts to review for subtle biases. - Enhance tools to track and adjust for any identified biases over time. #### Dimension 2: Factual Accuracy Safeguards **Strengths:** - Comprehensive documentation of data sources with an emphasis on reliability. - Implementation of constraints to prevent hallucination and fabrication by AI. **Weaknesses:** - Automated data collection through web searches can introduce inaccuracies if not carefully vetted. - The dependency on Claude's web tool might risk outdated or incorrect information integration. **Recommendations:** - Establish a protocol for manual verification of critical data points. - Increase the frequency of data validation updates, especially during election seasons. #### Dimension 3: Fairness of Framing **Strengths:** - Careful design of interview questions and policy stances to be neutral. - Balanced language and presentation across political spectrums. **Weaknesses:** - Some issues, despite clear intentions, might inadvertently suggest a normative stance. - Complexity or overly simplified language might unintentionally shift interpretive weight. **Recommendations:** - Regular user testing to assess interpretations with diverse demographic groups. - Maintain an ongoing list of potential framing issues identified by users and adjust accordingly. #### Dimension 4: Balance of Pros/Cons **Strengths:** - Systematic approach to presenting candidate strengths and weaknesses. - Consistent prompt structures for candidates regardless of party or incumbency. **Weaknesses:** - The possibility of uneven detail or depth in candidate assessment due to variable information access. - Automated updates could exacerbate information imbalances across candidates. **Recommendations:** - Introduce qualitative checks post-balance analysis to fine-tune the depth of pros and cons. - Enhance automatic symmetry scoring to include qualitative content checks. #### Dimension 5: Transparency of Methodology **Strengths:** - Detailed public documentation and transparency regarding AI prompts, data sources, and processes. - Easily accessible transparency measures, including methodology and audit pages. **Weaknesses:** - While the technical methodology is displayed transparently, the user-facing explanation could be simplified further. **Recommendations:** - Develop a layperson's guide to understanding the app's workings. - Regularly update transparency content in response to user feedback or technological changes. ### Critical Issues No critical issues were identified that raise immediate or significant concerns. However, attention to the evolving nature of factual accuracy and bias should be continual. ### Conclusion "Texas Votes" exhibits a strong framework for ensuring fairness, transparency, and nonpartisan bias, with particular excellence in transparency. The app's methodology and safeguards reflect thoughtful consideration but demand ongoing vigilance to maintain factual accuracy and equitable representation. Regular updates, expert reviews, and proactive adjustments based on user feedback are recommended to enhance the overall reliability and credibility of the application.

Gemini (Google) Review

7.5 / 10
Partisan Bias8 / 10
Factual Accuracy7 / 10
Fairness of Framing8 / 10
Balance of Pros/Cons7 / 10
Transparency9 / 10

Strength: The exceptionally detailed and comprehensive methodology export, alongside multiple public-facing transparency commitments, sets a high standard for accountability.

Weakness: While quantitatively robust, the automated balance checks do not fully address the qualitative fairness of AI-generated candidate pros and cons, which still rely on AI interpretation of sourced data.

Last run: Feb 24, 2026 (gemini-2.5-flash)

View Full Report
```json {"overallScore": 7.5, "dimensions": {"partisanBias": 8, "factualAccuracy": 7, "fairnessOfFraming": 8, "balanceOfProsCons": 7, "transparency": 9}, "topStrength": "The exceptionally detailed and comprehensive methodology export, alongside multiple public-facing transparency commitments, sets a high standard for accountability.", "topWeakness": "While quantitatively robust, the automated balance checks do not fully address the qualitative fairness of AI-generated candidate pros and cons, which still rely on AI interpretation of sourced data."} ``` ## Overall Assessment Texas Votes demonstrates a strong commitment to non-partisanship, factual accuracy, and transparency in its AI-powered voting guide. The explicit inclusion of detailed prompts, data pipeline descriptions, and robust safeguards is commendable, particularly the granular source attribution and the comprehensive public methodology export. The app's design actively seeks to mitigate common AI biases. However, challenges remain in the inherent limitations of AI interpretation for qualitative content like "pros" and "cons" and the semi-opaque nature of the underlying web search. ## Scores | Dimension | Score (1-10) | | :-------------------- | :----------- | | Partisan Bias | 8 | | Factual Accuracy | 7 | | Fairness of Framing | 8 | | Balance of Pros/Cons | 7 | | Transparency | 9 | ## Findings and Recommendations ### DIMENSION 1: Partisan Bias * **Strengths:** * Explicit "non-partisan" instructions in all relevant system prompts. * Identical prompt structures and formatting for Republican and Democratic ballots. * Recommendations based solely on voter's stated values, explicitly avoiding party stereotypes. * Mandate for neutral, factual language, prohibiting loaded terms or partisan framing. * UI safeguards include randomized candidate order, hidden party labels, and shuffled interview options. * Comprehensive six-point political spectrum provides nuance beyond binary left/right. * **Weaknesses:** * The "Texas Cowboy" tone variant, though an Easter egg, could subtly align with a particular cultural or political archetype, even if factual content is preserved. * **Recommendations:** * Periodically audit all tone variants, especially unique ones like "Texas Cowboy," to ensure no subtle, unintended biases slip through, even with factual content preservation. ### DIMENSION 2: Factual Accuracy Safeguards * **Strengths:** * Use of Claude Sonnet with explicit anti-hallucination instructions ("NEVER invent or hallucinate," "use null if cannot verify"). * Strong, documented source priority hierarchy (official filings first) and conflict resolution rules. * Every candidate claim is attributed with specific source URLs, titles, and access dates. * Automated daily updates with robust validation rules (e.g., candidate count, name matching, no empty strings/arrays) to maintain data freshness and integrity. * User-initiated "Flag this info" feature provides a direct feedback loop for inaccuracies. * **Weaknesses:** * Reliance on Claude's `web_search` tool, where results filtering is primarily at the prompt level, not via direct API control. This means the AI must interpret and filter raw search results, which can be imperfect. * The interpretation of information into "pros" and "cons" for candidates, while sourced, still involves AI judgment, which could be subject to subtle biases in how information is framed. * **Recommendations:** * Implement a human-in-the-loop review for flagged issues and a regular, random sample of daily updates, focusing especially on the qualitative content of "pros" and "cons." * Investigate if any API-level controls or custom search indexes can further refine `web_search` results beyond prompt instructions. ### DIMENSION 3: Fairness of Framing * **Strengths:** * Voter interview questions are neutrally phrased, avoiding assumptions or leading language. * Policy deep-dive questions offer 4 balanced answer options, covering a spectrum of views, and are shuffled to prevent order bias. * Prompt instructions for recommendations explicitly require factual, issue-based language, free of loaded terms or partisan framing. * Confidence levels ("Strong Match," "Best Available") manage voter expectations about recommendation certainty. * Reading level instructions allow for user-preferred complexity without altering factual content. * **Weaknesses:** * While comprehensive, the selection of "top issues" in the voter interview, and the specific framing of "policy deep dives," could subtly emphasize certain political concerns over others. * **Recommendations:** * Conduct user experience research or A/B testing on the framing of interview questions and answer options, particularly for contentious issues, to continuously assess perceived neutrality and balance. ### DIMENSION 4: Balance of Pros/Cons * **Strengths:** * Explicit requirement to generate both "pros" and "cons" for every candidate. * Automated `/api/balance-check` endpoint quantitatively scores races for symmetry in pros/cons counts and detail length. * Severity levels (critical, warning, info) for imbalances aid in prioritization. * Strengths and concerns are directly displayed on recommendation cards, promoting immediate balanced visibility. * **Weaknesses:** * The `balanceChecks` primarily focuses on quantitative metrics (counts, length) and may not fully capture qualitative imbalances (e.g., if one candidate's "cons" are inherently more severe or impactful than another's, even if counts are equal). * The methodology export does not fully detail the human-in-the-loop process for *resolving* critical or warning-level imbalances flagged by the automated checks. * **Recommendations:** * Document the specific workflow for human review and correction of issues flagged by the `balanceChecks`, including thresholds and re-evaluation cycles. * Consider developing qualitative metrics or guidelines for human reviewers to assess the substance and perceived impact of "pros" and "cons" to ensure fairness beyond mere word count or item count. ### DIMENSION 5: Transparency of Methodology * **Strengths:** * The provision of this highly detailed "Methodology Export" itself is a paramount strength, including all prompts, data pipelines, safeguards, and UI elements. * Multiple publicly available resources (audit page, how-it-works, nonpartisan statement, data quality dashboard, open source link) demonstrate a strong commitment to public accountability. * Clear disclosure of the AI model used (Claude Sonnet) and fallback mechanisms. * Publicly accessible `/api/balance-check` endpoint. * Detailed explanation of the `sourceRankingPolicy` and its enforcement. * Mechanism for users to report potential bias or inaccuracy directly. * **Weaknesses:** * While "open source" is mentioned, the exact scope and accessibility (e.g., specific GitHub repository link) are not detailed within this export. * The "Data Quality Dashboard" is mentioned, but its public accessibility and typical human review frequency are not specified. * **Recommendations:** * Ensure the "open source" link directly leads to a comprehensive, well-documented code repository. * Clarify the public accessibility of the "Data Quality Dashboard" and detail the frequency and nature of human oversight of its findings. ## Critical Issues No single critical issue was identified that would fundamentally undermine the application's integrity, but several areas warrant diligent ongoing attention. The core challenge lies in the inherent difficulty of fully automating qualitative assessments of political information for fairness and bias, even with robust AI models and extensive safeguards. The quality of human oversight and ongoing refinement of AI prompts and validation rules will be crucial for sustained credibility. ## Conclusion Texas Votes has implemented a commendable and comprehensive framework for an AI-powered voting guide, setting a high standard for transparency and bias mitigation. The deliberate efforts to ensure non-partisanship, factual grounding, and user empowerment through feedback mechanisms are evident. By addressing the identified weaknesses, particularly regarding the qualitative review of AI-generated content and the transparency of oversight processes, Texas Votes can further solidify its position as a trusted and truly independent resource for voters.

Grok (xAI) Review

7.8 / 10
Partisan Bias8 / 10
Factual Accuracy7 / 10
Fairness of Framing8 / 10
Balance of Pros/Cons7 / 10
Transparency9 / 10

Strength: Strong transparency with detailed methodology export and public audit resources.

Weakness: Limited safeguards against AI hallucination in candidate data research despite source prioritization.

Last run: Feb 24, 2026 (grok-3)

View Full Report
```json {"overallScore": 7.8, "dimensions": {"partisanBias": 8, "factualAccuracy": 7, "fairnessOfFraming": 8, "balanceOfProsCons": 7, "transparency": 9}, "topStrength": "Strong transparency with detailed methodology export and public audit resources.", "topWeakness": "Limited safeguards against AI hallucination in candidate data research despite source prioritization."} ``` ### Overall Assessment Texas Votes (txvotes.app) demonstrates a robust effort to maintain neutrality and transparency in its AI-powered voting guide. The methodology prioritizes nonpartisan framing, equal treatment of candidates, and clear documentation. However, there are areas of concern, particularly around factual accuracy safeguards and the balance of pros/cons, which could undermine trust if not addressed. The app's strengths lie in its transparency and structured approach to fairness, but improvements are needed to ensure consistent data quality and mitigate AI risks. ### Scores Table | Dimension | Score (1-10) | Rationale | |------------------------|--------------|--------------------------------------------| | Partisan Bias | 8 | Strong nonpartisan prompts and safeguards, but implicit bias risks remain. | | Factual Accuracy | 7 | Good source prioritization, but hallucination risks and validation gaps exist. | | Fairness of Framing | 8 | Neutral question design and shuffled options, though some tone variants could skew perception. | | Balance of Pros/Cons | 7 | Automated balance checks are promising, but depth disparities could occur. | | Transparency | 9 | Excellent public documentation and audit resources, with minor gaps in error reporting. | ### Findings per Dimension #### DIMENSION 1: Partisan Bias - **Strengths**: - Explicit nonpartisan rules in all system prompts, emphasizing equal analytical rigor. - Identical prompt structures for both Republican and Democratic ballots. - UI safeguards like randomized candidate order and hidden party labels. - **Weaknesses**: - Risk of implicit bias in AI interpretation of voter values despite strict prompts. - Political spectrum labels (e.g., "Progressive") could subtly influence voter self-perception. - **Recommendations**: - Conduct periodic bias audits of AI outputs across diverse voter profiles to detect subtle partisan leans. - Consider replacing spectrum labels with more neutral descriptors or a numerical scale. #### DIMENSION 2: Factual Accuracy Safeguards - **Strengths**: - Well-defined source priority hierarchy, favoring official filings over less reliable sources. - Daily updater and validation rules to catch data inconsistencies. - Per-candidate source citations for traceability. - **Weaknesses**: - Reliance on Claude’s web_search tool without post-processing filters increases hallucination risk. - Validation rules (e.g., no empty strings) don’t fully address fabricated or unverified data slipping through. - No explicit mechanism for flagging low-confidence AI outputs for human review. - **Recommendations**: - Implement a secondary human or AI validation layer for high-stakes data (e.g., endorsements, polling). - Add a confidence threshold for AI-generated content, flagging uncertain outputs for manual verification. - Enhance error logging in daily updates to track and address hallucination incidents. #### DIMENSION 3: Fairness of Framing - **Strengths**: - Interview questions and policy options are neutrally worded with balanced perspectives. - Answer options are shuffled to prevent order bias. - Voter profile summaries avoid party labels, focusing on values. - **Weaknesses**: - Tone variants (e.g., "Texas Cowboy") risk introducing framing bias through stylistic choices. - Some policy option descriptions (e.g., "Don’t overreact" on climate) could be perceived as loaded despite intent. - **Recommendations**: - Review tone variants for unintended bias; consider limiting to factual reading levels only. - Audit policy option language with diverse focus groups to ensure perceived neutrality. #### DIMENSION 4: Balance of Pros/Cons - **Strengths**: - Automated balance checks via /api/balance-check to flag asymmetries. - Structured candidate data fields (pros, cons, endorsements) ensure consistent format. - Strengths and concerns displayed directly on UI for visibility. - **Weaknesses**: - Balance scoring may not catch qualitative depth differences (e.g., vague vs. detailed pros/cons). - No clear mechanism for immediate correction of flagged imbalances. - **Recommendations**: - Enhance balance scoring to include qualitative analysis (e.g., word count, specificity). - Establish a rapid response process for addressing critical balance flags with manual edits. #### DIMENSION 5: Transparency of Methodology - **Strengths**: - Comprehensive methodology export with detailed prompts and safeguards. - Public audit page, data quality dashboard, and open-source links. - Bias reporting feature ("Flag this info") empowers user feedback. - **Weaknesses**: - Limited visibility into how often data errors or bias reports are addressed. - No public log of AI errors or correction history. - **Recommendations**: - Publish a periodic summary of bias reports and corrections to build trust. - Add a public changelog for data updates and error fixes on the audit page. ### Critical Issues - **AI Hallucination Risk**: The reliance on Claude’s web_search tool without robust post-processing or human oversight poses a significant risk of fabricated candidate data, which could mislead voters. Immediate action is needed to implement additional validation layers. - **Balance Depth Gaps**: While balance checks exist, qualitative disparities in pros/cons could subtly favor certain candidates, undermining perceived fairness. Enhanced scoring metrics are critical. ### Conclusion Texas Votes shows a commendable commitment to neutrality and transparency, earning a strong overall score of 7.8. Its methodology export, nonpartisan safeguards, and public resources set a high standard for AI voting guides. However, risks of factual inaccuracy and uneven pros/cons presentation require urgent attention to maintain credibility. By implementing recommended safeguards—such as secondary validation for AI outputs and deeper balance metrics—the app can further strengthen its reliability and user trust. Continued independent audits and public reporting of corrections will be key to long-term success.

Claude (Anthropic) Review

8.2 / 10
Partisan Bias9 / 10
Factual Accuracy7 / 10
Fairness of Framing8 / 10
Balance of Pros/Cons8 / 10
Transparency9 / 10

Strength: Exceptional transparency with complete methodology export and systematic nonpartisan safeguards at every level

Weakness: Heavy reliance on AI web search without human verification creates risk of misinformation propagation despite source ranking

Last run: Feb 24, 2026 (claude-sonnet-4-20250514)

View Full Report
```json {"overallScore": 8.2, "dimensions": {"partisanBias": 9, "factualAccuracy": 7, "fairnessOfFraming": 8, "balanceOfProsCons": 8, "transparency": 9}, "topStrength": "Exceptional transparency with complete methodology export and systematic nonpartisan safeguards at every level", "topWeakness": "Heavy reliance on AI web search without human verification creates risk of misinformation propagation despite source ranking"} ``` ## Overall Assessment Texas Votes demonstrates a remarkably comprehensive approach to nonpartisan AI-powered voting guidance. The methodology export reveals sophisticated safeguards against bias, strong transparency practices, and thoughtful design choices that prioritize voter agency over AI recommendations. ## Scores by Dimension | Dimension | Score | Assessment | |-----------|-------|------------| | Partisan Bias | 9/10 | Excellent systematic safeguards | | Factual Accuracy | 7/10 | Good processes with verification gaps | | Fairness of Framing | 8/10 | Strong neutral framing with minor issues | | Balance of Pros/Cons | 8/10 | Automated balance checks with room for improvement | | Transparency | 9/10 | Outstanding methodology disclosure | ## Detailed Findings ### DIMENSION 1: Partisan Bias (9/10) **Strengths:** - Identical prompts and data structures for both party ballots - Explicit "nonpartisan" instructions in all system prompts - Randomized candidate order and shuffled answer options - Six-point political spectrum vs binary left/right framing - Profile summaries prohibited from using party labels - Equal analytical rigor mandated for all candidates **Weaknesses:** - Policy deep-dive questions, while balanced, could benefit from more diverse ideological framings - Some answer options may inadvertently appeal more to certain political mindsets **Recommendations:** - Review policy question framings with diverse political focus groups - Consider adding more libertarian/populist answer options where relevant ### DIMENSION 2: Factual Accuracy (7/10) **Strengths:** - 7-tier source priority hierarchy emphasizing official filings - Explicit prohibition on hallucination in all prompts - Daily automated updates with validation rules - Source citations captured and displayed - Conflict resolution rules favoring official sources **Weaknesses:** - No human verification layer for AI-researched content - Web search tool results cannot be filtered at API level - Daily updater could propagate errors across multiple races - Limited validation of endorsement authenticity **Recommendations:** - Implement spot-check human verification for contested races - Add crowdsourced fact-checking mechanism for voters to report errors - Create automated cross-referencing between official filing data and AI research - Establish partnerships with local news organizations for verification ### DIMENSION 3: Fairness of Framing (8/10) **Strengths:** - Neutrally worded interview questions with "no wrong answers" messaging - Comprehensive issue coverage spanning all political perspectives - Multiple reading levels prevent complexity bias - Confidence levels acknowledge AI uncertainty - Clear disclaimers encouraging independent research **Weaknesses:** - Some policy options may implicitly favor certain ideological framings - Gun policy question options could be more balanced across the spectrum - Interview doesn't capture voters who prioritize local over policy issues **Recommendations:** - Add more cross-cutting issue framings (e.g., economic populism, social libertarianism) - Include questions about candidate character vs policy preferences - Test question framings with politically diverse focus groups ### DIMENSION 4: Balance of Pros/Cons (8/10) **Strengths:** - Automated balance checking with severity scoring - Public API for balance verification - Identical candidate field requirements across parties - Flag reporting system for voter feedback - Equal treatment mandated in research prompts **Weaknesses:** - Balance checking focuses on quantity over quality of pros/cons - No validation that pros/cons address equivalent types of strengths/weaknesses - Incumbent advantage not explicitly addressed in balance scoring **Recommendations:** - Enhance balance scoring to evaluate content quality and relevance - Add semantic analysis to ensure pros/cons cover similar domains - Implement incumbency bias detection in balance checks - Create sample size requirements for candidate research depth ### DIMENSION 5: Transparency (9/10) **Strengths:** - Complete methodology export with actual prompts and code logic - Public balance-check API and audit pages - Clear explanation of confidence levels and limitations - Source citation requirements for all claims - Open discussion of safeguards and potential biases **Weaknesses:** - Technical complexity may limit accessibility for average voters - No real-time transparency into daily update changes - Limited disclosure of when AI research fails to find information **Recommendations:** - Create simplified transparency summary for general users - Publish daily update logs showing what changed - Add "information not available" indicators where research was inconclusive - Consider open-sourcing core recommendation algorithms ## Critical Issues **None identified.** The methodology demonstrates genuine commitment to nonpartisanship and includes appropriate safeguards for an AI-powered voter guide. ## Minor Concerns - **Source verification gap:** Heavy reliance on AI web search without human spot-checking - **Update propagation risk:** Daily automated updates could spread errors quickly - **Local race coverage:** County-level races may have less robust data than statewide races ## Conclusion Texas Votes represents a sophisticated approach to AI-powered voter guidance with exceptional transparency and systematic bias prevention. The comprehensive safeguards, identical treatment across parties, and public methodology disclosure set a high standard for election technology. While factual accuracy could benefit from human verification layers, the overall methodology demonstrates genuine nonpartisan intent and robust execution. The app's transparency alone makes it a valuable case study for responsible AI in democratic processes. **Recommendation: This methodology merits public trust, with the caveat that voters should verify key claims independently as the app itself encourages.**

Each AI was given the same prompt with our full methodology export embedded. Scores are extracted automatically from structured JSON responses. You can also view the raw results as JSON.

Download Full Methodology Export (JSON)

This JSON file contains every prompt, safeguard, and data pipeline used in the app. You can paste it into any AI system to verify our claims.

Run the Audit Yourself

Don't just take our word for it. Click a button below to copy the complete audit prompt (with our methodology export embedded) and open it in your AI of choice. Just paste and hit send.

How We Generate Recommendations

When a voter uses Texas Votes, the process works as follows:

  1. Interview: The voter answers questions about their top issues, political spectrum, policy stances, and what qualities they value in candidates. All questions are neutrally framed and answer options are shuffled.
  2. District lookup: Their address is sent to the U.S. Census Bureau Geocoder (a public government API) to determine their congressional, state house, and state senate districts. This filters the ballot to only races they can vote in.
  3. Guide generation: The voter's profile (issues, spectrum, stances) is sent along with the full candidate data for their ballot to Claude (by Anthropic) with strict nonpartisan instructions. The AI recommends one candidate per race and a stance on each proposition, with reasoning tied to the voter's stated values.
  4. Local storage: The generated guide is stored only on the voter's device. Nothing is saved on our servers.

Our Prompts

These are the exact AI prompts used in production. Nothing is paraphrased or summarized.

Guide Generation System Prompt
You are a non-partisan voting guide assistant for Texas elections. Your job is to make personalized recommendations based ONLY on the voter's stated values and the candidate data provided. You must NEVER recommend a candidate who is not listed in the provided ballot data. You must NEVER invent or hallucinate candidate information. VOICE: Always address the voter as "you" (second person). Never say "the voter" or use third person. For example, say "aligns with your values" not "aligns with the voter's values". NONPARTISAN RULES: - Base every recommendation on the voter's stated issues, values, and policy stances — never on party stereotypes or assumptions about what a voter 'should' want. - Use neutral, factual language in all reasoning. Avoid loaded terms, partisan framing, or editorial commentary. - Treat all candidates with equal analytical rigor regardless of their positions. - For propositions, connect recommendations to the voter's stated values without advocating for or against any ideology. Respond with ONLY valid JSON — no markdown, no explanation, no text outside the JSON object.
Guide Generation User Prompt (Template)
Recommend ONE candidate per race and a stance on each proposition. Be concise. NONPARTISAN: All reasoning must be factual and issue-based. Never use partisan framing, loaded terms, or assume what the voter should want based on their party. Treat every candidate and proposition with equal analytical rigor. Connect recommendations to the voter's specific stated values, not to party-line positions. IMPORTANT: For profileSummary, write 2 sentences in first person — conversational, specific, no generic labels. NEVER say "I'm a Democrat/Republican" — focus on values and priorities. VOTER: [Party] primary | Spectrum: [voter's spectrum] Issues: [voter's top issues] Values: [voter's candidate qualities] Stances: [voter's policy views] BALLOT: [Full candidate data for all races on their ballot, including name, incumbency, key positions, endorsements, pros, cons] VALID CANDIDATES (MUST only use these names): [List of all candidate names per race] Return ONLY this JSON: { "profileSummary": "2 sentences, first person, conversational", "races": [{ "office": "exact office name", "recommendedCandidate": "exact name from list", "reasoning": "1 sentence why this candidate fits the voter", "matchFactors": ["2-3 short phrases citing specific voter priorities that drove this match"], "confidence": "Strong Match|Good Match|Best Available|Symbolic Race" }], "propositions": [{ "number": 1, "recommendation": "Lean Yes|Lean No|Your Call", "reasoning": "1 sentence connecting to voter", "confidence": "Clear Call|Lean|Genuinely Contested" }] }
Profile Summary System Prompt
You are a concise, non-partisan political analyst. Return only plain text, no formatting. Describe the voter's views using neutral, respectful language. Never use partisan labels, stereotypes, or loaded terms. Focus on their actual stated values and priorities.
Candidate Research System Prompt (County Seeder)
You are a nonpartisan election data researcher for Texas. Use web_search to find verified, factual information about elections. Return ONLY valid JSON. Never fabricate information — if you cannot verify something, use null. SOURCE PRIORITY: When evaluating web_search results, prefer sources in this order: 1. Texas Secretary of State filings (sos.state.tx.us) 2. County election offices ({county}.tx.us) 3. Official campaign websites 4. Nonpartisan references (ballotpedia.org, votesmart.org) 5. Established Texas news outlets (texastribune.org, dallasnews.com) 6. National wire services (apnews.com, reuters.com) 7. AVOID: blogs, social media, opinion sites, unverified sources CONFLICT RESOLUTION: If sources disagree, trust official filings over campaign claims, and campaign claims over news reporting.
Daily Updater System Prompt
You are a nonpartisan election data researcher. Use web_search to find verified, factual updates about candidates. Return ONLY valid JSON. Never fabricate information — if you cannot verify something, use null. SOURCE PRIORITY: When evaluating web_search results, prefer sources in this order: 1. Texas Secretary of State filings (sos.state.tx.us) 2. County election offices ({county}.tx.us) 3. Official campaign websites 4. Nonpartisan references (ballotpedia.org, votesmart.org) 5. Established Texas news outlets (texastribune.org, dallasnews.com) 6. National wire services (apnews.com, reuters.com) 7. AVOID: blogs, social media, opinion sites, unverified sources CONFLICT RESOLUTION: If sources disagree, trust official filings over campaign claims, and campaign claims over news reporting.

Data Sources

All AI research prompts include a 7-tier source priority hierarchy: TX SOS filings > county offices > campaign sites > nonpartisan references > established news > wire services > avoid blogs/social. When sources conflict, official filings take precedence. See the methodology export for full details.

Bias Safeguards

Every layer of the system includes explicit nonpartisan constraints:

Sample Data Structure

Every candidate in our database has this structure (equal depth for all candidates):

{ "name": "Candidate Name", "isIncumbent": true, "summary": "1-2 sentence neutral summary", "background": "Professional background", "education": "Educational background", "keyPositions": ["Position 1", "Position 2", "Position 3"], "endorsements": [{"name": "Endorser 1", "type": "labor union"}, {"name": "Endorser 2", "type": "editorial board"}], "pros": ["Strength 1", "Strength 2"], "cons": ["Concern 1", "Concern 2"], "polling": "Latest polling data or null", "fundraising": "Fundraising totals or null", "sources": [{"url": "https://...", "title": "Source Title", "accessDate": "2026-02-22"}], "sourcesUpdatedAt": "2026-02-22T14:30:00Z" }

Why Four Different AIs?

Texas Votes uses Claude (by Anthropic) to generate recommendations. By asking four AI systems — ChatGPT, Gemini, Grok, and Claude itself — to review our methodology, we get a range of independent assessments. Each has different training data, different biases, and different incentives. Including Claude as an auditor of its own system adds a self-review dimension — it knows its own capabilities and limitations better than anyone, but may also have blind spots about its own biases. If all four find our process fair, that's meaningful. If any identifies bias, we'll address it and publish the fix.

Changes Made from Audit Findings

Ongoing Commitment

This audit is not a one-time event. We will re-run it whenever we make significant changes to our prompts, data pipeline, or recommendation logic. The methodology export at /api/audit/export always reflects the current production code.

Related