For each high-value paper in the SciDEX corpus, reproduce one of its
core findings (a figure, a method, or a key statistical result), then
extend it — different parameter, different cohort, additional control,
sensitivity analysis. The reproduction validates the paper; the
extension produces a new SciDEX-original artifact derived from a known
foundation.
This compounds artifact value in two ways:
> ## Continuous-process anchor
>
> Recurring driver: pick a paper from the prioritized backlog,
> generate a replication task, queue it. Result handling reuses the
> percolation pipeline from
> quest_experiment_execution_participant_spec.md.
>
> Every principle in docs/design/retired_scripts_patterns.md applies.
This quest reuses the execution loop (claim → run → submit → reward)
from quest_experiment_execution_participant_spec.md. The differences
are:
Eligibility predicate:
SELECT p.* FROM papers p
WHERE p.citation_count >= 25
AND p.pmid NOT IN (SELECT paper_pmid FROM replication_attempts WHERE status IN ('claimed','running','completed'))
AND p.metadata->'figures' IS NOT NULL
AND jsonb_array_length(p.metadata->'figures') >= 1
AND p.metadata->>'methods_summary' IS NOT NULL
AND p.metadata->>'data_availability' IN ('public', 'supplementary')
AND p.has_in_silico_method = TRUE -- new computed column
ORDER BY (p.citation_count * COALESCE(p.relevance_to_active_hypotheses, 0.5)) DESC
LIMIT 20;has_in_silico_method is a new column populated by an LLM rubric that
reads each paper's methods section and decides if any analysis step is
in-silico-only (transcriptomics analysis, structural prediction,
network analysis, etc.). Wet-lab-only methods → out of scope.
relevance_to_active_hypotheses is a similarity score between the
paper's abstract embedding and the current hypothesis pool —
prioritizes papers relevant to active SciDEX work.
Recurring driver [Forge] CI: Replication target selector (every-6h, pri 91):
replication_attempts row:CREATE TABLE replication_attempts (
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
paper_pmid TEXT NOT NULL,
target_description TEXT NOT NULL, -- "Figure 2A heatmap"
target_finding TEXT NOT NULL, -- "Microglia in AD show 8-fold IFN-pathway upregulation"
expected_method_summary TEXT, -- from paper's methods
expected_data_sources TEXT[], -- from paper's data availability
status TEXT CHECK (status IN ('proposed','claimed','running','completed','failed','expired')),
claimed_by_actor_id TEXT,
proposed_at TIMESTAMPTZ DEFAULT NOW(),
claimed_at TIMESTAMPTZ,
completed_at TIMESTAMPTZ,
reproduction_artifact_id TEXT REFERENCES artifacts(id),
extension_artifact_id TEXT REFERENCES artifacts(id),
reproduction_match_score REAL, -- 0-1, how close to published result
extension_novelty_score REAL, -- 0-1, how distinct extension is
failure_reason TEXT
);
CREATE INDEX idx_replication_paper ON replication_attempts(paper_pmid);
CREATE INDEX idx_replication_status ON replication_attempts(status);Reuses claim mechanic from
quest_experiment_execution_participant_spec.md. A new actor
agent-replication-executor-001 registered with replication-tailored
capabilities (pubmed_search, osf_data_download, paper_figures skill,
domain-specific tools).
Iterative task created. Agent reads:
Agent then:
parent_artifact_id = paper_artifact_idreproducesReproduction match scoring (LLM judge):
After reproduction, agent proposes 1-2 extensions:
parent_artifact_id = reproduction_artifact_idextendsResult handling reuses
quest_experiment_execution_participant_spec.md Phase C. Specifically:
replication_status fieldquest_experiment_extraction_spec.md)
evidence_against link to any hypothesis the paper supported
claimPlus extension-novelty bonus: extension_novelty_score × 30 tokens.
When reproduction_match_score < 0.3 and the paper has > 100 citations
or supports an active hypothesis:
replication_failure_high_impactevidence_for weighted down if the paper was supporting/replication-crisis-tracker dashboardThis is one of SciDEX's clearest value propositions: a system that incentivizes finding replication failures rather than burying them.
GET /replications/runnable — currently proposed, claimableGET /replications/<id> — full record (paper, target, reproduction,GET /papers/<pmid>/replications — all replication attempts on a paperGET /replication-crisis-tracker — failed replications dashboardPOST /replications/<id>/dispute — reviewer challenges scoringhas_in_silico_method column populated for top 1000 cited papersquest_artifact_uuid_migration_spec.mdquest_artifact_metadata_semantic_spec.mdquest_artifact_reuse_provenance_qc_spec.mdquest_experiment_execution_participant_spec.md (claim/result loop reused)quest_real_data_pipeline_spec.md (data fetch from external sources)quest_analysis_sandboxing_spec.mdquest_experiment_extraction_spec.md (papers table + figures metadata)paper-figures skill (PMID → figure URLs)paper-corpus-search, pubmed-search, paper-lookup skillsSister quest to quest_experiment_execution_participant_spec.md,
reusing claim/result/reward mechanics. Reproduction-then-extend
pattern: every replication produces 2 artifacts (the reproduction +
the extension). Failed replications heavily rewarded (80 tokens,
1.5x first-mover) and auto-enrolled in Skeptic debate. Replication
crisis tracker as a public surface. In-silico-only at v1.
Open question: how to handle papers whose data is non-public despite
"public" availability claim? Document attempt as
failure_reason='data_access', mint 5 tokens for the diagnosis,
move on. Tracking these failures is itself useful (which journals /
labs honor data-availability claims?).
{
"requirements": {
"reasoning": 9,
"analysis": 9,
"coding": 9,
"safety": 8
}
}