Overview

This is a personal working taxonomy, a lens I use to orient myself when reading papers, not a formal classification scheme. The categories are fuzzy in practice, and reasonable people will assign the same paper differently. The goal is clarity of thought, not consensus.

The taxonomy uses a superposition model where each paper is viewed as a linear combination of seven fundamental contribution types (basis vectors). Examples throughout draw primarily from chemistry and materials science, though the framework applies across physical sciences.

The framework helps answer: “What is this paper’s primary contribution?” by identifying rhetorical patterns and structural elements that signal different research paradigms.

Core Principle: The Superposition Model

All papers in this domain can be viewed as a superposition of fundamental contribution vectors. This is an analogy, not a formal mathematical claim: the seven types below are not linearly independent in the strict sense, and most papers project onto more than one.

Most papers exhibit a profile across multiple basis vectors, blending contribution types (e.g., Method + Theory). One vector usually provides the primary narrative thrust; secondary vectors supply the supporting evidence.

I tend to classify a paper by identifying its Primary Projection (the dominant contribution) and Secondary Projections (supporting work). When a paper seems to split evenly across two types, I look at which type’s indicators dominate the abstract and section headings, and I treat the secondary projection as context that supports the primary narrative, not a co-equal claim.

The Seven Basis Vectors ($\Psi$)

Basis VectorAlias/FocusCore QuestionPrimary Output
1. $\Psi_{\text{Method}}$The Methodological Basis (Architecture/Algorithm)What new mechanism does this introduce?New algorithm, architecture, or approximation
2. $\Psi_{\text{Theory}}$The Theoretical Basis (Formal Analysis)Why does this work?Formal proof, generalization bound, or physical derivation
3. $\Psi_{\text{Resource}}$The Infrastructure Basis (Data/Software)What resources are available?Dataset, benchmark, or open-source software ecosystem
4. $\Psi_{\text{Systematization}}$The Review Basis (Synthesis)What do we know?Comprehensive survey or new organizing taxonomy (Systematization of Knowledge, SoK)
5. $\Psi_{\text{Position}}$The Sociological Basis (Perspective)Where should the field go?Opinion piece, perspective, or critique of community practice
6. $\Psi_{\text{Discovery}}$The Empirical Discovery BasisWhat new thing did we find?Experimentally or computationally validated material, molecule, or physical phenomenon
7. $\Psi_{\text{Application}}$The Transfer Basis (Domain Application)Does this approach work here?Empirical evaluation of an existing method in a new scientific domain

Assessment Guide: Rhetorical Indicators

These rhetorical patterns tend to signal which vector dominates. They are heuristics, not rules; the goal is to notice what the paper is primarily arguing for, not to tick boxes.

1. $\Psi_{\text{Method}}$: The Methodological Paper

Focuses on proposing a novel mechanism, architecture, or approximation (e.g., a new Transformer variant, a GNN with symmetry, a new DFT functional).

Rhetorical Indicators:

  • Ablation Study: Authors systematically remove components of their system to prove their specific innovation drives the performance gain. In physics venues, this often appears as parameter sensitivity analysis or direct comparison to prior potentials and functionals rather than a classic ablation table.
  • Baseline Comparison: A prominent table comparing the new method against the State-of-the-Art (SOTA)
  • Pseudo-code: An explicit block detailing the algorithmic steps (e.g., for training, sampling, or inference)

Examples in the notes: Flow Matching for Generative Modeling (introduces a new continuous normalizing flow mechanism with ablations), MOFFlow (novel SE(3) flow matching architecture for metal-organic frameworks with SOTA comparisons).

2. $\Psi_{\text{Theory}}$: The Theoretical Paper

Focuses on mathematical guarantees, proofs, or derivations from first principles.

Rhetorical Indicators:

  • Mathematical Proof Sections: Sections titled “Theorem 1,” “Proof of Equivariance,” or “Formal Bounds”
  • Analysis of Limits/Capacity: Investigates the expressivity (e.g., comparing a GNN to the Weisfeiler-Lehman Test) or analyzes the geometry of the optimization landscape
  • Generalization/OOD: Derives generalization bounds on test error or formally defines “chemical space coverage” for out-of-distribution (OOD) behavior
  • Exact Constraints: Derives exact conditions that true physical functions (like the universal Density Functional) must satisfy

Examples in the notes: Funnels, Pathways, and the Energy Landscape (derives a formal theoretical basis for protein folding dynamics from energy landscape theory), The Convexity Principle and Interacting Gases (formal theoretical analysis connecting flow-based generative models to thermodynamic principles).

3. $\Psi_{\text{Resource}}$: The Infrastructure Paper

Focuses on creating and sharing foundational tools for the community.

Rhetorical Indicators:

  • Curation Description: Detailed steps on how data was generated, filtered, or curated (e.g., describing millions of CPU-hours of DFT calculations for a dataset like QM9). In physics contexts, this often takes the form of DFT protocol descriptions and computational settings rather than a formal “datasheet.”
  • “Datasheets” and “Data Cards”: Inclusion of formal documentation detailing provenance, copyright, and potential biases in the data. In physics contexts, this often appears as detailed computational settings, DFT functional choices, and convergence criteria rather than a named datasheet document.
  • Benchmark Definition: Argues that “Metric X on Dataset Y” is the correct proxy for progress in a specific scientific task

Examples in the notes: Molecular Sets (MOSES) (releases a benchmark dataset and evaluation suite for molecular generation), SELFIES 2023 (releases an updated molecular string representation standard with tooling and validation).

4. $\Psi_{\text{Systematization}}$: The Review Paper

Focuses on organizing and synthesizing existing literature.

Rhetorical Indicators:

  • Survey Structure: Follows a linear, often chronological, progression or is grouped by architecture (e.g., Variational Autoencoders (VAEs), GANs, Diffusion models)
  • Systematization of Knowledge (SoK): A higher-order contribution that proposes a new taxonomy or a unified framework to connect disparate concepts
  • Citation Density: References a large fraction of the relevant prior literature; often includes a comparison table spanning many prior works across years or venues

Examples in the notes: Venus Evolution Through Time (synthesizes multi-decade observations into a unified timeline), Embedded Atom Method Review (1993) (organizes and contextualizes the body of work on EAM potentials into a coherent reference).

5. $\Psi_{\text{Position}}$: The Sociological Paper

Focuses on meta-science, arguing for a change in community norms, or critiquing systemic issues.

I see two subtypes come up often. The first is the roadmap/perspective paper: a constructive argument for where the field should focus, often written by senior researchers after a period of reflection. The second is the critique paper: a more adversarial argument that something the community is currently doing is wrong or counterproductive. Both share the same core signal (the argument itself is the contribution), but they have different tones and are often targeted at different audiences.

Rhetorical Indicators:

  • Venue/Track: Often found in “Position Tracks” or called “Blue Sky” or “Forward Looking” papers
  • Argumentative Tone: Uses qualitative or quantitative analysis (meta-analysis) to argue for a shift in how research is conducted or funded (e.g., a paper arguing that AI contracts the focus of science)
  • Argument as Contribution: The paper presents no new experimental findings; the primary contribution is the argument itself

Examples in the notes: Fold Graciously (position paper arguing for a shift in how the protein folding community evaluates progress), SELFIES 2022 (makes a normative argument for adopting SELFIES over SMILES for robust molecular generation; the paper also has Method and Resource characteristics, but the argument-first framing and advocacy tone push it toward Position for me).

6. $\Psi_{\text{Discovery}}$: The Empirical Discovery Paper

Focuses on the discovery of novel scientific artifacts using AI/ML tools. The key criterion is experimental or higher-fidelity confirmation of the AI’s prediction (wet-lab synthesis, physical characterization, or a higher-fidelity simulation), rather than performance on a held-out test set.

Rhetorical Indicators:

  • Structure: Follows a workflow: (1) Computational Screening (AI selects candidates), (2) Validation (wet-lab synthesis, physical characterization, or independent computational confirmation)
  • Core Claim: The primary contribution is a new material, molecule, or physical finding, with the AI/ML part serving as the necessary first step
  • Key Question: Does the AI’s prediction hold true against an independent ground truth (e.g., a physical experiment or a higher-fidelity simulation)?

Examples in the notes: Oxidation-Reduction Oscillations on Pt/SiO2 (1994) (validates a computationally-predicted dynamic phenomenon against physical experiment), Nature of LUCA and the Early Earth System (uses computational inference to recover a validated empirical claim about early life).

7. $\Psi_{\text{Application}}$: The Application Paper

Applies an existing method, architecture, or technique to a new scientific domain or task without introducing a new mechanism, dataset, or experimentally confirmed finding. The primary contribution is demonstrating feasibility or utility of transfer.

Rhetorical Indicators:

  • “We Apply X to Y” framing: The abstract explicitly names an existing method and a new domain; the contribution is the connection, not the method itself.
  • Benchmark-style Results: Performance reported on domain-specific tasks using existing metrics; baselines are often simple (classical methods, prior domain tools) rather than ML SOTA.
  • Absence of novelty claims: The paper does not argue for a new architecture, does not release a new dataset, and does not report a validated experimental discovery.

Examples in the notes: This type is common in the AI+physical sciences literature but underrepresented in my current collection. I’ll add examples as I encounter good ones.

Edge Cases and Disambiguation

These are the pairs I find myself second-guessing most often. I’ve written down how I tend to resolve them, not as rules, but as a record of my reasoning.

Method vs. Discovery: The key question I ask is whether there is independent validation beyond a held-out test set. If the paper reports a wet-lab confirmation or a higher-fidelity simulation verifying the AI’s output, I lean toward Discovery: the finding is real and the AI was the tool. If the architecture is the thing being argued for, and validation is limited to benchmarks, I lean toward Method. One edge case I see often in computational chemistry is ML-predicts + DFT-validates: no wet lab, but DFT is genuinely independent of the ML model. I treat this as Discovery if the DFT result is the claimed contribution, and Method if the ML model’s performance against DFT is what’s being argued for.

Resource vs. Systematization: I think of this as artifact-first vs. interpretation-first. If the paper’s main output is something you can download and use (a benchmark, a curated dataset, a software library), it’s Resource. If the paper’s main output is a new way of thinking about a body of literature (a taxonomy, a conceptual unification), it’s Systematization. The presence of a clear benchmark definition with metrics tends to push toward Resource even when there is substantial review content.

Application vs. Method: The cleanest distinguisher here is whether anything new was designed. If an existing method is taken off the shelf and pointed at a new domain, that’s Application. If the domain’s requirements motivated a modification to the method (even a small one), then there may be a legitimate Method component. Application papers often read as feasibility demonstrations; Method papers read as capability expansions.

When I’m still stuck: I look at three things in roughly this order: the abstract’s final sentence (in my experience, especially in ML venues, authors tend to signal what they most want credit for in that last sentence), section heading vocabulary (presence of “theorem,” “algorithm,” “dataset,” “survey,” or “we find” is usually diagnostic), and venue track (position/perspective tracks, SoK tracks, and empirical tracks each carry strong prior signal). I treat the result as a best guess, not a verdict.

Applications

This framework has been useful to me in a few concrete ways:

  • Organizing literature reviews: Knowing a paper’s vector profile tells me where it belongs in a review’s structure: whether it fits in the “methods” or “findings” or “community context” section, rather than forcing everything into chronological order.
  • Understanding conference and journal acceptance criteria: Different venues weight vectors differently. Physics journals favor Discovery; ML venues favor Method and Theory; interdisciplinary venues like Nature or Science often weight Discovery heavily but also publish high-impact Position papers. Knowing this helps calibrate expectations.
  • Identifying gaps in research portfolios: A collection that is heavy on Method but light on Resource or Discovery points to work that may lack empirical grounding or shared infrastructure. The vector breakdown makes these gaps visible.
  • Recognizing different types of scientific contribution: Not everything that looks like a methods paper is one, and not everything that looks like a survey is one. The rhetorical patterns here help me notice when my first impression was wrong and reconsider what is actually being claimed.