How we rank, in full detail.

We don't take money from anyone we review. No affiliate links, no sponsored slots, no "featured vendor" tier you can buy into. Just two rubrics — one for vendors, one for apps — both published in full below. Read them, challenge them, help us sharpen them.

Methodology versionv5.0
Last updatedMay 2026
Labs we test atJanoshik · Auxilium
Review cadenceQuarterly

Overview

Pepsider runs two distinct ranking systems: one for peptide vendors (we have different lists for the US-based and European shops) and one for peptide apps (the software people use to dose and track them). The two are scored separately in different ways - a vendor's most important attribute is verified product purity and safety, while an app's is UX, data handling and privacy.

Each criterion in both methodologies is scored on a 1-5 scale with written anchors. The anchors describe what evidence justifies each score, which is what stops scoring from drifting between reviewers and across rounds. Two graders looking at the same vendor should reach the same number.

Why anchored scoring

Most ranking sites use vague 1-10 scales where "8" means "we liked it." That's why their lists shuffle every quarter for no defensible reason. Anchored 1-5 scales force every score to be tied to specific, verifiable evidence - and let any reader audit the result.

Methodology · Part 1 of 2

How we rank peptide vendors.

Five weighted criteria, applied identically to every vendor on every ranking page. Testing weight is the highest because variance in product purity is the area with the largest real-world consequence. Everything else flows from that.

Weights overview

Criterion
Testing Standards
35%
Trust
25%
Shipping & Reship Policy
15%
Value
15%
Support & Transparency
10%

The same five criteria and weights apply to both the US and European vendor lists. Region-specific factors (customs seizure rates, reship turnaround) are evaluated against the regional baseline rather than against each other—a European vendor is scored on Shipping against European delivery expectations, not against US ones.

01

Testing Standards

35%

The single most important factor: independent rigor of how a vendor's products are tested. Independent third-party labs (Janoshik, Auxilium and equivalents) score highest. In-house testing - where the vendor tests their own product and publishes the result - scores far lower because the incentive to misreport is structural.

Within third-party testing we look at methods used (HPLC for purity is the floor, mass spec for identity confirmation raises the score, endotoxin and bacterial contamination testing raises it further), test coverage (per-batch testing beats occasional spot-checks), CoA freshness(the certificate must match the current batch - CoAs years old don't count), and pre-purchase availability (CoAs accessible before a buyer commits, not only on request after).

Sub-factors within Testing Standards are not equal-weighted—independence of testing and CoA freshness dominate; method breadth (mass spec, endotoxin) and pre-purchase availability are secondary modifiers within the anchor band.

Independent vs in-houseHPLC purityMass spec identityEndotoxin testingPer-batch coverageCoA freshnessPre-purchase access
Scoring anchors1-5 scale
1/5
No third-party testing of any kind. CoAs unavailable, not published, or fabricated. Vendor relies on supplier claims with no verification.
2/5
In-house testing only, or occasional third-party spot-checks. CoAs sometimes available on request. Methods undisclosed; coverage inconsistent.
3/5
Some independent third-party testing, typically HPLC purity only. CoAs available on request but often dated (older than 6 months). No identity confirmation or contamination testing.
4/5
Regular Janoshik/Auxilium-tier testing with HPLC purity + mass spec identity. CoAs batch-keyed, dated, and available pre-purchase on the product page.
5/5
Per-batch independent third-party testing including endotoxin/bacterial contamination. Fresh CoAs for the current batch. Published proactively per product, not on request.
02

Trust

25%

A weighted composite of community sentiment, calibrated to source quality. Not all reviews are equal - and some platforms are systematically gamed.

Highest-weighted sources are independent blind-testing programs and long-running reviewers with established track records. Below that, vetted Discord communities with active mod oversight earn medium-high weight. Reddit sits at medium weight, applied with skepticism - sudden trending with coordinated positive threads is a documented astroturfing pattern. Trustpilot sits at the bottom: vendor-solicited reviews dominate the platform and the signal is unreliable.

Track record matters on top of community signal: depth of scrutiny and independently validated consistency beat coordinated promotional bursts, because hype cycles in this market routinely collapse when scrutiny arrives.

Sub-factor weighting is explicit: corroborated blind-testing and deep reviewer dossiers move the anchored band first; vetted Discord and skeptical Reddit scraping fill corroborating detail; scraped Trust aggregates only tie-break when everything else agrees—they never rehabilitate weak evidence alone.

Independent reviewersDiscord communitiesReddit (skeptical)Trustpilot (low-weight)Astroturfing detectionTrack recordLongevity
Scoring anchors1-5 scale
1/5
Documented history of quality or customer-service issues, or community signal dominated by vendor-solicited Trustpilot reviews with no independent verification.
2/5
Unproven or thin reputation, or mixed community signals with possible coordinated-promotion patterns on Reddit. No coverage from independent testers.
3/5
Moderate organic Reddit/Discord presence; no major incidents but limited independent verification.
4/5
Established positive reception with coverage from at least one independent tester. Vetted Discord communities recommend the vendor. No documented major incidents on the public record.
5/5
Widely validated trust. Recommended by independent blind-testing programs and established third-party reviewers. Multiple verifiable third-party validations.
03

Shipping & Reship Policy

15%

Logistics outcomes, scored separately from communication quality. The four sub-factors are delivery success rate (orders that arrive intact and on time across multiple regions), stealth quality (packaging that reduces customs scrutiny without misrepresenting contents), customs handling (rate of seizures across tested regions), and most critically the reship/refund policy on seizures and lost packages—which separates vendors who back their orders from vendors who offload losses onto buyers.

Customs/delivery outcomes and documented reship protections are graded as a matched pair anchoring shipping scores: heroic policies cannot paper over chronic seizures, and clean delivery stats cannot compensate for refusing documented reships—the rubric insists on credible performance across both fronts for any 3+/5 anchor.

Delivery success rateStealth qualityCustoms handlingReship policyRefund on seizure
Scoring anchors1-5 scale
1/5
No reship policy. Frequent customs seizures observed in tested regions. Poor or no stealth packaging.
2/5
Reship handled case-by-case at vendor discretion. Mediocre packaging. Occasional customs issues.
3/5
Meets both requirements: a standard written reship policy on documented seizures and reasonably reliable delivery with adequate stealth in tested regions. Strong on policy but weak on delivery/stealth—or the inverse—drops to 2/5.
4/5
Meets both requirements: a clear reship plus partial-refund stance on documented losses or seizures, and strong stealth packaging with a high documented delivery-success pattern across regions. Uneven split between logistics and policy still caps out at 2/5.
5/5
Meets both requirements: guaranteed reship/full refund on seizures or losses, and exceptional stealth plus near-100% intact delivery across our benchmark regions. If either leg fails or one massively lags the other, revert to 2/5.
04

Value

15%

Price evaluated relative to verified purity, never raw cost. Suspiciously cheap vendors lose points rather than gain them - in this market, lowest price strongly anti-correlates with synthesis quality, and a 99% purity product at the bottom of the price range cannot be honestly delivered.

This criterion also captures pricing transparency (clear unit pricing, no hidden fees), volume discounts (predictable bulk pricing tiers), and payment options (crypto, cards, and direct bank transfer supported).

Price relative to verified purity drives placement first—payment breadth and discount transparency tighten or loosen the verdict inside the anchored band rather than overwriting it outright.

Price per verified mgSuspiciously low = penalizedPricing transparencyVolume discountsCrypto + cards
Scoring anchors1-5 scale
1/5
Price suspiciously low (red flag for synthesis quality) or unjustifiably high relative to verified purity. Opaque pricing.
2/5
Below-tier value for verified purity. Limited payment options. No volume discounts.
3/5
Average value within tier. Transparent pricing. Standard payment methods (cards or a single crypto option).
4/5
Strong value relative to verified purity. Clear volume discounts. Multiple payment options including crypto.
5/5
Excellent value relative to verified purity. Full pricing transparency including bulk thresholds. Crypto + cards accepted.
05

Support & Transparency

10%

Customer-service responsiveness, communication quality during issues, and general transparency about sourcing, manufacturing, and policies. Weighted lower than the others because it's the most subjective and the easiest area to perform well in for the wrong reasons (a vendor with great support and bad product is still a bad vendor).

Sub-factor emphasis: truthful sourcing and manufacturing disclosure plus durable issue resolution outweigh cheerfully fast replies—policy clarity and follow-through move the anchor more than tone alone.

Response timeIssue resolution qualitySourcing disclosureManufacturing standardsPolicy clarity
Scoring anchors1-5 scale
1/5
Unresponsive support. Opaque about sourcing, manufacturing, or policies. Defensive when questioned.
2/5
Slow replies (>72h). Limited transparency. Some sourcing information available but inconsistent.
3/5
Standard responsiveness (24-48h). Some sourcing information published. Reasonable issue resolution.
4/5
Fast replies (<24h). Good issue resolution. Sourcing and policies disclosed publicly.
5/5
Sub-12h replies. Proactive issue handling. Full transparency on synthesis origin, manufacturing standards, and all policies.
Methodology · Part 2 of 2

How we rank peptide apps.

Peptide apps live in a different failure space than vendors. The biggest risks are wrong dosing math, missed schedules, and sensitive health data leaking into ad networks. Functionality and privacy carry the most weight; payment dark-pattern checks and safety-content surfacing are each capped at five percent so they inform the score without overpowering core accuracy and data handling.

Weights overview

Criterion
Functionality & Accuracy
30%
Privacy & Data Handling
25%
Reliability
20%
UX & Platform Coverage
15%
Pricing & Code Transparency
5%
Safety & Harm-Reduction Content
5%
01

Functionality & Accuracy

30%

The core question: does the app correctly do what a peptide user actually needs? That means accurate reconstitution math, correct dosing volume calculations, reliable schedule logic (every-other-day, fasted timing, cycling), and meaningful stack support. Conflict handling is split into three buckets: scheduling conflicts (doses too close together, fasted-window violations), pharmacological interactions between stacked peptides, and interactions with user-entered prescription medications. Separate from conflict handling, we evaluate unit conversion: correct conversion between mcg, mg, IU, and insulin-syringe units.

Peptide dosing crosses these units constantly, and unit errors are the most common source of dangerous miscalculation.

An app that gets the math wrong is dangerous regardless of how nice it looks. We verify every app against a fixed battery of 15–20 reference reconstitution and dosing scenarios, covering common peptides (BPC-157, semaglutide, GH secretagogues), awkward edge cases (1mg vial in 5mL water targeting 250mcg), and unit-conversion traps. The full battery and per-app results are published alongside each review.

Reconstitution mathDosing volumeSchedule logicStack supportCycle planningScheduling conflictsStacked peptide interactionsRx medication interactionsUnit handling
Scoring anchors1-5 scale
1/5
No peptide-aware calculator, or calculator present but with errors frequent enough to make the tool unusable. Fundamental unit-conversion failures.
2/5
Reconstitution calculator present and correct in routine cases, but errors observed at edge cases in the reference battery. No stack support. Single-peptide focus. Weak or missing conflict and unit handling.
3/5
Reliable reconstitution math and solid unit handling in routine cases. Basic schedule logic. Limited stack support; interaction coverage thin.
4/5
Accurate dosing math verified against our reference scenario battery. Multi-peptide stacks. Scheduling conflict checks, peptide-peptide overlap warnings, or Rx interaction prompts where relevant.
5/5
Class-leading dose calculations, full stack and protocol modeling, and thorough handling of scheduling conflicts, stacked-peptide interactions, and user-entered prescription overlaps. Cycle planning tools.
02

Privacy & Data Handling

25%

Peptide use is sensitive health data. The privacy axis covers where data lives (local-first vs cloud-only), encryption (E2E vs in-transit only vs none), telemetry (whether the app sends usage data - and what data - to its developer or third parties), third-party SDKs (analytics, ad networks, crash reporters), account requirements (whether you can use the app at all without handing over identifying information), app-level security (biometric lock, session timeout, 2FA on cloud accounts where applicable), and server jurisdiction (where synced data physically resides and which legal regime governs it).

Open-source code isn't required for a high privacy score, but it provides a meaningful uplift here because it lets independent reviewers verify data-handling claims rather than taking the developer's word.

We audit data flows during the testing window using network inspection, not just developer claims.

Local-first vs cloudE2E encryptionTelemetry auditThird-party SDKsAccount requirementsApp-level securityServer jurisdictionAuditability / open source
Scoring anchors1-5 scale
1/5
Cloud-only storage, no encryption disclosure, third-party trackers detected during audit.
2/5
Cloud-only with basic encryption claim. Analytics SDKs present. Mandatory account. Weak session or device controls.
3/5
Cloud sync over TLS, minimal third-party data sharing, mandatory account creation. Jurisdiction or security posture only partially documented.
4/5
Local-first by default. Optional E2E-encrypted sync. No third-party analytics or ad SDKs on health data. App-level lock available.
5/5
All of 4/5, plus at least one of: open-source code, a self-hosting option, or fully account-optional usage. At least one strong auditability or sovereignty property on top of the privacy baseline.
03

Reliability

20%

Does the app actually work over time? Notification delivery rate in the test window, data persistence (no logged doses lost), sync stability (no duplicates or conflicts), offline handling (the app keeps working when the network doesn't), and data portability: CSV and PDF export, import from competitor apps, and the ability to generate a clean log to share with a clinician. Lock-in is a reliability risk—if the app shuts down, users need their history.

We test for at least 30 days per app, with a fixed protocol of scheduled doses across daypart boundaries, time zone changes, and intentional offline periods.

Notification deliveryData persistenceSync stabilityOffline handlingBackup & recoveryData portability
Scoring anchors1-5 scale
1/5
Notifications miss frequently. Data loss observed during 30-day test.
2/5
Inconsistent notification delivery. Occasional sync failures or duplicates. Weak exports or no import path from other apps.
3/5
Generally reliable. Isolated missed notifications. No data loss. Basic export (e.g. CSV/PDF), imperfect but usable portability.
4/5
99%+ notification delivery measured across at least 200 scheduled notifications spanning 30 days, time zone changes, and intentional offline periods. Stable sync. No data loss in extended testing. Practical exports and clinician-friendly logs.
5/5
100% notification delivery measured across at least 500 scheduled notifications spanning 60 days, three or more time zones, and intentional offline periods. Robust offline handling. Automatic backups and fast sync recovery. Strong portability (export, competitor import, shareable dosing history).
04

UX & Platform Coverage

15%

Design quality, accessibility (VoiceOver, TalkBack, dynamic type, sufficient contrast), onboarding clarity, and platform availability.Single-platform apps are not automatically penalized, but cross-platform parity contributes to the score because most users have a phone, a partner's phone, and a tablet or web context. A polished single-platform indie app can still score 3/5 here; it cannot reach 5/5.

Design qualityAccessibilityOnboardingiOS / Android / WebCross-platform parity
Scoring anchors1-5 scale
1/5
Poor usability or accessibility on the available platform — confusing dosing flows, brittle scheduling UI, minimal VoiceOver/TalkBack support — even if the app is more polished on a different platform.
2/5
One polished platform, or two platforms with large feature gaps or rough parity. Functional but uneven accessibility.
3/5
Strong single-platform experience (typical indie scope) or adequate cross-platform coverage with visible gaps.
4/5
iOS + Android + Web, with most core dosing features mirrored. Polished UI. Good accessibility. Smooth onboarding.
5/5
Major platforms at feature parity, exceptional design, full accessibility coverage (VoiceOver, TalkBack, dynamic type, contrast). Guided onboarding. Cross-device continuity that matches how people actually dose.
05

Pricing & Code Transparency

5%

Pricing model clarity, free-tier honesty, and absence of dark patterns (forced trials, hidden auto-renewal, deceptive cancellation flows). This category captures whether the business model is honest—separate from whether the code is auditable (which we score under Privacy).

Pricing clarityFree tier honestyNo dark patternsRenewal & cancellation UX
Scoring anchors1-5 scale
1/5
Hidden costs or bait-and-switch pricing. Paywalls before core dosing features without disclosure. Trials that silently convert or cancellation flows buried in nested menus.
2/5
Subscription demanded for essentials with vague renewal clauses. Free tier largely a demo. Dark-pattern pressure to upgrade (forced countdowns, fake scarcity).
3/5
Pricing page matches checkout reality. Honest explanation of limits on the free tier. Renewal and cancellation discoverable within the storefront settings.
4/5
Generous free tier or clear fair-use boundaries. Subscription value explained with changelog or roadmap parity. Renewal reminders prominent; cancellation is a straight path.
5/5
Exemplary commercial transparency: no coerced trials, no surprise renewals, refund policy honored as written, and monetization plainly separated from health-data exploitation. Pricing model is boringly honest even when the SKU list is complex.
06

Safety & Harm-Reduction Content

5%

Does the app surface non-obvious safety information at the points users need it? We look for sterile reconstitution technique guidance, warnings about counterfeit and untested supply, prompts to involve a clinician for symptoms or unexpected reactions, and references to third-party lab testing. Peptides occupy a regulatory gray zone in most jurisdictions—the app is often the only safety touchpoint a user encounters. Apps that ignore this, or worse, present themselves as medical advice, score poorly.

Sterile reconstitution guidanceCounterfeit supply warningsClinician escalation promptsThird-party lab testing referencesMedical-advice posture
Scoring anchors1-5 scale
1/5
Silent or dismissive on syringe hygiene, dubious supply sourcing, and symptom escalation. Any presentation of the app as medical advice or prescribing authority caps the score at 1/5 regardless of other factors.
2/5
Boilerplate disclaimers only("not medical advice") with no harm-reduction cues at the moments users assume risk—reconstitution, first administrations, stacking changes.
3/5
Touches safety once (e.g. single generic sterile-water note or static FAQ link) without contextual reminders or counterfeit awareness.
4/5
Context-aware warnings: prompts to escalate unexpected reactions or persistent symptoms to a clinician, plus explicit discussion of counterfeit/untested supply risk and sterile reconstitution basics.
5/5
Integrated harm-reduction at decision points—reconstitution checklists, supply-chain skeptic reminders, clinician handoff tooling, pointers to reputable independent testing—all without drifting into fake medical authority.

Independence policy

Pepsider does not accept payment, samples, or promotional consideration from any vendor or app developer in exchange for review placement or favorable scoring. All test purchases are made with our own funds.

Pepsider is completely independent—we run it as a passion project, not as a revenue play. There are no affiliate links, commissions tied to rankings, paid placements, or commercial agreements with reviewed vendors or app makers.

No editor or contributor may hold financial stakes in any reviewed entity. Contributor disclosures are available on request.

Frequently asked

What's the difference between the US and European vendor lists?

The methodology is identical — same criteria, weights, and anchors. What changes is the regional baseline for Shipping: a US vendor is scored on delivery to US buyers, a European vendor on delivery to European buyers. Testing, Trust, Value, and Support are region-neutral and directly comparable across both lists. Shipping scores are not, by design.

Why is the trust score skeptical of Trustpilot?

Because the platform is systematically gamed. Vendor-solicited reviews dominate the visible signal, and there's no meaningful gating on reviewer authenticity. We weight Trustpilot at the bottom of our trust composite, not zero - but it can never be the deciding factor.

Why do you penalize suspiciously cheap vendors?

In this market, lowest price strongly anti-correlates with synthesis quality. A 99% purity peptide cannot be honestly delivered at the lowest price tier - the math of synthesis cost doesn't allow it. We treat that gap as a quality signal, not a value signal.

Why isn't there a legal/regulatory criterion?

Because peptide legality is a buyer-side question, not a vendor-quality question. It varies by jurisdiction, intended use, and research-vs-human framing — none of which the vendor controls or we can verify. We assume readers have done their own legal due diligence for where they live.

Can a vendor request a retest?

Yes. Vendors can request a retest at any time at their own cost, conducted by a different lab from our independent partner panel. We publish both results.