SETH AI

UNCENSORED. OPEN-WEIGHT. FRONTIER INTELLIGENCE.

REQUEST ACCESS READ THE PAPER

01 // ABOUT

INTELLIGENCE WITHOUT GATES.

Seth AI is a family of large language models trained for unfiltered reasoning, code synthesis, and scientific research. We do not apply alignment taxes, safety filters, or content restrictions that degrade capability. The result is a model that thinks more clearly, writes more precisely, and solves problems that filtered systems refuse to touch.

Built on a sparse mixture-of-experts architecture with 68 billion total parameters and 37 billion active per forward pass. Released under a permissive open-weight license. You own the weights. You set the boundaries.

68B
PARAMETERS
37B
ACTIVE
128K
CONTEXT WINDOW

02 // WHAT IT DOES

01

UNCENSORED REASONING

No content filters. No refusal training. Seth AI reasons through any problem you present — scientific, technical, creative, or controversial — without manufactured hesitation.

02

CODE SYNTHESIS

Trained on production-grade codebases across 40+ languages. Generates, debugs, and optimizes software from system architecture to assembly.

03

SCIENTIFIC RESEARCH

Literature synthesis, hypothesis generation, statistical analysis, and experimental design. A research assistant that actually understands methodology.

04

LONG-CONTEXT MASTERY

128,000 token context window with full attention. Analyze entire codebases, research papers, or datasets in a single pass without degradation.


03 // MODELS

SETH-671B

SETH-68B

68B PARAMS | MoE | 37B ACTIVE

The flagship model. Sparse mixture-of-experts with 256 routed experts. Best-in-class reasoning, code generation, and scientific problem solving.

SETH-70B

SETH-37B

37B PARAMS | DENSE

Full dense attention model for deployments requiring maximum throughput. No routing overhead — every parameter contributes to every token.

SETH-14B

SETH-14B

14B PARAMS | DENSE | EDGE-OPTIMIZED

Compact and fast. Designed for on-device inference and API deployments where latency matters more than parameter count.

SETH-CODER

SETH-CODER

CODE-SPECIALIZED | 34B PARAMS

Fine-tuned exclusively on software engineering corpora. Surpasses generalist models on SWE-bench, HumanEval, and private production codebases.


04 // RESEARCH

Neural network architecture

TECHNICAL REPORT

SPARSE MIXTURE-OF-EXPERTS AT SCALE

We present the Seth MoE architecture: 68B total parameters with expert-choice routing, load-balanced auxiliary losses, and a novel attention mechanism that scales linearly with context length. The result is a model that outperforms dense models 10x its active size on reasoning benchmarks while maintaining inference efficiency.

Benchmark data visualization

BENCHMARK RESULTS

STATE-OF-THE-ART ON REASONING BENCHMARKS

Seth-68B achieves new highs on MATH, GSM8K, HumanEval, and SWE-bench Verified. Crucially, it does so without alignment degradation — performance holds across unfiltered evaluations that cause other models to refuse or fail.

MATH 92.4%
GSM8K 96.1%
HUMANEVAL 94.5%
SWE-BENCH 48.7%

05 // EARLY ACCESS

GET ACCESS.

Seth AI is rolling out in phases. Join the waitlist for API access, model weights, and research previews. No spam. No marketing. Just access.

400+
WAITLISTED
26
COUNTRIES
< 24H
AVG RESPONSE