How @cognition /
@windsurf new SWE-1.5 model was built (probably), based on piecing together bits of information that was shared. Base model by: @Zai_org possibly GLM-4.6
RL Training: @nvidia on 'thousands' GB200 NVL72
Inference: @cerebras at 950 tks/sec
SWE-1.5 Model Architecture: Base Model, RL Training, Inference
By
–
Leave a Reply