Skip to content

Use fused softmax kernel in llama attention layer #3584

Use fused softmax kernel in llama attention layer

Use fused softmax kernel in llama attention layer #3584

Re-run triggered October 23, 2024 17:43
Status Success
Total duration 1m 49s
Artifacts

book.yml

on: pull_request
Test candle-book
1m 39s
Test candle-book
Fit to window
Zoom out
Zoom in