@Deliverator@kbin.social
@Deliverator@kbin.social avatar

Deliverator

@Deliverator@kbin.social

Tales from the Terrordrome

Este perfil es de un servidor federado y podría estar incompleto. Explorar más contenido en la instancia original.

devious,

Well longer term allowing the community to focus all it’s hate on the CEO, rather than the platform (and it’s direction e.g. the API and gold changes) allows for an easy PR fix in future - when spez is replaced all the bad press goes away, but all the changes remain intact and forgotten.

Retentive Network: A Successor to Transformer for Large Language Models (arxiv.org) en

This is an exciting new paper that replaces attention in the Transformer architecture with a set of decomposable matrix operations that retain the modeling capacity of Transformer models, while allowing parallel training and efficient RNN-like inference without the use of attention (it doesn't use a softmax)....

  • Todo
  • Suscrito
  • Moderado
  • Favoritos
  • random
  • noticiascr
  • CostaRica
  • Todos las revistas