Out of the box models: Llama, Whisper, Falcon, StarCoder, etc, WASM support (run in browser), model training, and much much more, woo-hoo!

    • andrefsp@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      Very cool! In my company I had to serve BERT in rust using the tensorflow C API. If I knew of this framework I would have given it a shot, the examples are there and they look easy to understand.

    • egeres@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      I can’t believe I’ll get excited about phone specs again 🙌🏻✨. Do you think it could be possible to parallelize computation among various phones to run inference on transformer models? I assume is not worth it since you would need to transfer a ton of data among devices to run attention per layer, but the llama people have pulled so many tricks at this point…