SD Upscaler
A diffusion-based model for upscaling images to higher resolution, trained by Katherine Crowson in collaboration with Stability AI.
A diffusion-based model for upscaling images to higher resolution, trained by Katherine Crowson in collaboration with Stability AI. It is capable of upscaling both generated and non-generated images.
https://colab.research.google.com/drive/1o1qYJcFeywzCIdkfKJy7cTpgZTCM2EI4
Polyglot-Ko
A series of Korean autoregressive language models made by the EleutherAI polyglot team. We currently have trained and released 1.3B, 3.8B, and 5.8B parameter models.
Polyglot-Ko is a series of Korean autoregressive language models made by the EleutherAI polyglot team. We currently have trained and released 1.3B, 3.8B, and 5.8B parameter models.
CLIP-Guided Diffusion
A technique for doing text-to-image synthesis cheaply using pretrained CLIP and diffusion models.
A technique for doing text-to-image synthesis cheaply using pretrained CLIP and diffusion models.
https://colab.research.google.com/drive/12a_Wrfi2_gwwAuN3VvMTwVMz9TfqctNj#scrollTo=1YwMUyt9LHG1
Cloob-Conditioned Latent Diffusion
A highly efficient text-to-image model that can be trained without captioned images.
John David Pressman, Katherine Crowson
RWKV
RWKV is an RNN with transformer-level performance at some language modeling tasks. Unlike other RNNs, it can be scaled to tens of billions of parameters efficiently.
RWKV is an RNN with transformer-level performance at some language modeling tasks. Unlike other RNNs, it can be scaled to tens of billions of parameters quite efficiently.
GPT-NeoX-20B
An open source English autoregressive language model trained on the Pile. At the time of its release, it was the largest publicly available language model in the world.
GPT-NeoX-20B is a open source English autoregressive language model trained on the Pile,. At the time of its release, it was the largest publicly available language model in the world.
CARP
A CLIP-like model trained on (text, critique) pairs with the goal of learning the relationships between passages of text and natural language feedback on those passages.
A CLIP-like model trained on (text, critique) pairs with the goal of learning the relationships between passages of text and natural language feedback on those passages.
GPT-J
A six billion parameter open source English autoregressive language model trained on the Pile. At the time of its release it was the largest publicly available GPT-3-style language model in the world.
GPT-J is a six billion parameter open source English autoregressive language model trained on the Pile. At the time of its release it was the largest publicly available GPT-3-style language model in the world.
VQGAN-CLIP
A technique for doing text-to-image synthesis cheaply using pretrained CLIP and VQGAN models.
VQGAN-CLIP is a methodology for using multimodal embedding models such as CLIP to guide text-to-image generative algorithms without additional training. While the results tend to be worse than pretrained text-to-image generative models, they are orders of magnitude cheaper and can often be assembled out of pre-existing independently valuable models. Our core approach has been adopted to a variety of domains including text-to-3D and audio-to-image synthesis, as well as to develop novel synthetic materials.