If you haven't played with VQGAN / CLIP programs that draw images based on a text you enter, you're missing quite a bit.
I've made a collection of tags I'm using to change CLIP rendering behavior.
There it is with the code to parse tags: https://gist.github.com/6r1d/fd3dca3579cf15c9b3868e61ed0cf00b
For the context: I'm using two CLIP instances, EleutherAI's and Google Colabs by Katherine Crowson.