- Find old datasets (hn.algolia.com "datasets", use huggingface, search arxiv)
- Use weird search engines, e.g. exa.ai (searches based on embeddings vs. google's pagerank/keywords) + google dorks. weird input = weird output
- mix 2 ideas you see - look at showcase channels on discord and pages on random frameworks, things people are building at Buildspace
- Find old facebook groups with lots of grumpy members posting regularly
These are just random strategies I use before I make things I post on twitter (https://twitter.com/joshelgar), but they work pretty well for coming up with fun projects.
1. GPT4 "500 tough professions w/ hardship involved"
2. Multiply that with "500 shifts in <profession> or <industry> regulations> since 2000"
3. Get inspired to come up with - grumpy fisherman mad about mooring fees, dog walkers upset about walking limits, 10x increases in flytipping in blackpool
4. As your ideation improves, train a better gpt.
Frankly all those FB groups (min. names) are probably in the GPT training set, just a case of finding good prompts to get inspiration.
prior generations usually take fewer steps than vanilla SDXL to reach the same quality.
But yeah, the inference speed improvement is mediocre (until I take a look at exactly what computation performed to have more informed opinion on whether it is implementation issue or model issue).
The prompt alignment should be better though. It looks like the model have more parameters to work with text conditioning.
in my observation, it yields amazing perf at higher batch sizes (4 or better 8). i assume it is due to memory bandwith and the constrained latent space helping.
3. I used LLMs to caption a training set of 1 million Minecraft skins, then finetuned Stable Diffusion to generate minecraft skins from a prompt: https://multi.skin
Inspired by Neal's "Infinite Craft" (https://news.ycombinator.com/item?id=39205020), I made a food-based version. You can use the sidebar to create a new food and click or drag the example ones.