On the other hand, larger and larger transformer networks are constantly improving.
Assuming you are in the Northern Hemisphere so "last summer" means Julyish, then Google's T5[1] and Microsoft Turing-NLG[2] come to mind.
I find keeping an eye on HuggingFace's models list[3] is useful for this.
[1] https://arxiv.org/abs/1910.10683
[2] https://www.microsoft.com/en-us/research/blog/turing-nlg-a-1...
[3] https://github.com/huggingface/transformers#model-architectu...
On the other hand, larger and larger transformer networks are constantly improving.
Assuming you are in the Northern Hemisphere so "last summer" means Julyish, then Google's T5[1] and Microsoft Turing-NLG[2] come to mind.
I find keeping an eye on HuggingFace's models list[3] is useful for this.
[1] https://arxiv.org/abs/1910.10683
[2] https://www.microsoft.com/en-us/research/blog/turing-nlg-a-1...
[3] https://github.com/huggingface/transformers#model-architectu...