Hey
dazedanon good news here
Just recently was released
You must be registered to see the links
that work pretty much as Chat GPT with no exception, is a 4bit quantized model but it will reflect also the same performance as the original meaning that you can host it locally, now.. to host a 65b model even in 4bit you still need a large sum of vram (around 45gb +) but you can rent a server with these specs in a much cheaper way than to use the official api (eg Runpod will ask you 0.79$ /hr for 48gb), ofc if you not run it only for test.
Said that, this model is still censored so will need the usual tricks but it shouldn't be long anyway to get a full uncensored.