Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Here is their Chat Playground for PaLM 2 https://console.cloud.google.com/vertex-ai/generative/langua... (you have to be logged in to Google Cloud Console I think)

Anyone know what parameters are best for code generation? I tried something simple for Node.js and it wasn't horrible but not working. Maybe I used the wron parameters. I tried using 0 for the temperature and turning everything else down like I do with the OpenAI API.



I get this: „ERROR. Quota exceeded for aiplatform.googleapis.com/online_prediction_requests_per_base_model with base model: chat-bison. Please submit a quota increase request.“

Has anyone gotten this fixed?



Whoah it gives some seriously wrong answers to coding questions, to the point that they are dangerous!

I think we have to wait for the explanation is chat-bison PaLM 1 or 2


Isn't Chat-Bison-001 Palm 1?

Edit: It seems I can't use my free credits on Vertex APIs... Not nice.


Bison is apparently the second largest PaLM 2 model:

> Even as PaLM 2 is more capable, it’s also faster and more efficient than previous models — and it comes in a variety of sizes, which makes it easy to deploy for a wide range of use cases. We’ll be making PaLM 2 available in four sizes from smallest to largest: Gecko, Otter, Bison and Unicorn. Gecko is so lightweight that it can work on mobile devices and is fast enough for great interactive applications on-device, even when offline.

https://blog.google/technology/ai/google-palm-2-ai-large-lan...


Did anyone see what unicorn is capable of? Why is it not publicised? Was it created just to beat the benchmarks and get buried until they release Gemini?


It is definitely PaLM 2:

Versions Resource ID Release date Release stage Description text-bison@001 2023-05-10 Public Preview Quality improvements and restage -001 as the first stable base model release

https://console.cloud.google.com/vertex-ai/publishers/google...


I don't think so because the CEO mentioned Bison as one of the PaLM 2 models in the Keynote. If I remember correctly.


But would be interested to know if that was not the case. They seemed to be saying that PaLM 2 was rolling out. Also the pages say its a preview. So why would they be previewing the old model still?


https://cloud.google.com/blog/products/ai-machine-learning/g...

> Generative AI Studio, Model Garden, and PaLM 2 for Text and Chat are moving from trusted tester availability to preview, meaning everyone with a Google Cloud account has access.

> Codey, Imagen, Embeddings API for images, and RLHF are available in Vertex AI through our trusted tester program, and Chirp, PaLM 2, Embeddings API, and Generative AI Studio for text are available in preview in Vertex AI to everyone with a Google Cloud account.

It seems like you are right and general PaLM 2 is available. Fine-tuned code-generation model (Codey) is not publicly available yet.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: