shin wee chuang 庄欣伟’s Post

View profile for shin wee chuang 庄欣伟, graphic

Founder/CEO at Pand.ai

Gemini 1.5 Pro now comes with 2M context tokens. For context, GPT4 has 128K. What it means: 1️⃣ You can now put more data into your context window thereby increasing your output accuracy, all else being equal 2️⃣ At a text input price of USD0.70/1M tokens, this means a single prompt on Gemini 1.5 Pro could set you back by up to USD1.40, not including any output response Is there a way to optimise accuracy and cost? One way is to use context caching offered by Gemini (but not available on other LLM, such as GPT4o), which may reduce the price to pay under certain conditions (not all, yeah, it’s a bit complicated). Another way is to use RAG, which is applicable to all LLM, but its implementation is highly skill-dependent. Finally, and probably the easiest way, is to engage a professional and watch them do the magic, because a GenAI solution, when well-designed and well-developed, works like magic. 

Gemini 1.5 Pro 2M context window, code execution capabilities, and Gemma 2 are available today- Google Developers Blog

Gemini 1.5 Pro 2M context window, code execution capabilities, and Gemma 2 are available today- Google Developers Blog

developers.googleblog.com

To view or add a comment, sign in

Explore topics