Together AI

Together AI

Software Development

San Francisco, California 28,131 followers

The future of AI is open-source. Let's build together.

About us

Together AI is a research-driven artificial intelligence company. We contribute leading open-source research, models, and datasets to advance the frontier of AI. Our decentralized cloud services empower developers and researchers at organizations of all sizes to train, fine-tune, and deploy generative AI models. We believe open and transparent AI systems will drive innovation and create the best outcomes for society.

Website
https://together.ai
Industry
Software Development
Company size
51-200 employees
Headquarters
San Francisco, California
Type
Privately Held
Founded
2022
Specialties
Artificial Intelligence, Cloud Computing, LLM, Open Source, and Decentralized Computing

Locations

  • Primary

    251 Rhode Island St

    Suite 205

    San Francisco, California 94103, US

    Get directions

Employees at Together AI

Updates

  • View organization page for Together AI, graphic

    28,131 followers

    Today marks an inflection point for open-source AI with the launch of AI at Meta Llama 3.1 405B, the largest openly available foundation model, that rivals the best closed-source models in AI, rapidly accelerating the adoption of open-source AI with developers and enterprises. We are excited to partner with Meta to bring all the Llama 3.1 models (8B, 70B, 405B, and LlamaGuard) to Together Inference and Together Fine-tuning. Together Inference delivers horizontal scalability with industry-leading performance of up to 80 tokens per second for Llama 3.1 405B and up to 400 tokens per second for Llama 3.1 8B, which is 1.9x to 4.5x faster than vLLM while maintaining full accuracy with Meta’s reference implementation across all models. Together Turbo endpoints are available at $0.18 for 8B and $0.88 for 70B, 17x lower cost than GPT-4o. This empowers developers and enterprises to build Generative AI applications at production scale in their chosen environment – Together Cloud (serverless or dedicated endpoints) or on private clouds. As the launch partner for the Llama 3.1 models, we're thrilled for customers to leverage the best performance, accuracy, and cost for their Generative AI workloads on the Together Platform while allowing them to keep ownership of their models and their data secure. Function calling is supported natively by each of the models, and JSON mode is available for the 8B and 70B models (coming soon for the 405B model). Together Turbo endpoints empower businesses to prioritize performance, quality, and price without compromise. It provides the most accurate quantization available for Llama-3.1 models, closely matching full-precision FP16 models. These advancements make Together Inference the fastest engine for NVIDIA GPUs and the most cost-effective solution for building with Llama 3.1 at scale. https://lnkd.in/gFwBNQhJ

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
      +1
  • View organization page for Together AI, graphic

    28,131 followers

    🎉 Congratulations to AI at Meta on the incredible adoption of Llama! At Together AI, we believe that the future of GenAI relies on open research, and fostering trust among researchers, developers, and enterprises. As the launch partner for Llama 3.1, we’re proud to support over 130,000 developers, and enterprises like Zomato, The Washington Post, and DuckDuckGo running in production – with the best combination of speed, accuracy and price. 

    View organization page for AI at Meta, graphic

    858,978 followers

    Open source AI is the way forward and today we're sharing a snapshot of how that's going with the adoption and use of Llama models. Read the full update here ➡️ https://go.fb.me/mfc5ki 🦙 Highlights • Llama is approaching 350M downloads on Hugging Face. • Our largest cloud service providers have seen Llama token usage more than double since May. • Llama models are being adopted across the industry with great examples from Accenture, AT&T, DoorDash, Goldman Sachs, Infosys, KPMG, Niantic, Inc., Nomura, Shopify, Spotify and Zoom as just a handful of strong examples. Open source AI is how we ensure that the benefits of AI extend to everyone, and Llama is leading the way.

    • No alternative text description for this image
  • Together AI reposted this

    View organization page for Lux Capital, graphic

    31,588 followers

    We're back! Lux Capital is running a Bio x ML Hackathon with EvolutionaryScale and Enveda Biosciences Bio from Oct 10-20! By bringing together the world’s top minds, we’re driving forward the next frontier of science with AI-driven tools and scientific imagination. Build on top of the latest foundation models like ESM3 (98B parameters, GPU access) and proprietary datasets to predict protein activity, model across therapeutic modalities, and develop next-gen biology applications. Model APIs, datasets & compute powered by Amazon Web Services (AWS), OpenAI, NVIDIA, DigitalOcean, EvolutionaryScale, Modal, LatchBio, Together AI, Enveda Biosciences, and RunPod. Open globally. Get started here: hackathon.bio

    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,131 followers

    Introducing the Together Rerank API: a new serverless endpoint for enhancing enterprise search and RAG systems. As the exclusive launch partner for Salesforce AI Research's LlamaRank, we're bringing a powerful reranker model to developers everywhere. This partnership combines Salesforce's AI expertise and enterprise search knowledge with Together's production-ready infrastructure, featuring fast inference and developer-friendly APIs. Key features of the Together Rerank API with LlamaRank: ▶ 8K context length for documents ▶️ Handles semi-structured data (JSON, email, tables, code) ▶️ Improves search relevance and reduces LLM compute costs in RAG systems ▶️ Compatible with Cohere's Rerank API for easy migration Developers can start using the Together Rerank API with LlamaRank today. Read more in our blog post: https://lnkd.in/g2Me2x7E We can’t wait to see how developers will enhance their search and RAG applications with this new API.

    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,131 followers

    Congratulations to our partners at AI21 Labs on the release of their Jamba 1.5 models – powerful long context models for the enterprise. This language model family introduces: ▶️ 256K context window - longest among open models  ▶️ Up to 2.5X performance boost on long contexts  ▶️ Benchmark scores surpassing larger proprietary models  ▶️ Developer features: structured JSON output, function calling, citation mode  ▶️ Multilingual capabilities across 9 languages We’re proud to announce the upcoming addition of Jamba 1.5 to our platform, furthering our commitment to providing developers with access to state-of-the-art, open source AI models. 👏 Incredible work from AI21 Labs Jamba 1.5 – Coming Soon on Together.ai

  • View organization page for Together AI, graphic

    28,131 followers

    Announcing Together Analytics in beta! You can now view your usage (requests, latency, and TPM) over time. This was a highly requested feature from our customers to be able to see their usage in a more granular level. You can now see usage over time and see successful requests, errors, latency, and TPM for all your usage on Together in one place. We're also already working to make it even better! Available today in beta for all customers, directly in your Together dashboard.

  • Together AI reposted this

    View profile for Varun S., graphic

    🇫🇮🇮🇳 | CPTO at Daily | Built and ran @callstatsio | Developed multimedia networks | WebRTC | DASH | Advisor | Investor

    We made it simpler to run voice ai and partnered with Anthropic Cartesia Deepgram Together AI for this launch. Our goal is to combine the best tools, best developer ergonomics, and best infrastructure for real-time AI into a single platform. Daily will host and launch the bots for you, these bots are connected to STT, LLM, and TTS. We have several options under each category, plus you can bring your own API Keys. In addition, we offer function calling to third-party, or to your own infra (using webhooks and post requests). Further, since this is built on RTVI, you as a developer have the ability to augment pr change the config settings and services on the fly, mid-conversation. On the SDK front daily bots comes with a client component and a thin server component. Clients are an RTVI client app and we have web, Android, and iOS SDKs and sample apps. The server side is basically an init, the first request from the client initiates /bot/start. The server is also helpful in the case of telephony, wherein there isn’t a rtvi client per-say and you need a good default config, service and prompt. In an upcoming update you maybe be able to bind dtmf tones to external function calls, build a dynamic IVR for example.

    View organization page for Daily, graphic

    2,934 followers

    Today we’re launching Daily Bots, the ultra low latency Open Source cloud for voice, vision, and video AI. Build voice and video AI bots, with any LLM, at conversational latencies as low as 500ms. Developers can: ✳️ build with Open Source SDKs ✳️ mix and match the best Generative AI models for specific use cases ✳️ run at scale on Daily’s real-time global infrastructure We've partnered with Anthropic Cartesia Deepgram Together AI for this launch. Our goal is to combine the best tools, best developer ergonomics, and best infrastructure for real-time AI into a single platform. Daily Bots is the culmination of the last 18 months of work we've done with customers and partners. The two fastest growing Open Source real-time AI projects came out of this work: Pipecat and RTVI. *️⃣ Daily Bots apps are built using the RTVI Open Source SDKs for the Web, iOS, and Android. Your Daily Bots code will run anywhere that supports the RTVI standard. (Or you can run your own infrastructure.) *️⃣ Your Daily Bots can also answer the phone. (You can buy a phone number from us with a single curl command.) *️⃣ Bring your own API Keys and use any inference provider that supports OpenAI-compatible APIs. Or run your own models and connect your bots to your infrastructure. For a highly non-serious take on the kinds of things we’ve been creating for ourselves, as we’ve worked on Daily Bots, check out a fun demo video 👇 and more links in thread. We'd love to hear what multi-modal, real-time AI directions are most interesting to you, and excited to support developers building! #conversationalAI #voicetovoice #ai #claude #llama #generativeai

  • View organization page for Together AI, graphic

    28,131 followers

    Developers now can build voice-to-voice AI with Llama models! We’re excited to partner with Daily to bring the power of Llama models to ultra low latency voice and video agents.

    View organization page for Daily, graphic

    2,934 followers

    Today we’re launching Daily Bots, the ultra low latency Open Source cloud for voice, vision, and video AI. Build voice and video AI bots, with any LLM, at conversational latencies as low as 500ms. Developers can: ✳️ build with Open Source SDKs ✳️ mix and match the best Generative AI models for specific use cases ✳️ run at scale on Daily’s real-time global infrastructure We've partnered with Anthropic Cartesia Deepgram Together AI for this launch. Our goal is to combine the best tools, best developer ergonomics, and best infrastructure for real-time AI into a single platform. Daily Bots is the culmination of the last 18 months of work we've done with customers and partners. The two fastest growing Open Source real-time AI projects came out of this work: Pipecat and RTVI. *️⃣ Daily Bots apps are built using the RTVI Open Source SDKs for the Web, iOS, and Android. Your Daily Bots code will run anywhere that supports the RTVI standard. (Or you can run your own infrastructure.) *️⃣ Your Daily Bots can also answer the phone. (You can buy a phone number from us with a single curl command.) *️⃣ Bring your own API Keys and use any inference provider that supports OpenAI-compatible APIs. Or run your own models and connect your bots to your infrastructure. For a highly non-serious take on the kinds of things we’ve been creating for ourselves, as we’ve worked on Daily Bots, check out a fun demo video 👇 and more links in thread. We'd love to hear what multi-modal, real-time AI directions are most interesting to you, and excited to support developers building! #conversationalAI #voicetovoice #ai #claude #llama #generativeai

  • View organization page for Together AI, graphic

    28,131 followers

    Our guide: Optimizing GPU Clusters for Generative AI Model Training was featured on blockchain.news. Click to read a summary of insights into our comprehensive acceptance testing processes for new GPU clusters: https://lnkd.in/ek8t4amF Or check out our detailed process (with links to github repos and code snippets) in the comment below ↓

    Optimizing GPU Clusters for Generative AI Model Training: A Comprehensive Guide

    Optimizing GPU Clusters for Generative AI Model Training: A Comprehensive Guide

    blockchain.news

Similar pages

Funding