Together AI

Together AI

Software Development

San Francisco, California 28,421 followers

The future of AI is open-source. Let's build together.

About us

Together AI is a research-driven artificial intelligence company. We contribute leading open-source research, models, and datasets to advance the frontier of AI. Our decentralized cloud services empower developers and researchers at organizations of all sizes to train, fine-tune, and deploy generative AI models. We believe open and transparent AI systems will drive innovation and create the best outcomes for society.

Website
https://together.ai
Industry
Software Development
Company size
51-200 employees
Headquarters
San Francisco, California
Type
Privately Held
Founded
2022
Specialties
Artificial Intelligence, Cloud Computing, LLM, Open Source, and Decentralized Computing

Locations

  • Primary

    251 Rhode Island St

    Suite 205

    San Francisco, California 94103, US

    Get directions

Employees at Together AI

Updates

  • View organization page for Together AI, graphic

    28,421 followers

    🚀 NVIDIA H200 and the Together Kernel Collection (TKC) are coming to Together GPU Clusters to deliver accelerated performance, efficiency, and reliability for your AI training, fine-tuning, and inference workloads. Reserve your H200 cluster today: https://lnkd.in/g6dAtUiq Key features of the Together Kernel Collection (TKC): • Reduces GPU hours, cutting costs and accelerating time-to-market • Up to 24% speedup for frequently used AI training operators • 75% faster inference on FP8 operations • Seamless PyTorch compatibility for easy integration Supercharge your AI workloads with the NVIDIA H200: • 141GB HBM3e memory and 4.8TB/s bandwidth for massive data processing and scalability • Eliminates bottlenecks with long sequences and data-intensive applications • NVLink and Infiniband enable ultra-high-speed communication between GPUs 🔗 Discover how the Together Kernel Collection with NVIDIA H200 and H100 can accelerate your AI infrastructure in our latest blog: https://lnkd.in/gzGkXVSH

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,421 followers

    New guide: How to build a Claude Artifacts clone with Llama 3.1 405B. We'll go over how we built our LlamaCoder example app and got it to 150k visitors. In this tutorial, we'll build the core features of LlamaCoder from scratch including: - Generating an app from a single prompt - Streaming the response using Together AI - Running the code live in the browser with SandPack Read the guide here: https://lnkd.in/eRCmm4Ym

    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,421 followers

    Explore the new TEAL method on our latest blog by James Liu, Pragaash P., Tianle Cai, Han Guo, Yoon Kim, and Ben Athiwaratkun! TEAL, a training-free method that boosts activation sparsity in large language models like Llama-2/3 and Mistral. TEAL achieves up to 50% model-wide activation sparsity, enhancing decoding speeds by effectively managing zero-valued activations. This enables end-to-end speedups of 1.53x-1.8x when combined with a hardware-aware kernel!TEAL capitalizes on the natural clustering of activations around zero throughout the model. By strategically clipping entries close to zero, TEAL reintroduces significant sparsity with minimal impact on model performance. Dive deeper into how TEAL redefines efficiency in model performance, and explore our full findings in the blog here: https://lnkd.in/gd_xeWqP

    TEAL: Training-Free Activation Sparsity in Large Language Models

    TEAL: Training-Free Activation Sparsity in Large Language Models

    together.ai

  • View organization page for Together AI, graphic

    28,421 followers

    🎉 Congratulations to AI at Meta on the incredible adoption of Llama! At Together AI, we believe that the future of GenAI relies on open research, and fostering trust among researchers, developers, and enterprises. As the launch partner for Llama 3.1, we’re proud to support over 130,000 developers, and enterprises like Zomato, The Washington Post, and DuckDuckGo running in production – with the best combination of speed, accuracy and price. 

    View organization page for AI at Meta, graphic

    861,022 followers

    Open source AI is the way forward and today we're sharing a snapshot of how that's going with the adoption and use of Llama models. Read the full update here ➡️ https://go.fb.me/mfc5ki 🦙 Highlights • Llama is approaching 350M downloads on Hugging Face. • Our largest cloud service providers have seen Llama token usage more than double since May. • Llama models are being adopted across the industry with great examples from Accenture, AT&T, DoorDash, Goldman Sachs, Infosys, KPMG, Niantic, Inc., Nomura, Shopify, Spotify and Zoom as just a handful of strong examples. Open source AI is how we ensure that the benefits of AI extend to everyone, and Llama is leading the way.

    • No alternative text description for this image
  • Together AI reposted this

    View organization page for Lux Capital, graphic

    31,726 followers

    We're back! Lux Capital is running a Bio x ML Hackathon with EvolutionaryScale and Enveda Biosciences Bio from Oct 10-20! By bringing together the world’s top minds, we’re driving forward the next frontier of science with AI-driven tools and scientific imagination. Build on top of the latest foundation models like ESM3 (98B parameters, GPU access) and proprietary datasets to predict protein activity, model across therapeutic modalities, and develop next-gen biology applications. Model APIs, datasets & compute powered by Amazon Web Services (AWS), OpenAI, NVIDIA, DigitalOcean, EvolutionaryScale, Modal, LatchBio, Together AI, Enveda Biosciences, and RunPod. Open globally. Get started here: hackathon.bio

    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,421 followers

    Introducing the Together Rerank API: a new serverless endpoint for enhancing enterprise search and RAG systems. As the exclusive launch partner for Salesforce AI Research's LlamaRank, we're bringing a powerful reranker model to developers everywhere. This partnership combines Salesforce's AI expertise and enterprise search knowledge with Together's production-ready infrastructure, featuring fast inference and developer-friendly APIs. Key features of the Together Rerank API with LlamaRank: ▶ 8K context length for documents ▶️ Handles semi-structured data (JSON, email, tables, code) ▶️ Improves search relevance and reduces LLM compute costs in RAG systems ▶️ Compatible with Cohere's Rerank API for easy migration Developers can start using the Together Rerank API with LlamaRank today. Read more in our blog post: https://lnkd.in/g2Me2x7E We can’t wait to see how developers will enhance their search and RAG applications with this new API.

    • No alternative text description for this image
  • View organization page for Together AI, graphic

    28,421 followers

    Congratulations to our partners at AI21 Labs on the release of their Jamba 1.5 models – powerful long context models for the enterprise. This language model family introduces: ▶️ 256K context window - longest among open models  ▶️ Up to 2.5X performance boost on long contexts  ▶️ Benchmark scores surpassing larger proprietary models  ▶️ Developer features: structured JSON output, function calling, citation mode  ▶️ Multilingual capabilities across 9 languages We’re proud to announce the upcoming addition of Jamba 1.5 to our platform, furthering our commitment to providing developers with access to state-of-the-art, open source AI models. 👏 Incredible work from AI21 Labs Jamba 1.5 – Coming Soon on Together.ai

  • View organization page for Together AI, graphic

    28,421 followers

    Announcing Together Analytics in beta! You can now view your usage (requests, latency, and TPM) over time. This was a highly requested feature from our customers to be able to see their usage in a more granular level. You can now see usage over time and see successful requests, errors, latency, and TPM for all your usage on Together in one place. We're also already working to make it even better! Available today in beta for all customers, directly in your Together dashboard.

  • Together AI reposted this

    View profile for Varun S., graphic

    🇫🇮🇮🇳 | CPTO at Daily | Built and ran @callstatsio | Developed multimedia networks | WebRTC | DASH | Advisor | Investor

    We made it simpler to run voice ai and partnered with Anthropic Cartesia Deepgram Together AI for this launch. Our goal is to combine the best tools, best developer ergonomics, and best infrastructure for real-time AI into a single platform. Daily will host and launch the bots for you, these bots are connected to STT, LLM, and TTS. We have several options under each category, plus you can bring your own API Keys. In addition, we offer function calling to third-party, or to your own infra (using webhooks and post requests). Further, since this is built on RTVI, you as a developer have the ability to augment pr change the config settings and services on the fly, mid-conversation. On the SDK front daily bots comes with a client component and a thin server component. Clients are an RTVI client app and we have web, Android, and iOS SDKs and sample apps. The server side is basically an init, the first request from the client initiates /bot/start. The server is also helpful in the case of telephony, wherein there isn’t a rtvi client per-say and you need a good default config, service and prompt. In an upcoming update you maybe be able to bind dtmf tones to external function calls, build a dynamic IVR for example.

    View organization page for Daily, graphic

    2,940 followers

    Today we’re launching Daily Bots, the ultra low latency Open Source cloud for voice, vision, and video AI. Build voice and video AI bots, with any LLM, at conversational latencies as low as 500ms. Developers can: ✳️ build with Open Source SDKs ✳️ mix and match the best Generative AI models for specific use cases ✳️ run at scale on Daily’s real-time global infrastructure We've partnered with Anthropic Cartesia Deepgram Together AI for this launch. Our goal is to combine the best tools, best developer ergonomics, and best infrastructure for real-time AI into a single platform. Daily Bots is the culmination of the last 18 months of work we've done with customers and partners. The two fastest growing Open Source real-time AI projects came out of this work: Pipecat and RTVI. *️⃣ Daily Bots apps are built using the RTVI Open Source SDKs for the Web, iOS, and Android. Your Daily Bots code will run anywhere that supports the RTVI standard. (Or you can run your own infrastructure.) *️⃣ Your Daily Bots can also answer the phone. (You can buy a phone number from us with a single curl command.) *️⃣ Bring your own API Keys and use any inference provider that supports OpenAI-compatible APIs. Or run your own models and connect your bots to your infrastructure. For a highly non-serious take on the kinds of things we’ve been creating for ourselves, as we’ve worked on Daily Bots, check out a fun demo video 👇 and more links in thread. We'd love to hear what multi-modal, real-time AI directions are most interesting to you, and excited to support developers building! #conversationalAI #voicetovoice #ai #claude #llama #generativeai

Similar pages

Funding