Groq

Groq

Semiconductor Manufacturing

Mountain View, California 59,847 followers

Driving the cost of compute to zero.

About us

Groq builds the world’s fastest AI inference technology. The LPU™ Inference Engine by Groq is a hardware and software platform that delivers exceptional compute speed, quality, and energy efficiency. Groq, headquartered in Silicon Valley, provides cloud and on-prem solutions at scale for AI applications. The LPU and related systems are designed, fabricated, and assembled in North America.

Website
https://groq.com/
Industry
Semiconductor Manufacturing
Company size
51-200 employees
Headquarters
Mountain View, California
Type
Privately Held
Founded
2016
Specialties
ai, ml, artificial intelligence, machine learning, engineering, hiring, compute, innovation, semiconductor, llm, large language model, gen ai, systems solution, generative ai, inference, LPU, and Language Processing Unit

Locations

Employees at Groq

Updates

  • Groq reposted this

    View profile for Jakub Neander, graphic

    Software Engineer, Tech Lead, Entrepreneur

    This shifted my view entirely 🤯 transporting data to a data center for processing is better than doing it locally! Data centers are more energy-efficient for tasks like AI inference. Transporting data to a data center for processing and then sending the result back is faster and uses less energy than processing it locally. This approach is not only faster (sic! 🔴) but also more cost-effective and environmentally friendly due to the efficient use of large-scale compute resources Interview with Jonathan Ross from Groq

  • View organization page for Groq, graphic

    59,847 followers

    We're thrilled to share two more hackathon winners🥇🥈 powered by Groq!

    View organization page for Vapi (YC W21), graphic

    1,832 followers

    Drum roll, please… 🥁 The results of the Vapi Hackathon are here! 🏆 Congratulations to all the developers who showcased their skills and participated in our event 💙 Announcing our winners ↓ 🥇 Amulya • Training and testing voice AI tools for any team Project demo: https://usephone.ai GitHub repo: https://lnkd.in/dXtQmrVk 🏆 $5,000 in AI credits, swag and a Opal Camera! 🥈 Attila • Voice AI app that comes to your rescue when you need a witty escape from any situation Project demo: https://lnkd.in/ddm-3fHJ GitHub repo: https://lnkd.in/duMB7GVd 🏆 $3,000 in AI credits, swag and a mechanical keyboard! 🥉 Pranav • Angela, a voice AI that can provide information about everything in his portfolio Project demo: https://lnkd.in/dVQm_Pet GitHub repo: https://lnkd.in/demU-Z_p 🏆 $2,000 in AI credits, swag, and an ergonomic mouse for developers Also, thanks to our partners who were in this awesome event with us Deepgram, Daily, Outerbase, Vercel and Groq 👾 #hackathon #voiceAI #AIhackathon

    • No alternative text description for this image
  • Groq reposted this

    View profile for Shelby Hritz, graphic

    Vice President, Aerospace and Defense, Groq Public Sector | xNASA xGoogle xVMware xIntel

    👉 "Generative AI is not about who has the Data - it's who has the Compute" - Jonathan Ross, CEO/Founder Groq 💡 Those who succeed in harnessing GenAI will understand this paradigm shift. Spend 🔟 seconds with Jonathan here

  • View organization page for Groq, graphic

    59,847 followers

    Thanks for the summary!

    View profile for Saritha Prasad Vrittamani, graphic

    Visionary Technology Leader leveraging Gen AI to revolutionize the Industrial Metaverse and create immersive, intelligent experiences | Mixed Reality Solutions| Product Leader | Innovator | D&I Advocate | Coach | Mentor

    Either AI Is Happening to You, or You Are Making AI Happen - By Jonathan Ross, CEO & Founder at Groq, at the #GenAISummit San Francisco 2024! The AI revolution is here, and it's about to get a whole lot faster. Groq, a trailblazing startup, has developed a groundbreaking AI inference chip called the LPU that outperforms GPUs in speed and efficiency. By focusing on optimizing their compiler and building hardware around it, Groq has created a platform that can deploy new AI models with lightning speed, attracting over 200,000 developers in just 11 weeks. Groq's LPUs are not only faster but also more energy-efficient, using up to 90% less energy per token compared to GPUs. This breakthrough allows for unprecedented inference speeds, with Groq currently testing at 30,000 tokens per second on an 8 billion parameter model and aiming to reach an astonishing 25 million tokens per second by the end of the year. The implications are vast. Groq's technology enables real-time multimodal models, including large language models, image recognition, and speech-to-text, opening up a world of possibilities for AI applications. As Groq rapidly expands its computing infrastructure globally, it's poised to revolutionize industries and empower developers to create groundbreaking AI-powered solutions. The future is here, and it's driven by Groq's unwavering commitment to pushing the boundaries of AI performance. Brace yourself for an exhilarating ride as Groq unleashes the full potential of AI, one breakthrough at a time. PC - Twitter

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • View organization page for Groq, graphic

    59,847 followers

    📈 🚀 Thanks for the benchmark Artificial Analysis.

    View organization page for Artificial Analysis, graphic

    1,672 followers

    Groq extends its lead and is serving Llama 3 8B at almost 1,200 output tokens/s! We can confirm Groq's Llama 3 8B speed improvements seen in their chat interface are reflected in the performance of their API. This represents the fastest language model inference API performance that we benchmark. Groq has let us know this is driven by compiler optimizations and they are still far off the theoretical performance potential of their LPUs. This is particularly impressive when considering their current generation LPU is manufactured with a 14nm process node which is an older process technology compared to newer, higher density nodes including the 5nm node which Nvidia’s H100 (4N) is manufactured on. Link to our Llama 3 8B API providers analysis: https://lnkd.in/g8TEVZt4

    • No alternative text description for this image

Similar pages

Browse jobs

Funding