At NVIDIA’s AI Summit, I met Anshul Khandelwal, Co-founder and CTO of invideo, and attending his insightful session on their lip-sync model. They serve 25 million users, the model leverages NVIDIA’s StyleGAN with audio-driven style learning to deliver realistic, high-quality lip sync while maintaining facial identity, which stood out. Anshul also shared the challenges of scaling the model and the key lessons learned during its development. To efficiently handle their large user base, they utilize TensorRT and Triton Inference Server to optimize performance and reduce costs. The session dived into how AI is transforming video content creation, empowering creators to produce high-quality multimedia experiences with ease. It was inspiring to learn how InVideo’s platform, powered by E2E Cloud Network’s GPUs, is driving innovation in digital content. Thank you, Anshul, for sharing your journey and insights! It was a pleasure meeting you. Hoping to stay connected. Cheers! #invideo #nvidia #ai #virtualhumans #dcverse
Abhigyan Suman’s Post
More Relevant Posts
-
Paraphrased from NVIDIA's earnings call: Inferencing, which is fast token generation at massive scale, has become incredibly complex. Generative AI is driving a full stack platform shift in computing that will transform every computer interaction. We're shifting from an information retrieval to an answers and skills generation model of computing. AI will understand context and our intentions, be knowledgeable, reason, plan and perform tasks. As such, token generation will drive multiyear build-out of AI factories. Beyond cloud service providers, generative AI has expanded to consumer Internet companies and enterprise, Sovereign AI, automotive, and health care customers, creating multiple multibillion-dollar vertical markets.
To view or add a comment, sign in
-
🚀Meta Advances Generative AI with Llama 3.3 70B Meta’s Llama 3.3 70B is here, offering the performance of its largest AI model (405B) at a fraction of the cost. 🔍Why it matters: • Outperforms Google, OpenAI, and Amazon on key benchmarks like MMLU. • Accessible for download on Hugging Face and Meta’s official platforms. • Supports commercial applications with flexible licensing for most users. Meta is doubling down on AI with a $10B data center in Louisiana to train Llama 4, leveraging a cluster of over 100,000 Nvidia GPUs. 💡Impact: Llama powers Meta AI, already serving 600M+ active users monthly, and continues to expand its influence across industries. What opportunities do you see for businesses leveraging Llama’s capabilities? #AI #MetaAI #Llama3 #GenerativeAI #Innovation
To view or add a comment, sign in
-
-
I wonder if people understand how epic the NVIDIA announcement about the Blackwell AI technology is. On top of that the announcements about the NIMs and Omniverse are substantial. The fact that we can simulate virtually anything and then connect the simulation to the real world will change the development economy. Going from writing apps to connecting AI actions and APIs will fundamentally change how people develop code. Garrett Davis Anthony Alicea Donald Ladwig ai.nvidia.com is a new cloud platform for doing this type of development. #nvidia #ai #findinfinite
To view or add a comment, sign in
-
🤖Microsoft #VASA-1 is the future of lifelike talking faces! 🕋The VASA Framework introduces a revolutionary way of creating authentic and expressive talking face videos using only a single image and speech audio. The VASA-1 model produces natural head motions, synchronized lip movements, and facial nuances, making it the perfect tool for creating lively and engaging content. What sets #VASA-1 apart is its innovative technology, which uses a holistic model for facial dynamics and head movement to generate expressive and disentangled face latent space using videos. This technology allows for a more natural and lifelike talking face experience. Furthermore, the potential and ethical considerations of this technology are vast and promising. The responsible development of AI, like #VASA-1, can have a positive impact on human well-being. Join us in exploring the possibilities of this exciting technology! https://lnkd.in/eb86uZZr #MicrosoftIA #VASA #MicrosoftResearch #Azure #LLM
To view or add a comment, sign in
-
-
Big Serverless Update! 🚨 Exciting news! Meta Llama 3 405B is now part of the Hugging Face x NVIDIA NIM API (serverless) with pay-as-you-go pricing at $0.0023/second/gpu. This update is exclusively available for all Enterprise Hub organizations! 😍🚀 How to transition to Meta Llama 3 405B: 🔹 Be a member of an Enterprise Hub organization. 🔹 Create fine-grained tokens with organization scope. 🔹 Replace GPT-4o in your code with meta-llama/Meta-Llama-3.1-405B-Instruct-FP8. 🔹 Run Inference and kickstart your Generative AI applications using open models. #meta #huggingface #llama #ai #llm #gpu #genai #nvidia #llama3
To view or add a comment, sign in
-
-
NVIDIA just dropped Llama 3.1-Nemotron-51B, a super smart language model that's fast and accurate! It's like a better, quicker version of Meta's Llama-3.1-70B but uses fewer resources. This means it runs smoother on a single GPU and is much cheaper! 💸 • 2.2x Faster than previous models • Super Accurate without losing quality • Lower Costs for running AI models • Fits on a single GPU for big workloads Key Benefits: • Better accuracy at a lower cost • Optimized for inference and deployment • Works smoothly on cloud and data centers For Tech teams needing speed and efficiency and companies looking to scale AI services this is a game changer. The table below lists all the benchmarks we evaluated – comparing our model and the reference model – Llama3.1-70B. The ‘Accuracy preserved’ is the ratio between our model’s score and that of the teacher. ---- Follow for more Daily dose of AI! #AI #Chatbots #TechNews #LLM
To view or add a comment, sign in
-
-
🚀 **Tech News Roundup: Nvidia's AI Agents, The Best of CES, & TikTok's Day in Court!** 🚀 Hello DataxLogic community! 📊✨ We’re here to keep you updated on the latest happenings in the tech world, all of which are invaluable for businesses leveraging AI solutions. First up is **Nvidia’s AI agents**, which have taken the tech scene by storm. These sophisticated AI agents showcased at CES are redefining the way we think about automation and intelligent interaction. With their ability to understand and process natural language, they promise to enhance user experiences across various platforms. At DataxLogic, we are excited about their potential and are continuously improving our own AI voice agents to ensure that your business benefits from cutting-edge technology. Imagine having AI agents that can seamlessly integrate with your operations, improving efficiency and customer satisfaction! Speaking of CES, this year’s showcase revealed some groundbreaking innovations that are setting the bar for the technology industry. From smart home devices to advanced robotics, the best of CES is an inspiration for our ongoing mission to deliver top-tier AI solutions to businesses. Stay tuned as we adopt some of these trends within our own offerings! Lastly, we can’t overlook TikTok’s day in court regarding privacy and data concerns. The outcome of this case could shape future regulations affecting numerous tech companies, including those in the AI space. It's a reminder for all businesses to prioritize data safety and compliance when deploying AI solutions. At DataxLogic, we're dedicated to keeping you informed and equipped with the best AI voice agents tailored to your needs. Let’s embrace AI responsibly and innovatively! Keep an eye on our page for more updates! 💬👩💻 #DataxLogic #AI #Nvidia #CES2023 #TechNews #BusinessSolutions Source: https://lnkd.in/dSccdq-f
To view or add a comment, sign in
-
We have been busy the past few months and we are thrilled to share an exciting update! 🎉 We have a new version of Reka Flash ⚡, our powerful 21B model that supports interleaved multimodal inputs (text 📄, image 🖼, video 📽, audio 🎧). This update now brings you significant capability improvements on: ☑ multimodal understanding, ☑ general reasoning, ☑ agent building blocks, and ☑ speech input-output. This makes our models more useful across a wide range of practical use cases. The new Reka Flash ⚡ is available today on Reka Chat and Reka API. It can also be deployed on-premises or in a private cloud setting. 👀 Check our blog for more details. https://lnkd.in/gqk_EbEX Try Reka Flash ⚡ Reka Chat: https://chat.reka.ai/ Reka API: https://lnkd.in/gV9ZfF6Q ➕ In addition, we are partnering with NVIDIA to package this model as NVIDIA NIM, which optimizes the model for higher throughput and lower latency. The NIM microservice will soon be available on ai.nvidia.com.
To view or add a comment, sign in
-
Huge update with our newest version of Reka Flash ⚡, our powerful 21B model that supports interleaved multimodal inputs (text 📄, image 🖼, video 📽, audio 🎧). This update now brings you significant capability improvements on: ☑ multimodal understanding, ☑ general reasoning, ☑ agent building blocks, and ☑ speech input-output. This makes our models more useful across a wide range of practical use cases. The new Reka Flash ⚡ is available today on Reka Chat and Reka API. It can also be deployed on-premises or in a private cloud setting. 👀 Check our blog for more details. https://lnkd.in/gqk_EbEX Try Reka Flash ⚡ Reka Chat: https://chat.reka.ai/ Reka API: https://lnkd.in/gV9ZfF6Q ➕ In addition, we are partnering with NVIDIA to package this model as NVIDIA NIM, which optimizes the model for higher throughput and lower latency. The NIM microservice will soon be available on ai.nvidia.com.
We have been busy the past few months and we are thrilled to share an exciting update! 🎉 We have a new version of Reka Flash ⚡, our powerful 21B model that supports interleaved multimodal inputs (text 📄, image 🖼, video 📽, audio 🎧). This update now brings you significant capability improvements on: ☑ multimodal understanding, ☑ general reasoning, ☑ agent building blocks, and ☑ speech input-output. This makes our models more useful across a wide range of practical use cases. The new Reka Flash ⚡ is available today on Reka Chat and Reka API. It can also be deployed on-premises or in a private cloud setting. 👀 Check our blog for more details. https://lnkd.in/gqk_EbEX Try Reka Flash ⚡ Reka Chat: https://chat.reka.ai/ Reka API: https://lnkd.in/gV9ZfF6Q ➕ In addition, we are partnering with NVIDIA to package this model as NVIDIA NIM, which optimizes the model for higher throughput and lower latency. The NIM microservice will soon be available on ai.nvidia.com.
To view or add a comment, sign in
-
Kinetica is featured alongside Cohesity DataStax and NetApp as a leading NVIDIA Data Platform Partner working with their NeMo Retriever NIM microservices to boost AI models’ accuracy and throughput. "Kinetica will use NVIDIA NeMo Retriever to develop LLM agents that can interact with complex networks in natural language to respond more quickly to outages or breaches — turning insights into immediate action." This integration allows our users to invoke embedding and inferencing models provided by NIM directly within Kinetica, simplifying the development of production-ready generative AI applications that can converse with and extract insights from enterprise data. By combining these services with Kinetica’s robust compute and vector search capabilities, developers can easily build data copilots that meet rigorous performance and security requirements. Check out this tutorial to see how to connect Kinetica to #NIM with just a few SQL statements: https://lnkd.in/egXQrYwj
AI, Go Fetch! 🐶 https://nvda.ws/4daZRTq The new NVIDIA NeMo Retriever NIM #inference microservices can help enterprises unlock the value of business data by boosting model accuracy and throughput.
To view or add a comment, sign in
-