In the realm of real-time applications, Large Language Models (LLMs) have long dominated language-centric tasks, while tools like OpenCV have excelled in the visual domain. However, the future (maybe) lies in the fusion of LLMs and deep learning, giving birth to the revolutionary concept of Large Action Models (LAMs).
Imagine a world where AI not only comprehends language but mimics human actions on technology interfaces. For example, the Rabbit r1 device presented at CES 2024, driven by an AI operating system and LAM, brings this vision to life. It executes complex commands, leveraging GUIs with unprecedented ease.
In this presentation, join me on a journey as a software engineer tinkering with WebRTC, Janus, and LLM/LAMs. Together, we’ll evaluate the current state of these AI technologies, unraveling the potential they hold for shaping the future of real-time applications.
Alberto is from Barcelona, Spain and is currently living in Miami, USA. He holds a degree in Telecommunication Engineering from UPC and a Masters degree in Information Technology and Management from IIT. He is a consultant and CTO at WebRTC.Ventures in Miami, FL, and brings over 8 years of software and telecom engineering expertise. As a Senior Software Developer, his impact spans architecting IoT, VoIP, and WebRTC projects, implementing over 45 real-time solutions.
His telecom engineering background and interest in tech have led him to build many projects for small and big enterprises, projects including: modern call centers, telehealth video chats, telepresence remote control, speech analytics, broadcasting and many others Real Time applications.
Beyond the tech world, Alberto finds joy in kiteboarding and skiing. And trying to dance salsa.