Pinned straw:
Thanks @Arizona and @Shapeshifter for the AI-Media links. I found them useful for fleshing out my understanding of their products, and their marketing activities.
I don't speak German or Spanish, so can anyone comment on the quality of the voice translation in the first video? Also, the voice translation was pre-recorded, so not a thorough test of a live broadcast.
Also, have to admit I was disappointed with the quality of the live captioning in the third video with Mark Skehan. At times it was accurate, but at other times it was completely wrong, and at times it just seemed to make up new unrelated content (AI hallucination?). I didn't run the numbers, but didn't seem much better than around 50% accurate. Much worse than existing readily available translation tools. Again, not knowing Spanish, I can't judge the live text translation, but given it involves another layer of AI to translate, it wouldn't surprise me if it was near useless.
Maybe the transcription demo wasn't a best case scenario. Perhaps a business trade stall isn't ideal for voice quality. Still, the voice quality when watching the video was ok, so why couldn't Lexi do better? And Tony and Russ are spruiking a wide range of use cases, many which won't have noise-free studio-quality audio recording.
Makes me question where they are getting their accuracy numbers. It definitely didn't outperform what a human could do. In their H1FY25 results presentation they said Lexi was achieving 98.8% accuracy - no way did they achieve anything like that in the demo video. Where are they getting those accuracy numbers? Or why was the demo so much below what Lexi can achieve on average?
The demo felt more like a proof of concept rather than a finished product. Of course, AI is improving rapidly, so maybe Lexi improves rapidly. Or maybe it doesn't - past transcription and translation tools quickly reached 90% usefulness, but couldn't then improve further.
Any thoughts from others on this?
Thanks @Shapeshifter and @Arizona for the flag on AIM's Infocomm 2025 involvement and the links to the video’s. Still trying to find a reliable way to get notified of these events and links directly!
I viewed the 3 interviews and actually took away quite a few new insights.
In summary,
Discl: Held IRL and in SM and topped up again at 48.5c
My take away points:
TONY INTERVIEW WITH PETE CONNAN
Tony summed up AIM's capability nicely with this:
“We are delivering “AI” into every possible workflow for our customers, live and recorded, using the best of public information, private information, that is scaled within that organisation to be coherent and clear, and then live information on top of that. Think about it as the best of ChatGPT, but with your private information and updated live with the thousands of Lexi feeds that we have going on around the world. A Bloomberg service, tailored to your organisation”.
The other insight in that interview was Tony mentioned that both his 2 co-founders were deaf. The problem they were trying to solve was originally not about broadcast, but how do deaf people get an education. In 2003, 50% of deaf kids did not complete high school. This drove the finding of a way to deliver live captioning, at scale, in classrooms by 2007., At peak, AIM has 800 re-speakers, using/checking AI to deliver live captioning
I did not understand this before, but I can now make better sense of Tony’s comments in earlier interviews in describing the transition from human to AI when I started deep diving AIM.
DANTE INTEGRATION
Technically, AIM is about orchestrating live video and voice feeds through the AIM iCap ecosystem/Lexi Cloud, which then opens up all the Lexi services to those feeds and all the good revenue that comes with it.
My immediate reaction to the Dante integration was thus “wow, that makes good sense”.
Enabling the hookup from AIM to Dante would be a “one-time” integration - do it once platform-to-platform, and theoretically, all customers on Dante would thereafter have a standard technical vehicle to connect a Dante-enabled AV setup to the Lexi Cloud. This is very much aligned to AIM’s vision of growth via platform-to-platform integration.
Once there is a live voice and/or audio feed from a Dante site to Lexi Cloud, then that Dante site should be able to access all Lexi Services - Lexi Voice, captioning etc, probably not much differently from the major broadcasters
AD8/Dante is primarily focussed on corporate/office/education institutions/government AV installations - a Dante pipe would then allow AIM penetration into the non-broadcast world, which makes up 6 of AIM’s 9 squares
As I also like and hold AD8, this should be a good thing for AD8 as it should add another out-of-the-box capability to the Dante platform

So, pursuing a Dante integration makes great sense, given how it would open up non-broadcast via an immediate Dante-enabled ecosystem. This does not feel like a random opportunity. Something to absolutely cheer on, in my view!
BACKBONE NETWORK
This was interesting to get low latency audio translation from Lexi Voice to mobile devices in a live event setting for huge audiences where the delivery of the real time voice translation is controlled by AIM and is not dependent on the infrastructure of the source environment
This should theoretically improve the delivery of broadcast content to live events as well as/and/or mobiles. Very nice!
MARK SKEHAN INTERVIEW
I thought a couple of interesting things came out of the AVIXA interview.
AIM are holding conversations with Audinate to look at intergrating with Dante.
They have partnered with Backbone Networks to develop ultra low latency translation for mobiles.
Basically they are throwing heaps of mud at the walls and hoping some sticks!