Logo
Datadrifters Blog Header Image

Are We Building Skynet? Can’t-Miss Robotics and AI Updates from Neuralink, Amazon, Salesforce, Disney, Nvidia, Unitree and More!

2024-08-27


We’re inching closer to building Skynet, and I’m not even hyping this up.


It’s crazy how fast robotics and AI are evolving, every time I turn around, there’s another jaw-dropping breakthrough.


Companies are doing things we only dreamed of a few years ago.


These are wild times but I’m not just fascinated by the tech; I’m curious about what it really means for us, for our future.


The latest updates from big names like Neuralink, Amazon, Salesforce, Disney, Nvidia, and others show where we’re headed.


And there’s no denying that we’re in the middle of something huge.


So, let’s unpack it all and see what’s coming next:



Let’s GO! (Also some practitioner guides to build with AI at the end)



Neuralink’s Groundbreaking BCI Interface: A New Era of Human-Computer Interaction


Neuralink’s latest update on their second human trial participant, Alex, is seriously cool.


Alex got his Neuralink implant just last month, and is now playing Counter-Strike 2 using nothing but his brain and the Link.



Not only that, but he shattered the world record for brain-controlled cursor speed on day one.



The surgery itself was smooth, and Alex bounced back quickly, going home the next day.


Since then, he’s been diving deep into what the Link can do.


For example, he’s picked up CAD software to design 3D objects. This is a big deal because, before his injury, Alex was all about fixing and building things.


The fact that he’s now using brain power to design and 3D print custom mounts for his Neuralink charger is like watching someone regain a piece of their identity.



But what really stands out here is how quickly Alex adapted.


Within minutes of linking up to his computer, he was already controlling a cursor with more precision than any other assistive tech he’d used before.


The Link makes it all feel natural and fluid.


Neuralink is working on fine-tuning the interface to let Alex control different types of mouse clicks with his thoughts.



Neuralink learned from the first participant, Noland, who had some issues with the threads retracting, which messed with his brain-computer interface (BCI) performance for a bit.


They’ve now refined the procedure to avoid that in future implants, and there are no such problems with Alex.


A very clear example of how customer experience (in this case, from the patients) can drive product improvements.


In short, what we’re seeing here is not just about playing games or designing gadgets — it’s about creating a seamless, high-performance interface that can redefine autonomy for people with disabilities.


And from a developer’s perspective, the sky’s the limit on where this technology can go.


Imagine the tools, apps, and experiences we can build when the interface itself is just a thought away.


Unitree’s G1: A Leap Forward in Robotic Flexibility and Precision


Unitree, a robotics company from China, just dropped something that’s next-level.


They’ve got this G1 bot ready for mass production that’s supposedly a notch above what we’ve seen in terms of functionality and looks.


This is a robot with flexibility that could actually put most of us to shame — seriously, it’s got joint movement that’s way beyond what humans can do, powered by a crazy number of joint motors, anywhere between 23 and 43, depending on the configuration.



The control here is precise and dynamic, thanks to a combination of imitation learning and reinforcement learning.



It’s got a very dexterous hand.


We’re talking about a hand that not only mimics human movements but does so with such precision that it can handle objects almost as well as we can, which means it can adapt to whatever it’s handling with remarkable sensitivity and reliability.


From a product dev standpoint, this kind of tech isn’t just iterative — it’s a game-changer.


Imagine the use cases, from delicate surgical tools to intricate assembly lines.



Then there’s the UnifoLM!


It’s basically their unified large model for robotics.


They’re pushing this as a collaborative platform to develop intelligent systems together.


The way I see it, this isn’t just about building robots; it’s about creating an entire ecosystem where AI and robotics co-evolve with input from the community.


That’s a powerful idea, especially for those of us focused on building products that aren’t just smart, but can truly adapt and learn from real-world interactions.



With the race between the US and China heating up, we’re going to see an insane amount of innovation in this space.


This thing is apparently close to mass production and could be hitting the market for around $16,000.


That’s incredibly competitive for what it offers. If this goes mainstream, it could seriously democratize access to advanced robotics, opening the door for all sorts of innovative applications.


What do you think?



The G1 is something to keep on your radar. I’m definitely going to dig into this more — there’s a lot we could potentially learn and apply from what Unitree is doing here.


Before we continue, let me add a brief and kind request here:


Stay in the loop


AGIBOT’s Revolutionary Humanoid Robots: Open-Source Disruption


So, there’s also this startup in China called AGIBOT, who just dropped a bombshell in the robotics scene.


They rolled out five new humanoid robots, and what’s crazy is they’re going open-source with the designs!



Yup, you heard that right — open-source.


Now, why should you care?


Well, these bots aren’t just tech demos.


Each one is built with a specific set of tasks in mind, ranging from helping around the house to doing heavy-duty industrial work.



What’s really impressive is how quickly they’ve made this happen. We’re talking about just 18 months of hustle, and bam — they’re already giving Elon Musk’s Tesla a run for its money with these new bots.


It feels like a real David vs. Goliath situation, and AGIBOT isn’t shying away from it.


Their flagship bot, Yuanzheng A2, is like the Swiss Army knife of robots.


It’s got the dexterity to thread a needle — plus, it’s got a human-like height and weight, so it feels more relatable, it’s designed to be part of our everyday environment.


AGIBOT plans to start shipping 300 units by the end of 2024.



They’re saying they’ve got the commercialization and cost-control game down better than Tesla, which is a bold claim.


But if they pull it off, this could be a game-changer in making advanced robotics more accessible.


ETH Zurich and Disney’s AI: Redefining Robot Motion


I think you’ll appreciate this one, especially if you’re into the whole intersection of AI, robotics, and animation.



ETH Zurich and Disney have teamed up to create this new AI system that can make a robot dance, walk, or even do flips just by feeding it a simple text or image input.


The system is a two-stage process, and here’s how it works:



In fact, it’s doing a better job than anything else out there right now in terms of accuracy and generalization.


It can transfer these movements directly to real robots without losing balance or style, which is pretty impressive.


With personal robots possibly hitting the market as soon as 2025, this tech could be the key to making them feel less like machines and more like companions.


The idea that we might be living alongside robots sooner than we thought isn’t just sci-fi anymore — it’s becoming a reality, and this kind of tech is a big part of why.


Would you like to have such a companion?


Salesforce’s Einstein AI Agents: Transforming Sales Automation


Salesforce has unleashed these two new AI-powered sales agents:



They’re fully autonomous, meaning they handle a lot of the grunt work that typically eats up time, but they do it in a way that’s actually useful and smart.



First up, the Einstein SDR Agent goes beyond the typical chatbot stuff.


Instead of just spitting out pre-programmed responses, it actually makes decisions on the fly.


Whether that’s answering a question about a product, handling an objection, or setting up a meeting, it’s all done automatically, and it’s based on real, contextual data from your CRM.


It’s also not just limited to one language or one channel.


This thing can juggle multiple leads at the same time, across different platforms and languages, without breaking a sweat.


They can offload those repetitive, top-of-the-funnel tasks and actually spend time on what really matters — building relationships and closing deals.


Einstein Sales Coach Agent is all about training and leveling up your sales team.


It’s basically an AI coach that runs through role-plays with your reps.


But here’s where it gets interesting — it tailors these role-plays to the specific deal at hand.


So, imagine you’re about to pitch to a tough client.



The Einstein Sales Coach will simulate a real buyer, complete with objections and all, grounded in actual data from your CRM.


After the session, it gives personalized feedback, helping the rep improve their pitch or negotiation skills before they even meet the customer.


It’s like having a practice run, but with a coach who knows exactly what’s at stake.



For the sales teams, this is something that give them an edge even before they step into a real conversation with a prospect.


Luma Labs’ Dream Machine 1.5: Pushing the Boundaries of AI Video Generation


Luma Labs dropped Dream Machine 1.5, and there’s some cool stuff under the hood that you’ll want to know about.



If you’ve ever had to tweak prompts endlessly to get something usable, you’ll appreciate this. They’ve seriously improved the realism and how the model understands and follows prompts.


The new version handles text-to-video like a pro — more accurate, more fluid motion, and way better at sticking to the vibe you’re going for with just a few words.


With 1.5, they’re clearly aiming to stay ahead of the pack (e.g., Runway, Kling, Haiper, Pika).



The text rendering has gotten a significant boost, too.


Now, you can generate clear, readable text in videos, which was always a bit of a pain point.



You will be able to create entire video sequences, including text elements like logos or end screens, just from a simple prompt.


It’s almost like having a design tool baked into your video generator.


They’ve made it so intuitive that getting legible text is as easy as using double quotes around your words — just like you would in any AI image generator like Midjourney.


If you’re looking for something that can seamlessly integrate into your workflow, save you time, and still deliver top-notch results, Dream Machine 1.5 is definitely worth a look.


Ideogram v2: Elevating Text-to-Image AI


Unlike other tools out there like Midjourney, Ideogram v2 has seriously nailed the ability to generate almost perfect text within images.


I know, it sounds like a small thing, but this opens up a ton of new possibilities (e.g., thumbnails, posters, newsletter graphics, memes).


Anything that needs sharp text integrated into visuals, this model can handle like a pro.



The best part is that it’s free for everyone on their website and their new iOS app!


Of course, if you want some extra goodies, there are premium features available through subscriptions.


For us developers, they’ve also rolled out a beta API, so you can start integrating Ideogram 2.0’s capabilities into your own projects right away.



What’s really impressive about this release is how much it outperforms the competition. Ideogram 2.0 isn’t just good; it’s set a new bar in the industry for generating realistic images, typography, and overall image-text alignment.


The team built this model from scratch, and it shows in the quality — it blows other models out of the water in most metrics that matter.



Oh, and they’ve got this thing called Ideogram Search now, where you can sift through over a billion images that users have created over the past year. That’s a treasure trove of inspiration just waiting to be tapped into.



In Ideogram 2.0, you’ve got Realistic, Design, 3D, and Anime, each crafted for specific types of images.


For instance, the Realistic style is so good that the images could easily pass as real photos. Textures, skin, hair — all look super lifelike.



And if you’re into customization, this version lets you generate images in any aspect ratio, even ultra-wide or tall ones like 3:1 or 1:3.


Plus, you can now control the color palette, which is clutch for maintaining brand consistency or just nailing that specific vibe you’re going for.


Oh, and speaking of Midjourney, they’ve finally got a web-based AI image editor now, and they’re offering free trials. It’s nice to see them moving away from Discord and making their tools more accessible.



Amazon’s AI Assistant Q: A Game-Changer in Software Development


I just had to share this because it’s wild — in the best possible way.


So, Andy Jassy shared some pretty insane numbers about Amazon’s new AI assistant, Q, and how it’s totally flipping the script on how we handle software development.


Q has apparently saved Amazon what would’ve taken 4,500 years of developer time. Yeah, you heard that right — thousands of years.


Upgrading an application to Java 17, which used to be a massive pain, taking around 50 developer days, now gets done in just a few hours.



Think about how much more you can get done with that kind of efficiency. We’re talking about moving from dreading those tedious, “necessary evil” tasks to having them knocked out almost automatically.


Another crazy stat Jassy threw out there was that 79% of AI-generated code reviews are shipped without anyone needing to touch them afterward.


But it’s not just about saving time — the upgrades facilitated by Q are also making the software more secure and cutting down on infrastructure costs.


They’re talking about $260 million in annual savings— coding agents are really changing the businesses as a whole.


What’s really cool about all this is that it lets us focus on the fun stuff — the new features, the creative challenges — without getting bogged down by the boring but necessary maintenance work.


Jassy even admitted that this kind of task is usually dreaded and often pushed aside for more exciting projects.


But now, with AI tools like Q, we can knock out the mundane tasks quickly and efficiently, leaving us more time and energy for the things that actually move the needle.


Nvidia and Mistral’s Minitron 8B: Achieving Efficiency and Accuracy in AI Models


Finally, if you’re into optimizing performance without sacrificing quality (who isn’t?), Nvidia and Mistral dropped a new model called Mistral-NeMo-Minitron 8B, and it’s worth paying attention to.


You know how we’re always battling that trade-off between model size and accuracy?


We either get a massive model that nails the accuracy but is a beast to run, or we go smaller and sacrifice some of that precision.


Well, Minitron 8B is flipping the script on that.


It’s a compact model — small enough to run on a decent workstation or even a high-end laptop — but it still holds its own against the big guys like Mistral-7B and Meta’s LLaMA 3.1–8B in benchmarks.



Nvidia combined two optimization techniques — pruning and distillation — to shrink down the Mistral NeMo 12B into something more manageable, all while keeping that top-tier accuracy.


This delivers the firepower of a 12B model, but at a fraction of the size and computational cost.


It means you can integrate powerful AI into applications without needing to rely on massive cloud infrastructures or worrying about latency.


Plus, running these models locally on edge devices doesn’t just cut down on costs — it also boosts security since you’re not shipping data back and forth to a server.


If you’re building anything that involves chatbots, virtual assistants, or content generation, this could be a viable option.


And if you’re thinking about deploying it, Nvidia’s made it super straightforward.


Minitron 8B comes packaged as an Nvidia NIM microservice with a standard API, so you can get up and running quickly.


They’re also planning to release a downloadable version that you can slap onto any GPU-accelerated system in minutes.


One more thing: if the 8B model is still too hefty for what you’re working on, Nvidia’s AI Foundry lets you prune and distill it even further.


You can create a custom, smaller version of the model tailored for specific use cases, whether that’s for mobile devices, embedded systems, or whatever else you’re building.


And you’re not sacrificing much in terms of accuracy thanks to the way they’ve optimized the training process.


Oh, and they also rolled out another model this week — Nemotron-Mini-4B-Instruct.


It’s designed for low memory use and fast responses, specifically for NVIDIA RTX-powered PCs and laptops.


Definitely worth exploring if you’re looking to leverage generative AI without breaking the bank or the CPU.


Bonus Content: Building with AI


Have a look at some practitioner resources that we published recently:


Say Hello to ‘Her’: Real-Time AI Voice Agents with 500ms Latency, Now Open Source

Fine-Tune Meta’s Latest AI Model: Customize Llama 3.1 5x Faster with 80% Less Memory

Fine Tuning FLUX: Personalize AI Image Models on Minimal Data for Custom Look and Feel

Data Management with Drizzle ORM, Supabase and Next.js for Web & Mobile Applications


Thank you for stopping by, and being an integral part of our community.


Happy building!