Leveraging AI

102 | Claude introduces Projects a better version of GPTs, Google Gemini 1.5 Pro with 2 million tokens is now publicly available, Suno and Udio are being sued, and other important AI news for the week ending on June 28

• Isar Meitis • Season 1 • Episode 102

🔥 Special Offer Alert! 🔥
Transform Your Business with AI
egister Here (https://multiplai.ai/ai-course/) and use Promo Code LEVERAGINGAI100 to get $200 off – ONLY AVAILABLE until the end of June!

Is AI Transforming Your Business? 
Here's What's Happening Right Now.

AI advancements are rolling out at a rapid pace. Are you keeping up? Discover how the latest updates from AI giants like Anthropic and OpenAI are set to reshape the business landscape.

In this episode of Leveraging AI, host Isar Meitis dives into the latest AI news, covering major announcements and their implications for businesses and executives. If you missed our landmark 100th episode, you’re in luck—catch up on invaluable tips from 18 AI experts!

This week's highlights include:

  • Anthropic’s Claude 3.5 Sonnet: Discover the speed and capabilities of this powerful model, including its innovative new feature, Artifacts.
  • Anthropic Projects: Learn about this workspace toolset, Claude's answer to GPTs, which is set to revolutionize how businesses automate and collaborate.
  • OpenAI’s Direct Sales Surge: Understand why OpenAI is now selling more models directly to customers rather than through Microsoft.
  • Google’s Gemini Models: Get the scoop on Gemini 1.5 Flash and Pro, including the game-changing 2 million token context window.
  • AI Critiques AI: OpenAI’s fascinating approach to reducing hallucinations by having GPT-4 critique its own outputs.


Don’t miss out on transforming your business with our AI Business Transformation Course. Register now with the promo code LEVERAGINGAI100 for a $200 discount—available only until the end of June!

Register here: https://multiplai.ai/ai-course/ 

About Leveraging AI

If you’ve enjoyed or benefited from some of the insights of this episode, leave us a five-star review on your favorite podcast platform, and let us know what you learned, found helpful, or liked most about this show!

Isar:

Hello and welcome to Leveraging AI, the podcast that shares practical, ethical ways to improve efficiency, grow your business and advance your career. This is Isar Metis, your host, and this is another weekend news edition. I actually thought of maybe not doing the news edition this week because we released two episodes already, episode 100 part one and 100 part two, which is on your player episodes, 100 and 101. If you miss those, they are out amazing. You want to go back and listen to them because we had 18 different AI experts. Each and every one of them sharing the number one tip that they have on how to leverage AI in a business context. So it's. an amazing list of people providing an amazing, wealth of AI knowledge that you don't want to miss. So if you miss those goes back and listen to those, but there were so many big things happening this week that I couldn't skip that and not let you know about it. So here we go. I'm going to start with, we're going to start with Anthropic. Anthropic made some really big announcements in the past two weeks. So last week, they released Clod 3. 5 Sonnet, which is their middle scale model. And we talked a little bit about this. it's a very powerful model. It is twice as fast as the previous one that they released only three months ago, and it can do some cool stuff. One of the capabilities that released last week is called Artifacts, which allows you to see the results of what you're doing on the right side of the screen, whether it's code generation or images or stuff like that. is an amazing capability that allows companies and individuals deploy pieces of code on the fly for any need. One of my clients showed me a really cool example where they took a matrix that they had for HR evaluation purposes and uploaded to Claude and ask it to code a web user interface that will allow people to give these reviews that they were doing manually for years on a web form. And within five minutes, they had the form ready that they can now copy the code and embed it on their internal employee website for them to evaluate each other. And that would have cost them X number of thousands of dollars previously to hire somebody, to do this kind of work. So endless numbers of opportunities where you can code your own stuff, see the results right there and then deploy it in any way you want. A lot of people are developing games and so on, but that was last week. I'm just sharing with you a lot of cool use cases, but this week they've announced projects. And projects is Claude's version of GPTs, which is the automation platform created by OpenAI a while back. So what you can do is you can create dedicated workspaces where users can chat with Claude, upload files, work together, collaborate on documents, code, images, and more, and build these recurring automations that will do the same thing consistently again and again. Similar to the things you can do with GPTs, you can upload files like PDFs and images and code snippets, and then use cloud to analyze the data and generate new outcome based on the data that you've uploaded. And as I mentioned, using the same instructions again and again, so you don't have to re prompt to get the same results. As with anything Anthropic, they put a big emphasis on privacy and security, so they are claiming that the data that's uploaded to projects is encrypted and is not used to train their models, which means you can, in theory, upload any data that you have, even if it's somewhat sensitive without worrying where that data is going to go. Now, that being said, the only thing it's still missing that drives me fricking crazy is the fact that it still does not have internet access. Meaning you still are limited to the data that you're uploading and you cannot query anything from the web. The workaround for that is you can use both Claude Sonnet 3. 5, as well as Claude 3 Opus, their largest model within Perplexity, if you have the paid version. So I found a lot of success in doing things that were not possible before just running Claude within Perplexity. And that gives it internet access still with the power of Claude, which I absolutely love as a model to review, summarize, and create content. Staying on Anthropic, their CEO, Dario Amodei has held an interview this past week. And he talked a lot about the potential negative income of AI on society as a whole. And he's claiming that UBI Universal Basic Income is just not sufficient to address the potential economic inequality that AI is going to cause, meaning a lot of companies are going to make stupid amounts of money, like we've never seen before, and many many people are going to be out of jobs and will not be able to make money. So he's saying that UBI could provide a safety net, but as a society, we need to come up with proactive solutions that will help people get purpose in life and fulfillment once their job is displaced by the AI. I agree with him 100%. I think the only way to do this is some kind of an international collaboration that has to start now, because this is right around the corner and that will find ways to help people find purpose in life and do things when a lot of jobs are going to disappear. He suggested stuff like national service programs, lifelong education and training programs, and incentives for people to pursue creative and care work that he claims will take AI longer to automate. So in a nutshell, I agree with him 100%. And that's one of my biggest fears as far as the negative impacts of AI on our society. I want to pause the news for just a second to share with you an exciting piece of information. We are opening the registration for the July cohort of our highly successful AI business transformation course. It's a course I've been personally teaching since April of last year, we're teaching two courses every single month. So hundreds of individuals and companies have went through that course and are literally transforming their businesses because the information that they have gathered. The course includes four sessions of two hours, starting on July 8th, four weeks in a row at noon Eastern. So so it fits anybody, whether you're in Europe or in the East Coast or on the West Coast or anywhere in between. These courses sell out every single time we run them. So if you want to change the trajectory of your career by becoming the AI expert of your company, or if you are in a leadership position and you want to transform your business, then go and check it out. The link is going to be in the show notes, so you can open it right now on your phone and click through and see what is the information. In addition, if you're listening to this episode still on Saturday or Sunday, meaning in the month of June, you are in luck because to celebrate the episode 100 of the Leveraging AI podcast, we created a promo code for$200 off, which we have never done before and probably will never do again, at least not before episode 200. So maybe that will happen. But if you are listening to this, As I mentioned, before the end of June, which is when this promo code expires, You can use promo code LEVERAGINGAI100, all uppercase, to get 100 off. So if you can make a decision quickly, you are in luck and you can save 200. If not a big deal. Come and join us. I can guarantee you that the money you're going to invest is going to save you that tenfold in the first month during the course, probably. So it's worth every cent and in the landing page in the link that is shared in the notes, you can see what people who took the course are saying about it. I think that will help you make up your mind. And now back to the news. and From Anthropic and Claude to obviously their biggest competitor, OpenAI, an interesting piece of news surfaced this week that claims that OpenAI is now selling more of their own models directly to customers than through its partnership with Microsoft. Why is that interesting? Part of the partnership basically gave OpenAI a channel to get faster, bigger distribution Because Microsoft is connected to most or a lot of the corporations, at least on the Western hemisphere. And through them, they were supposed to sell a lot more licenses. But apparently now the tables have turned openAI is selling its licenses directly to corporations more than it's selling through Microsoft. That obviously builds on a lot of moves that we've seen from OpenAI by hiring a lot of senior people from the product side and distribution side and sales, and trying to focus a lot more on distribution and productization. Of what started as a research project and that is definitely seems to be paying dividends. what does that mean for Microsoft and how they feel about this new thing? I don't know, I can imagine. But we've seen this frenemies relationship going on for a while where Microsoft announced that they're building their own models while in parallel open AI are developing more and more capabilities for the enterprise level. So it will be interesting to see how this evolves in the future. Staying on OpenAI, OpenAI a few weeks ago, gave us the most amazing demo of a voice capable AI model that they call gPT 4. 0 for Omni and the demos were nothing short of incredible and allowed the users to have a normal human conversation across any topic while interrupting the model in the middle of the sentence and going back and forth and gaining benefits across Business and education and personal life and so on. And then they released some of the features, but not the voice feature. And they said that the voice feature is going to be delayed through the fall. And yet a few users suddenly had access to it in this past week, these users already shared with excitement, their capability to now use the tool and have a conversation with it and even create some really cool creative outcomes with role playing with the models across a story. But OpenAI quickly turned this off, and now apparently nobody has access to it. So was that a mistake? Maybe. Was that a test? We don't really know. There's no clear feedback on that from OpenAI themselves. But the reality is the models are not available right now, and they are still claiming that they're going to be released around the fall. We've seen this pattern from a lot of the big players, including open AI, Microsoft and Google, where they all in the recent announcement showed really amazing things. And most of them are not available to us in the near future. I really hope that they will release them as soon as possible because some of these capabilities are game changers when it comes to business efficiency and our ability to engage with computers like this changes everything because we'll be able to literally have a conversation with any computer, whether it's our watch, our car, things that run our house, the computers that work, anything will just won't need the keyboard anymore. And we'll be able to have conversations with computers and they will understand our needs and the background and the context and what we're trying to do. And we'll be able to complete the tasks and help us in being a lot more efficient and avoid tedious tasks. Another interesting report from OpenAI this week showed in a blog post that they are now leveraging GPT 4 to critique GPT 4. So what they're doing is they're taking the outputs of the regular model and sending it to the same model, asking it to find errors and issues with the original content that was created. And it's actually doing a really good job. So the blog post shows examples of mistakes caught by GPT 4 from GPT 4, such as incorrect calculations and misinterpreted context and outdated information, as well as hallucinations and inaccuracies and biases in the original content that was created. This is probably going to be a built in mechanism in most or all large language models to reduce hallucinations to a bare minimum. I don't think we can avoid them completely, but this will dramatically reduce the number of errors and issues that we have with the content generated by these large language models. Now, what does that mean as far as cost and energy and so on. That's a big question because basically every prompt you run is running twice. So it will be interesting to see how they overcome that issue. Maybe a lot of it is be done through training or through some other mechanisms that will reduce the doubling of the cost on every prompt that we run. From open AI to another giant that we don't talk about a lot. And that's Databricks. Databricks has been around as a data and AI company, and they'd be doing a very good and successful work in the corporate world, allowing companies to dump their data into their databases and to engage with it through different AI tools. And they're now announcing a new architecture. They're called compound AI systems that they're claiming can challenge the dominance of open AI and Anthropic and Gemini from Google. What compound AI systems is it combines multiple AI models and techniques to create more powerful and capable systems than just a single model. So think about how a large complex tasks is done in a business today. There are multiple people doing multiple things because they have different expertise and this works in exactly the same way. So basically there's multiple models and the task moves around from one model to the other to do more specialized aspects and type of tasks in order to complete the barrier task, they're claiming faster, better, and more efficiently. The other benefit of that is obviously cost flexibility because you can pick specific models for specific tasks that was going to be cheaper than running the same behemoth model if you're running, let's say, Cloud 3 Opus. And this way you can save both time and money. money, electricity, and so on. So I think it's an interesting approach. I think in the future, that's what we're going to see from more or less from everyone. And obviously we cannot pass a week without talking about Google. Google made a big announcement this week. Google has released Gemini 1. 5 flash and Gemini 1. 5 pro models to the public. Both these models were released previously to a short wait list of companies that were able to use it. But now they're opening it to anybody that are Google Cloud customers through their Google AI studio in Vertex. And what these models are, the Flash 1. 5 model has a relatively short context window of 8, 000 tokens, but it's designed for fast, efficient processing of shortened text inputs. So anything that needs a quick response and is not too complicated, this is perfect for it. And the other one is 1. 5 pro model that has a 2 million tokens context window. So 1. 5 Pro previous model had a million tokens context window, which was still the largest in the industry. The second one being Claude 3 with 200, 000. So that was five X and now it's 10 X the next model. So it allows to run a huge amount of data into a single prompt and keep it consistent and coherent and get really good in detail results from that. I am certain that a lot of people will jump all over this for more complex analysis of company data. In addition, they have announced they're releasing GEMMA 2. GEMMA is their line of open source models. And GEMMA 2 is a multi modal AI model that can generate and understand images, audio, and video enabling the analysis and creation of sophisticated real world kind of scenarios. And as I mentioned, that's an open source model that anybody can manipulate and use. But maybe the biggest announcements from Google this week is that they are more or less forcing Gemini into workspace apps like Gmail, Drive, Docs, Sheets, Slides, et cetera. Meaning previously you had to sign up for it and activate it in order to have it available. And now they're allowing anyone to get access to this. It's just about a matter of getting the licenses and activating them. And these tools will be available on a back on a panel on the side of all the apps that we know in Google. As I mentioned, such as slide and drive and docs, etc. And it will allow people to analyze the existing data, summarize it, create calculations, et cetera, et cetera, everything we know from large language models without leaving the native apps themselves. We saw the same exact thing done by Microsoft in their Office 365 environment. So I expect this competition to continue going and develop more and more capabilities and integrations that will not only allow you to work within one of these tools, but actually to share data and knowledge across all of these. So you can ask general questions about your business and find anything you want, regardless of which application you were in. And there's zero doubt in my mind, that's the direction it's all going. Now, speaking of new models and capabilities and who's leading and which one is better, Hugging Face, which is the biggest community for AI platforms, has launched a new leaderboard to track the performance of open source large language models. So they call it OpenLLM Leaderboard and it evaluates open source models across multiple benchmarks, including language understanding and content generation and reasoning tasks, etc. And the interesting thing is that the leading models on the leaderboard right now are all from China. So models like Wudao2 and Pangu Alpha and CPM2 have all achieved really high scores outperforming the open source models that was released by OpenAI and Google. So the competition between the US and China or the West and China, if you want, is definitely burning. And I don't expect that to slow anytime soon. Now, speaking on interesting tools and releases, Figma, which is the world's most popular collaborative design platform, has announced a significant redesign that integrates AI to more or less everything that they do. So the AI powered features allows designers to complete various tasks that previously were manual, using only natural language. And it also can make a lot of suggestion and automate a lot of processes. It can do things like create images and icons and illustrations, but it can also create and suggest complete designs and color patterns and palettes and typography pairing for the relevant style that you're trying to get, and really helping designers be more creative a lot faster. Now. In addition, you can upload an existing design and figma. I will give you insights about how accessible this website is going to be. Is it built well for responsiveness? What's the consistency across multiple pages and aspects of the website and really help with creating a high quality, consistent design across everything that we're doing. This is not a surprise that we're talking about this for a while, but now it's released and available. And I expecting that it's going to drive a very dramatic change in the way websites are designed, how they're Mostly because Figma are a huge company that has a huge distribution with, as I mentioned, most of the web design is done on that platform today. And so we will see dramatic changes in that industry as well. Speaking of companies that has all the right components in order To deploy successful AI tools, McKinsey introduced this week a comprehensive suite of 20 plus AI tools within their environment developed by their in house R& D team. These products combine advanced AI capabilities and machine learning techniques combined with a huge amount of data and domain expertise that McKinsey has gathered through the years by helping thousands of companies across multiple industries run their businesses more efficient. These tools are going to be available on McKinsey's Horizon platform, which incorporates best practices and risk guardrails for security and privacy. Now, going back to connecting this with Figma, both these companies have the three things that company needs in order to be successful in the AI era. They have a huge I'm out of data that is proprietary that almost nobody else has. So on the McKinsey side is all the data that they've collected from working, as I mentioned, with thousands of companies for generations. And because both these companies has the three things that are required in order to be successful in deploying AI tools on a large scale. They have the data. They have the resources to train the models and build the right technology. And they have the relationships and the distribution with their target audience. So in the Figma example, they have the data from millions of designs and the relationships with the designers themselves. And on the McKinsey side, it's they have the data from years of supporting thousands of companies across multiple industries. And obviously the current relationships that they have with their clients in order to embed this and benefit from it. So I expect both these tools to be highly successful and help both these companies grow faster and provide more value to their clients. And now two news about the creative world. So the first one, which is cool and interesting is Toys R Us, apparently the company still exists somehow in some shape or form has collaborated with a creative studio called Native Foreign to produce the first ever brand film using open AI's. Sora tool. So those of you who don't know, Sora came as a complete surprise early this year, I think February as an incredible tool to convert text to video. So short proms turned into a minute long, high definition, highly consistent videos that is very hard to differentiate from real life. That being said, open AI never released this to the public. They only gave access to people from the industry in order to evaluate and learn and improve it. And now we're seeing the first example of what can be done with it. So the movie called The Origin of Toys R Us tells the story of their founder, Charles Lazarus, and his vision on how he thought he could transform, and did for a while, the toy store industry. And the video is actually a pretty cool, cute video. I suggest you watch it. It's a very short video. And it's featuring the iconic mascot, Jeffrey the Giraffe. Now This film was premiered on the 2024 Cannes Film Festival, showcasing the innovation in technology and the capabilities to do that. And the really crazy thing about it is that using Sora, the film was created in just a few weeks. So what they were able to do is to generate hundreds of iteration shots and then edit them together into one cohesive film in a time frame that it was absolutely impossible with legacy systems. And I mentioned that before, I think that's the direction that we're going. I think we're going to see an explosion of creativity and a lot more video content than we've ever seen because anybody will be able to create really high end videos with AI. I'm predicting that by the end of this year, but worst case scenario, Q1 of next year, these AI tools will be 100 percent consistent, will create content that will be indifferentiable from real life or from traditional cartoons. And will allow the users to edit and control the camera views, the zooms, the transitions and even edit stuff after it's created and change things in specific scenes and hence replacing a lot of the manual work across the entire film and video generation industry. And that's going to be a huge revolution that has significant implications to people in that industry, but way beyond because, as I mentioned, the amount of video content will be exposed to is going to be insane. And the other piece of news from the Creative World Universal Music Group, Sony Music Entertainment and Warner Music Group and a bunch of other major record labels are filing a lawsuit against the AI generation companies like Suno and Udio. Both these tools enable to create really high end music from just a short prompt and you can choose your style and genre and what the topic needs to be about. And it creates the lyrics and the music and the singer and everything else. And it's absolutely magical. And I love it. But these companies are arguing that both companies, Suno and Udio are trained on copyright information, which they do not have the rights to train on. Now they've shared several examples on how they can generate songs from these tools that sound very much like existing songs. The examples included Chuck Berry's Johnny be good. And Mariah Carey's, all I want for Christmas is you and other similar songs. But here's where we are. we've seen lawsuits like this against all the big players, and this is just another variation of this. Now, one thing that is similar and one thing that is highly different. The things that is similar is question number one. Can they prove it? Can they prove that it was trained on this data? Don't think they can claim that based on assumptions. Oh, this sounds a lot like X, but a lot of songs that people create sound a lot like X, which leads to item number two, can they prove that this is not fair use? So in all the previous corporate infringement cases with AI, What the A. I. Companies claim is that training on the data is not a problem because this is what artists do. So one artist listens to music. He gets ideas to a new kind of music, and then they create the music and you can sue them for the new music that they created, despite the fact that we're listening to another piece of music. So that's the claim that all these companies are making, and the courts will have to decide whether training on data is fair use or not. Not once they decide that on one case, I assume this is going to be a domino effect, and that will trickle to all the other cases. Now the third question is, can they prove damages? if you're suing somebody to get money, you need to prove that the fact that now there is a song that sounds like Johnny B. Goode is going to reduce the amount of times people listen to Johnny B. Goode, and that will be, I think, very hard to prove. That being said, I think as a society, we have to find ways to compensate creators in a different way than we did so far and come up with completely different mechanisms in order to allow people to continue being creative with or without AI and get compensated for it without having their work stolen in five seconds and reproduce in a million other variations because the song was successful. So where's this going? I don't know, but I hope we're going towards an era where there's going to be a fair solution for that, that will enable us to experience the explosion in creativity while still allowing people to benefit from it. Be back on Tuesday with another fascinating interview that is going to show you seven Powerful use cases you need to know on how to use chat GPT in a business environment. It is an incredible episode with a huge amount of value. Again, seven very detailed use cases with how to across multiple aspects of the business from very basic to highly complex. And I'm sure you're going to absolutely love it, that's going to be released on Tuesday until then, as I mentioned, if you haven't listened to episode 100, Go and check it out. It's an amazing episode and have an amazing weekend.

People on this episode