Ā© 2024 Texti Newsletter
May 06
2024
Texti Newsletter #25: Rabbit's got a little poop, Nothing but AI Enabled Earbuds š§
|
Apr 29
2024
Texti Newsletter #24: Amazon's Genius Move, Meta-glasses - going mainstream.
|
Apr 22
2024
Texti Newsletter #23, Amazon's a fluke, Zucker + š¦ = ā¤ļø
|
Apr 16
2024
Texti Newsletter #22: AI Pin is here and Udio is the deal breaker
|
Apr 08
2024
Texti Newsletter #21: Black Jack
|
Apr 01
2024
Texti Newsletter #20: All about video
|
Mar 25
2024
Texti #19: Meet IRL Professor X, Sam and Lex on a life trip
|
Mar 18
2024
Texti Newsletter #18: Devin is Killing jobs, Robots are near.
|
Mar 11
2024
Texti Newsletter #17
|
Mar 04
2024
Texti Newsletter #16: Incredible next generation AI images are HERE!
|
Feb 26
2024
Texti News #15: Stable Diffusion 3, Deep Fakes to Back to School
|
Feb 19
2024
Texti #14: Sora and Gemini these are the topics of the week.
|
Feb 12
2024
Texti Newsletter #13
Texti Newsletter #13 |
All righty, 2 days until everybody falls in love, or breaks-up again because the present is not love-able enough, are you ready? The highlights of the week:
Boeing + AI (āļø + š¤) = š¤TLDR: Boeing has a patent to create an AI agent to help inspectors find malfunctions in their planes. Long story This week Boeing decide to make love to their engineers ā¤ļø, results were not let to be awaited for long, they are now officially pregnant. They filled a patent, to create an AI agent, that will be used to help mainly inspectors to detect malfunctions or elements that need attention in their airplanes. A trivial job like a plane inspection might be boring and very dull, it's a routine after all, but when it's same thing again and again, you get the eye fatigue. As you know a robot is never tired ā Boeing's plan is to make a robot that will do exactly that. I think that's super cool! ā A company seeking to replace humans is meant to fail, a company seeking to help humans, is meant to succeed. ā Peter Thiel Instead of jumping into, "...and then this AI will replace the human that does the job! š¤¬", I firmly believe we are not there yet, and we won't be anytime soon. During the industrial revolution, in the 1900s, many people lost their jobs, mainly because their job got either obsolete, or automated. However instead of old jobs, new jobs appeared. Machines are smart, powerful, and crazy good, but still all of them need human interaction and guidance. Jobs became more intellectual, more skilled, and more safe. This ultimately increased the human expectancy to life, pay, and intellect. Was that a bad change? oh initially yes, later hell, no! Otherwise we wouldn't be able to waste our life scrolling through TikTok. Duh! Back to the quote, I honestly and from the depth of my heart believe, that all AI tools that are now born on a daily basis, are here to aid, and improve humanity. We are not going to destroy ourselves! ...only AGI might. Brain2MusicTLDR: Google Software Engineers collaborated with a group of researchers from Singapore, and converted MRI waves into audio waves using, MusicLM. Long Story: Last year a group of researchers from Singapore analyzed brain activity of a human, while showing him an image. It seems that the brain has a certain combination of neurons active while watching images, and that could be converted into a dataset for the AI to learn and try and replicate. The result were somewhat close, not to say exactly close, but good enough IMO for the first version. Google Researchers saw this and they just launched their MusicLM, which basically converts text to music, you give it a prompt and it will try and create a song based on your input. As they thought why wouldn't we try to do the same thing with music. As in let humans listen to music in an MRI machine, and we'll prompt the AI to replicate the music, based on brain activity. The results are I'd say in a testing phase, as it has similarities and but it's not there yet. You can check the video presentation bellow. Demo at minute 3:19. āI know this is in it's incipient days, but imagine the DJ's of the future thinking about music, and then making a sample directly with their brains, no manual tuning just brain it out. As said, in the previous topic, I truly believe AI will empower us to do things we have only dreamed about. 150 years ago we dreamed of flying like birds, tomorrow we dream of talking to the birds, and ask them to shut-up in the morning. P.S: I know you're curious to try out text 2 music. And you can do it for free here (https://huggingface.co/spaces/facebook/MusicGen) Tips of the dayMeta published their guide to LLM and how to run them locally but also they shared the prompt engineering guide along the way. You can find it here. In essence though here are the main take-aways:
ā Britney Spears and the Ballad of Bin Ladenās Beat In a world veiled in secrecy and shrouded in shadows, the most feared man on the planet, the sultan of subterfuge: Osama bin Laden, harbored a clandestine secretāa secret so audacious, so out of character, that had it been known, it would have sent shockwaves through the annals of espionage and international intrigue. This is the tale of how Britney Spears, pop queen extraordinaire, unwittingly became the muse of the mastermind behind the most labyrinthine of lairs. Our adventure begins with bin Ladenās unexpected encounter with the irresistible draw of Western pop culture. As he clandestinely surfed the airwaves, evading spy satellites and coded broadcasts, his ears caught the siren song of Britney Spears. Her mellifluous voice pierced the veil of his austere existence. With every āOops!ā and every beat that demanded he āhit it one more time,ā his heart beat in syncopated rhythm with the music. The adventure took a twist when bin Laden, a man cloaked in the garb of terror, began to cloak himself in fandom. He amassed a treasure trove of Britney memorabiliaābootleg cassettes, rare posters smuggled across continents, and even a trove of glittering paraphernalia that spangled under cave lanterns. His devotion to Britneyās dance anthems was such that he commissioned a forbidden satellite dish, perched precariously atop his mountain fastness, all to catch glimpses of Britneyās spellbinding performances. In the throes of his Britney-inspired reverie, bin Laden orchestrated an operation so bizarre it was deemed fictitious by his own confounded cadre. He dubbed it Project Pop Princess, and it was here that the adventure took a comical turn. By night, the cavernous chambers of his hideout echoed with the sounds of Toxic as bin Laden, donning a makeshift headset, attempted to emulate the pop iconās legendary dance moves. Choreography was practiced, rewound, and practiced again, all under the guise of a new form of combat training. The comic caper reached its crescendo when word of bin Ladenās obsession leaked to the worldās intelligence agencies. Spymasters and codebreakers were left befuddled, pouring over Britneyās discography for potential messages. Was āStrongerā a call to arms? Was āPiece of Meā a taunt to the West? they mused, as they scoured every lyric and every beat. Our adventure concludes with an unexpected twistāa twist where pop culture and geopolitical tensions collide and coalesce in the most fantastical of ways. As Britneyās music infiltrates the airwaves of friend and foe alike, the beats begin to bridge gaps and build bonds. Armies lay down their arms, secret agents dance in unison, and the world unites under the banner of Britney. And thus, in the pages of our comic adventure, Britney Spears is immortalized as the unwitting harbinger of harmonyāa pop star whose catchy choruses delivered an encore of peace. This is the legend of how the Princess of Pop swayed the heart of darkness, not with might, but with the magnetic pull of melody and the power of an adventurous spirit. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Love & Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Feb 05
2024
Texti Newsletter #12
Texti Newsletter #12 |
When I hear February I hear shortest month of the year, I hear 29 is once in 4 years, and this year is one of them, and I hear love is in the air! āThe highlights of the week:
Neuralink Implants is in the first HumanTLDR: Neuralink successfully implanted a micro-computer into a human brain this week. Extended: If you're not aware of Neuralink ā it's another Elon Musk's baby company, that aims to push us into cyber space and closer to cyber punk reality. The sole purpose of this company is to create micro-computers which can be implanted into human brains. Micro-computers aim to assist people in controlling computer and mobile devices with the power of their thought, but also help disabled people with their illness. This week the first implant operation has been done successfully and have granted a human super powers ā that isn't proved yet as the patient is still recovering and twitter is a little silent about it. But the chip won't be limited to this abilities, the plans are really grandiose and promise to allow humans to learn languages and other things in minutes through the chip. Just imagine how cool that would be, no more job interviews, no more years of training, no more lazy days, and deprivation because you can't understand the how the multi-verse works, just download it from wikipedia, and voilĆ” you're smart enough for any company! Basically 2 things will be left:
āMobile Subsecond: Diffusion from GoogleāGoogle just announced this Friday, that they are preparing a project that, will allow to generate images on device with their relatively small, for current standards, model with just about 520 Million parameters. It promises to generate images in a relatively low resolution, just 512x512px, however the kicker is that this is going to be done on your device, without the ping of the internet and the bomb is it's going to be done as you type! Honestly I can't wait to replace my abusing usage of giphy with something something that generates images on the fly. Seriously though, I think it's gonna be a fantastic tool, and will allow people to increase their prompting abilities exponentially, simply because you can literally polish your image as you type. You don't need to wait 30seconds to regenerate your images, and experiment with them, or even worse pay subscriptions that limit you in your usage, because you're using too much hardware! Damn this is annoying. *Cough* Chat-GPT *Cough* Anyways, I was saying that this could make your life better, but you'll need a high end device for that capable of running mobile diffussion models on your processor. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://newsletter.texti.app/ā |
{{ address }} |
Jan 29
2024
Texti: Google Lumiere, OpenAI's own media chips and how to Dall-e 3.
Texti Newsletter #11 |
First month of the year is almost out! How's your new year resolution progress going on? Still on the starting point? Well my advice is ā stop procrastinating, just do it. The highlights of the week:
āGoogle LumiereāTLDR AI Video generator that can:
Extended Version As promised at the beginning of the year, 24 is all about videos. You all know Google ain't gonna sit and watch how everybody's kicking their ass. They have the best fucking engineers in the world! So what was actually announced: Text-2-Video Exactly as it sounds, write a prompt and generate a video out of it. Just because I've previously tried Pika, and well to say the least it was something really well experimental rather than actually useful, my hopes aren't really high on this one. Image-2-Video This one is new to me and I didn't see this one yet. Upload a pic and make a short video out of it. This is literally live-photos on your iPhone or Motion Photos on your Pixel. I do keep my live photos on at all times, but it's only because it can capture hilarious moments, and add context. AI can't do that ā so to me I'm a bit unsure about the potential of this feature. Video-2-Video This one I'm most stoked about to be honest with you. I've seen it in practice in a different startup, DomoAI. Source Origami Style For example this is how Domoai is converting the videos today, and you can surely try it out as well. Source Anime Styled Is it perfect? Not by any means, but hey it's already looking nice. And it managed to preserve all the shapes and things I actually wanted to see. Let's see how Google's team will show-off their take on this. Fill-the-gap This one might be able to help you out and fill in your old pictures and vids where you want your ex removed from the photo - because he fucking sucks and must burn in hellā¤ļøāš„! Here's a little less innocent example provided by google. We can see here how the AI managed to fill in the gap and maintain the motion. Now this - is more of an interesting and useful purposeful usage of AI. Also isn't this dog a cutie š„° š„ŗ Sam Altman to join the AI Hardware raceTLDR: OpenAI are looking for more hardware for AI, and are considering TSMC (Taiwan Semiconductor Manufacturing) to produce their own branded chips to stay competent in the current market. Follow their stocks it might hit the sky soon. Not so Extended Version There is no doubt that the current gold rush is AI, AI is also powered by hardware and it order to make this parrot speak all you need to do is get those powerful chips. Any big tech company either buys the chips commercially form Nvidia, making em a big pile of cash, either build themselves like Apple and Google does. OpenAI, noticed they have a gap in their business plan, if they continue to expand they're going to run out of free money eventually, but by the time they need to be established on the market. The next logical move is to build their own product, to become fully self sustained and independent and keep pushing the prices they want. ā To build a great company, you have to be a monopoly and dominate the market. ā Zero to One, Peter Thiel Which is exactly what OpenAI is slowly crippling towards. Building a monopoly, so that they could control all the aspects of this industry! Well done guys! Tips and tricks for image promptsToday's tips and tricks come from Chase Lean, the author My Logo Creator GPT, that I was praising in the last letter, the ice-cream generator. He wrote a nice twitter thread about this. OpenAI's dall-e works a little more different than the standard prompt generators out there like midjourney.
That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Jan 22
2024
Texti Newsletter #10
Texti Newsletter #10 |
We have our mini jubile, it's the 10th edition of this newsletter. I'd like to welcome and be grateful to everyone who joined in the last 10 editions. You are the ones I make this for, and you make this a reality. Thank you. š The highlights of the week:
On Jan 20, in a short video posted on his instagram, Mark Zuckerberg said that he is preparing a new intergalactic infrastructure for the world. He wants to make it available for all developers out there. He said that Meta will make available to everyone another 350k Nvidia H100s, summing up to a total of 600k ($35k a piece) which means they'll make available to everyone about $21 Billion worth of hardware. What the internet has taught us over the years, is that there's no such thing as free. There's always a catch, and the problem is that it ain't clear really what is Meta's purpose here, they promise AGI (Artificial General Intelligence), but skeptics say we're still years away from it, I'm on the skeptic's side, independent computer thinking is still very far from a reality. Current AI Models are purely optimal guesses which means it can't make decisions independently, it's just tries to copy existing information. Coming back to Meta's hidden purposes, here are my theories:
Ultimately these are just speculations, I have no idea of the true purpose behind the generosity of Meta, maybe they don't even know it yet on their own. However, I doubt it. Zuckerberg is and always was a genius, he always had a plan for facebook, it did evolve with time, but he build his startup correctly ā he started with a local community, and then slowly expanded it across the world. He always had a purpose and he always knew where he was going with it. But man he's a weirdo. Samsung's new AI SaaSThis week right after the CES 2024, Samsung in it's usual habit, has presented their new flagship phones. Samsung couldn't stand the temptation of AI, so the flooded the presentation with AI! Let's dive into the AI Features:
I am trying to be truly objective here, Samsung did an Apple thing. Same - same, but different, where they literally upgrade the cameras a bit, and add a bunch of buzz words, and say they've built the best Phone ever. The kicker, Samsung has the audacity, to convert your phone features into a SaaS! On their website, all the AI features are marked with an asterisks. Life taught me, if you see an asterisks, find the worst visible text in the page, and read it out loud! It'll likely f*ck-up your life. ā * Galaxy AI features will be provided for free until the end of 2025 on supported Samsung Galaxy devices. Different terms may apply for AI features provided by third parties. This likely means the models don't run on your local hardware, but rather in the cloud, and if you're on low connectivity all your AI Features, are just unavailable. SaaS businesses make this world a worse palce. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Jan 17
2024
Texti Newsletter #9
Texti Newsletter #9 |
Holy this week was rich in events! CES, Rabbit r1, GPT Store, I launched my own mini gpt to compete and conquer the entire store š. Texti Newsletter got it's own page where you can read through out previous news!ā The highlights of the week:
GPT StoreAs I have told you already last time, GPT Store was about to launch this week, and it did! On January 10th, the one and only Mr. Sam Altman, tweeted that the GPT Store is live! The show is on. Now everybody can build their own mini-gpt and publish to the store. I couldn't stand aside and tried a couple of them. Some of them are genuinely fantastic, while some are just pure garbage, but that's to be expected, ultimately normal distribution predominates everywhere! For example Consensus, this one I think is a masterpiece, and I'm jealous I didn't come up with this idea. It is basically a context aware search throughout research papers. Why do I think it is super cool? Well it finally answers to questions that everybody craved an answer for, but were too lazy to research. For example: Question: Is it healthy to eat salt? If it depends, how much salt should a human consume on a daily basis? Isn't his amazing? Now I can show throw that apple into my doctor's face! š š§āāļø Now you could ask question that burned you for years, like:
And many more, if there's a research paper, then there's an answer. ā Next one on the list is Logo Creator. This one is super nice, it managed to create a logo as described, and even with multiple variations and requests it obeyed to all my orders, make it bigger, make it smaller, make it lighter, make it simpler etc. It went through 4 iterations until I was happy with the result and it did well. It is a cool tool and can replace a junior brand designer, if you know what do you want, if you don't, well then a human is required. ā āScreen Shot to Code is the next one, and oh boy it sucks. It literally created a sample boilerplate website that looked always the same. You'd better use any free templates available on the web, than this tool. Maybe one day, it'll be useful, but currently it is an absolutely unusable piece of wasted internet. ā Rule 34 and the Internet. What would be the internet community without trying to make porn out of everything that exists in this world. Well not long awaited was the moment until gpt store was flooded with virtual girlfriends, and everyone trying to achieve the same result as achieved in the movie HER. Just look how many options are there to promote love. Well it seems the next generation won't be able to speak but type until the age of 16. I know for sure that OpenAI's policy don't allow any nudity and violent pictures, as this goes against their policy, but I'm sure there isn't too long to wait until somebody's going to create a third party image generation that will allow nudity and porn of all types. Just look how many girlfriends are out there in chat gpt. Ultimately as a classic broadway song, the internet is for porn. āSummary In summary I think the Chat GPT store launch was a success, I also built and launched a small motivational mini gpt, F-ing Great Advice, which gives you a fucking great advice. It will likely land soon in Texti.app as well. CES 2024One of the most famous and most important Hardware events that happen annually is the Consumer Electronic Show in Las Vegas. It is full of concepts, and sometimes almost real products that are there to impress you and predict what's going to appear in the upcoming years. Well this year, everything was AI related, every company release at least a couple of products that were called using this brilliant formula: AI + noun = unmatched representation of the future At CES 2024 there was everything AI.
And many more, basically AI + anything you can imagine was already presented as a prototype at CES. Generally I ain't that excited about the upcoming things, it seems that the hologram is just a projector in a case. Car prototypes are literally trying to compete over who manages to put the biggest screen in a car. The mirror is just a Philips hue with proximity sensor. The robots are pretty interesting though and I think this has the most work behind the scene which should be the more interesting thing in the long run. Lastly at CES 2024 there was a hit, which got all the HYPE! Rabbit r1 Rabbit R1Probably one of the most hyped products from this year was the Rabbit R1 AI companion. In less than a week they sold about 40 000 units and now another 10 000 units are on sale as of today January 17th. First of all, I'm extremely happy that in 2024, someone managed to build a startup that is successful, and I'd like to congratulate them! Great fucking beginning into the Year 2024. What's this all about? In their keynote, this device was presented as the next generation of mobile phones. It promises to be the much awaited revolution in the UI controls. I personally think they fired all the designers, added a floating logo, and said that this is now going to be our interface ā . As you've already guessed it's all AI based. In fact it is a voice activated interface for ChatGPT, that can do everything chat gpt can. You're wondering why do you need a device for that huh? Well they have a couple secret ingredients to make their sauce taste good. The Secret Sauce First, the Voice processing is much faster than any existing alternatives, which likely means that they have a voice processor and an LLM model on device. Based on the demo, it looks quite good, and compared to existing voice processors I've tested before this feels instant. But Large Language Models aren't new to us, we want more! Taking actions and actually doing something is the next possibile thing! LAM - Large Action Model, an invention from the team at rabbit, is basically a robot that can go to a web page, open it, find necessary buttons and click on them to accomplish an action. This sounds soo cool! It is literally an assistant that expects direct orders! Hey at least somebody's listening! In an utopian future, we won't need interfaces any more, just schematics which will be accessed by robots, to find the most relevant and useful buttons, the next level of SEO will arise and we'll design just low fidelity prototypes. In their Demo, they planned, and booked a trip to Paris, bought flight tickets, booked a hotel, a restaurant and planed three days of activities. š¤Æ Extremely promising! The future is now! I said to myself. This is not all, they also showed, a way to teach rabbit do actions he doesn't know yet. You literally record a video of doing it, feed it to the software, then rabbit r1 will know how to execute the process from now one. All of this at only 200$! Boom š„, Rabbit R1 blew up on social media, it was sold out 4 days in a row š¤. When we start breathing again š§š» As we start breathing again, and the initial buzz calms down, the hype mellows down. We start reading between the lines. Real questions start popping up:
Speaking of the founders, reddit found out about them. Apparently they are ex-crypto founders that raised a bunch of capital and never delivered a proper product. Rabbit.inc was previously called Cybermanufacture Co., which was renamed from a NFT company called GAMA. So these are just guys running on hype fuel and trying to find that jar full of gold. It seems they hit the jack pot this time, at 40k sales and 200$ per unit that's $8 million - on hands šø. The real question, will they deliver? Is the device going to be as fast as demo-d? or this is just gonna be a, oh yeah it's buggy we know, but we'll fix it in a couple of patches, and a couple of patches later they're gone with customers money and investors money. A crypto story once again. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Jan 08
2024
Texti Newsletter #8
Texti Newsletter #8 |
Welcome back to our weekly AI updates. Hope you have recovered from the food comma and you're back on track to become the next most fit person the world have ever seen! As new year resolutions are set in place we dive into the highlights of the week! The highlights of the week:
Chat GPT opens up the GPT store š¤©As promised during Openai's developer conference, the Chat GPT will have it's own store of plugins, which they call mini-gpts. Think of them as enhancers of AI's abilities, which are also specialized in a particular task, and could yield way greater results than crafting topics and asking specific questions. The greatest part of this is OpenAI's take on this, instead of taking a technology and making private and closed, they're basically "open-sourcing", taking the hardest part on themselves, which is hardware and resource management, allowing developers do what are they most effective at ā imagination. I know calling it open-source is a stretch, however I truly think launching a centralized place where you can try-out a mini-gpts in one click, and having it always at your hand is crucial. They will also destroy out the competition and will make it one of a kind place. I can easily foresee people integrating tooons of third party SaaS as mini gpts. I do believe this has the potential to make a small industrial revolution among existing software. Just imagine, ordering pizza via chat-gpt at your home. Just type here's my address, I'm organizing a little party for me and my friends and we want a variety of pizzas from this vendor. Please make sure to order it by 7pm. And you're just done, you don't even need to care about anything else. Minigpt can talk consciously with a delivery place and fulfill the order. The best part is OpenAI will share a part of their revenue with the developers, not sure yet how much, but the revenue will be shared! Money is one of the most important motivation for humans, as everything is built around them. Sharing the revenue is a killer feature, that will make a massive impact on the evolution and distribution of mini-gpts. An important step for humanity, and a great moment and unique moment in our history. Here's why:
We live in a world where everything is possible, and anybody has the ability to shape the future of our world. This is yet another place that has the potential to allow you to change your live dramatically! I live in excitement to the grand opening and can't wait to tell you what's going to be there the next week! Pope Francis has an AI advisor.I'll be completely honest with you, I had no idea The Vatican is so progressive, and has any idea about AI š¤·š¼āāļø, not only that but the Pope has had an AI Advisor since last summer, and his name is Paolo Benanti. More than that I don't really know what his job is and what does he do on a daily basis, does he pray to the AI gods? š It's super hilarious that the Pope has an opinion about AI, moreover it has a dedicated person to research it. I did find out though that Mr. Benanti has had a considerable impact on the Artificial Intelligence ethics which have been adopted by IBM, Microsoft and OpenAI. He does actually have a large impact on the world and the church never misses an opportunity to remind about itself. Last week on 2nd of January, during the 57th anniversary, of the Day of Peace, Pope Francis called for Regulation of AI to Safeguard Humanity, and it seems he does this for a while now. It seems that he's concerns are related to the potential takeover of the machines. I would also be worried about that though, Arnold is really old by now and he won't be able to carry the weight of the Terminator again. š„ š That said, ethics are extremely important topic, especially when we're talking about Artificial Intelligence and preparing upfront for truly independent thinking artificial general intelligence. It has to be fundamentally one of the core pillars of it's belief and thinking. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Jan 02
2024
Texti Newsletter #7
Texti Newsletter #7 |
Happy New Year Everybody! Welcome to a new challenge and a new year! It's officially 2024 almost everywhere (China sorry you live in a different dimension). I am really excited about the following year! We have grandious plans for the following year and we look forward to it! From everybody at Texti, we are wishing you a sparkling New Year filled with bright moments! May the coming year wrap you in the warmth of love and the joy of laughter. Here's to new adventures, health, and happiness in 2024. Happy New Year! šāØš The highlights of the week are a little short (everybody's on vacation)
āVideo PoetāAs predicted, this year is going to be all about text 2 video. Google is not ready to stand aside, and it has been in the works for a while. Google has posted their own research paper on Text2Video, unfortunately no hands-on yet, but there are plenty of examples on their web page. They even a pretty cool demo where present a little story about a raccoon and his travel across space via portals. The major differentiation between Video Poet and competition like Pika, is that it is meant to produce a continuous stream of video, with a narrative and an actual story line, rather than an arbitrary image. As Google describes it, it's a state of art model that uses under the hood a bunch of other individual components like MAGVIT V2 for video generation, SoundStream for audio matching and an autoregressive language model to orchestrate all of this together and provide a unique content format. This indeed looks like a game changer for the media creation and future tiktoks. As described it currently targets short content format. It even allows you to add stylization, in video modifications and generate video from static images Ā š¤Æ. All we need to do is just work on our imagination to create original content ideas. I think we're getting closer to fully self-sustained AI generated content creators. At the beginning of last year somebody said that in the next 5-10 years, software engineers won't exist. In 2023 we were able to make a bot do anything just with a set of simple instructions, simply by using Auto-GPT. By 2025 we should be able to create a bot that will create short content form videos, targeted to every specific ethnicity, that will be fully automated, follow the news and current trends. I think this is already wild - the future is today, we live it and we just need accept that this is here and now, and enjoy the really unlimited possibilities we currently have! The New York Times Sues OpenAI!It ain't a new thing that OpenAI has been training it's data on the free data available on the internet. Ultimately all this information that AI is aware of, is just an approximation algorithm of words provided by humans, and written over and over again, in the last couple of decades we published on the interwebs. That's why Reddit API is now paid, that's why OpenAI has had conversations with many publications on paying for their models to train their dataset. Apparently it seems that the discussions haven't had any advancement since August, because The Times and OpenAI couldn't agree on a specific amount to pay (hint this seems the root cause of the scandal). The Times complains that OpenAI is making and raising billions while they produce the raw material that has to chew up by a language model. Note to say, OpenAI has been signing similar agreements with a multitude of media companies out there, to make it right for them and fair or at least somewhat. This is the first major lawsuit against OpenAI, and I'm extremely curious to see how this is going to end out. After-all OpenAI promised that they will protect and step-up for every company that has been using OpenAI's API and has had a law-suit filled against them. That's exciting and interesting, as this is going to be a really decisive moment in the history of AI. If this is going to be solved with money, the decisive moment will be postponed for a while, otherwise, we'll know how the further evolution of large language models (LLM) will follow through. Stepping out a little, and looking at the image from top, we can see a classic battle of the industrial revolution, a business that has been at it's peak about a hundred years ago, fights against the new emerging technology, that will replace it. The fact that The Times complains in clear text that OpenAI's raising money is a clear sign they don't understand the Open AI's investors view over the company and why the times doesn't get money poured in. The Times's value is clear as day, they hit their ceiling, Open AI's just starting. We have yet to see it thrive! Your Go-To Topic Tip for Daily Discovery!I love Texti.app and I love generating content with it, however I often encounter issues and it has limited abilities so it keeps answering with answers like:
This is really annoying and it makes the answer really long and pointless from time to time, especially because I don't really need this answer. I need to get straight to the answer. Well there's a really cool trick you can do. If you apply the following topic to a conversation AI model will make sure to answer the questions way better. If you're using Texti.app you can apply the topic to any conversation by clicking on the cog āļø in the chat menu, and selecting General -> Everyday Conversation. Or alternatively just paste the following in your topic area:
That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Dec 25
2023
Texti Newsletter #6
Texti Newsletter #6 |
Merry Christmasšš š· to everybody! This is by far the most beautiful time of the year in the Northern Hemisphere! People are drinking hot wine outside, eating all sorts of delicious and winter exclusive foods! What a beautiful time to be alive! With this occasion we decided to celebrate and warm your house with a unique discount for subscribers. Use the code MERRYCHRISTMAS23 at checkout and get 50% off next 3 months on any of our products! The highlights of the week:
Beta launch of Midjourney v6Twitter is exploding with Midjourney v6's launch! If you browse twitter just a little, you can see everybody show-off their ultimate skills at generating art with Midjourney v6. As claimed by the developers itself in their blog article the capabilities of the model are exceeding their own expectations. I don't think anybody was ready to see this coming from an AI model yet, it is still just a noise vector approximation, however with some clever algorithms and the right prompt the results are indeed pretty fascinating. In my opinion the key difference here, is that you reallyĀ don't need to use the obnoxious keywords we used to mandatory add to the prompt for better results. Finally no more 4k, ultra-hd, high-res, hyper-realistic, ultra-high-super-duper-puper-muper-grouper-high-resolution-instruction. This noise really is annoying and you focus on the prompt itself and forget the keywords, this really is annoying š In case you were wondering if the model managed to learn how to generate text, the answer is somewhat. As described in their own article, you need to get a little lucky and also make sure to include the desired words in "quotes", this way the algorithm understands which words needs to be properly included and written on an image. On the example bellow, you can see how to stylize the font in the generated image. This is getting closer to an a photo editor that used to take time and effort, now it's just a matter of seconds. Of course the designers have a place in this world, but their niche is getting narrower day by day. Example:
I you look down further in the blog post Midjourney has posted, you can even instruct the AI copy styles of text or image, or even replace the style of text, that is actually extremely valuable because it allows you to consistently edit the desired output and even combine results it for perfect image. This alone is insane, and creates a high amount of possibilities. My only concern is: will it use licensed fonts? And if so, how many cases of font copyright infringements will there be. To my eye this is very impressive, and midjourney is getting now the king of image generation market again! With the text prompting this could really disrupt the graphic designers market, as ad generation becomes way cheaper and affordable. Some go further and even say that this Midjourney 6 is a job killer. After such a demo and hands-on experience, I really must say. I agree. But hey this a Christmas edition, and I can't leave you on a negative note, so worry not, people are too lazy to generate images by themselves, select and maintain a style, if you do this on your own, you're most likely a poor entrepreneur. Designers are still going to have a job, yet their job specific is going to change, responsibilities will include preserving the style and brand of the company. Job perspective is going to be rather analytical than artistic, and I think this really interesting. Hands on Pika.ArtāThis week I've got invited to try out Pika Art, I was telling you about in the previous newsletter. Hint, you've already seen an example of the results produced by it. Just after the Christmas congratulations of this email. Yes that is purely AI generated. After generating about 50 videos and experimenting with all sort of prompt variations, I can safely assure you that our guess was absolutely correct. Video Generations is literally generating about 72 sequential images one by one, and combining them in a 3seconds long video of 24 FPS. Is it any good? It's better than nothing š , but the results may vary a lot from case to case, it's closer to a DALL-E 2, level than midjourney v5. Images are not very crisp, details are a little lacking, and artifacts are very frequent. It obviously struggles with fingers, text and for video there's a new dimension to struggle about -- motion. We tried 9 different prompts to make a wheel rotate, and it failed to do it, see the list bellow Prompt: car wheel rotating very fast close-up Probably the best result we had, was the following video, where the wheels are barely visible, and the give you a sense of rotation. Prompt: ford mustang, moving sideways and smoking rotating rear tires on a track at the sunset. take a 3 quarters angle from the back of the car Probably the best motion result I've got was making this bunny turn around, although the input was different to the output Prompt: In an enchanted world, create a high definition, high resolution video capturing the magical essence of a mythical garden. The star of this scene is a tiny, fluffy bunny with sparkling eyes and a twitching nose, full of energy and curiosity. This bunny is not ordinary, its fur shimmers with an array of pastel hues under the sunlight, reflecting the captivating charm of its surrounding. However other prompts yielded worse and way more unpleasant results. Check out this surfer: Prompt: cinematic short film, about a giant wave of nazare, portugal, with a surfer riding the wave. Or this dolphins Prompt: cinematic film of many dolphins swimming in a blue lagoon In most of these examples we can clearly see that there are algorithms working on creating sequential images, however motion is still lacking. There's plenty of improvement yet to make, but at the moment this is still an alpha test, rather than a ready product. The output is far from consistent and it really depends on the input and the user skills. I am not the best prompt engineer out there yet, but I'm learning and will guide your thought it. Let me know if you're experience differs. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! ā Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Dec 19
2023
Texti Newsletter #5
Texti Newsletter #5 |
Christmas is around the corner, this means everybody's wearing a funny Christmas sweater, and everybody's rushing to complete all the project in time. Even if you need to cut the corners. The highlights of the week:
āPika 1.0 is officially the new head turner!If you didn't know what Pika is, it's the new kid on the block! It turns head like no-one else. You tell Pika: time lapse photography of a rose bud unfurling, blooming, red petals expanding You tell Pika: a lion fish swimming underwater Pika: a cinematic film still of a breaching whale, falling down into the water, splashes, sea foam Well isn't this impressive? It sure is, but that was to be expected. A video is just a series of pictures following a little progression with each frame. It was just a matter of time until we learn how to make pictures progress. To be completely frank with you I think it's not the how to do it, but rather drop the cost. Currently image generation still costs a fortune if you're not a billion dollar company. NVIDIA is selling their hardware for hefty price tag and it's still heavily unaffordable. Yet everybody's predicting that the next year it will be the year of Text to Video, which makes everyone super excited, but at same time the price tag for this will be pretty astronomical. Gemini Demo was all a lie!Last week, Google Announced Gemini, we've talked about it in the last Newsletter. We did say that the Demo was mighty impressive, and I still stand to this words! It was really impressive, so impressive in fact that it raised questions to many! To be fair, Google said at the end of the document where they presented the demo, that there's an in depth analysis of the demo -- in a separate document how they actually have build every piece of the demo, however, the demo looked flawless and all in real time. Now everybody got really disappointed that this presentation box, was kept together with tape and not actual structure. But worry not, the internet is broad, there are a bunch of nerds out there who can react promptly to this. Here's one: A guy called Greg, just remade the Google Demo in real time, with OpenAI. This is a closer representation of the current state of AI. It first needs time to analyze and process the information it has, process the images, convert it into text, and then convert it into speech, hence, you can see in the video that there's still a substantial delay in the communication between the AI and Human. Still is very impressive, and the days were life's is going to be visually assisted is really close. You can watch it right here: āFor now we can still say OpenAI is the one and only King of the current race! Sam Altman is still holding that crown really tight on his head! Hint of the week š”If you are struggling with the AI responses, and AI still can't give you a proper answer, there's a simple solution. Split the task into smaller pieces, and feed them one by one into the chat bot. AI is dumb, even if it seems smart, complex tasks are hard to manage for him, as it's lacking background knowledge, and can't see the big picture. If you feed it granular tasks, it will accomplish them with large success. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |
Dec 09
2023
Texti Newsletter #4
Texti Newsletter #4 |
Holy this week is on fire! Google doesn't know how naming works, Meta launches ImagineAI that learned from your personal pictures posted on facebook and instagram. AI Alliance wants to keep security tight on AI. The highlights of the week:
Meta Launched Imagineā It was just a matter of time until Meta was about to launch their version of image generation models. Well This week they officially launched Imagine which is an AI trained model that can generate images for you. Interesting to know is that it is gradually integrated in every of their own products. It'll soon be available in instagram, messenger, facebook and they are planning on allowing customers to extend their landscape images, to generate vertical content for stories. What is curious to know is that Meta will likely use the images dataset they have uploaded during years to their own platform. So if you ever thought why people post pictures of food in their instagram, now you know, these images will be used to train AI models and generate you the perfect launch, visually. We tried to access and try out imagine, but it's still not available in all countries, and for now only US citizens can access the model and attempt to play with it. But with a little desire you can still access it, just use any VPN software available and access the website and voila! The closed doors are now open. That said we have tried and ran a simple prompt:Hyper photo-realistic photo of an eastern european male, of about 30 years old. Include the full body, make sure to avoid body imperfections, however add slight hair imperfections. Which generated a decent enough result. An impressive perspective on Eastern European Males. But we must say it looks pretty decent. Feel free to play around with it. Let us know how feel about it. Note to say, we of course tried to ask the AI for nude pictures, but this keyword was rejected. However, if you ask to keep the clothing to minimum, or reduce their visibility to 0, you might get some NSFW results. I'll let you do the exploration there. āAI Allianceā You're probably asking the same question as I did. What the hell is this, and what do they want from our beloved AI. Well if you read the IBM announcements website, you can think of them like they want to be safety net of the AI developments. In a couple of words without the fancy wording they want to:
Who are they and who's going to be part of this Alliance? Well it seems IBM and Meta are the main coordinators, but they invited into the Alliance a very imposing list of companies that are trending in AI. Worth mentioning would be: HuggingFace (largest model library so far), Sony (owning a very imposing list of trademarks), AMD, Intel, Dell, Oracle, and Stability AI. Seems like the first year of AI was wild and free, so will be the next one, but control is coming over really soon. Google Launches Geminiā It seems Google doesn't know how naming works. Just about 6 months ago google announced bard, and it was claiming it's going to outperform any existing competition. This week they announced Gemini and it is promised to be extremely fast. Regardless of performance numbers, the demo looks mighty impressive!Google announced that Gemini comes in 3 flavors:
Performance numbers are also very impressive. Google says that Gemini Ultra (their high-end model) outperformed Human experts on MMLU (this is a set of multiple disciplines, as math, physics, history, law, medicine and ethics). We love to see this, finally Google is catching the competition! Given that Google is light-years ahead of competition in terms of data centers, it should be always ahead of the competition. But we can see from the image that Google didn't even compare their model to our biggest head-runner from OpenAI, GPT-4 turbo. Which obviously raises the competition is Gemini that good or are they just better at marketing? In my subjective opinion, Gemini is still behind, but what's cool is that a small start-up has turned upside down the industry in just a year, and made the well established companies struggle with competition! Our Golden Boy Sam Altman made the impossible, made everybody stand-up and actually start working! Among other interesting announced things were. Gemini Nano is being launched on Google Pixel 8 Pro running in Recorder app for summaries, and Smart Reply in Google board. This is also very exciting, because once AI Models start running on device, this means a whole lot more for mobile devices. That's it folks, see you next week ā¤ļøļøļøļøļøļøļø Happy Prompting! Remember to invite your friends to subscribe at https://texti.app/newsletterā |
{{ address }} |