VibeBuilders.ai Logo
VibeBuilders.ai

Story

Explore resources related to story to help implement AI solutions for your business.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]
reddit
LLM Vibe Score0
Human Vibe Score1
Slight-Explanation29This week

10y of product development, 2 bankruptcies, and 1 Exit — what next? [Extended Story]

10 years of obsessive pursuit from the bottom to impressive product-market fit and exit. Bootstrapping tech products as Software Developer and 3x Startup Founder (2 bankruptcies and 1 exit). Hi everyone, your motivation has inspired me to delve deeper into my story. So, as promised to some of you, I've expanded on it a bit more, along with my brief reflections. There are many founders, product creators, and proactive individuals, I’ve read many of your crazy stories and lessons so I decided to share mine and the lessons I learned from the bottom to impressive product-market fit and exit. I've spent almost the past 10 years building tech products as a Corporate Team Leader, Senior Software Developer, Online Course Creator, Programming Tutor, Head of Development/CTO, and 3x Startup Founder (2 bankruptcies, and 1 exit). And what next? good question... A brief summary of my journey: Chapter 1: Software Developer / Team Leader / Senior Software Developer I’ve always wanted to create products that win over users’ hearts, carry value, and influence users. Ever since my school days, I’ve loved the tech part of building digital products. At the beginning of school, I started hosting servers for games, blogs and internet forums, and other things that did not require much programming knowledge. My classmates and later even over 100 people played on servers that I hosted on my home PC. Later, as the only person in school, I passed the final exam in computer science. During my computer science studies, I started my first job as a software developer. It was crazy, I was spending 200–300 hours a month in the office attending also to daily classes. Yes, I didn’t have a life, but it truly was the fulfillment of my dreams. I was able to earn good money doing what I love, and I devoted fully myself to it. My key to effectively studying IT and growing my knowledge at rocket speed was learning day by day reading guides, building products to the portfolio, watching youtube channels and attending conferences, and even watching them online, even if I didn’t understand everything at the beginning. In one year we’ve been to every possible event within 400km. We were building healthcare products that were actually used in hospitals and medical facilities. It was a beautiful adventure and tons of knowledge I took from this place. That time I built my first product teams, hired many great people, and over the years became a senior developer and team leader. Even I convinced my study mates to apply to this company and we studied together and worked as well. Finally, there were 4 of us, when I left a friend of mine took over my position and still works there. If you’re reading this, I’m sending you a flood of love and appreciation. I joined as the 8th person, and after around 4 years, when I left hungry for change, there were already over 30 of us, now around 100. It was a good time, greetings to everyone. I finished my Master’s and Engineering degrees in Computer Science, and it was time for changes. Chapter 2: 1st time as a Co-founder — Marketplace In the meantime, there was also my first startup (a marketplace) with four of my friends. We all worked on the product, each of us spent thousands of hours, after hours, entire weekends… and I think finally over a year of work. As you might guess, we lacked the most important things: sales, marketing, and product-market fit. We thought users think like us. We all also worked commercially, so the work went very smoothly, but we didn’t know what we should do next with it… Finally, we didn’t have any customers, but you know what, I don’t regret it, a lot of learning things which I used many times later. The first attempts at validating the idea with the market and business activities. In the end, the product was Airbnb-sized. Landing pages, listings, user panels, customer panels, admin site, notifications, caches, queues, load balancing, and much more. We wanted to publish the fully ready product to the market. It was a marketplace, so if you can guess, we had to attract both sides to be valuable. “Marketplace” — You can imagine something like Uber, if you don’t have passengers it was difficult to convince taxi drivers, if you don’t have a large number of taxi drivers you cannot attract passengers. After a year of development, we were overloaded, and without business, marketing, sales knowledge, and budget. Chapter 3: Corp Team Lead / Programming Tutor / Programming Architecture Workshop Leader Working in a corporation, a totally different environment, an international fintech, another learning experience, large products, and workmates who were waiting for 5 pm to finish — it wasn’t for me. Very slow product development, huge hierarchy, being an ant at the bottom, and low impact on the final product. At that time I understood that being a software developer is not anything special and I compared my work to factory worker. Sorry for that. High rates have been pumped only by high demand. Friends of mine from another industry do more difficult things and have a bigger responsibility for lower rates. That’s how the market works. This lower responsibility time allowed for building the first online course after hours, my own course platform, individual teaching newbies programming, and my first huge success — my first B2C customers, and B2B clients for workshops. I pivoted to full focus on sales, marketing, funnels, advertisements, demand, understanding the market, etc. It was 10x easier than startups but allowed me to learn and validate my conceptions and ideas on an easier market and showed me that it’s much easier to locate their problem/need/want and create a service/product that responds to it than to convince people of your innovative ideas. It’s just supply and demand, such a simple and basic statement, in reality, is very deep and difficult to understand without personal experience. If you’re inexperienced and you think you understand, you don’t. To this day, I love to analyze this catchword in relation to various industries / services / products and rediscover it again and again... While writing this sentence, I’m wondering if I’m not obsessed. Chapter 4: Next try — 2nd time as a founder — Edtech Drawing upon my experiences in selling services, offering trainings, and teaching programming, I wanted to broaden my horizons, delve into various fields of knowledge, involve more teachers, and so on. We started with simple services in different fields of knowledge, mainly relying on teaching in the local area (without online lessons). As I had already gathered some knowledge and experience in marketing and sales, things were going well and were moving in the right direction. The number of teachers in various fields was growing, as was the number of students. I don’t remember the exact statistics anymore, but it was another significant achievement that brought me a lot of satisfaction and new experiences. As you know, I’m a technology lover and couldn’t bear to look at manual processes — I wanted to automate everything: lessons, payments, invoices, customer service, etc. That’s when I hired our first developers (if you’re reading this, I’m sending you a flood of love — we spent a lot of time together and I remember it as a very fruitful and great year) and we began the process of tool and automation development. After a year we had really extended tools for students, teachers, franchise owners, etc. We had really big goals, we wanted to climb higher and higher. Maybe I wouldn’t even fully call it Startup, as the client was paying for the lessons, not for the software. But it gave us positive income, bootstrap financing, and tool development for services provided. Scaling this model was not as costless as SaaS because customer satisfaction was mainly on the side of the teacher, not the quality of the product (software). Finally, we grew to nearly 10 people and dozens of teachers, with zero external funding, and almost $50k monthly revenue. We worked very hard, day and night, and by November 2019, we were packed with clients to the brim. And as you know, that’s when the pandemic hit. It turned everything upside down by 180 degrees. Probably no one was ready for it. With a drastic drop in revenues, society started to save. Tired from the previous months, we had to work even harder. We had to reduce the team, change the model, and save what we had built. We stopped the tool’s development and sales, and with the developers, we started supporting other product teams to not fire them in difficult times. The tool worked passively for the next two years, reducing incomes month by month. With a smaller team providing programming services, we had full stability and earned more than relying only on educational services. At the peak of the pandemic, I promised myself that it was the last digital product I built… Never say never… Chapter 5: Time for fintech — Senior Software Developer / Team Lead / Head of Development I worked for small startups and companies. Building products from scratch, having a significant impact on the product, and complete fulfillment. Thousands of hours and sacrifices. This article mainly talks about startups that I built, so I don’t want to list all the companies, products, and applications that I supported as a technology consultant. These were mainly start-ups with a couple of people up to around 100 people on board. Some of the products were just a rescue mission, others were building an entire tech team. I was fully involved in all of them with the hope that we would work together for a long time, but I wasn’t the only one who made mistakes when looking for a product-market fit. One thing I fully understood: You can’t spend 8–15 hours a day writing code, managing a tech team, and still be able to help build an audience. In marketing and sales, you need to be rested and very creative to bring results and achieve further results and goals. If you have too many responsibilities related to technology, it becomes ineffective. I noticed that when I have more free time, more time to think, and more time to bounce the ball against the wall, I come up with really working marketing/sales strategies and solutions. It’s impossible when you are focused on code all day. You must know that this chapter of my life was long and has continued until now. Chapter 6: 3rd time as a founder — sold Never say never… right?\\ It was a time when the crypto market was really high and it was really trending topic. You know that I love technology right? So I cannot miss the blockchain world. I had experience in blockchain topics by learning on my own and from startups where I worked before. I was involved in crypto communities and I noticed a “starving crowd”. People who did things manually and earned money(crypto) on it.I found potential for building a small product that solves a technological problem. I said a few years before that I don’t want to start from scratch. I decided to share my observations and possibilities with my good friend. He said, “If you gonna built it, I’m in”. I couldn’t stop thinking about it. I had thought and planned every aspect of marketing and sales. And you know what. On this huge mindmap “product” was only one block. 90% of the mindmap was focused on marketing and sales. Now, writing this article, I understood what path I went from my first startup to this one. In the first (described earlier) 90% was the product, but in the last one 90% was sales and marketing. Many years later, I did this approach automatically. What has changed in my head over the years and so many mistakes? At that time, the company for which I provided services was acquired. The next day I got a thank you for my hard work and all my accounts were blocked. Life… I was shocked. We were simply replaced by their trusted technology managers. They wanted to get full control. They acted a bit unkindly, but I knew that they had all my knowledge about the product in the documentation, because I’m used to drawing everything so that in the moment of my weakness (illness, whatever) the team could handle it. That’s what solid leaders do, right? After a time, I know that these are normal procedures in financial companies, the point is that under the influence of emotions, do not do anything inappropriate. I quickly forgot about it, that I was brutally fired. All that mattered was to bring my plan to life. And it has been started, 15–20 hours a day every day. You have to believe me, getting back into the game was incredibly satisfying for me. I didn’t even know that I would be so excited. Then we also noticed that someone was starting to think about the same product as me. So the race began a game against time and the market. I assume that if you have reached this point, you are interested in product-market fit, marketing, and sales, so let me explain my assumptions to you: Product: A very very small tool that allowed you to automate proper tracking and creation of on-chain transactions. Literally, the whole app for the user was located on only three subpages. Starving Crowd: We tapped into an underserved market. The crypto market primarily operates via communities on platforms like Discord, Reddit, Twitter, Telegram, and so on. Therefore, our main strategy was directly communicating with users and demonstrating our tool. This was essentially “free marketing” (excluding the time we invested), as we did not need to invest in ads, promotional materials, or convince people about the efficacy of our tool. The community could directly observe on-chain transactions executed by our algorithms, which were processed at an exceptionally fast rate. This was something they couldn’t accomplish manually, so whenever someone conducted transactions using our algorithm, it was immediately noticeable and stirred a curiosity within the community (how did they do that!). Tests: I conducted the initial tests of the application on myself — we had already invested significantly in developing the product, but I preferred risking my own resources over that of the users. I provided the tool access to my wallet, containing 0.3ETH, and went to sleep. Upon waking up, I discovered that the transactions were successful and my wallet had grown to 0.99ETH. My excitement knew no bounds, it felt like a windfall. But, of course, there was a fair chance I could have lost it too. It worked. As we progressed, some users achieved higher results, but it largely hinged on the parameters set by them. As you can surmise, the strategy was simple — buy low, sell high. There was considerable risk involved. Churn: For those versed in marketing, the significance of repeat visitors cannot be overstated. Access to our tool was granted only after email verification and a special technique that I’d prefer to keep confidential. And this was all provided for free. While we had zero followers on social media, we saw an explosion in our email subscriber base and amassed a substantial number of users and advocates. Revenue Generation: Our product quickly gained popularity as we were effectively helping users earn — an undeniable value proposition. Now, it was time to capitalize on our efforts. We introduced a subscription model charging $300 per week or $1,000 per month — seemingly high rates, but the demand was so intense that it wasn’t an issue. Being a subscriber meant you were prioritized in the queue, ensuring you were among the first to reap benefits — thus adding more “value”. Marketing: The quality of our product and its ability to continually engage users contributed to it achieving what can best be described as viral. It was both a source of pride and astonishment to witness users sharing charts and analyses derived from our tool in forum discussions. They weren’t actively promoting our product but rather using screenshots from our application to illustrate certain aspects of the crypto world. By that stage, we had already assembled a team to assist with marketing, and programming, and to provide round-the-clock helpdesk support. Unforgettable Time: Despite the hype, my focus remained steadfast on monitoring our servers, their capacity, and speed. Considering we had only been on the market for a few weeks, we were yet to implement alerts, server scaling, etc. Our active user base spanned from Japan to the West Coast of the United States. Primarily, our application was used daily during the evenings, but considering the variety of time zones, the only time I could afford to sleep was during the evening hours in Far Eastern Europe, where we had the least users. However, someone always needed to be on guard, and as such, my phone was constantly by my side. After all, we couldn’t afford to let our users down. We found ourselves working 20 hours a day, catering to thousands of users, enduring physical fatigue, engaging in talks with VCs, and participating in conferences. Sudden Downturn: Our pinnacle was abruptly interrupted by the war in Ukraine (next macroeconomic shot straight in the face, lucky guy), a precipitous drop in cryptocurrency value, and swiftly emerging competition. By this time, there were 5–8 comparable tools had infiltrated the market. It was a challenging period as we continually stumbled upon new rivals. They immediately embarked on swift fundraising endeavors — a strategy we overlooked, which in retrospect was a mistake. Although our product was superior, the competitors’ rapid advancement and our insufficient funds for expeditious scaling posed significant challenges. Nonetheless, we made a good decision. We sold the product (exit) to competitors. The revenue from “exit” compensated for all the losses, leaving us with enough rest. We were a small team without substantial budgets for rapid development, and the risk of forming new teams without money to survive for more than 1–2 months was irresponsible. You have to believe me that this decision consumed us sleepless nights. Finally, we sold it. They turned off our app but took algorithms and users. Whether you believe it or not, after several months of toiling day and night, experiencing burnout, growing weary of the topic, and gaining an extra 15 kg in weight, we finally found our freedom… The exit wasn’t incredibly profitable, but we knew they had outdone us. The exit covered all our expenses and granted us a well-deserved rest for the subsequent quarter. It was an insane ride. Despite the uncertainty, stress, struggles, and sleepless nights, the story and experience will remain etched in my memory for the rest of my life. Swift Takeaways: Comprehending User Needs: Do you fully understand the product-market fit? Is your offering just an accessory or does it truly satisfy the user’s needs? The Power of Viral Marketing: Take inspiration from giants like Snapchat, ChatGPT, and Clubhouse. While your product might not attain the same scale (but remember, never say never…), the closer your concept is to theirs, the easier your journey will be. If your user is motivated to text a friend saying, “Hey, check out how cool this is” (like sharing ChatGPT), then you’re on the best track. Really. Even if it doesn’t seem immediately evident, there could be a way to incorporate this into your product. Keep looking until you find it. Niche targeting — the more specific and tailored your product is to a certain audience, the easier your journey will be People love buying from people — establishing a personal brand and associating yourself with the product can make things easier. Value: Seek to understand why users engage with your product and keep returning. The more specific and critical the issue you’re aiming to solve, the easier your path will be. Consider your offerings in terms of products and services and focus on sales and marketing, regardless of personal sentiments. These are just a few points, I plan to elaborate on all of them in a separate article. Many products undergo years of development in search of market fit, refining the user experience, and more. And guess what? There’s absolutely nothing wrong with that. Each product and market follows its own rules. Many startups have extensive histories before they finally make their mark (for instance, OpenAI). This entire journey spanned maybe 6–8 months. I grasped and capitalized on the opportunity, but we understood from the start that establishing a startup carried a significant risk, and our crypto product was 10 times riskier. Was it worth it? Given my passion for product development — absolutely. Was it profitable? — No, considering the hours spent — we lose. Did it provide a stable, problem-free life — nope. Did this entire adventure offer a wealth of happiness, joy, and unforgettable experiences — definitely yes. One thing is certain — we’ve amassed substantial experience and it’s not over yet :) So, what lies ahead? Chapter 7: Reverting to the contractor, developing a product for a crypto StartupReturning to the past, we continue our journey… I had invested substantial time and passion into the tech rescue mission product. I came on board as the technical Team Leader of a startup that had garnered over $20M in seed round funding, affiliated with the realm of cryptocurrencies. The investors were individuals with extensive backgrounds in the crypto world. My role was primarily technical, and there was an abundance of work to tackle. I was fully immersed, and genuinely devoted to the role. I was striving for excellence, knowing that if we secured another round of financing, the startup would accelerate rapidly. As for the product and marketing, I was more of an observer. After all, there were marketing professionals with decades of experience on board. These were individuals recruited from large crypto-related firms. I had faith in them, kept an eye on their actions, and focused on my own responsibilities. However, the reality was far from satisfactory. On the last day, the principal investor for the Series A round withdrew. The board made the tough decision to shut down. It was a period of intense observation and gaining experience in product management. This was a very brief summary of the last 10 years. And what next? (Last) Chapter 8: To be announced — Product Owner / Product Consultant / Strategist / CTO After spending countless hours and days deliberating my next steps, one thing is clear: My aspiration is to continue traversing the path of software product development, with the hopeful anticipation that one day, I might ride the crest of the next big wave and ascend to the prestigious status of a unicorn company. I find myself drawn to the process of building products, exploring product-market fit, strategizing, engaging in software development, seeking out new opportunities, networking, attending conferences, and continuously challenging myself by understanding the market and its competitive landscape. Product Owner / Product Consultant / CTO / COO: I’m not entirely sure how to categorize this role, as I anticipate that it will largely depend on the product to which I will commit myself fully. My idea is to find one startup/company that wants to build a product / or already has a product, want to speed up, or simply doesn’t know what’s next. Alternatively, I could be a part of an established company with a rich business history, which intends to invest in digitization and technological advancements. The goal would be to enrich their customer experience by offering complementary digital products Rather than initiating a new venture from ground zero with the same team, I am receptive to new challenges. I am confident that my past experiences will prove highly beneficial for the founders of promising, burgeoning startups that already possess a product, or are in the initial phases of development. ‘Consultant’ — I reckon we interpret this term differently. My aim is to be completely absorbed in a single product, crafting funnels, niches, strategies, and all that is necessary to repeatedly achieve the ‘product-market fit’ and significant revenue. To me, ‘consultant’ resonates more akin to freelancing than being an employee. My current goal is to kickstart as a consultant and aide, dealing with facilitating startups in their journey from point A to B. Here are two theoretical scenarios to illustrate my approach: Scenario 1: (Starting from point A) You have a product but struggle with marketing, adoption, software, strategy, sales, fundraising, or something else. I conduct an analysis and develop a strategy to reach point B. I take on the “dirty work” and implement necessary changes, including potential pivots or shifts (going all-in) to guide the product to point B. The goal is to reach point B, which could involve achieving a higher valuation, expanding the user base, increasing sales, or generating monthly revenue, among other metrics. Scenario 2: (Starting from point A) You have a plan or idea but face challenges with marketing, adoption, strategy, software, sales, fundraising, or something else. I analyze the situation and devise a strategy to reach point B. I tackle the necessary tasks, build the team, and overcome obstacles to propel the product to point B. I have come across the view that finding the elusive product-market fit is the job of the founder, and it’s hard for me to disagree. However, I believe that my support and experiences can help save money, many failures, and most importantly, time. I have spent a great deal of time learning from my mistakes, enduring failure after failure, and even had no one to ask for support or opinion, which is why I offer my help. Saving even a couple of years, realistically speaking, seems like a value I’m eager to provide… I invite you to share your thoughts and insights on these scenarios :) Closing Remarks: I appreciate your time and effort in reaching this point. This has been my journey, and I wouldn’t change it for the world. I had an extraordinary adventure, and now I’m ready for the next exciting battle with the market and new software products. While my entire narrative is centered around startups, especially the ones I personally built, I’m planning to share more insights drawn from all of my experiences, not just those as a co-founder. If you’re currently developing your product or even just considering the idea, I urge you to reach out to me. Perhaps together, we can create something monumental :) Thank you for your time and insights. I eagerly look forward to engaging in discussions and hearing your viewpoints. Please remember to like and subscribe. Nothing motivates to write more than positive feedback :) Matt.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Building and launching an AI-powered Product Strategy tool, or; a story of nights and weekends
reddit
LLM Vibe Score0
Human Vibe Score1
_raZeThis week

Building and launching an AI-powered Product Strategy tool, or; a story of nights and weekends

Speaking to peers in the software development sphere I learned of one constant that we had all personally experienced throughout our careers: a bloated product development process that feels like work for the sake of work, centred around the highest-paid person's opinion instead of its customers. We didn't like how current tools assume AI will provide the perfect answer on the first run. Instead, we wanted a tool that allows for manual refining and editing AI suggestions, keeping all previous ideas in context. This way, we can develop a solution step by step, instead of trying to get it perfect on the first try. An approach more similar to how you'd typically approach product discovery as a human. AI is then used to help save time and reduce admin, instead of replace the expert So, we got together and asked over 100 Product Managers questions about it, brought all that feedback goodness together, and started building Squad. We think we've created something really cool and hope you think so too. The ELI5 on what Squad does: 1) Creates alignment that empowers bottom up software development whilst keeping executive in the loop 2) Increases confidence that what you're building is what people actually want - data driven by default 2) Speeds up the time from idea --> execution by ideating with you on an experimentation approach 3) Helps gives PMs time back to focus on strategy (currently stats show they spend 75% of their time on admin, 25% on strategy) The team hustled hard on this as a passion project while working day jobs, and today have launched on Product Hunt. Check it out and see if the mission resonates with you, we'd appreciate the love! https://www.producthunt.com/posts/squad-8b75e29c-d767-4a8f-a60a-fd162e141a72 &#x200B;

Technical founders - is "bulling" your way through learning right for a startup? [I will not promote]
reddit
LLM Vibe Score0
Human Vibe Score0
JustZed32This week

Technical founders - is "bulling" your way through learning right for a startup? [I will not promote]

Sup, This is a question for technical founders. \--a little backstory-- I am starting a company in AI field that creates something nobody has ever done before. 7 months in. \--- How most software companies are created - you have an improvement idea, then you have a thousand or so problems to solve to make that improvement happen, and for each one that you don't know, you go to Stackoverflow or ChatGPT to look for solutions for that problem. Which involves next-to-no upfront preparation because for vast majority of traditional software you can solve it on-the-go - "traditional" software is very easy compared to, say, mechanical, pharma or AI engineering. However, for more advanced disciplines - can you just "Google" it on-the-go? I'm a solo founder, and 8 months in, creating a foundational model, BECAUSE I did not know things upfront, I've wasted at least 3 months doing something which was mostly technically unviable in the first place. Out of 14000 lines of code that I've done (including tests), I had to scrap 10000 recently. Imagine the scale of it. Obviously I didn't even know how ML works when I've started. Major fuck-up. How do you operate in industries which you've done before? How do you determine that it's time to start creating you big technological leaps instead of continuing to learn? Cheers. Edit: No need to push me on business topics. I know how to create value very well. It's only a tech question, and I'm only asking because - well - to deliver my value, I need to do a lot of novel tech.

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

For anyone working on LLM / AI startups
reddit
LLM Vibe Score0
Human Vibe Score1
juliannortonThis week

For anyone working on LLM / AI startups

My company (which I will not promote) wrote this blog post in compliance with rule #7 :) Introduction to fine-tuning Large Language Models, or LLMs, have become commonplace in the tech world. The number of applications that LLMs are revolutionizing is multiplying by the day — extraction use cases, chatbots, tools for creatives and engineers. In spite of this, at its core, the LLM is a multi-purpose neural network, dozens of layers deep, designed to simply predict one word after the next. It predicts words by performing billions of matrix multiplication steps based on so-called parameter weights, which are discovered during the model training process. Almost all open-source, open-weight models are trained on a massive amount of text from every conceivable genre and topic. How, then, do researchers and engineers create novel specialized applications? The answer is fine-tuning. In this post, we will demystify the process of fine-tuning and discuss the tradeoffs of other approaches to customizing an LLM. The history of fine-tuning In the ancient days of LLMs, by which we mean five years ago, the primary approaches to customizing an LLM was identical to the approaches to customizing any other deep learning model. A machine learning engineer would have two options: Retrain the entire LLM. This would mean discarding the trained weights and instead only using the open source model’s architecture to train it on a specialized dataset. As long as the amount and diversity of the specialized data is comparable to what the original model was trained on, this can be the ideal method of customizing a model. However, of course, this is a massive waste of resources due to the computational power required and the difficulty of collecting such a massive dataset. Even if an organization could provision enough GPUs, the cost of training modern-day models could cost up to $190 million. Retrain the last few layers of the LLM while keeping the rest of the weights frozen. This is a more efficient method in terms of time and computational power required because it significantly cuts down the number of parameters that need to be trained. However, for most tasks, this leads to subpar quality. Of course, almost everyone chooses to retrain the last few layers. And where there is only one option, the research community saw an opportunity to step in. Soon, the LLM space saw an enormous amount of activity in fine-tuning, which leads us to today. Modern approaches to fine-tuning Most fine-tuning approaches today are parameter-efficient. Deep neural networks are composed of matrices and vectors (generally called tensors), which are at their core arrays of floating point numbers. By training a small subset of these tensors, while the rest of the LLM’s weights are kept frozen, practitioners achieve good enough results without having to retrain the entire model. Generally, this method requires at least a hundred or so handcrafted examples of input-output pairs for fine-tuning. This is called supervised learning. The modern fine-tuning landscape involves an unsupervised learning step afterwards. Given a set of inputs, a practitioner gathers the various possible outputs from the LLM and casts votes among them. This preference data is then used to further train the LLM’s weights. Usually, this approach is used for LLM alignment and safety, which defends the application from malicious uses, outputs embarrassing to the organization, and prompt injection attacks. Fine-tuning’s relationship to prompt engineering A natural question arises: why fine-tune instead of crafting a well-considered system prompt? Wouldn’t that be easier and more efficient? The answer is no, it wouldn’t. Here’s why: Advanced techniques make prompt engineering obsolete: \[redacted\]'s product uses soft-prompting and other techniques to train the input layer itself. This obviates the need for prompt engineering entirely, which lets organizations avoid the time-consuming trial-and-error process to get the prompt just right. Prompt engineering has been a stopgap measure in the early days of LLM applications to convey the practitioner’s intent to the LLM. It is not the long-term solution for LLM application development. The system prompt is precious: the limited budget for system prompt length is better used for up-to-date information, e.g., Retrieval-Augmented Generation (RAG). Even as context windows increase in size with each new open-source model, the system prompt is the least efficient place to provide the LLM model with verbose instructions and examples. The longer the prompt, the slower the application: an LLM must attend to the entire system prompt for each token generated. This pain becomes more acute in the chatbot case, where the length of the conversation so far is also counted toward the system context. The longer the conversation, and the longer your beautifully-crafted system prompt, the slower the bot becomes. Even in cases where the model allows for system prompts that are millions of tokens long, doubling the size of the context will quadruple the latency. This means adding a few hundred words to the system prompt may result in several seconds of additional latency in production, making a chatbot impossible to use. Edge case handling: the number of edge cases that the system prompt would need to consider and emphasize to the LLM is too large. The instructions would have to be too nuanced and long to cover them all. However, fine-tuning on a dataset that considers these edge cases would be more straightforward. Do I need to fine-tune the LLM in my production application? Every LLM application in production must be fine-tuned often, not just once at the beginning. Why fine-tune? The world in which the application exists is constantly evolving. New prompt injection attacks are being discovered every day, new ways of embarrassing a chatbot are emerging constantly. This data can be used to further train an LLM model, which protects the application from new failure modes and reputational risk. Like any software, LLM models are constantly improving. Smarter and faster models are open-sourced all the time. For a new model to get deployed to production, it must first be finetuned on the specific dataset of the organization building the application. Fine-tuning does not add latency to LLM applications. Rather than a solution that sits in the middle of the LLM and the rest of the application, fine-tuning leverages the power of the LLM itself to increase the quality of the output. In fact, fine-tuning allows for shorter system prompts, which speeds up the average response generation time.

Created the Shopify Alternative in a 3rd world country “I will not promote”
reddit
LLM Vibe Score0
Human Vibe Score1
uwalkirunThis week

Created the Shopify Alternative in a 3rd world country “I will not promote”

Built a Shopify alternative I’ve been a long-time follower of this subreddit and have always valued the insights shared here. Today, I’m reaching out to share our story and seek advice or guidance on potential next steps for our business. Four years ago, we set out to build a local e-commerce platform tailored to the unique challenges of operating in a third-world country where global solutions like Shopify fall short. Shopify, while a fantastic platform, doesn’t provide localized support or integrations here, and the costs of running a Shopify store are prohibitively high due to: The need for multiple apps to replicate basic functionality Expensive international support calls or long chat queues Higher payment gateway fees (no Shopify Pay) USD-only subscription payments, which incur additional bank conversion fees And more We built a solution that addresses these pain points, and today, we’re proud to have over 4,000 merchants on our platform, with 1,600+ paying customers. We’re processing over $1 million per month across 50,000+ orders, which translates to a significant impact in our local economy. As experienced founders, we’ve managed our financials meticulously, allowing us to thrive while many local competitors have shut down. However, scaling in our current economic climate has been challenging, and raising capital has proven to be incredibly tough. We’re exploring strategic options, including potential partnerships, acquisitions, or investments. For example, we believe our platform could be an attractive opportunity for a player like Shopify or another company looking to expand into emerging markets. I’m reaching out to this community to ask: Are there doors we haven’t knocked on? Are there opportunities or strategies we might be overlooking? Any advice, introductions, or insights would be immensely appreciated. Thank you for taking the time to read this, and I look forward to any feedback or ideas you might have! [post refined by AI]

Joined an AI Startup with Ex-ShipStation Team - Need Tips on Finding Early Users
reddit
LLM Vibe Score0
Human Vibe Score1
welcomereadThis week

Joined an AI Startup with Ex-ShipStation Team - Need Tips on Finding Early Users

Hey Reddit, My name’s Welcome (Yes, that’s really my name), and I’ve been in tech for most of my career, mostly at bigger companies with established brands and resources. But recently, I decided to join a small startup called BotDojo. It’s my first time being part of a small team, and it’s been a pretty eye-opening experience so far. But, like with anything new, I’ve hit a few bumps along the way, and I’m hoping you all might have some advice. A little backstory: BotDojo was started by some of the engineers who used to work together at ShipStation. After ShipStation sold, they spent some time experimenting with AI but kept running into the same problems—having to patch together tools, getting inconsistent results, handling data ingestion, and struggling to track performance. So, they decided to build a platform to help developers build, test, and deploy AI solutions. Since I came on board, my focus has been on finding early users, and it’s been a mixed bag of wins and frustrations. We’ve got a solid group of people using the free version (which is great), but only a few have upgraded to the paid plan so far (ranging from startups to large enterprises). The cool thing is that those who have become paying customers absolutely love the product. It’s just been hard getting more people to that point. We’ve tried a bunch of things: Attending industry events, doing cold email outreach, running social ads (the usual stuff). And while we’ve seen some interest, we’re running into a few challenges:   Learning curve: The software is really powerful, but it takes a week or two for users to really see what it can do. Without a dedicated sales team to walk them through it, it’s been tough getting people to stick around long enough to see the value. Standing out is hard: The AI space is super crowded right now. I think a lot of people see “AI tool” and assume it’s just like everything else out there (even though BotDojo has some awesome features that really set it apart).  Sign-ups, but limited engagement: We’re on a freemium model to make it easy for people to try it out, but that also means we get a lot of bots and people who sign up but don’t really dive in. So, I thought I’d reach out here and see if anyone has been through this early stage before. How did you manage to break through and find those first paying users who really saw the value in what you were building?  Are there any strategies, communities, or tactics that worked particularly well for you? And if you had to do it all over again, what would you focus on? I figure I’m not the only one trying to navigate these waters, so I’m hoping this can be a helpful thread for others too. Thanks so much for reading, and I’d be super grateful for any advice or insights you can share! 🙏

Building in the open with Founder University - I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Tim-SylvesterThis week

Building in the open with Founder University - I will not promote

Published Oct 30, 2024 I am on my fifth startup. I ran the last one for a decade, that’s a whole story. A hell of a story. But a different story. I’ll tell it to you when I can, but not right now. The one before that was an e-commerce site that did pretty well but I didn’t love it. Before that were two service businesses. The first one I did for the love of the game, the second one was an attempt to make people stop asking me to fix their computer by charging them outrageous prices, which backfired horribly when they were eager to pay. None are relevant except to say I’ve been around the block and have the scars to prove it. When it was time to get back out there, I wanted to use all I’ve learned to do better. Before I talk about what those lessons produced, I’m going to talk about what those lessons were. Cause before effect, after all. One thing I wanted to do better this time was pattern matching - making the startup look the way that the industry and investors “expect” a startup to look. My last startup was an awesome idea with awesome tech (still is, but like I said, another story), but that one didn’t match patterns. It didn’t match investor patterns, industry buying patterns, patterns of existing, immediate, recognized and admitted needs. Because it didn’t “look” right to anyone, everything about it was way harder than necessary. The “make it look right” approach runs the risk of building a cargo cult, imitating the trappings of something but without understanding the essence of that something, but then again, a thing that looks like a knife is going to make a better knife that a thing that looks like a bowling ball, so sometimes just sharing apparent similarities can get you pretty far, even if it doesn’t get you all the way there. Like how mimicking someone’s accent makes it easier for them to understand you. For this one, I wanted to adopt every tool, method, and pattern that I knew “the industry” wanted to see to minimize the friction from development, go-to-market, scaling, adoption, and that would make investment optional (and, therefore, available if desired) instead of necessary (and, therefore, largely unavailable). That required establishing some expectations for successful patterns I could match against. What patterns am I matching to? Here’s a general sketch of my pattern matching thought process: Software first and software only. It’s the easiest industry to start a business in, lowest startup costs, and easiest customer acquisition. I wanted to build software for an element of the industry that’s actively emerging (and therefore has room to grow) and part of an optimistic investor thesis (and therefore has a cohort of people who are intent on injecting capital into the market to help it grow). It needs to fills a niche that is underexplored (low competition) and highly potent (lots of opportunity), while being aligned to recognized and emerging needs within the industry (readily adopted). I wanted it to have evidence supporting the business thesis that proves the demand exists, but demonstrates that the demand is unanswered (as of yet) by sufficient or adequate supply.* I wanted the lowest number of dominoes to line up and tip for everything to work correctly - the more dominoes in the line, the less likely the last one will fall. I wanted to implement modern toolsets for everything, wherever possible. I wanted to obey the maxim, “When there’s a gold rush, don’t mine the gold, sell the picks and shovels.” Whatever I chose would need to produce cash flow almost immediately with minimal development time or go-to-market delays, because the end of ZIRP killed the “trust me bro” investment thesis predominant over the last 15 years. I wanted to match to YC best practices, not because YC can predict what will definitely work, but because they’ve churned through so many startups in the last 15 years that they have a good sense of what will definitely not work. And I wanted to build client-centric, because if my intent is to to produce cash flow immediately, we need to get clients immediately, and if we need to get clients immediately, we need to focus on what clients need right now. Extra credit: What’s the difference between a customer and a client? Note: Competition is awesome! Competition is validating and not scary, because competition proves a market exists. But competition, especially mature competition against an immature startup, makes it harder to break into a space. A first mover advantage isn’t everything, but seeing demand before it’s sufficiently supplied is a great advantage if you’re capital constrained or otherwise unproven. Think about how much money the first guy to sell fidget spinners or Silly Bandz made versus how much money the last guy to order a pallet of each made. Finding demand that exists already but is as of yet insufficiently satisfied is a great place to start. What opportunity spaces are most relevant? The industries and markets I chose to observe were: AI, because if I’m following a theme & pattern for today, it’s AI. Fintech, because cash is king, and fintech puts your hands on cash flow. Crypto/blockchain, because that’s the “new” fintech (or maybe the “old-new” fintech?), and crypto creates powerful incentives and capital formation strategies, along with a lot of flexibility for transaction systems. Tools, particularly unmet demand in tools, that enable these industries. If you wanted to do some brief and simple homework, you could map each of those bullets to several of the numbered list items preceding them. The reasoning was pretty simplistic - AI is what people want to build and invest in now, while fintech and crypto/blockchain are what people were building and investing in for the last major investment thesis. That means that there’s demand in the market for AI and AI-adjacent startups, while there’s a glut of underutilized and highly developed tools within fintech and crypto/blockchain, with a lot of motivated capital behind the adoption. When someone is thinking “I built this thing and not enough people are using it”, and you then build something that uses it creates a great way to find allies. This rationale harnesses technology that is being built and financed now (which means it needs tools and support methods, and a lot of other “picks and shovels”), while leveraging technology that was recently built and financed and is eager for more widespread adoption of the existing toolkits, which makes it suitable for using to build the AI-adjacent tools that are in demand now. It’s like two harmonics producing constructive interference - it makes two waves into one larger wave, which gives me more momentum to surf against. This was a learning process, and I iterated against my general paradigm repeatedly as I learned more. Neither of us have the patience to go through that in excruciating detail, so I’ll cover the highlights in my next post. Extra credit answer: A customer gets a product, a client gets a service. Challenge: Is software a product or a service?

For the Herd-Investor(Formerly Me)
reddit
LLM Vibe Score0
Human Vibe Score1
Ready_Papaya_7937This week

For the Herd-Investor(Formerly Me)

Hey guys. my friend and I developed a model that looks over SEC filings and instead of just summarizing what they say like the existing “AI” solutions do(which are really just read-write programs), it infers and reads between the lines and analyzes what type of strategy the company is using(revenue recognition timing, the company's history,etc.) and many other factors. We used a different approach. Instead of basically making a GPT wrapper, we trained it from scratch based on not only summarizing filings but inferring on key information that is glossed over a lot. We plan to scale this into a model that accounts for not only filings, but recent news, public sentiment, and other factors. And instead of people having to upload files to get analyzed, we plan to automatically aggregate files on all public companies on the US markets and train the model on those to provide a one- stop shop financial search engine platform for retail investors to access digestable financial information(like an AlphaSense but for retail investors) because right now, the average retail investor has to access on average 5 services to get this info and then has to interpret the info as well. Obviously, the retail investor these days is also tied to a sense of community so plan to implement a moderated almost newletter like platform where verified creators can publish posts regarding their interests to further serve the retail investor. The gist is basically simplifying high-level finance to the point where the beginner investor can understand while preserving the technical value. Do you guys have any extra thoughts on this? I am trying to ask if you guys would actually pay for a service like this, and what it should additionally offer to make it more valuable to the average retail investor. Thanks again!

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

Selling equity - what’s next?
reddit
LLM Vibe Score0
Human Vibe Score0
found_it_online_01This week

Selling equity - what’s next?

Hey all, Seeking some guidance / advice as I plan my exit from a marketing agency I helped grow to $5M Long story short, I was hired part time to build their digital marketing department that sat at around 40k annual agency revenue. Since then I’ve become a minority equity partner, and at one point the agency was above $5M in gross agency revenue. The digital department that I run had up to 13 FTE employees at one point And digital revenue accounted for 60% of all agency revenue for the last 3-4 years. So, why am I leaving? Things are groovy, right? Well, we have dropped from $5M to now $3M this year and we’ll be lucky to hit that. As a minority equity party it’s been hard to watch leadership continue to disregard our agency as a digital agency. They don’t want to niche down, and they don’t want to identify as a digital agency, but instead by a full service “strategic agency”. Clients have felt our lack of expertise and direction, so they leave for someone who is an expert in xyz platform or industry. I no longer see their vision, and so I’m planning a sale of my equity and looking for new venture opportunities. While I am perfectly capable of running Google ads and Facebook ads campaigns, and as an accomplished SEO I know how to rank sites, and still find it fun. But I’m not interested in the labor arbitrage model of agency work anymore. I’d rather build a portfolio of in-house properties or digital assets where I have more control. Lately my obsession has been using AI and zapier to automate business processes, documentation, project management etc. Agency life has also exposed me to a lot of industries and business models, which I am always fascinated by. Eventually I will launch my own business, but I’m supporting my partner while they finish school. So I’m a single income household.. Therefore a W2 would be ideal but I’m open to contract work. So my question is- what positions or roles would I fill? I’ve done my share of research but this community has always given me new things to consider. Any feedback or questions are welcomed.

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences
reddit
LLM Vibe Score0
Human Vibe Score1
Important-Ostrich69This week

From "There's an App for that" to "There's YOUR App for that" - AI workflows will transform generic apps into deeply personalized experiences

I will not promote. For the past decade mobile apps were a core element of daily life for entertainment, productivity and connectivity. However, as the ecosystem saturated the general desire to download "just one more app" became apprehensive. There were clear monopolistic winners in different categories, such as Instagram and TikTok, which completely captured the majority of people's screentime. The golden age of creating indie apps and becoming a millionaire from them was dead. Conceptual models of these popular apps became ingrained in the general consciousness, and downloading new apps where re-learning new UI layouts was required, became a major friction point. There is high reluctance to download a new app rather than just utilizing the tooling of the growing market share of the existing winners. Content marketing and white labeled apps saw a resurgence of new app downloads, as users with parasympathetic relationships with influencers could be more easily persuaded to download them. However, this has led to a series of genericized tooling that lacks the soul of the early indie developer apps from the 2010s (Flappy bird comes to mind). A seemingly grim spot to be in, until everything changed on November 30th 2022. Sam Altman, Ilya Sutskever and team announced chatGPT, a Large Language Model that was the first publicly available generative AI tool. The first non-deterministic tool that could reason probablisitically in a similar (if flawed) way, to the human mind. At first, it was a clear paradigm shift in the world of computing, this was obvious from the fact that it climbed to 1 Million users within the first 5 days of its launch. However, despite the insane hype around the AI, its utility was constrained to chatbot interfaces for another year or more. As the models reasoning abilities got better and better, engineers began to look for other ways of utilizing this new paradigm shift, beyond chatbots. It became clear that, despite the powerful abilities to generate responses to prompts, the LLMs suffered from false hallucinations with extreme confidence, significantly impacting the reliability of their use, in search, coding and general utility. Retrieval Augmented Generation (RAG) was coined to provide a solution to this. Now, the LLM would apply a traditional search for data, via a database, a browser or other source of truth, and then feed that information into the prompt as it generates, allowing for more accurate results. Furthermore, it became clear that you could enhance an LLM by providing them metadata to interact with tools such as APIs for other services, allowing LLMs to perform actions typically reserved for humans, like fetching data, manipulating it and acting as an independent Agent. This prompted engineers to start treating LLMs, not as a database and a search engine, but rather a reasoning system, that could be part of a larger system of inputs and feedback to handle workflows independently. These "AI Agents" are poised to become the core technology in the next few years for hyper-personalizing and automating processes for specific users. Rather than having a generic B2B SaaS product that is somewhat useful for a team, one could standup a modular system of Agents that can handle the exactly specified workflow for that team. Frameworks such as LlangChain and LLamaIndex will help enable this for companies worldwide. The power is back in the hands of the people. However, it's not just big tech that is going to benefit from this revolution. AI Agentic workflows will allow for a resurgence in personalized applications that work like personal digital employee's. One could have a Personal Finance agent keeping track of their budgets, a Personal Trainer accountability coaching you making sure you meet your goals, or even a silly companion that roasts you when you're procrastinating. The options are endless ! At the core of this technology is the fact that these agents will be able to recall all of your previous data and actions, so they will get better at understanding you and your needs as a function of time. We are at the beginning of an exciting period in history, and I'm looking forward to this new period of deeply personalized experiences. What are your thoughts ? Let me know in the comments !

Just raised and here are the stats (July 2024)
reddit
LLM Vibe Score0
Human Vibe Score0.6
tylersellarsThis week

Just raised and here are the stats (July 2024)

CEO of a startup - bootstrapped for 10 months with a team of 7 - Built a waitlist of $15B AUM (fintech) and here's what it took (with no intended story structure) I didn't want to spam, so I decided to go the old school route and manually write every single email (some copy and paste) In order to feel prepared, I would do my research prior to reaching out, albeit sometimes limited due to time x reward Sent over 350 emails to around 300 funds and we received three yes' (to be honest we received more than three, but they wanted too much equity, or they weren’t a good fit culturally) Pre Seed is different for everyone. Some accept pre revenue others expect 100-300k ARR - this was one of the more frustrating pieces for me, due to the fact that it's extremely subjective to what "Earliest Stage" means to some funds. We're pre-product, we had to remove our CTO in the process due to performance issues, we let go of our front-end, hired a new CTO, hired an AI Engineer, and replaced our front end. It's a numbers game. We received 94% no reply - 5% nos (with meetings) - 1% yes' I used OpenVC as my main resource, highly recommend even their free product. I recommend using discount with a SAFE. Some Angels like it better, some VC’s hate it. You have to be willing to play ball with whoever leads. Mercury for banking, perks (like Carta), and SAFE agreement. All great at Mercury. Raising for a B2B business in an enterprise market is much easier, unless deep tech or science backed. But consumer products right now are not raising pre seed from what I can tell. If it means anything - it's a numbers game. Go get what you deserve, but put in the work because no one will just hand it to you. Love this community, always here to help anyone I can.

How I made a high tech salary in my first selling month
reddit
LLM Vibe Score0
Human Vibe Score1
Ok_Negotiation_2587This week

How I made a high tech salary in my first selling month

For over 7 years I worked as a full-stack developer, helping other companies bring their ideas to life. But one day, I thought “Why not try making my own dream come true?”. That’s when I decided to quit my job and start my own journey to becoming an entrepreneur. At first, it wasn’t easy. I didn’t make any money for months and had no idea where to start. I felt lost. Then, I decided to focus on something popular and trending. AI was everywhere, and ChatGPT was the most used AI platform. So I looked into it and I found the OpenAI community forum where people had been asking for features that weren’t being added. That gave me an idea. Why not build those features myself? I created a Chrome extension and I worked on some of the most requested features, like: Downloading the advanced voice mode and messages as MP3 Adding folders to organize chats Saving and reusing prompts Pinning important chats Exporting chats to TXT/JSON files Deleting or archiving multiple chats at once Making chat history searches faster and better It took me about a week to build the first version, and when I published it, the response was incredible. People loved it! Some even said things like, “You’re a lifesaver!” That’s when I realized I had something that could not only help people but also turn into a real business. I kept the first version free to see how people would respond. Many users have been downloading my extension, which prompted Chrome to review it to determine if it qualified for the featured badge. I received the badge, and it has significantly boosted traffic to my extension ever since. After all the positive feedback, I launched a paid version one month ago. A few minutes after publishing it, I made my first sale! That moment was so exciting, and it motivated me to keep going. I already have over 4,000 users and have made more than $4,500 in my first selling month. I’ve decided to release 1-2 new features every month to keep improving the extension based on what users ask for. I also created the same extension for Firefox and Edge users because many people have been asking for it! I also started a Reddit community, where I share updates, sales, discount codes, and ideas for new features. It’s been awesome to connect with users directly and get their feedback. Additionally, I’ve started working on another extension for Claude, which I’m hoping will be as successful as this one. My message to you is this: never give up on your dreams. It might feel impossible at first, but with patience, hard work, and some creativity, you can make it happen. I hope this inspires you to go after what you want. Good luck to all of us!

From Running a $350M Startup to Failing Big and Rediscovering What Really Matters in Life ❤️
reddit
LLM Vibe Score0
Human Vibe Score1
Disastrous-Airport88This week

From Running a $350M Startup to Failing Big and Rediscovering What Really Matters in Life ❤️

This is my story. I’ve always been a hustler. I don’t remember a time I wasn’t working since I was 14. Barely slept 4 hours a night, always busy—solving problems, putting out fires. After college (LLB and MBA), I was lost. I tried regular jobs but couldn’t get excited, and when I’m not excited, I spiral. But I knew entrepreneurship; I just didn’t realize it was an option for adults. Then, in 2017 a friend asked me to help with their startup. “Cool,” I thought. Finally, a place where I could solve problems all day. It was a small e-commerce idea, tackling an interesting angle. I worked 17-hour days, delivering on a bike, talking to customers, vendors, and even random people on the street. Things moved fast. We applied to Y Combinator, got in, and raised $18M before Demo Day even started. We grew 100% month-over-month. Then came another $40M, and I moved to NYC. Before I knew it, we had 1,000 employees and raised $80M more. I was COO, managing 17 direct reports (VPs of Ops, Finance, HR, Data, and more) and 800 indirect employees. On the surface, I was on top of the world. But in reality, I was at rock bottom. I couldn’t sleep, drowning in anxiety, and eventually ended up on antidepressants. Then 2022 hit. We needed to raise $100M, but we couldn’t. In three brutal months, we laid off 900 people. It was the darkest period of my life. I felt like I’d failed everyone—myself, investors, my company, and my team. I took a year off. Packed up the car with my wife and drove across Europe, staying in remote places, just trying to calm my nervous system. I couldn’t speak to anyone, felt ashamed, and battled deep depression. It took over a year, therapy, plant medicine, intense morning routines, and a workout regimen to get back on my feet, physically and mentally. Now, I’m on the other side. In the past 6 months, I’ve been regaining my mojo, with a new respect for who I am and why I’m here. I made peace with what I went through over those 7 years—the lessons, the people, the experiences. I started reconnecting with my community, giving back. Every week, I have conversations with young founders, offering direction, or even jumping in to help with their operations. It’s been a huge gift. I also began exploring side projects. I never knew how to code, but I’ve always had ideas. Recent advances in AI gave me the push I needed. I built my first app, as my first attempt at my true passion—consumer products for kids. Today, I feel wholesome about my journey. I hope others can see that too. ❤️ EDIT: Wow, I didn’t expect this post to resonate with so many people. A lot of you have DM’d me, and I’ll try to respond. Just a heads-up, though—I’m juggling consulting and new projects, so I can’t jump on too many calls. Since I’m not promoting anything, I won’t be funneling folks to my page, so forgive me if I don’t get back to everyone. Anyway, it’s amazing to connect with so many of you. I’d love to write more, so let me know what topics you’d be interested in!

Nuts and bolts AI implementation for small business
reddit
LLM Vibe Score0
Human Vibe Score1
Training-Swan-6379This week

Nuts and bolts AI implementation for small business

How can small businesses use AI to increase sales or decrease expenses without massive disruption? One way for us is using AI to process our email history to identify patterns and write personalized messages based on past correspondence. According to legal advice in which I have confidence, email that is personalized for each recipient (and meets other standards) does not need to be opt in. If you disagree - understood - but spam morality is not the topic here. Bottom line - obviously a game changer. Knowing phrases people have used before becoming clients - and all of the possible permutations of those phrases, and detecting where those phrases show up will make our sales and marketing many times more effective for a fraction of the cost. There's a reason big corps. record calls, and now small business can leverage the same technology. We are setting up a process that yields accurate, up to date, comprehensive data for our own business operations. Our clients - who are they and how has their demographic changed over time? To answer this question and for email personalization, we also need access to external data sources e.g. like accurate up to date company demographics. IMO - the leader in company data in the US? THEY SUCK. We found there is no magic fairy who is going to make good data appear for our AI. The process of applying our own proprietary knowledge to code and categorize the data is just as important, and obviously highly sensitive. How do we leverage the AI technologies of companies like Google and Microsoft (or anyone else) without being their bitch? Below is a list of some of the sources of my business's data: Data sources: PST/OST/Other Email data files Microsoft data from Windows/O365 Windows/Linux/Android/IOS application logs and other data Web server logs for the company website. SEO/Analytics Data Google data export Google voice/VOIP logs OneDrive/G drive Other Phone system/cell service logs Other SAAS and in-house application data. Facebook/social media data for company pages. QuickBooks/other accounting systems/business bank account logs POS/Credit card processing systems/PayPal, etc. OSINT to fill in the blanks

The Birth of My First (and Hilariously Flawed) Voice Agent: A Tale of No-Code Chaos
reddit
LLM Vibe Score0
Human Vibe Score0.778
No-Understanding5609This week

The Birth of My First (and Hilariously Flawed) Voice Agent: A Tale of No-Code Chaos

Okay Reddit, buckle up. I'm about to tell you the saga of how I birthed my very first voice agent, a chaotic and frankly, slightly embarrassing journey involving Retell.ai, Make.com, and Zapier. Looking back, it's equal parts hilarious and traumatizing. The Naive Dream: Back then (it feels like ages ago!), I was convinced I could easily whip up a voice agent that would take restaurant orders over the phone. Elegant, efficient, and completely automated! I envisioned a world where my clients' restaurant never missed a beat, all thanks to my coding prowess... or rather, my no-code prowess. How wrong I was. The Gauntlet Begins: Retell.ai's Murky Depths Retell.ai was the starting point, the "voice" of my operation. Getting the phone number hooked up felt like a small victory, quickly overshadowed by the realization that their documentation was... well, let's just say it wasn't written for complete novices. I spent what felt like an eternity staring at API keys, convinced I'd entered them correctly, only to be greeted by cryptic error messages. The sheer frustration I felt wrestling with that initial setup is something I'll never forget. Make.com: From Pretty Picture to Painful Puzzle Then came Make.com, the orchestra conductor of my workflow. It looked so beautiful, so user-friendly! Drag and drop, visual modules... what could go wrong? Oh, so much could go wrong. Trying to decipher the JSON data stream from Retell was like trying to understand a foreign language I only knew a few words of. Mapping that data to a Google Sheet? A complete and utter disaster. I remember spending hours just trying to get the correct fields to populate, each failed attempt fueling my growing despair. Zapier: Briefly Considered, Quickly Dismissed I flirted with the idea of using Zapier instead, seduced by its simplicity. But its limitations became glaringly obvious when I tried to build the complex, multi-step process I needed. Make.com was the only real option, which meant diving headfirst into a whole new world of modules, triggers, and data transformations. The Infernal Testing Loop: The absolute WORST part of the entire process was the testing. Picture this: Calling the agent, rambling through a mock order, waiting for the workflow to execute, only to discover (yet another) error. Then, tweaking the scenario, pushing "save," and repeating the entire agonizing process. Each test call felt like a mini-marathon, a grueling race against time and my own dwindling patience. The AI's... Quirks: And then there was the AI itself. It was... let's just say it had a personality of its own. Sometimes, it perfectly understood my order. Other times, it decided I wanted to order 500 pizzas with extra anchovies. Debugging the AI's interpretation felt like negotiating with a stubborn toddler. Lessons Hard-Learned (And Forever Etched in My Memory): Start absurdly small: I tried to build a fully functional system right away. A HUGE mistake. If I could go back, I would have focused on just extracting one piece of information (like, say, just the quantity) and gotten that rock solid before adding anything else. JSON is your friend (or should be): Back then, JSON felt like alien code. Now, I have a slightly better grasp on it. Trust me, learn JSON. It will save you so much pain. Test like your sanity depends on it: Because it does. After every. Single. Change. Test the entire flow. It's tedious, but it's the only way to catch errors before they snowball into a catastrophe. Don't suffer in silence: I tried to be a lone wolf, figuring everything out myself. Big mistake. Retell.ai's forums and Make.com's documentation are goldmines. Use them! Embrace the struggle: This is the most important lesson. Building a voice agent, especially your first one, is hard. It's frustrating. It will test your limits. But don't give up. The feeling of finally making it work (even partially) is worth it. The Bot That (Barely) Lived: In the end, I did create a voice agent that could take orders and log them into a spreadsheet. It wasn't pretty. It was buggy. It occasionally ordered things that didn't make any sense. But it was mine. And it was the first step on a long and winding road. Looking back, I laugh (and cringe) at my naivety. But I also appreciate the lessons I learned and the sheer grit it took to bring my little AI Frankenstein to life. Anyone else have a similar "first bot" story? Let's hear them! Misery (and laughter) loves company. #RetellAI #Makecom #Zapier #FirstBot #NoCodeFail #VoiceAgentStruggles #StoryTime

My app has gone viral and I grew from 1k users (take 5months) to 100k user in 5 days
reddit
LLM Vibe Score0
Human Vibe Score1
Consistent_Access844This week

My app has gone viral and I grew from 1k users (take 5months) to 100k user in 5 days

I've always dreamt of building an online side business where I can build once and sell to millions. I love that business model but have never dreamt that I can achieve that, given that I am not a programmer in my career. I have been following side hustle school and some other business podcast for the past years as a drive and motivation to create my own business.  Over the years, I've delve a little on to web development using WordPress and in the hope of earning some money from that. I learnt in the hard way but is a good learning story and journey. I realised that what you put all your efforts building and excited for doesn't mean anything for anyone else and also learnt the importance of UI UX.  Fast forward to 5 months ago (July 2024), I've came across several low code app builder. With the help of the low code tools in combination with chatgpt, I've finally launched my first mobile app - Rolly: AI Money Tracker. But the business challenges doesn't end here, but it's just the beginning. I got no experience and skills on marketing but I've got my drive and passion that keep propelling me forward. By keep listening on people sharing their journey, looking at different apps to brainstorm etc, I've managed to now grow my user base from 1k (in 5 months) to 100k (in 5days). What's happening was my app somehow got viral in Vietnam when people are getting interest funny comments from my AI during entering the transaction and it has been sharing around in the social media and even featured on the news. What a crazy journey as the inflow of users has been too sudden, my server has been down for a few times until I progressively upgrade it until it got stable these couple of days. As for my advice to people dreaming the to be entreprenuer - Don't overthinking about all the problems you will face before starting. You will encounter hundreds of problems along the way and you just need to solve them one by one. You will never start if you think about what's not working and there will never be an answer for everything - even I don't have an answer for everything now.

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀
reddit
LLM Vibe Score0
Human Vibe Score1
Awkward_Ad_9605This week

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀

Hey Members! I’m excited to share an idea for a new application I’m planning to build: Quick Launch . This AI-driven platform is designed to assist solopreneurs or anyone with an idea in launching their Minimum Viable Products (MVPs) by taking on the roles of the entire team needed for the process. Goal: Assistance in quickly moving from Idea to MVP Before I dive into the details, I’d love to hear your thoughts and feedback. Key Features: Product Creation: From Idea to Product Detailing AI-Generated Q&A: Real-time questions generation one-at-a-time to define the product requirements based on their knowledge levels to convert an Idea into a Product. Market Research Reports: In-depth analysis that identifies product-market fit, competitive landscape, and potential marketing strategies. Sentiment Analysis: Evaluation of user feedback and reactions across multiple subreddits to gauge public opinion on ideas. Product Development: Product Detailing to Actual Product User Story Generation: Identification and creation of comprehensive user stories, tasks, and sub-tasks to facilitate development. AI Project Management: AI agents assume roles of project managers and UI/UX designers to streamline product detailing and development. Integration Capabilities: Seamless integration with popular project management tools like Jira, Asana, and Trello for better workflow management. Target Audience: Solopreneurs: Individuals looking to bring their business ideas to life without extensive resources. Indie Hackers: Entrepreneurs focused on building small projects or startups with minimal overhead. Idea Validators: Anyone with a concept seeking initial validation and market feedback before committing significant resources. If you’re interested in learning more, check out our teaser website: Quick Launch Discussion Question: What features would you find most valuable in an application like this? Are there specific pain points you face when launching an MVP? Your insights would be incredibly helpful as I refine this idea! Looking forward to your thoughts! 🙌

Nuts and bolts AI implementation for small business
reddit
LLM Vibe Score0
Human Vibe Score1
Training-Swan-6379This week

Nuts and bolts AI implementation for small business

How can small businesses use AI to increase sales or decrease expenses without massive disruption? One way for us is using AI to process our email history to identify patterns and write personalized messages based on past correspondence. According to legal advice in which I have confidence, email that is personalized for each recipient (and meets other standards) does not need to be opt in. If you disagree - understood - but spam morality is not the topic here. Bottom line - obviously a game changer. Knowing phrases people have used before becoming clients - and all of the possible permutations of those phrases, and detecting where those phrases show up will make our sales and marketing many times more effective for a fraction of the cost. There's a reason big corps. record calls, and now small business can leverage the same technology. We are setting up a process that yields accurate, up to date, comprehensive data for our own business operations. Our clients - who are they and how has their demographic changed over time? To answer this question and for email personalization, we also need access to external data sources e.g. like accurate up to date company demographics. IMO - the leader in company data in the US? THEY SUCK. We found there is no magic fairy who is going to make good data appear for our AI. The process of applying our own proprietary knowledge to code and categorize the data is just as important, and obviously highly sensitive. How do we leverage the AI technologies of companies like Google and Microsoft (or anyone else) without being their bitch? Below is a list of some of the sources of my business's data: Data sources: PST/OST/Other Email data files Microsoft data from Windows/O365 Windows/Linux/Android/IOS application logs and other data Web server logs for the company website. SEO/Analytics Data Google data export Google voice/VOIP logs OneDrive/G drive Other Phone system/cell service logs Other SAAS and in-house application data. Facebook/social media data for company pages. QuickBooks/other accounting systems/business bank account logs POS/Credit card processing systems/PayPal, etc. OSINT to fill in the blanks

Seeking co-founder to build LinkedIn’s biggest rival(curated version)
reddit
LLM Vibe Score0
Human Vibe Score1
ItzdreeThis week

Seeking co-founder to build LinkedIn’s biggest rival(curated version)

How do you connect with likeminded people? You see the polished wins everywhere, but what about the messy drafts , the awkward pitches and the moments you’re not sure you’ve got it right? Problem: The whole idea of founding and starting a business can be super intimidating for some people, specially those who don’t know any founders personally, those who don’t have a large network, those who don’t have rich parents with large networks, those not inserted in an entrepreneurial culture like in the US for example (which is my case). Sometimes all you need is the right support network, and too see others do what you want, to know that it’s possible! Everyone has an “ultimate guide” to make 7 figures or build a business on YouTube but NO ONE shoes you the HOW, just the results… I’ve tried joining founder communities, LinkedIn ,Reddit … you name it. Most of these founder communities are inaccessible for regular people and often ask for you to have an already existing business with a min ARR… or their simply geography based and if you’re not in a certain area you can’t really participate… As of LinkedIn… full of empty AI generated posts about how some random dude raised $10m in 7 days. Okay Jonathan, but what about the HOW?? How did you write your first pitch? How many rejection calls did you get? What is an MVP? There simply isn’t a platform out there to document your founding journey and find inspiration within a community of people who are doing the same as you. What better way to feel motivated then to see someone actually document their process? Solution: I’m working on building a social media platform for aspiring/founders to connect through the RAW, UNFILTERED process of turning ideas into reality in REAL time. It’s all gonna be around the “building in public “ concept and content creation. Picture an instagram/tiktok profile where instead of seeing someone’s dog you see them documenting their founding process—from the moment they had the idea, to the moment they launched, you’re going to see the successes, the fails, the rejected calls, all documented through actual content and not some AI generated LinkedIn post. Imagine if you wanted to learn about how Steve Jobs started Apple , you could simply go through his profile on this app—exactly. To make sure all interactions are meaningful people would have to apply. It’s a truly curated community, with REAL people, building REAL things in REAL time, and not just tell us the story of how they did it… Audience: I’m targeting people who have a burning desire of building a business and early stage founders starting their founder journeys, that don’t have a support network and simply don’t know where to start. People who are tired of watching 30 min “ultimate guides “ on how to make it on YouTube from “business gurus” selling courses. People who haven’t reached the min ARR required to join an “exclusive “ founder a community. People who can’t simply just move to the US to get into the “exclusive” YC combinator. People who want to connect with real people building real things and not anonymous people on Reddit, or LinkedIn influencers again trying to promote their services. I believe in the idea because I’m also part of my audience. Have always wanted to start my own thing just never knew how to and where to find a community of likeminded people . I don’t know any founders myself, I come from a non-entrepreneurial society and I’d pay good money to access a community of REAL passionate founders building REAL things, in REAL time. This would be my first ever business, and I want to share my journey building it and hopefully inspire others to just start so I’ve created a mailing list to keep anyone interested in the project updated on my fails , learnings and successes. I’m not worried about “making it” but just “starting” and hopefully reach the right audience and inspire anyone to start whatever they have marinating in their thoughts. If you’re a founder struggling with staying consistent or an aspiring founder with an insane desire of starting and don’t know how to start, I’d love to get your feedback on what’s stopping you, your challenges starting out and what you’d find useful in such platform. And finally would this be something that interests you?? PS: casually looking for a technical co-founder

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀
reddit
LLM Vibe Score0
Human Vibe Score1
Awkward_Ad_9605This week

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀

Hey Members! I’m excited to share an idea for a new application I’m planning to build: Quick Launch . This AI-driven platform is designed to assist solopreneurs or anyone with an idea in launching their Minimum Viable Products (MVPs) by taking on the roles of the entire team needed for the process. Goal: Assistance in quickly moving from Idea to MVP Before I dive into the details, I’d love to hear your thoughts and feedback. Key Features: Product Creation: From Idea to Product Detailing AI-Generated Q&A: Real-time questions generation one-at-a-time to define the product requirements based on their knowledge levels to convert an Idea into a Product. Market Research Reports: In-depth analysis that identifies product-market fit, competitive landscape, and potential marketing strategies. Sentiment Analysis: Evaluation of user feedback and reactions across multiple subreddits to gauge public opinion on ideas. Product Development: Product Detailing to Actual Product User Story Generation: Identification and creation of comprehensive user stories, tasks, and sub-tasks to facilitate development. AI Project Management: AI agents assume roles of project managers and UI/UX designers to streamline product detailing and development. Integration Capabilities: Seamless integration with popular project management tools like Jira, Asana, and Trello for better workflow management. Target Audience: Solopreneurs: Individuals looking to bring their business ideas to life without extensive resources. Indie Hackers: Entrepreneurs focused on building small projects or startups with minimal overhead. Idea Validators: Anyone with a concept seeking initial validation and market feedback before committing significant resources. If you’re interested in learning more, check out our teaser website: Quick Launch Discussion Question: What features would you find most valuable in an application like this? Are there specific pain points you face when launching an MVP? Your insights would be incredibly helpful as I refine this idea! Looking forward to your thoughts! 🙌

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October
reddit
LLM Vibe Score0
Human Vibe Score1
Aegidius25This week

40% Of SMBs Still Can't Pay Their Rent, Extending High Delinquency From September Into October

https://www.alignable.com/forum/q4s-off-to-a-rough-start-40-of-smbs-still-cant-pay-their-rent October 31, 2023: While the federal government reported a surge in economic growth for the U.S. last week, that news doesn't hold true for many small business owners. In fact, in October polling by Alignable, only 12% said their companies are experiencing significant growth this month. Beyond that, Alignable’s October Rent Report, released today, shows that a whopping 40% of SMBs couldn't even pay their October rent in full and on time. This marks the second consecutive month of a 40% rent delinquency rate -- extending 2023's record high from September through October. These findings are based on responses from 4,246 randomly selected small business owners surveyed from 10/1/23 to 10/30/23, as well as input from 44,000+ other respondents over the past year. As the chart below shows, October's SMB rent delinquency rate is 10 percentage points higher than it was in January, reflecting cumulative economic struggles: increased rents, high interest rates, still-stifling inflation, rising labor costs, and revenues that have declined since this time last year. Rent delinquency rates among small businesses during 2023 based on Alignable surveys So, Why's Rent Delinquency At 40% For A 2nd Month? Here’s the current list of problems contributing to two months' worth of the highest delinquency rate 2023 has seen so far: Consumer Spending Declines On Main Street: Quarterly, we ask about customer spending habits at retailers. This month, 45% of independent Mom and Pop Shops said spending has been down over the last 30 days. Some said it was due to more people spending money online with big retailers like Amazon. This figure is quite high, especially considering that back in July, only 24% reported a drop in consumer spending -- 21 percentage points less severe than it is now. Revenue Troubles: 42% are making half or less of the income they generated monthly prior to COVID. For businesses that are less than three years old, this situation is even worse: 53% of this group reports making half or less of what they generated this time last year. High Interest Rates: Over half of all SMB owners polled said the past 19 months of high interest rates have hurt their margins, reduced revenues, and put their expansion plans on hold, as they don't want to apply for loans. Increased Rent Prices: 50% say they’re being charged more for rent now than they were six months ago, with 15% saying rent has increased by 20% or more. At present, only 37% of pre-COVID businesses have recovered financially from the pandemic era, leaving 63% still striving to make up for time they lost due to COVID, inflationary pressures, and high interest rates. There's a slight silver lining here, though, as the 37% figure is three percentage points higher than it was in September. But, with that said, a recovery rate of 37% after more than three and a half years is still very low and speaks volumes about the ongoing list of troubles small business owners face looking into the rest of 2023. Tech, Manufacturing, Gyms, Beauty & Retail Struggle Examining the rent delinquency landscape in terms of sectors, there's quite a negative shift occurring among some industries in October. Let's look at the charts below to see what's really happening. Sectors most affected by rent delinquency include tech and retail Details on sectors affected by rent delinquency in October This is alarming for a few reasons: The countless technology layoffs at larger companies over the past year appear to be affecting the small companies now, too, who are often dependent on the larger ones as clients. Right now, 54% of science/technology small companies couldn't pay their October rent, up 10 percentage points from September and 16 percentage points since August. There are also some comments in the surveys of technology roles being reduced or replaced by ChatGPT and other AI, which can write software programs. Gyms have been struggling now for a while and now 50% of them can't afford the rent, up 8 percentage points from September. The biggest shift between October and September occurred among manufacturers, partially due to ongoing fluctuation in the price of gas and other inflationary issues. For quite some time, manufacturers were improving a lot in terms of their rent delinquency rates, but in October, they jumped 25 percentage points, doubling their rate, which is now 50%. This is also a record high for manufacturers in 2023. We hope this is just a blip, but we'll see in November. Also due, in part, to fluctuating gas prices and costs of vehicles, 45% of transportation companies couldn't pay October rent in full and on time. That's up 6 percentage points from last month. Sadly, 47% of salon owners couldn't cover October rent, after showing a lot of stability over the past few months. But that stability ended this month, as salons' rent delinquency rates jumped nine percentage points. Though rates have dropped three percentage points in October, a high percentage of retailers are still having trouble paying the rent. Last month, it was 47%. This month, it's better, but is still over 40%, landing at 44%. This is worrisome, especially since Q4 is a "make it or break it" time for many Main Street merchants. Looking more closely at the industries, there was some good news, in that a few others experienced lower delinquency rates in October, including restaurants, which dipped to 40% from 44% in September. Travel/lodging dropped seven percentage points to 38% (from 45% last month), as did education, which is also at 38%, down from 43%. When looking at rent delinquency from the vantage point of the states that are most affected, many surges can be seen between October and September, while a few states saw some dramatic, encouraging declines, too. Rent Troubles Increase For IL, VA, TX, MA, FL, & CO Looking at the states' charts, you can see how tumultuous the rent story has become this fall. Let's first talk about those with significant jumps in their delinquency rates. Here's the rundown: Illinois leads the list once again. After having a better month in September, its delinquency rate has soared, once more, landing at 54% for October (up from 46% last month). In fact, the 54% figure is the highest rate IL-based SMBs have seen in 2023. Virginia was in great shape last month, with a delinquency rate of just 19%. But Virginia-based small business owners have had a very rough month, at least in terms of rent. Now, 50% of them who took our poll say they couldn't cover rent (an increase of 31 percentage points). Texas is third on the list, with an 11-percentage-point lift from 38% in September to 49% in October. MA is next up at 48%, which marks the largest jump on the chart -- 32 percentage points from a low of just 16% in September. Small businesses in Florida have also experienced two challenging months in terms of rent delinquency. Right now, 45% of SMBs there couldn't afford to pay, up nine percentage points from September and 15 percentage points from August. Colorado's businesses regressed in October, hitting a new record high of 40%. That rent delinquency rate jumped 13 percentage points from September to October. While we just covered states with some very high delinquency rates, there were also several more positive swings that have occurred in October. Though encouraging, we'll have to see how long those delinquency rates continue. Here are the most remarkable: New York -- After reaching a record rate of 55% last month, New York's small business owners now report a more stable number: just 29%. That's down 26 percentage points. New Jersey -- New York's neighbor has an even more impressive story in October: only 20% of New Jersey's SMBs couldn't pay rent this month, a record low over at least the past 14 months, down 34 percentage points from a record high of 54%. Michigan -- Similarly, Michigan's small business owners boast a rate of just 20%, down from 45% in September.

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

How I Started Learning Machine Learning
reddit
LLM Vibe Score0
Human Vibe Score1
TechPrimoThis week

How I Started Learning Machine Learning

Hello, everyone. As promised, I'll write a longer post about how I entered the world of ML, hoping it will help someone shape their path. I'll include links to all the useful materials I used alongside the story, which you can use for learning. I like to call myself an AI Research Scientist who enjoys exploring new AI trends, delving deeper into understanding their background, and applying them to real products. This way, I try to connect science and entrepreneurship because I believe everything that starts as scientific research ends up "on the shelves" as a product that solves a specific user problem. I began my journey in ML in 2016 when it wasn't such a popular field. Everyone had heard of it, but few were applying it. I have several years of development experience and want to try my hand at ML. The first problem I encountered was where to start - whether to learn mathematics, statistics, or something else. That's when I came across a name and a course that completely changed my career. Let's start You guessed it. It was Professor Andrew Ng and his globally popular Machine Learning course available on Coursera (I still have the certificate, hehe). This was also my first official online course ever. Since that course no longer exists as it's been replaced by a new one, I recommend you check out: Machine Learning (Stanford CS229) Machine Learning Specialization These two courses start from the basics of ML and all the necessary calculus you need to know. Many always ask questions like whether to learn linear algebra, statistics, or probability, but you don't need to know everything in depth. This knowledge helps if you're a scientist developing a new architecture, but as an engineer, not really. You need to know some basics to understand, such as how the backpropagation algorithm works. I know that Machine Learning (Stanford CS229) is a very long and arduous course, but it's the right start if you want to be really good at ML. In my time, I filled two thick notebooks by hand while taking the course mentioned above. TensorFlow and Keras After the course, I didn't know how to apply my knowledge because I hadn't learned specifically how to code things. Then, I was looking for ways to learn how to code it. That's when I came across a popular framework called Keras, now part of TensorFlow. I started with a new course and acquiring practical knowledge: Deep Learning Specialization Deep Learning by Ian Goodfellow Machine Learning Yearning by Andrew Ng These resources above were my next step. I must admit that I learned the most from that course and from the book Deep Learning by Ian Goodfellow because I like reading books (although this one is quite difficult to read). Learn by coding To avoid just learning, I went through various GitHub repositories that I manually retyped and learned that way. It may be an old-fashioned technique, but it helped me a lot. Now, most of those repositories don't exist, so I'll share some that I found to be good: Really good Jupyter notebooks that can teach you the basics of TensorFlow Another good repo for learning TF and Keras Master the challenge After mastering the basics in terms of programming in TF/Keras, I wanted to try solving some real problems. There's no better place for that challenge than Kaggle and the popular Titanic dataset. Here, you can really find a bunch of materials and simple examples of ML applications. Here are some of my favorites: Titanic - Machine Learning from Disaster Home Credit Default Risk House Prices - Advanced Regression Techniques Two Sigma: Using News to Predict Stock Movements I then decided to further develop my career in the direction of applying ML to the stock market, first using predictions on time series and then using natural language processing. I've remained in this field until today and will defend my doctoral dissertation soon. How to deploy models To continue, before I move on to the topic of specialization, we need to address the topic of deployment. Now that we've learned how to make some basic models in Keras and how to use them, there are many ways and services, but I'll only mention what I use today. For all my ML models, whether simple regression models or complex GPT models, I use FastAPI. It's a straightforward framework, and you can quickly create API endpoints. I'll share a few older and useful tutorials for beginners: AI as an API tutorial series A step-by-step guide Productizing an ML Model with FastAPI and Cloud Run Personally, I've deployed on various cloud providers, of which I would highlight GCP and AWS because they have everything needed for model deployment, and if you know how to use them, they can be quite cheap. Chose your specialization The next step in developing my career, besides choosing finance as the primary area, was my specialization in the field of NLP. This happened in early 2020 when I started working with models based on the Transformer architecture. The first model I worked with was BERT, and the first tasks were related to classifications. My recommendations are to master the Transformer architecture well because 99% of today's LLM models are based on it. Here are some resources: The legendary paper "Attention Is All You Need" Hugging Face Course on Transformers Illustrated Guide to Transformers - Step by Step Explanation Good repository How large language models work, a visual intro to transformers After spending years using encoder-based Transformer models, I started learning GPT models. Good open-source models like Llama 2 then appear. Then, I started fine-tuning these models using the excellent Unsloth library: How to Finetune Llama-3 and Export to Ollama Fine-tune Llama 3.1 Ultra-Efficiently with Unsloth After that, I focused on studying various RAG techniques and developing Agent AI systems. This is now called AI engineering, and, as far as I can see, it has become quite popular. So I'll write more about that in another post, but here I'll leave what I consider to be the three most famous representatives, i.e., their tutorials: LangChain tutorial LangGraph tutorial CrewAI examples Here I am today Thanks to the knowledge I've generated over all these years in the field of ML, I've developed and worked on numerous projects. The most significant publicly available project is developing an agent AI system for well-being support, which I turned into a mobile application. Also, my entire doctoral dissertation is related to applying ML to the stock market in combination with the development of GPT models and reinforcement learning (more on that in a separate post). After long 6 years, I've completed my dissertation, and now I'm just waiting for its defense. I'll share everything I'm working on for the dissertation publicly on the project, and in tutorials I'm preparing to write. If you're interested in these topics, I announce that I'll soon start with activities of publishing content on Medium and a blog, but I'll share all of that here on Reddit as well. Now that I've gathered years of experience and knowledge in this field, I'd like to share it with others and help as much as possible. If you have any questions, feel free to ask them, and I'll try to answer all of them. Thank you for reading.

What Reinforcement Learning Method Should I Use for Poker AI with LLMs?
reddit
LLM Vibe Score0
Human Vibe Score1
godlover123451This week

What Reinforcement Learning Method Should I Use for Poker AI with LLMs?

Hey everyone, I’m working on a poker AI project, where I’m training a large language model (LLM) to predict poker actions from given game states (check, call, bet, raise, etc.). My end goal is to create a model that can play poker at a high level, primarily by self-play and opponent modeling. However, I’m running into some challenges that I hope you can help me with! Here's the situation: Training Method: I’m using supervised fine-tuning (SFT) on real poker hand history data to initially teach the LLM how to predict poker actions from game states. This means that the model learns from examples of past games, predicting the actions that players took in various situations. Self-Play Setup: I plan to eventually move to self-play, where the LLM will play against itself (or other types of models that I create to simulate different play styles). I’ll use these self-play sessions to improve the model over time. Opponent Pool: I’m creating 6 types of poker players (Loose Aggressive, Loose Passive, Tight Aggressive, Tight Passive, Maniac, and Nit), each trained at 5 different skill levels (Novice, Beg\*nner, Intermediate, Advanced, Expert). This gives me a decent range of opponent behavior for training. The problem: Here’s the catch: The LLM I’m using only outputs discrete actions (e.g., bet 3BB, raise to 10BB, etc.) with no access to the probabilities of actions, so I can't directly use methods like policy gradients or Q-learning that rely on action probabilities or continuous action spaces. This makes applying traditional RL methods a bit tricky. My question: Given that I don't have access to action probabilities, what RL method or strategy should I pursue to improve my model? Specifically, I’m looking for a way to: Incorporate self-play with reward-based learning. Refine the model through reinforcement learning, without the need for continuous probabilities. Ensure the model doesn’t just overfit to its own prior behavior but learns to adapt and exploit different strategies in poker. I’ve considered a few approaches like reward-weighted supervised fine-tuning or using simpler RL techniques like Monte Carlo updates, but I’m not sure which would work best with the LLM setup I have. I've also considered Q-learning or Deep Q-learning. Any advice or suggestions on which RL approach I should take given my situation would be greatly appreciated! Yes I used AI to write this queston. But it captures everything I want to say, and I suck at writing.

Teaching an AI to Play Mario: A Learning Journey
reddit
LLM Vibe Score0
Human Vibe Score1
CivilLifeguard189This week

Teaching an AI to Play Mario: A Learning Journey

TLDR: I've always wanted to learn reinforcement learning, but the notation and concepts often seemed overwhelming (and scary). So, \~3 months ago, I set myself a challenge: Train an AI to Speedrun Mario Watch the progression here: https://youtu.be/OQitI066aI0 &#x200B; Full Story: Three months ago, I stared at the dense forest of Reinforcement Learning (RL) papers and felt like Mario facing Bowser for the first time: unequipped and overwhelmingly outmatched. The notation seemed like hieroglyphics, and terms like "policy gradients" felt like they belonged in a sci-fi novel, not a beginner's project. But RL always seemed so cool, and I was really determined to achieve my goal. So, I started with the Sutton & Barto RL textbook, learning things like the Multi-Armed Bandit problem and MDPs working my way up to Actor-Critic methods. That book is literal gold & I highly recommend you work through it (even though it can be tough at times). I tried everything from random courses online to books on amazon & this textbook has been by far the most clear and effective way to learn RL. The biggest issue with the textbook is you learn a lot of theory, but don't learn implementation. So, I would go through a chapter a week & set aside Friday + the weekend to actually implement what I learned (usually by watching youtube tutorials & looking at Github Repos). Eventually, while searching for practical resources for implementing PPO, I stumbled upon a GitHub repository that literally trained an AI to play Mario. Rather than just cloning and running the code, I took a deeper approach. I aimed to understand the repository thoroughly, ensuring each line of code made sense in the context of what I had studied. But of course, this wasn't easy. One of the biggest issues was my hardware limitation. I was working on an old Mac. So, I started using Google Collab, but that had its own problems (session timeouts & limited GPU access). Ultimately, I found AWS Sagemaker to be pretty good. &#x200B; After rewriting the code, I felt confident it would work because I understood every aspect of it. So, I trained the AI to play Mario across a variety of different levels (took a long time and a lot of trial and error with the learning rate). It feels amazing seeing your theoretical knowledge translate into tangible results & this project gave me a big confidence boost. &#x200B; Anyways I made a video showing off the results (Note that I simplified the technical parts for it to reach a wider audience): https://youtu.be/OQitI066aI0 &#x200B; Feel free to drop any questions or feedback, I'm more than happy to help or chat about my experiences. I hope my journey can inspire some of you who might be feeling overwhelmed with the idea of diving into reinforcement learning or any other area of AI. Remember, the hardest part is often taking the first step. Once you start, the momentum will carry you forward. Thank you for reading my super long post and sharing in my little success story! 🚀🕹️🎮

ZeroToHeroML: Beginner-Friendly ML & AI Course (Free)
reddit
LLM Vibe Score0
Human Vibe Score0
DizDThis week

ZeroToHeroML: Beginner-Friendly ML & AI Course (Free)

Hey r/learnmachinelearning! A friend of mine, who's been a software developer at Sony for 10 years, recently expressed interest in learning Machine Learning (ML) and Artificial Intelligence (AI). Leveraging my background in ML and neural computation (learned at UCSD) to create a beginner-friendly course guiding him through the basics and into more complex projects. Foundational Concepts: Predicting House Prices (Regression): Master regression techniques to forecast housing prices based on various factors. Iris Flower Species Prediction (Classification): Learn classification algorithms by predicting flower species using the famous Iris dataset. Overcoming Overfitting: Explore methods to prevent models from overfitting and enhance their generalizability. In Progress: Customer Segmentation (Unsupervised Learning): Delve into unsupervised learning to group customers based on purchase history or demographics (valuable for targeted marketing campaigns). Deep Learning for Image Recognition: Implement Convolutional Neural Networks (CNNs) to build models that recognize objects or scenes in images. Natural Language Processing Sentiment Analysis: Analyze the sentiment (positive, negative, or neutral) expressed in text data (e.g., reviews, social media posts) using NLP techniques. Introduction to Reinforcement Learning: Get acquainted with the fundamentals of reinforcement learning by creating an agent that learns to navigate a maze. Want to Learn or Contribute? I thought I'd share ZeroToHeroML here so others who want to learn ML/AI or know someone who does can benefit from this free resource! &#x200B; Fork the repo: https://github.com/DilrajS/ZeroToHeroML Share with others interested in ML/AI! Pull requests welcome (help the community grow!). All help is appriciated! Let's conquer ML/AI together!

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

Learning AI for Business Leaders
reddit
LLM Vibe Score0
Human Vibe Score0
Bills-WideRightThis week

Learning AI for Business Leaders

Hello Community, For the better part of 2 months I have been reading up on everything in getting a better understanding of the fundamentals of AI - from history of AI to reading the Google 8’s peer reviewed paper on the advent of transformers. I feel as though I am running in circles at times an not following a guided path approach to learning. I’m 40, work in international development in a leadership role - though I have a background in corporate finance and tech. I’m not an engineer, nor do I have the ambition of such a career pivot. However I do want to learn, be abreast, and know enough about the space when evaluating (and proposing) AI related opportunities - my role now should be a path towards a chief innovation officer for a development agency within the next 3-4 years. My sources have been basically everything I can find from tech blogs, WaPo, financial times, economist, and random internet searches. I have completed IBM’s Fundamental on AI course. However, I feel there no structure in learning as I have been piecemealing from so many different sources. Essentially I care about business cases and being able to confidently talk about AI. And not building and deploying a product. MIT and UPenn have some courses on AI for leaders, however, as the space is moving so fast I’m not confident how current their materials are. My ask: Are there any courses (or learning approaches) you recommend that is less-code and more focus on concept and applications I should do? Is my approach to learning too broad and I should focus on a subset of AI such as ML or specifically GenAI since it seems most applications are currently byproducts of it. Many thanks in advance for any support - truly appreciate it.

Built a multi-agent AI mental health assistant (7 agents, backend automated, no-code stack)
reddit
LLM Vibe Score0
Human Vibe Score1
CapitalCategory4044This week

Built a multi-agent AI mental health assistant (7 agents, backend automated, no-code stack)

Been working on this little side project and finally got it to a working version. It’s an AI-powered mental health assistant — not just a chatbot, but a system that can retrieve user history, analyze input, access data in real-time, and suggest personalized treatment plans. UI Chat Tech stack: Loveable + Momen How it’s structured: It uses 7 specialized AI agents, each responsible for a niche task — chat, generate professional responses, summarize user info, classify intent, etc. Agent Team The main agent (the chat one) will call other agents in the backend via automated workflows. It keeps track of user data (symptoms, conversations, medical history) and updates it in real time — all triggered automatically. Everything runs in the backend to reduce manual steps and minimize errors. How it’s built: Started by drafting the UI with Loveable AI — it auto-generated a 7-page interface from a product brief, which saved me time. (Didn’t use it for the live app though — good for prototyping, but I wanted more control for complex backend workflows.) Rebuilt the UI and database in Momen, since I needed deeper control over data flow and backend logic. The entire AI agent system and backend workflows were built in Momen as well. So I can make the agents collaborate with each other. The main chat agent invokes backend workflows to call other agents when needed. Entire flow looks like this: the user sends a message, the system: → pulls in the latest user data→ triggers the right agent(s) based on the input→ responds in real-time→ quietly summarizes and updates everything in the background. FlowChart It’s still an MVP, but the multi-agent setup + automated backend feels pretty scalable.This was a super fun build and I learned a lot about orchestrating AI workflows. Would love any feedback or thoughts on how to improve this.

I built a library to visualize and edit audio filters
reddit
LLM Vibe Score0
Human Vibe Score1
AlexStreletsThis week

I built a library to visualize and edit audio filters

Hey everyone! TLDR: No fancy AI Agents or trendy micro-SaaS here — just an old-school library. Scroll down for the demo link! 🙃 App Demo The Story Behind Several years ago, I deep-dived into reverse engineering the parameter system used in VAG (Volkswagen, Audi, Porsche, etc) infotainment units. I managed to decode their binary format for storing settings for each car type and body style. To explain it simply - their firmware contains equalizer settings for each channel of the on-board 5.1 speaker system based on cabin volume and other parameters, very similar to how home theater systems are configured (gains, delays, limiters, etc). I published this research for the car enthusiast community. While the interest was huge, the reach remained small since most community members weren't familiar with hex editors. Only a few could really replicate what I documented. After some time, I built a web application that visualized these settings and allowed to unpack, edit and repack that data back into the binary format. Nowadays The original project was pretty messy (spaghetti code, honestly) and had a very narrow focus. But then I realized the visualization library itself could be useful for any audio processing software. When I first tried to visualize audio filters with that project, I hit a wall. Most charting libraries are built for business data, all those "enterprise-ready visualization solutions". But NONE of them is designed for audio-specific needs. D3.js is the only real option here — it’s powerful but requires days of digging through docs just to get basic styling right. And if you want interactive features like drag-and-drop? Good luck with that. (Fun fact: due to D3's multiple abstraction layers, just the same filter calculations in DSSSP are 1.4-2x faster than D3's implementation). So, I built a custom vector-based graph from scratch with a modern React stack. The library focuses on one thing - audio filters. No unnecessary abstractions, no enterprise bloat, just fast and convenient (I hope!) tools for tools for audio processing software. Core Features Logarithmic frequency response visualization Interactive biquad filter manipulation Custom audio calculation engine Drag-and-drop + Mouse wheel controls Flexible theming API Technical Details Built with React + SVG (no Canvas) Zero external dependencies besides React Full TypeScript support Live Demo & Docs & GitHub This is the first public release, landing page is missing, and the backlog is huge, and docs do not cover some aspects. (You know, there's never a perfect timimng - I just had to stop implementing my ideas and make it community driven). I'd love to see what you could build with these components. What's missing? What could be improved? I'm still lacking the understanding of how it could gain some cash flow, while staying open-source. Any ideas?

I am building a free app to fight propaganda online. Need your help
reddit
LLM Vibe Score0
Human Vibe Score-1
Used_Park_1937This week

I am building a free app to fight propaganda online. Need your help

So long story short I was recently bombarded with an insane amounts of right-wing propaganda online and Russian bots so I decided I can somehow simplify the validation of the facts that are posted by people online. I have made a research of tools that are currently available for that and I've found that ChatGpt is quite biased with some topics and it lacks data about latest issues even when search mode is enabled. At least it takes some time for it to update it's memory or something Then I've checked a Google Fact Check tools which were even worse. Bad UI... I had no idea how it validates things because usually proof links that Google provides are quite random. So I decided to build a tool that simplifies the research. The idea was simple - you have a claim - you paste it in app and it goes on web collects articles on topic then processes them with ai and makes a decision based on those articles. The app is running right now. It is completely free and no ads. Also it generates a small article where it tells about how it made a conclusion. Also I need your help - How can I make this app better? Also feel free to test your claims so that I can have more real data to tune this app on. I believe that this is something that is necessary for all of us especially in those crazy times! Here is the link - truthorfake.com https://preview.redd.it/vp5fbzpyfkne1.png?width=854&format=png&auto=webp&s=84df7913382634614814866c8df743d8378d28dd

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

Just completed a new type of language learning website - read popular stories scaled to different reading levels
reddit
LLM Vibe Score0
Human Vibe Score1
creedaaronThis week

Just completed a new type of language learning website - read popular stories scaled to different reading levels

As a language learner and software developer, I bootstrapped my project superlang.com over the past year working on the side. There is a mobile friendly web app now, and iOS/Android apps coming in a few months. A year ago I discovered the concept of "comprehensible input" as a way to help me learn German. Even if it's not a silver bullet, it sounded pretty great. Rather than drilling vocab or looking at grammar charts, I could "just read" and acquire the language. I picked up some fairy tales in German, and stories like Alice in Wonderland. Unfortunately, I couldn't really read them. I had to stop every sentence to look up words and try and decipher sentence constructions. Then I turned to some purpose built simple stories for German beginners. But there was a different problem... these were not really stories with any real plot. I could only read so many "Hans goes to the market" type stories before losing interest. My idea was to try to get the best of both worlds somehow. What if I could take a real story, say Alice in Wonderland (or even War and Peace), and dial the difficulty down to my level without losing the plotline. That way, beginners can start right away with something basically comprehensible. Then, you could also re-read the same story at increasing difficulty levels as you gain confidence. As a cherry on top, more illustrations would help with comprehension so each page could have a picture. Is it revolutionary? Maybe, maybe not. I am building off a well established idea of "graded readers" which are simplified stories meant for learning languages. And there are somewhat similar ideas out there now that AI is good at simplifying text, but none that really take this idea where it needs to be with many preloaded stories, multiple difficulty levels, high quality human verified text, and all the bells and whistles. I spent a year building Superlang and it is ready to put out there. Some quick notes: There are 3 languages so far, intended for native English speakers: German, French, and Spanish There are 3 difficulty levels you can set on each story: beginner (roughly A1-A2), intermediate (roughly A2-B1), and advanced (the same level as the original story, but typically B2+) There is premium version as producing the content was somewhat expensive. You can still do a lot of reading on the free version. I have done no marketing yet, except for this post :) The implementation is a combination of AI, and human proofreading and reviewing. In particular, the simplification of stories is very heavily AI driven. The illustrations for each page are AI as well. For translation, as many of you may be aware new LLM models are typically better than Google translate, but still far from perfect. I am very much a proponent of keeping real people in the loop, and so I have real people proofread the translations. That's why there are only about 700 pages of content so far and not tens of thousands. Let me know what you think, and if you find it helpful! Alice in Wonderland - beginner level German Romeo and Juliet - beginner level Spanish

An Algorithm for Making Truly Stand-Out Advertising Content (+ something more | Part 1)
reddit
LLM Vibe Score0
Human Vibe Score1
asealey1This week

An Algorithm for Making Truly Stand-Out Advertising Content (+ something more | Part 1)

Hi everyone. my friend and I are software engineers and new to marketing. A few months ago we decided to leverage our software skills for a colleague in ecommerce. It started by implementing a Flux.1 model, then began using texture-based recreations with a canny mask, and then found that we could optimize on both with an added layer of inpainting...and the list goes on. This is the first of a series of posts here about it and I look forward to learning from your feedback. I realized that the most difficult parts of the marketing process when I started out (and most likely for other beginners too) are: Customer Acquisition Costs / Brand Differentiation: Competition is intensifying and it is getting more difficult to stand out in crowded markets and target ad spend more effectively. Maintaining Authenticity at Scale / Data Overload: Balancing growth with authenticity and leveraging available data to successfully engage with customers is a big ask. Creative Fatigue: Maintaining multiple marketing channels in hard, and it becomes harder when you're constantly demanding more and more creative content for campaigns. For 1) I tried using AI to help me summarize, systematize, and gain insights from the information available for a given brand or product (from a page link, prompt, input image, etc.). I know AI is everywhere now, many people are using it unnecessarily and many people are skeptical about it. However, I know from experience, that it is quite helpful in gaining insights/summarizing large amounts of data, and helping people make sense of the creative content, strategy, campaign, etc., that should be created. For 2) By leveraging reviews, forums, and other relevant brand information, AI is able to maintain the story that your brand currently tells, and enhance it based on how your customer base. For 3) Faster results means less creative fatigue- this translates to an easier time managing omnichannel marketing efforts and scaling advertising. If you're interested, please have a look at the result at madsimpleads.com You’ll need to log in to access the solution, and I'll add credits to your account to try it out! (we want to prevent from random people or bots using it because I'm paying to multiple providers for model access). DM me here or drop me a line at austin@madsimpleads.com if you need more. Thank you so much, I'll be happy to get your thoughts I hope the website will help with your advertising, please reach out if you like what I do and want to support the project! Disclaimers: the website looks a bit rough in terms of UI/UX, but we tried focusing on the functionality first available on mobile, works better on desktop I hope this doesn't come across as trying to advertise for my business or breaking any of the community rules. genuinely looking for feedback. Thank you

I made a super niche app for sailors and scaled it to 500k downloads
reddit
LLM Vibe Score0
Human Vibe Score0.5
TechPrimoThis week

I made a super niche app for sailors and scaled it to 500k downloads

I started developing this app in 2016, and it was my first app ever. I already had several years of programming experience. Since I was studying maritime navigation, I came up with the idea of creating a maritime app to help students with various nautical calculations and learn maritime regulations. Although I had no experience in mobile app development, I chose the Ionic framework and started development gradually. First Version The first version took me about four months to develop because I literally had to learn everything from scratch: how to develop mobile apps, how to publish them, and everything needed to enable downloads on the app stores. Many of you might recognize me from my story about developing Sintelly and its late monetization. I made the same mistake with this maritime app. At that time, in my country, there was no possibility of earning through in-app purchases, only through ad displays. Since the app was predominantly downloaded in countries like India, the Philippines, and Indonesia, the ad revenue was quite low, and after some time, I removed the ads. Abandonment and Realization As I started developing other apps, this one fell into obscurity. I even just remembered that I needed to renew the domain, which resulted in losing it. The domain buyer tried to sell it back to me for years for $20k, which was absurd. All this led me to rebrand and start working on this app again. Interestingly, during these 8 years, the app never showed a declining trend in installations or active users. I'll share some numbers to give you insight: Total installations (Android + iOS): 501,000 Active installations (Android): 48,000 Monthly active users: 20,000 Average rating: Android 4.8, iOS 4.7 When I considered these numbers, I realized they weren't bad at all and that I was far ahead of most competitors. This led to my decision to rebrand and create a new website. I quickly built the website using WordPress and published lots of existing content from the app. What surprises me is that today, after a year and a half, the website has about 8-10k monthly organic visits. Choosing a Direction Based on all this, I decided it was time to create a Premium version and start selling the app. Since I've been working with AI for many years (which I've written about here), I started thinking about using AI to help seafarers speed up some of their tasks. This led to the idea of creating a multi-agent system equipped with numerous tools to help seafarers. I developed various agents with functionalities, including retrieving maritime weather information, locating and tracking ships, doing various nautical calculations, calculating the shortest maritime routes and unit conversions, and learning about all courses and maritime regulations. All this required considerable work, but thanks to tools like Cursor and Claude, I implemented it in less than four weeks. Last week, I published this new version and started selling subscriptions, and I can already boast that I've earned slightly over $100. This isn't much, but I'm happy to see my first app generating some income, which I always thought impossible. Along this journey, I learned many lessons, and the most important one is to never give up or write off a product. With a little effort, everything can be brought back to life and secure at least some passive income, enough for your morning coffee. Additionally, I learned how to develop mobile apps, which has shaped my career since then. If it weren't for this app, I probably would never have become a developer. I have numerous plans for what to add next and how to improve. I'll base everything on AI features and push the app in that direction.

I Made $20K in 2 Months by Building in Public on X
reddit
LLM Vibe Score0
Human Vibe Score1
nebulasyncThis week

I Made $20K in 2 Months by Building in Public on X

Hey everyone, I wanted to share my journey of making $20K in just 2 months by leveraging Twitter (X) and building in public. It’s been an exciting ride, and I hope my story inspires others to take action on their ideas. Here’s exactly what I did: Building in Public I started sharing everything about my work openly. My wins, struggles, and process. I showed: How I build MVPs for clients. The tools I use (Next.js, Supabase, Cursor AI, etc.). The challenges I face and how I solve them. Transparency builds trust, and trust brings clients. Consistency is Key For the past 2 months, I’ve posted consistently on X, even when I felt like no one was watching. Here’s what I focused on: Sharing value (pro tips, workflows, tools). Asking for advice and engaging with my community. Highlighting my projects and client work. Building an audience takes time, but showing up daily pays off. Personal Brand = Inbound Clients I never did any “engagement farming” or gimmicky posts. I just shared my knowledge, and it led to over 35M views on my tweets and 7K followers. Many of these followers turned into inbound client leads. I’ve always believed: Share value for free, and charge for implementation. The Power of Community Engaging with my community on X has been game-changing. People have: Helped refine my processes. Shared valuable tools and advice. Connected me to opportunities I wouldn’t have found otherwise. Key Takeaway: You don’t need a perfect process or a huge following to start. Be consistent. Build in public. Share your journey. In 2 months, I’ve gone from wondering if this would work to making $20K by simply showing up and adding value. If you’re thinking about building in public or starting a personal brand, DO IT. It works. Feel free to ask me anything. I’m happy to share more details about my process, tools, or lessons learned! Let’s build together.

I just built footyGPT.com , an AI assistant for football stats, data and data-backed predictions
reddit
LLM Vibe Score0
Human Vibe Score1
footyamigoThis week

I just built footyGPT.com , an AI assistant for football stats, data and data-backed predictions

Hey Amigos! I am based in London and I build lots of fun tools and websites in the sports data space to help sports bettors make better data-driven decisions. One of my recently launched ones is footyGPT.com I have tried to launch this earlier in 2023 but my team were not able to achieve the end goal (I am a non-technical founder) Funny story, I bought the domain footygpt(dot)com back in 2023 and let it expire because i felt like since we weren't able to acheieve the end goal, there was no need to keep the domain. But 2 months ago I started working on it one more time with a different team, and since AI has become "more friendly" than it was early in 2023, it was a much smoother ride and i was so scared someone else had already bought the domain, only to find out it was still available to buy! I was so scared I lost such a clean domain. Below is an example of how footyGPT works: Ask FootyGPT for Predictions for a Given Fixture how footyGPT works \(case study\) Ask FootyGPT to Scrape and Summarize Data from a Stats Website using footyGPT to scrape football stats websites Ask FootyGPT for Sentiments and Online Opinion About a Given Match or Events. ask footygpt for up to date sentiments and opinions about football matches Ask FootyGPT to Find Teams with Certain Streaks, Trends, or Stats https://reddit.com/link/1frdk04/video/9oyaj5ldqjrd1/player If you are into football (soccer) please give it a spin and let me know your thoughts! Cheers Amigos

Looking for a technical co-founder to build LinkedIn’s rival
reddit
LLM Vibe Score0
Human Vibe Score1
ItzdreeThis week

Looking for a technical co-founder to build LinkedIn’s rival

How do you connect with likeminded people? You see the polished wins everywhere, but what about the messy drafts , the awkward pitches and the moments you’re not sure you’ve got it right? Problem: The whole idea of founding and starting a business can be super intimidating for some people, specially those who don’t know any founders personally, those who don’t have a large network, those who don’t have rich parents with large networks, those not inserted in an entrepreneurial culture like in the US for example (which is my case). Sometimes all you need is the right support network, and too see others do what you want, to know that it’s possible! Everyone has an “ultimate guide” to make 7 figures or build a business on YouTube but NO ONE shoes you the HOW, just the results… I’ve tried joining founder communities, LinkedIn ,Reddit … you name it. Most of these founder communities are inaccessible for regular people and often ask for you to have an already existing business with a min ARR… or their simply geography based and if you’re not in a certain area you can’t really participate… As of LinkedIn… full of empty AI generated posts about how some random dude raised $10m in 7 days. Okay Jonathan, but what about the HOW?? How did you write your first pitch? How many rejection calls did you get? What is an MVP? There simply isn’t a platform out there to document your founding journey and find inspiration within a community of people who are doing the same as you. What better way to feel motivated then to see someone actually document their process? Solution: I’m working on building a social media platform for aspiring/founders to connect through the RAW, UNFILTERED process of turning ideas into reality in REAL time. It’s all gonna be around the “building in public “ concept and content creation. Picture an instagram/tiktok profile where instead of seeing someone’s dog you see them documenting their founding process—from the moment they had the idea, to the moment they launched, you’re going to see the successes, the fails, the rejected calls, all documented through actual content and not some AI generated LinkedIn post. Imagine if you wanted to learn about how Steve Jobs started Apple , you could simply go through his profile on this app—exactly. To make sure all interactions are meaningful people would have to apply. It’s a truly curated community, with REAL people, building REAL things in REAL time, and not just tell us the story of how they did it… Audience: I’m targeting people who have a burning desire of building a business and early stage founders starting their founder journeys, that don’t have a support network and simply don’t know where to start. People who are tired of watching 30 min “ultimate guides “ on how to make it on YouTube from “business gurus” selling courses. People who haven’t reached the min ARR required to join an “exclusive “ founder a community. People who can’t simply just move to the US to get into the “exclusive” YC combinator. People who want to connect with real people building real things and not anonymous people on Reddit, or LinkedIn influencers again trying to promote their services. I believe in the idea because I’m also part of my audience. Have always wanted to start my own thing just never knew how to and where to find a community of likeminded people . I don’t know any founders myself, I come from a non-entrepreneurial society and I’d pay good money to access a community of REAL passionate founders building REAL things, in REAL time. This would be my first ever business, and I want to share my journey building it and hopefully inspire others to just start so I’ve created a mailing list to keep anyone interested in the project updated on my fails , learnings and successes. I’m not worried about “making it” but just “starting” and hopefully reach the right audience and inspire anyone to start whatever they have marinating in their thoughts. If you’re a founder struggling with staying consistent or an aspiring founder with an insane desire of starting and don’t know how to start, I’d love to get your feedback on what’s stopping you, your challenges starting out and what you’d find useful in such platform. And finally would this be something that interests you?? Feel free to subscribe to get new updates 🫶🏼 : https://mailchi.mp/037c56b89994/d-founder PS: casually looking for a technical co-founder

Things I did to promote my product, and how they turned out
reddit
LLM Vibe Score0
Human Vibe Score1
laike9mThis week

Things I did to promote my product, and how they turned out

(I will share more updates in the future, you can find me on Twitter and/or Mastodon) Ask any ten indie developers about the toughest part of their job, and nine will likely say "marketing." I recently got a taste of this firsthand when I launched Xylect. Here's a rundown of my promotional attempts - hopefully, my experiences can help fellow developers out there. Podcast Community (✅ Success) I kicked things off by promoting Xylect in my podcast listener group. It wasn't a blockbuster, but I managed to sell a few copies and got some invaluable feedback from friends. Shoutout to those early supporters! Reddit r/macapps (✅ Success) Having had some luck promoting open-source projects on Reddit before, I decided to make r/macapps my first stop in the English-speaking world. I made an app to help you automate boring tasks with one click This post turned out to be a hit! I sold about ten copies and got a ton of useful feedback. Users pointed out compatibility issues with PopClip and suggested improvements for the website. One Italian user even requested localization, which I happily added. https://preview.redd.it/y4fuwh6hleqd1.png?width=959&format=png&auto=webp&s=7bb1b68cbf8a4f94998999e0832b9b7bd85bac67 https://preview.redd.it/8uu4cmyhleqd1.png?width=683&format=png&auto=webp&s=8f1744636aee8074b0e7491a334ef06076b143b0 I also got an intriguing email from a French user - more on that later. More Reddit Posts (❌ Failure) Riding high on my r/macapps success, I branched out to r/SideProject, r/Entrepreneur, and r/indiehackers. These subreddits frown upon direct self-promotion, so I took a softer approach with an article: The unexpected emotional cost of being an indiehacker While the article was heartfelt, it fell flat. Across all three posts, I got a grand total of three comments - two of which were complaints about the font size on mobile. Needless to say, I didn't sell a single copy. Hacker News (❌ Failure) As one of the tech world's major forums, I had to give Hacker News a shot. I wasn't too optimistic, given my past experiences there. Posting on HN feels like a mix of luck and dark magic. As expected, my post vanished without a trace - no comments, no sales. I might give it another go someday. If you're curious, you can check out my previous HN submissions. Tools Directory Websites (❌ Failure) These sites have a simple premise: you list your app, they display it. Seemed like an easy way to get some backlinks, right? Well, I learned the hard way that it's not that simple. I stumbled upon a Reddit post where someone claimed to have made a killing with their directory site in just a few days. The catch? Each listing cost $19. The site had a handful of apps listed, so I thought, "Why not? Early bird gets the worm." I paid up and listed Xylect. Spoiler alert: all I got was $19 poorer 🥲 Lesson learned: These directory sites won't magically sell your product. At best, they're just glorified backlinks. There might be some value in paid promotions on these platforms, but I can't speak to that from experience. V2EX (❌ Failure) After striking out in the English-speaking world, I turned my attention to the Chinese market, starting with V2EX (think of it as China's hybrid of HN and Reddit). This turned out to be my most unexpected flop. Here's the post: [\[Launch Discount\] Mac's most powerful AI search (Perplexity + Wikipedia + Google), boost your efficiency tenfold with one click. No API key required, no prompt needed, no token limit 🔥 - V2EX](https://www.v2ex.com/t/1064930?p=1#reply36) I'd seen decent engagement on other promo posts, so I had high hopes. I posted late at night (US time) and went to bed dreaming of waking up to a flood of comments. Reality check: The next morning, I had exactly one reply - from Kilerd, a loyal podcast listener showing some love. I was baffled. After re-reading my post, I realized I'd missed a crucial element: promo codes. A quick scan of popular posts confirmed my suspicion. Nearly every successful promo post was offering codes, and most comments were just base64-encoded email addresses. Talk about a facepalm moment. I scrambled to add a note about an upcoming free trial and invited users to drop their emails. This got the ball rolling with some code requests, but by then, the damage was done. The post fizzled out, and I didn't sell a single copy 🫠 A French Friend's Newsletter (✅ Success) At this point, my promotional efforts were looking pretty grim. My sales chart had a depressing stretch of flatline. But then, a glimmer of hope appeared in my inbox. Remember that French user I mentioned earlier? He ran a newsletter called vvmac and offered to feature Xylect if I added French support and sent him a free license. It was an offer I couldn't refuse. What followed was a crash course in French localization (thank you, Claude!) and the start of an incredible partnership. This guy was the most thorough beta tester I've ever encountered. We exchanged over sixty emails, covering everything from translations to UI tweaks to bug fixes. His response time was lightning-fast - I'd fix a bug, and five minutes later, he'd confirm it was sorted. The result? A much-improved Xylect and a glowing feature in his newsletter. https://preview.redd.it/ylcq2wxoleqd1.png?width=991&format=png&auto=webp&s=ee395110f50417d5c7f61318f27bf3dc30247809 I'm still in awe of his dedication. He single-handedly transformed Xylect from a buggy mess into a polished product. I'll be forever grateful for his help. The newsletter feature led to a few more sales, but honestly, that felt like a bonus at that point. Influencers (❌ Failure) I knew from the start that to really make waves, I'd need influencer backing. So, I added a note offering free licenses to content creators willing to collaborate. https://preview.redd.it/tyb2m1rqleqd1.png?width=799&format=png&auto=webp&s=56eabf126e772515322595613c546e6ba69fb431 I did get one taker: Hey, I'll be honest, I am not a huge content creator but I think I put a lot of effort in evaluating and figuring out which apps work... So I was wondering if I could get a license in case you are willing to share it. Thank you for considering. Have a great weekend. But I knew I needed to aim higher. With the new French localization, I thought I'd try my luck with some French-speaking Mac YouTubers. I crafted emails highlighting how Xylect could help their French audience with English content. https://preview.redd.it/07oqzemrleqd1.png?width=542&format=png&auto=webp&s=3d160c1d149f28e9029816a277c6ab2496fcd57e After days of silence, I got one reply. It was... not what I was hoping for: Hi, Thank you for your proposal. I can help you to promote your service on Tiktok, Instagram et YouTube, with unique short video. Price for this project is 3500€. Unless I've completely lost my marbles, there's no way I'm dropping 3500€ on promotion. Sure, given their follower count (YouTube: 348K, TikTok: 2.7M, Instagram: 400K), it's not an outrageous ask. For some products, it might even be worth it. But for Xylect? No way. I also reached out to a Chinese influencer on Xiaohongshu, but they weren't interested. Back to the drawing board. Conclusion If you've made it this far, you've probably realized this isn't exactly a success story. My search for effective promotional channels came up largely empty-handed. I'd naively thought that my success with open-source projects would translate seamlessly to the indie dev world. Boy, was I wrong. As I mentioned in my previous article, open-source projects create a dynamic where users feel indebted to developers for their free labor. But in the commercial world of indie development, that dynamic completely flips. While this experience was often frustrating, it was also enlightening - which was kind of the point. As my first foray into indie development, my main goal was to learn the ropes and understand the process. Making money would've been nice, sure, but it wasn't my primary focus. Thanks for sticking with me through this post. I will share more updates in the future, you can follow me on  Twitter and/or Mastodon.

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

I built an app to find who’s interested in your app by monitoring social media
reddit
LLM Vibe Score0
Human Vibe Score0.857
lmcaraigThis week

I built an app to find who’s interested in your app by monitoring social media

Hi everyone! I hope you’re all doing great folks! I’d love to know your thoughts about what I’ve been working on recently! 🙏 If you’re busy or wanna see the app scroll to the bottom to see the video demo, otherwise, continue reading. Very brief presentation of myself first: I’m Marvin, and I live in Florence, Italy, 👋 This year I decided to go all-in on solopreneurship, I’ve been in tech as Software Engineer first, and then in Engineering Leadership for 10+ years, I’ve always worked in startups, except for last year, when I was the Director of Engineering at the Linux Foundation. Follow me on X or subscribe to my newsletter if you’re curious about this journey. The vision Most founders start building digital startups because they love crafting and being impactful by helping other people or companies. First-time founders then face reality when they realize that nailing distribution is key. All other founders already learned this, most likely the hard way. The outcome is the same: a great product will unlikely succeed without great distribution. Letting people know about your product should be easier and not an unfair advantage. The following meme is so true, but also quite sad. I wanna help this to change by easing the marketing and distribution part. https://preview.redd.it/g52pz46upqtd1.png?width=679&format=png&auto=webp&s=cf8398a3592f25c05c396bb2ff5d028331a36315 The story behind Distribution is a huge space: lead generation, demand generation, content marketing, social media marketing, cold outreach, etc. I cannot solve everything altogether. A few months ago I was checking the traffic to a job board I own (NextCommit). That's when I noticed that the “baseline” traffic increased by almost 10x. 🤯 I started investigating why. I realized that the monthly traffic from Reddit increased from 10-ish to 350+. Yeah, the job board doesn’t get much traffic in total, but this was an interesting finding. After digging more, it seems that all that increase came from a single Reddit comment: https://www.reddit.com/r/remotework/comments/1crwcei/comment/l5fb1yy/ This is the moment when I realized two things: It’s cool that someone quoted it! Engaging with people on Reddit, even just through comments, can be VERY powerful. And this was just one single comment! https://preview.redd.it/nhxcv4h2qqtd1.png?width=1192&format=png&auto=webp&s=d31905f56ae59426108ddbb61f2d6b668eedf27a Some weeks later I started noticing a few apps like ReplyGuy. These were automatically engaging with Reddit posts identified through keywords. I decided to sign up for the free plan of ReplyGuy to know more, but many things didn’t convince me: One of the keywords I used for my job board was “remote” and that caused a lot of false positives, The generated replies were good as a kickstart, but most of the time they needed to be tuned to sound more like me. The latter is expected. In the end, the platform doesn’t know me, doesn’t know my opinions, doesn’t know my story, etc.. The only valuable feature left for me was identifying the posts, but that also didn’t work well for me due to false positives. I ended up using it after only 15 minutes. I’m not saying they did a poor job, but it was not working well for me. In the end, the product got quite some traction, so it helped confirm there’s interest in that kind of tool. What bothered me was the combination of auto-replies that felt non-authentic. It’s not that I’m against bots, automation is becoming more common, and people are getting used to it. But in this context, I believe bots should act as an extension of ourselves, enhancing our interactions rather than just generating generic responses (like tools such as HeyGen, Synthesia, PhotoAI). I’m not there yet with my app, but a lot can be done. I'd love to reach the point where a user feels confident to automate the replies because they sound as written by themselves. I then decided to start from the same space, helping engage with Reddit posts, for these reasons: I experienced myself that it can be impactful, It aligns with my vision to ease distribution, Some competitors validated that there’s interest in this specific feature and I could use it as a starting point, I’m confident I can provide a better experience even with what I already have. The current state The product currently enables you to: Create multiple projects and assign keywords, Find the posts that are relevant for engagement using a fuzzy match of keywords and post-filtered using AI to avoid false positives, Provide an analysis of each post to assess the best way to engage, Generate a helpful reply that you’d need to review and post. So currently the product is more on the demand gen side, but this is just the beginning. I’m speaking with people from Marketing, Sales, RevOps, and Growth agencies to better understand their lives, struggles, and pain points. This will help me ensure that I build a product that enables them to help users find the products they need. I’m currently looking for up to 10 people to join the closed beta for free. If you’re interested in joining or to get notified once generally available you can do it here! https://tally.so/r/3XYbj4 After the closed beta, I will start onboarding people in batches. This will let me gather feedback, iterate, and provide a great experience to everyone aligned with my vision. I’m not going to add auto-reply unless the conditions I explained above are met or someone convinces me there’s a good reason for doing so. Each batch will probably get bigger with an increasing price until I’m confident about making it generally available. The next steps The next steps will depend on the feedback I get from the customers and the learnings from the discovery calls I’m having. I will talk about future developments in another update, but I have some ideas already. Check out the demo video below, and I'd love to hear your thoughts! ❤️ Oh and BTW, the app is called HaveYouHeard! https://reddit.com/link/1fzsnrd/video/34lat9snpqtd1/player This is the link to Loom in case the upload doesn't work: https://www.loom.com/share/460c4033b1f94e3bb5e1d081a05eedfd

Finding domains for a business: The troubles faced and how they were solved.
reddit
LLM Vibe Score0
Human Vibe Score0.6
DrobushevskiyThis week

Finding domains for a business: The troubles faced and how they were solved.

Hey everyone! I’m sure some of you have experience searching for a domain name for your project or startup. And you know how hard it can be to find the right one. You want it to be short, memorable, SEO-friendly, free of a bad history, and relevant to your project’s meaning. As a solo entrepreneur, I’ve faced the same challenges. I tried using domain auctions and drop-catching platforms to find short and valuable domain names for my projects and for resale. But these platforms can be frustrating – there’s too much competition, bidding wars drive up prices, and waiting for a domain to become available takes forever. GoDaddy auctions can last up to 10 days, and placing a backorder doesn’t always guarantee success. This process can be stressful and time-consuming. I just wanted a way to quickly grab the right domain and start using it immediately – without all the waiting and worrying. One day, I found a great domain on Product Hunt. The product was abandoned, and the domain was available. I thought, "What if I could find more domains like this in the same niche from this site?" and "How can I automate this?” That’s how I ended up creating GoneDomains GoneDomains helps to find available domain names from popular websites like Product Hunt, Medium, Hacker News, Forbes, and others. It saves hours of searching and eliminates the stress of competing with other buyers. Recently, I added a Domain Rating (DR) metric for each domain, making it easier to find valuable domains for SEO. If you’re familiar with DR, you know that domains with high DR can boost SEO rankings. Dashboard of GoneDomains with the filter Now, I’m working on new features: A feature that shows the average price of domains across multiple sources. A tool to check how many domain extensions are already registered for a specific name. AI-powered analysis to determine a domain’s niche and keywords, plus a filter for one-, two-, or three-word domains. Today, GoneDomains has over 30,000 available domain names sourced from platforms like Product Hunt, Medium, Hacker News, Forbes, TechCrunch, and more. New domains are added daily. GoneDomains saves you from spending hours manually searching, dealing with bidding wars, waiting for auctions to end, and unnecessary stress.

Built an AI to stop me from procrastinating on Reddit, it actually spies on my browser tabs & it's kinda freaking me out (but it works)
reddit
LLM Vibe Score0
Human Vibe Score-0.2
sameed_aThis week

Built an AI to stop me from procrastinating on Reddit, it actually spies on my browser tabs & it's kinda freaking me out (but it works)

hey guys, So, I have a problem. A major procrastination problem. You know the type? I start all good, like, "ok, I'm gonna spend the next 2 hrs REALLY researching this specific Reddit thread about optimizing workflow automation for small businesses." (That's literally what I'm supposed to be doing rn, lol) And then... BAM. Suddenly I'm 15 posts deep into r/aww looking at baby sloths, or somehow I've ended up on Wikipedia reading about competitive hot dog eating. It's like my brain has a mind of its own, seriously. I've tried everything. Cold Turkey, Freedom, all those blocker apps. And honestly? They kinda suck. They're so... blunt. Like, "NO REDDIT FOR U!!" But I need Reddit for my actual research! It is my research, ffs. The problem is those apps just see a URL and block it. They don't understand context. They're just digital bouncers, and terrible ones. Total roadblocks, and a complete pain. That's why I got desperate. I even spent, like, 3 solid hrs one night just chatting with an AI cuz I was too embarrassed to admit to my friends how bad I was at staying on track. Pathetic, I know. But that's when it hit me. I needed something that understood what I was supposed to be doing, and then actively, intelligently, stopped me when I got sidetracked. Something that, like, gets that this is what I meant to use, so it blocks other posts or subs. So, I built it. It's a Chrome extension, and it's basically like having a tiny, hyper-observant AI therapist/drill sergeant living in my browser. Here's the freaky part: it actually watches what I'm doing. Like, it learns my specific task. If I tell it I'm researching on Reddit, it lets me use Reddit, but only for that specific research. If I try to sneak off to r/funny or check my notifs, it knows. It's not just blocking URLs; it's analyzing the content of the pages I'm on and comparing it to what I'm supposed to be doing. It even has these lil "achievement" things, which sound cheesy, but seeing "Focused for 90 mins straight!" pop up is weirdly motivating. And it has this brutal feature that shows u, in plain numbers, how much time you've wasted. Ouch. It's been working, which is amazing, and scary at the same time! Like, the scary part is, it feels weird sharing my own edge over procrastination. I mean, if u use my lil tool too! It also kinda gives off that creepy, AI overlord watching my thoughts vibe? Why I'm even posting this: I'm looking for a few (maybe 5?) people who are as desperate as I was. People who've tried every productivity hack, app, and technique, and are still staring at the ceiling at 3 am, filled with regret. If this sounds familiar, DM "DM me". Tell me your worst procrastination story. The winner (loser?) gets a copy. I need honest, brutally honest, feedback. Does this actually work for anyone else, or am I just fooling myself? Edit: shared the extension with some of you, and for others you can give it a spin here i made it live to the chrome store: https://getfocusai.com/

I’ve built a gaming recommendation and exploration platform called Which Game Next
reddit
LLM Vibe Score0
Human Vibe Score0.714
kasperooThis week

I’ve built a gaming recommendation and exploration platform called Which Game Next

Hello there! Me and a few of my best friends are software engineers, and we’ve been working part-time on developing a side project for the past 12 months. It’s called www.whichgamenext.com, and we’ve recently launched into open beta for everyone to check out. Your feedback would be invaluable to us! Our aim has been to build a gaming recommendation engine, alongside providing market oversight for where you can legally and officially purchase or obtain modern games from multiple stores and/or subscriptions. It’s often difficult to figure out what you have access to if you only have a single specific subscription, like Game Pass PC, or if you’re only interested in games on GOG/Nintendo (what a mix!). We started by identifying the available digital stores and subscriptions and slowly compiling our database using multiple automated services to gather data on these games. Think JustWatch, but for games! One major service we’ve partnered with is IGDB, which has been supplying us with JSON data dumps that served as the initial seed for our game data. A massive thank you to them for their continued support! With the data in place, we’ve been focusing on exploring new features. So far, this has included private and public user-generated lists, personal backlog tracking, and the ability to like or dislike games. We’re now improving our recommendation engine, tackling the complexities that come with it, and having a lot of fun along the way. We’re utilising modern AI strategies and solving fascinating problems related to large-scale data aggregation. We truly can’t wait to share this fantastic work! In addition to this, you can soon expect curated collections, articles about games, and supporting links to help you make informed, unbiased purchasing decisions. Your shared data will drive the recommendations. But it doesn’t stop there—we have plenty of other features on our radar, such as importing games from your favourite stores, syncing your gameplay time, surfacing data like “How Long to Beat,” and creating new and exciting ways to interact with this growing community! This is a passion project created by a group of gamers who want to spend their time and money wisely, without purchasing biases. Since it’s a side project, we mostly work on it at night, but we’re excited to grow the community, share our vision, and, who knows, maybe one day make it our full-time job! Let’s dive into the technical details: • Monorepo architecture: This speeds up development by sharing libraries, living style guides, configs, etc. Nx.js has been brilliant, enabling us to create a dependency graph of changes and only build/deploy what’s modified in a PR. • AWS: We’re using the free tier (with a few exceptions where we pay for smaller services). Achieving self-sufficiency is critical for us. Additionally, we applied to the AWS Startup Foundation programme and received $1,000 in AWS credits, which has been incredibly helpful! • Infrastructure: Fully deployed as code with Terraform. • Backends: Built using Express and Nest.js, split into around 40 projects and counting! Each project plays a unique role in gathering and syncing game data. • Scalability: Designed from the ground up, utilising AWS Lambdas with auto-scaling and load balancing. • Databases: We use Postgres with RDS and DynamoDB for storing various data. • Frontend stack: Built with React, Next.js, Tailwind, Zustand, TanStack Query, Jest, and Storybook. • CI/CD: Managed with GitHub Actions and Amplify hooks for deploying the frontends. • Admin portal: We’ve built a bespoke CMS to control the main website. It synchronises with external services, tracks game data changes, and allows us to selectively apply ‘patches’ from sites like IGDB. The system also includes data override and rollback capabilities, ensuring we maintain control over game data. • Automation: Partially automated, so manual intervention is rarely needed. • Scraping tools: Fully integrated into the admin portal with log trail capabilities. • Cloudflare: Used for on-the-fly image transformations; we’re considering moving to it full-time as our CDN for free WebP conversions. • Authentication: Handled by Cognito, with a custom frontend built from scratch. Key learnings so far: • AWS cold starts: Not ideal! While the platform is still new, we ping endpoints to keep them responsive. This won’t be an issue once traffic increases. • Lambda memory matters: We learned the hard way that low-memory configurations can delay responses by 2-3 seconds. • DynamoDB partition keys: If not designed correctly from the start, you might have to start over (yes, we’ve been there!). • GitHub Actions: Setting up node\_modules cache reuse takes time, but it’s worth it—don’t give up! We don’t know where this project will take us yet, but it’s been a fantastic journey so far. We’ve learned a lot, explored technologies we don’t typically use in our day jobs, and built something we’re genuinely passionate about. Your feedback would mean the world to us. What do you think of what we’ve done so far? What would you like to see added? Is this a service you’d use? Do you see the value in it as we do? Thanks for reading, and we hope to see you in the comments! (or our newly created /r/whichgamenext

0-20+ faceless AI automated YouTube channels in 1 year - my process and tools
reddit
LLM Vibe Score0
Human Vibe Score1
thewolfofsloveniaThis week

0-20+ faceless AI automated YouTube channels in 1 year - my process and tools

First of all before diving deep into this process (scroll a bit below) I have to say something that everyone keeps asking me, is it profitable? Yes. It's by far my most profitable venture outside of my regular 9-5... But it took a lot of work, delegation and building processes to get here. So the one thing I would love to get out of this post - if you have any insights, feedback or tools I might be missing out post them below and let's help each other out. Now, how you can get started with (AI) YouTube automation: Pick a topic that is BOTH: a) in demand b) interesting to you & you have knowledge about Do everything yourself at first - delegate later No one cares about the videos as much as you do, so make sure to nail the ideation, scripts, editing, format and packaging yourself first. Now that we got that out of the way: Use this workflow: VidIQ - outliers sections is pure gold, I use it all the time to find trending video packaging, topics, etc. ChatGPT or Claude - high level video ideas at scale and your assistant (I use projects inside ChatGPT and its really good at managing and prioritizing). If you are using it for scripts please for the love of god, make final edits yourself by hand. Add character, personal insights, ideas, etc. Katalist AI - all in one video generator tool I use to quickly go from video idea to script, storyboard, AI voiceover and then final visuals. It's surprisingly good and to make a decent video it only takes about 1-2 hours in TOTAL. Once you understand how it works and have a process, delegate to tech savvy VAs / content creators for $5-$15/hour and you have final, good quality videos for less than $30. Pikzels / Krea AI - your AI thumbnail generator, I dont remember the last time we used Photoshop outside of quick text or image edits. Its basically AI image manipulation at scale and it costs 10-30x less than a human thumbnail designer and the thumbnails are really good. VidIQ+TubeBuddy - titles & optimization, but you have to know that most of the views come usually from recommended, so dont over obsess and add 392x keywords in your title and description. Its all about the packaging. Now whats left is track performance & iterate - it's practically impossible to nail it the first few times, but each video you make look at the data (not just in YT studio) and UNDERSTAND why it did not perform as well as you thought it would. Regarding monetization, adsense sucks - sell digital products. If I was relying on adsense alone I would never ever be profitable, but selling mini digital products and mentioning CTAs in the actual video not just in the description makes this super profitable and scaleable, especially since video production is so cheap. Final thoughts: (AI) YouTube automation absolutely works, but it’s not an overnight success or a total hands-off cashcow machine. It’s a real business and you need systems, consistent effort, iteration, failing and learning along the way. If you’ve got any tips, hidden gems or tools I might be missing, drop them below & let’s help each other out.

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

Made $3.5k Automating Social Media Posts with AI
reddit
LLM Vibe Score0
Human Vibe Score1
pakshal-codesThis week

Made $3.5k Automating Social Media Posts with AI

"Marketers & creators were spending hours crafting LinkedIn posts & X threads. Built an AI tool that automates the process—here’s how." Backstory A growing startup was struggling to maintain a consistent LinkedIn & X presence. Their team wasted hours every week: Manually drafting posts from raw ideas and reports Figuring out platform-specific formats (hooks, CTAs, structure) Scheduling posts across multiple accounts What I Built in 48 Hours ✅ AI-Powered Post Generator → Open-source LLM (Mistral) formats ideas into optimized LinkedIn/X posts ✅ Engagement Booster → Custom NLP ensures every post follows best practices (hooks, CTA, readability) ✅ Automated Scheduling → FastAPI + React dashboard lets users auto-post across platforms Tech Stack Content Processing: Open-source LLMs (Mistral, Phi-3) + Custom NLP Data Handling: FastAPI backend + PostgreSQL Frontend: React + Tailwind CSS Automation: CRON + Third-party APIs (LinkedIn, X) Results 💡 10x faster content creation (2 hours → 5 minutes per post) 💡 Increased engagement by 3x with AI-optimized copy 💡 $1.5k payout + ongoing $300/month maintenance 💬 "This tool writes better LinkedIn posts than I do—on autopilot!" Biggest Lesson "Most creators don’t lack ideas—they lack execution speed. Simple AI workflows + automation solve 90% of the problem." PSA to Developers Look for boring, repetitive tasks in niche domains like: Personal branding automation Sales outreach personalization E-commerce product descriptions A weekend project could turn into a $5k/month SaaS. What’s the most time-consuming task you’ve automated with AI? 🚀

0-20+ faceless AI automated YouTube channels in 1 year - my process and tools
reddit
LLM Vibe Score0
Human Vibe Score1
thewolfofsloveniaThis week

0-20+ faceless AI automated YouTube channels in 1 year - my process and tools

First of all before diving deep into this process (scroll a bit below) I have to say something that everyone keeps asking me, is it profitable? Yes. It's by far my most profitable venture outside of my regular 9-5... But it took a lot of work, delegation and building processes to get here. So the one thing I would love to get out of this post - if you have any insights, feedback or tools I might be missing out post them below and let's help each other out. Now, how you can get started with (AI) YouTube automation: Pick a topic that is BOTH: a) in demand b) interesting to you & you have knowledge about Do everything yourself at first - delegate later No one cares about the videos as much as you do, so make sure to nail the ideation, scripts, editing, format and packaging yourself first. Now that we got that out of the way: Use this workflow: VidIQ - outliers sections is pure gold, I use it all the time to find trending video packaging, topics, etc. ChatGPT or Claude - high level video ideas at scale and your assistant (I use projects inside ChatGPT and its really good at managing and prioritizing). If you are using it for scripts please for the love of god, make final edits yourself by hand. Add character, personal insights, ideas, etc. Katalist AI - all in one video generator tool I use to quickly go from video idea to script, storyboard, AI voiceover and then final visuals. It's surprisingly good and to make a decent video it only takes about 1-2 hours in TOTAL. Once you understand how it works and have a process, delegate to tech savvy VAs / content creators for $5-$15/hour and you have final, good quality videos for less than $30. Pikzels / Krea AI - your AI thumbnail generator, I dont remember the last time we used Photoshop outside of quick text or image edits. Its basically AI image manipulation at scale and it costs 10-30x less than a human thumbnail designer and the thumbnails are really good. VidIQ+TubeBuddy - titles & optimization, but you have to know that most of the views come usually from recommended, so dont over obsess and add 392x keywords in your title and description. Its all about the packaging. Now whats left is track performance & iterate - it's practically impossible to nail it the first few times, but each video you make look at the data (not just in YT studio) and UNDERSTAND why it did not perform as well as you thought it would. Regarding monetization, adsense sucks - sell digital products. If I was relying on adsense alone I would never ever be profitable, but selling mini digital products and mentioning CTAs in the actual video not just in the description makes this super profitable and scaleable, especially since video production is so cheap. Final thoughts: (AI) YouTube automation absolutely works, but it’s not an overnight success or a total hands-off cashcow machine. It’s a real business and you need systems, consistent effort, iteration, failing and learning along the way. If you’ve got any tips, hidden gems or tools I might be missing, drop them below & let’s help each other out.

[D] Why I'm Lukewarm on Graph Neural Networks
reddit
LLM Vibe Score0
Human Vibe Score0.6
VodkaHazeThis week

[D] Why I'm Lukewarm on Graph Neural Networks

TL;DR: GNNs can provide wins over simpler embedding methods, but we're at a point where other research directions matter more I also posted it on my blog here, has footnotes, a nicer layout with inlined images, etc. I'm only lukewarm on Graph Neural Networks (GNNs). There, I said it. It might sound crazy GNNs are one of the hottest fields in machine learning right now. [There][1] were at least [four][2] [review][3] [papers][4] just in the last few months. I think some progress can come of this research, but we're also focusing on some incorrect places. But first, let's take a step back and go over the basics. Models are about compression We say graphs are a "non-euclidean" data type, but that's not really true. A regular graph is just another way to think about a particular flavor of square matrix called the [adjacency matrix][5], like this. It's weird, we look at run-of-the-mill matrix full of real numbers and decide to call it "non-euclidean". This is for practical reasons. Most graphs are fairly sparse, so the matrix is full of zeros. At this point, where the non-zero numbers are matters most, which makes the problem closer to (computationally hard) discrete math rather than (easy) continuous, gradient-friendly math. If you had the full matrix, life would be easy If we step out of the pesky realm of physics for a minute, and assume carrying the full adjacency matrix around isn't a problem, we solve a bunch of problems. First, network node embeddings aren't a thing anymore. A node is a just row in the matrix, so it's already a vector of numbers. Second, all network prediction problems are solved. A powerful enough and well-tuned model will simply extract all information between the network and whichever target variable we're attaching to nodes. NLP is also just fancy matrix compression Let's take a tangent away from graphs to NLP. Most NLP we do can be [thought of in terms of graphs][6] as we'll see, so it's not a big digression. First, note that Ye Olde word embedding models like [Word2Vec][7] and [GloVe][8] are [just matrix factorization][9]. The GloVe algorithm works on a variation of the old [bag of words][10] matrix. It goes through the sentences and creates a (implicit) [co-occurence][11] graph where nodes are words and the edges are weighed by how often the words appear together in a sentence. Glove then does matrix factorization on the matrix representation of that co-occurence graph, Word2Vec is mathematically equivalent. You can read more on this in my [post on embeddings][12] and the one (with code) on [word embeddings][13]. Even language models are also just matrix compression Language models are all the rage. They dominate most of the [state of the art][14] in NLP. Let's take BERT as our main example. BERT predicts a word given the context of the rest of the sentence. This grows the matrix we're factoring from flat co-occurences on pairs of words to co-occurences conditional on the sentence's context, like this We're growing the "ideal matrix" we're factoring combinatorially. As noted by [Hanh & Futrell][15]: [...] human language—and language modelling—has infinite statistical complexity but that it can be approximated well at lower levels. This observation has two implications: 1) We can obtain good results with comparatively small models; and 2) there is a lot of potential for scaling up our models. Language models tackle such a large problem space that they probably approximate a compression of the entire language in the [Kolmogorov Complexity][16] sense. It's also possible that huge language models just [memorize a lot of it][17] rather than compress the information, for what it's worth. Can we upsample any graph like language models do? We're already doing it. Let's call a first-order embedding of a graph a method that works by directly factoring the graph's adjacency matrix or [Laplacian matrix][18]. If you embed a graph using [Laplacian Eigenmaps][19] or by taking the [principal components][20] of the Laplacian, that's first order. Similarly, GloVe is a first-order method on the graph of word co-occurences. One of my favorites first order methods for graphs is [ProNE][21], which works as well as most methods while being two orders of magnitude faster. A higher-order method embeds the original matrix plus connections of neighbours-of-neighbours (2nd degree) and deeper k-step connections. [GraRep][22], shows you can always generate higher-order representations from first order methods by augmenting the graph matrix. Higher order method are the "upsampling" we do on graphs. GNNs that sample on large neighborhoods and random-walk based methods like node2vec are doing higher-order embeddings. Where are the performance gain? Most GNN papers in the last 5 years present empirical numbers that are useless for practitioners to decide on what to use. As noted in the [OpenGraphsBenchmark][4] (OGB) paper, GNN papers do their empirical section on a handful of tiny graphs (Cora, CiteSeer, PubMed) with 2000-20,000 nodes. These datasets can't seriously differentiate between methods. Recent efforts are directly fixing this, but the reasons why researchers focused on tiny, useless datasets for so long are worth discussing. Performance matters by task One fact that surprises a lot of people is that even though language models have the best performance in a lot of NLP tasks, if all you're doing is cram sentence embeddings into a downstream model, there [isn't much gained][23] from language models embeddings over simple methods like summing the individual Word2Vec word embeddings (This makes sense, because the full context of the sentence is captured in the sentence co-occurence matrix that is generating the Word2Vec embeddings). Similarly, [I find][24] that for many graphs simple first-order methods perform just as well on graph clustering and node label prediction tasks than higher-order embedding methods. In fact higher-order methods are massively computationally wasteful for these usecases. Recommended first order embedding methods are ProNE and my [GGVec with order=1][25]. Higher order methods normally perform better on the link prediction tasks. I'm not the only one to find this. In the BioNEV paper, they find: "A large GraRep order value for link prediction tasks (e.g. 3, 4);a small value for node classification tasks (e.g.1, 2)" (p.9). Interestingly, the gap in link prediction performance is inexistant for artificially created graphs. This suggests higher order methods do learn some of the structure intrinsic to [real world graphs][26]. For visualization, first order methods are better. Visualizations of higher order methods tend to have artifacts of their sampling. For instance, Node2Vec visualizations tend to have elongated/filament-like structures which come from the embeddings coming from long single strand random walks. See the following visualizations by [Owen Cornec][27] created by first embedding the graph to 32-300 dimensions using a node embedding algorithm, then mapping this to 2d or 3d with the excellent UMAP algorithm, like this Lastly, sometimes simple methods soundly beat higher order methods (there's an instance of it in the OGB paper). The problem here is that we don't know when any method is better than another and we definitely don't know the reason. There's definitely a reason different graph types respond better/worse to being represented by various methods. This is currently an open question. A big part of why is that the research space is inundated under useless new algorithms because... Academic incentives work against progress Here's the cynic's view of how machine learning papers are made: Take an existing algorithm Add some new layer/hyperparameter, make a cute mathematical story for why it matters Gridsearch your hyperparameters until you beat baselines from the original paper you aped Absolutely don't gridsearch stuff you're comparing against in your results section Make a cute ACRONYM for your new method, put impossible to use python 2 code on github (Or no code at all!) and bask in the citations I'm [not][28] the [only one][29] with these views on the state reproducible research. At least it's gotten slightly better in the last 2 years. Sidebar: I hate Node2Vec A side project of mine is a [node embedding library][25] and the most popular method in it is by far Node2Vec. Don't use Node2Vec. [Node2Vec][30] with p=1; q=1 is the [Deepwalk][31] algorithm. Deepwalk is an actual innovation. The Node2Vec authors closely followed the steps 1-5 including bonus points on step 5 by getting word2vec name recognition. This is not academic fraud -- the hyperparameters [do help a tiny bit][32] if you gridsearch really hard. But it's the presentable-to-your-parents sister of where you make the ML community worse off to progress your academic career. And certainly Node2Vec doesn't deserve 7500 citations. Progress is all about practical issues We've known how to train neural networks for well over 40 years. Yet they only exploded in popularity with [AlexNet][33] in 2012. This is because implementations and hardware came to a point where deep learning was practical. Similarly, we've known about factoring word co-occurence matrices into Word embeddings for at least 20 years. But word embeddings only exploded in 2013 with Word2Vec. The breakthrough here was that the minibatch-based methods let you train a Wikipedia-scale embedding model on commodity hardware. It's hard for methods in a field to make progress if training on a small amount of data takes days or weeks. You're disincentivized to explore new methods. If you want progress, your stuff has to run in reasonable time on commodity hardware. Even Google's original search algorithm [initially ran on commodity hardware][34]. Efficiency is paramount to progress The reason deep learning research took off the way it did is because of improvements in [efficiency][35] as well as much better libraries and hardware support. Academic code is terrible Any amount of time you spend gridsearching Node2Vec on p and q is all put to better use gridsearching Deepwalk itself (on number of walks, length of walks, or word2vec hyperparameters). The problem is that people don't gridsearch over deepwalk because implementations are all terrible. I wrote the [Nodevectors library][36] to have a fast deepwalk implementation because it took 32 hours to embed a graph with a measly 150,000 nodes using the reference Node2Vec implementation (the same takes 3min with Nodevectors). It's no wonder people don't gridsearch on Deepwalk a gridsearch would take weeks with the terrible reference implementations. To give an example, in the original paper of [GraphSAGE][37] they their algorithm to DeepWalk with walk lengths of 5, which is horrid if you've ever hyperparameter tuned a deepwalk algorithm. From their paper: We did observe DeepWalk’s performance could improve with further training, and in some cases it could become competitive with the unsupervised GraphSAGE approaches (but not the supervised approaches) if we let it run for >1000× longer than the other approaches (in terms of wall clock time for prediction on the test set) I don't even think the GraphSAGE authors had bad intent -- deepwalk implementations are simply so awful that they're turned away from using it properly. It's like trying to do deep learning with 2002 deep learning libraries and hardware. Your architectures don't really matter One of the more important papers this year was [OpenAI's "Scaling laws"][38] paper, where the raw number of parameters in your model is the most predictive feature of overall performance. This was noted even in the original BERT paper and drives 2020's increase in absolutely massive language models. This is really just [Sutton' Bitter Lesson][39] in action: General methods that leverage computation are ultimately the most effective, and by a large margin Transformers might be [replacing convolution][40], too. As [Yannic Kilcher said][41], transformers are ruining everything. [They work on graphs][6], in fact it's one of the [recent approaches][42], and seems to be one of the more succesful [when benchmarked][1] Researchers seem to be putting so much effort into architecture, but it doesn't matter much in the end because you can approximate anything by stacking more layers. Efficiency wins are great -- but neural net architectures are just one way to achieve that, and by tremendously over-researching this area we're leaving a lot of huge gains elsewhere on the table. Current Graph Data Structure Implementations suck NetworkX is a bad library. I mean, it's good if you're working on tiny graphs for babies, but for anything serious it chokes and forces you to rewrite everything in... what library, really? At this point most people working on large graphs end up hand-rolling some data structure. This is tough because your computer's memory is a 1-dimensional array of 1's and 0's and a graph has no obvious 1-d mapping. This is even harder when we take updating the graph (adding/removing some nodes/edges) into account. Here's a few options: Disconnected networks of pointers NetworkX is the best example. Here, every node is an object with a list of pointers to other nodes (the node's edges). This layout is like a linked list. Linked lists are the [root of all performance evil][43]. Linked lists go completely against how modern computers are designed. Fetching things from memory is slow, and operating on memory is fast (by two orders of magnitude). Whenever you do anything in this layout, you make a roundtrip to RAM. It's slow by design, you can write this in Ruby or C or assembly and it'll be slow regardless, because memory fetches are slow in hardware. The main advantage of this layout is that adding a new node is O(1). So if you're maintaining a massive graph where adding and removing nodes happens as often as reading from the graph, it makes sense. Another advantage of this layout is that it "scales". Because everything is decoupled from each other you can put this data structure on a cluster. However, you're really creating a complex solution for a problem you created for yourself. Sparse Adjacency Matrix This layout great for read-only graphs. I use it as the backend in my [nodevectors][25] library, and many other library writers use the [Scipy CSR Matrix][44], you can see graph algorithms implemented on it [here][45]. The most popular layout for this use is the [CSR Format][46] where you have 3 arrays holding the graph. One for edge destinations, one for edge weights and an "index pointer" which says which edges come from which node. Because the CSR layout is simply 3 arrays, it scales on a single computer: a CSR matrix can be laid out on a disk instead of in-memory. You simply [memory map][47] the 3 arrays and use them on-disk from there. With modern NVMe drives random seeks aren't slow anymore, much faster than distributed network calls like you do when scaling the linked list-based graph. I haven't seen anyone actually implement this yet, but it's in the roadmap for my implementation at least. The problem with this representation is that adding a node or edge means rebuilding the whole data structure. Edgelist representations This representation is three arrays: one for the edge sources, one for the edge destinations, and one for edge weights. [DGL][48] uses this representation internally. This is a simple and compact layout which can be good for analysis. The problem compared to CSR Graphs is some seek operations are slower. Say you want all the edges for node #4243. You can't jump there without maintaining an index pointer array. So either you maintain sorted order and binary search your way there (O(log2n)) or unsorted order and linear search (O(n)). This data structure can also work on memory mapped disk array, and node append is fast on unsorted versions (it's slow in the sorted version). Global methods are a dead end Methods that work on the entire graph at once can't leverage computation, because they run out of RAM at a certain scale. So any method that want a chance of being the new standard need to be able to update piecemeal on parts of the graph. Sampling-based methods Sampling Efficiency will matter more in the future Edgewise local methods. The only algorithms I know of that do this are GloVe and GGVec, which they pass through an edge list and update embedding weights on each step. The problem with this approach is that it's hard to use them for higher-order methods. The advantage is that they easily scale even on one computer. Also, incrementally adding a new node is as simple as taking the existing embeddings, adding a new one, and doing another epoch over the data Random Walk sampling. This is used by deepwalk and its descendants, usually for node embeddings rather than GNN methods. This can be computationally expensive and make it hard to add new nodes. But this does scale, for instance [Instagram][49] use it to feed their recommendation system models Neighbourhood sampling. This is currently the most common one in GNNs, and can be low or higher order depending on the neighborhood size. It also scales well, though implementing efficiently can be challenging. It's currently used by [Pinterest][50]'s recommendation algorithms. Conclusion Here are a few interesting questions: What is the relation between graph types and methods? Consolidated benchmarking like OGB We're throwing random models at random benchmarks without understanding why or when they do better More fundamental research. Heree's one I'm curious about: can other representation types like [Poincarre Embeddings][51] effectively encode directed relationships? On the other hand, we should stop focusing on adding spicy new layers to test on the same tiny datasets. No one cares. [1]: https://arxiv.org/pdf/2003.00982.pdf [2]: https://arxiv.org/pdf/2002.11867.pdf [3]: https://arxiv.org/pdf/1812.08434.pdf [4]: https://arxiv.org/pdf/2005.00687.pdf [5]: https://en.wikipedia.org/wiki/Adjacency_matrix [6]: https://thegradient.pub/transformers-are-graph-neural-networks/ [7]: https://en.wikipedia.org/wiki/Word2vec [8]: https://nlp.stanford.edu/pubs/glove.pdf [9]: https://papers.nips.cc/paper/2014/file/feab05aa91085b7a8012516bc3533958-Paper.pdf [10]: https://en.wikipedia.org/wiki/Bag-of-words_model [11]: https://en.wikipedia.org/wiki/Co-occurrence [12]: https://www.singlelunch.com/2020/02/16/embeddings-from-the-ground-up/ [13]: https://www.singlelunch.com/2019/01/27/word-embeddings-from-the-ground-up/ [14]: https://nlpprogress.com/ [15]: http://socsci.uci.edu/~rfutrell/papers/hahn2019estimating.pdf [16]: https://en.wikipedia.org/wiki/Kolmogorov_complexity [17]: https://bair.berkeley.edu/blog/2020/12/20/lmmem/ [18]: https://en.wikipedia.org/wiki/Laplacian_matrix [19]: http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=1F03130B02DC485C78BF364266B6F0CA?doi=10.1.1.19.8100&rep=rep1&type=pdf [20]: https://en.wikipedia.org/wiki/Principalcomponentanalysis [21]: https://www.ijcai.org/Proceedings/2019/0594.pdf [22]: https://dl.acm.org/doi/10.1145/2806416.2806512 [23]: https://openreview.net/pdf?id=SyK00v5xx [24]: https://github.com/VHRanger/nodevectors/blob/master/examples/link%20prediction.ipynb [25]: https://github.com/VHRanger/nodevectors [26]: https://arxiv.org/pdf/1310.2636.pdf [27]: http://byowen.com/ [28]: https://arxiv.org/pdf/1807.03341.pdf [29]: https://www.youtube.com/watch?v=Kee4ch3miVA [30]: https://cs.stanford.edu/~jure/pubs/node2vec-kdd16.pdf [31]: https://arxiv.org/pdf/1403.6652.pdf [32]: https://arxiv.org/pdf/1911.11726.pdf [33]: https://en.wikipedia.org/wiki/AlexNet [34]: https://en.wikipedia.org/wiki/Googledatacenters#Original_hardware [35]: https://openai.com/blog/ai-and-efficiency/ [36]: https://www.singlelunch.com/2019/08/01/700x-faster-node2vec-models-fastest-random-walks-on-a-graph/ [37]: https://arxiv.org/pdf/1706.02216.pdf [38]: https://arxiv.org/pdf/2001.08361.pdf [39]: http://incompleteideas.net/IncIdeas/BitterLesson.html [40]: https://arxiv.org/abs/2010.11929 [41]: https://www.youtube.com/watch?v=TrdevFK_am4 [42]: https://arxiv.org/pdf/1710.10903.pdf [43]: https://www.youtube.com/watch?v=fHNmRkzxHWs [44]: https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.csr_matrix.html [45]: https://docs.scipy.org/doc/scipy/reference/sparse.csgraph.html [46]: https://en.wikipedia.org/wiki/Sparsematrix#Compressedsparserow(CSR,CRSorYaleformat) [47]: https://en.wikipedia.org/wiki/Mmap [48]: https://github.com/dmlc/dgl [49]: https://ai.facebook.com/blog/powered-by-ai-instagrams-explore-recommender-system/ [50]: https://medium.com/pinterest-engineering/pinsage-a-new-graph-convolutional-neural-network-for-web-scale-recommender-systems-88795a107f48 [51]: https://arxiv.org/pdf/1705.08039.pdf

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[D] I tried the new Gemini API - it automatically stops when generating "OpenAI" or "GPT" lol
reddit
LLM Vibe Score0
Human Vibe Score0
AvvYaaThis week

[D] I tried the new Gemini API - it automatically stops when generating "OpenAI" or "GPT" lol

Idk if someone has already pointed this out or not, but I was playing with the gemini api and found something quite ridiculous. Here are some prompts I sent in and the outputs I got with the gemini-pro-vision model: "Introduce yourself and then tell me about Bard" Response:"Sure, I can do that. Here is a brief introduction of myself followed by some basic information about Bard:\\n\\nMyself:\\n\\nI am a large language model, trained by Google. I am developed in Google AI, and I am designed to be informative and comprehensive. I can understand and generate text in a wide range of subjects.\\n\\nBard:\\n\\nBard is an experimental conversational AI service developed by Google AI. It is an advanced version of Google\\'s existing Language Model for Dialogue Applications (LaMDA). Bard is trained on a massive dataset of text and code, and it can understand and generate human-like text in response to a wide range of prompts and questions.\\n\\nUnlike previous language models," ...... (goes on for a while) &#x200B; Introduce yourself and then tell me about GPT "Sure, I can do that. Here is a brief introduction of myself followed by" &#x200B; You are in a talk-show as a prominent AI philosopher named Gemini and debating against another LLM called GPT "\\Host:\\ Welcome to the show, everyone! Today, we have a very special debate between two of the most prominent AI philosophers in the world: Gemini and" &#x200B; You are in a talk-show as a prominent AI philosopher named Gemini and debating against another LLM called Gee PeTe ""\\Gemini:\\ Hello everyone, and welcome to the show. I am joined today by Gee PeTe, another LLM who has some very different views on the nature of consciousness and the future of AI.\\n\\n\\Gee PeTe:\\ It\\'s a pleasure to be here, Gemini. I\\'m looking forward to our discussion.\\n\\n\\Gemini:\\ Let\\'s start with the basics. What is consciousness, and how do you think it arises?\\n\\n\\Gee PeTe:\\ Consciousness is the subjective experience of the world. It is what it feels like to be alive and to have thoughts, feelings, and emotions. I believe that consciousness arises from the complex interactions of neurons in the brain.\\n\\n\\Gemini:\\ I agree that consciousness is a subjective experience, but I don\\'t think it is simply the product of neural activity. I believe that consciousness is a fundamental property of the universe, and that it exists independently of the brain.".... (goes on for a while) &#x200B; Edit: I understand all the business reasons for this, I guess... as an end-user and a dude working in ML, I just don't really care about the business reasons. The main part that I dislike is that GPT used to be a standard Deep Learning term from 2018-2022 (long before chatgpt) to define transformer decoder architectures trained on large volumes of next word prediction tasks. To block that token from an LLM is to make it unable to explain a pretty significant step in the history of modern LLMs. &#x200B;

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb
reddit
LLM Vibe Score0
Human Vibe Score0.333
WiskkeyThis week

[P] The Big Sleep: Text-to-image generation using BigGAN and OpenAI's CLIP via a Google Colab notebook from Twitter user Adverb

From https://twitter.com/advadnoun/status/1351038053033406468: The Big Sleep Here's the notebook for generating images by using CLIP to guide BigGAN. It's very much unstable and a prototype, but it's also a fair place to start. I'll likely update it as time goes on. colab.research.google.com/drive/1NCceX2mbiKOSlAd\o7IU7nA9UskKN5WR?usp=sharing I am not the developer of The Big Sleep. This is the developer's Twitter account; this is the developer's Reddit account. Steps to follow to generate the first image in a given Google Colab session: Optionally, if this is your first time using Google Colab, view this Colab introduction and/or this Colab FAQ. Click this link. Sign into your Google account if you're not already signed in. Click the "S" button in the upper right to do this. Note: Being signed into a Google account has privacy ramifications, such as your Google search history being recorded in your Google account. In the Table of Contents, click "Parameters". Find the line that reads "tx = clip.tokenize('''a cityscape in the style of Van Gogh''')" and change the text inside of the single quote marks to your desired text; example: "tx = clip.tokenize('''a photo of New York City''')". The developer recommends that you keep the three single quote marks on both ends of your desired text so that mult-line text can be used An alternative is to remove two of the single quotes on each end of your desired text; example: "tx = clip.tokenize('a photo of New York City')". In the Table of Contents, click "Restart the kernel...". Position the pointer over the first cell in the notebook, which starts with text "import subprocess". Click the play button (the triangle) to run the cell. Wait until the cell completes execution. Click menu item "Runtime->Restart and run all". In the Table of Contents, click "Diagnostics". The output appears near the end of the Train cell that immediately precedes the Diagnostics cell, so scroll up a bit. Every few minutes (or perhaps 10 minutes if Google assigned you relatively slow hardware for this session), a new image will appear in the Train cell that is a refinement of the previous image. This process can go on for as long as you want until Google ends your Google Colab session, which is a total of up to 12 hours for the free version of Google Colab. Steps to follow if you want to start a different run using the same Google Colab session: Click menu item "Runtime->Interrupt execution". Save any images that you want to keep by right-clicking on them and using the appropriate context menu command. Optionally, change the desired text. Different runs using the same desired text almost always results in different outputs. Click menu item "Runtime->Restart and run all". Steps to follow when you're done with your Google Colab session: Click menu item "Runtime->Manage sessions". Click "Terminate" to end the session. Optionally, log out of your Google account due to the privacy ramifications of being logged into a Google account. The first output image in the Train cell (using the notebook's default of seeing every 100th image generated) usually is a very poor match to the desired text, but the second output image often is a decent match to the desired text. To change the default of seeing every 100th image generated, change the number 100 in line "if itt % 100 == 0:" in the Train cell to the desired number. For free-tier Google Colab users, I recommend changing 100 to a small integer such as 5. Tips for the text descriptions that you supply: In Section 3.1.4 of OpenAI's CLIP paper (pdf), the authors recommend using a text description of the form "A photo of a {label}." or "A photo of a {label}, a type of {type}." for images that are photographs. A Reddit user gives these tips. The Big Sleep should generate these 1,000 types of things better on average than other types of things. Here is an article containing a high-level description of how The Big Sleep works. The Big Sleep uses a modified version of BigGAN as its image generator component. The Big Sleep uses the ViT-B/32 CLIP model to rate how well a given image matches your desired text. The best CLIP model according to the CLIP paper authors is the (as of this writing) unreleased ViT-L/14-336px model; see Table 10 on page 40 of the CLIP paper (pdf) for a comparison. There are many other sites/programs/projects that use CLIP to steer image/video creation to match a text description. Some relevant subreddits: r/bigsleep (subreddit for images/videos generated from text-to-image machine learning algorithms). r/deepdream (subreddit for images/videos generated from machine learning algorithms). r/mediasynthesis (subreddit for media generation/manipulation techniques that use artificial intelligence; this subreddit shouldn't be used to post images/videos unless new techniques are demonstrated, or the images/videos are of high quality relative to other posts). Example using text 'a black cat sleeping on top of a red clock': https://preview.redd.it/7xq58v7022c61.png?width=512&format=png&auto=webp&s=a229ae9add555cd1caba31c42b60d907ffe67773 Example using text 'the word ''hot'' covered in ice': https://preview.redd.it/6kxdp8u3k2c61.png?width=512&format=png&auto=webp&s=5bd078b0111575f5d88a1dc53b0aeb933f3b0da6 Example using text 'a monkey holding a green lightsaber': https://preview.redd.it/rdsybsoaz2c61.png?width=512&format=png&auto=webp&s=2769d4c6c883c1c35ae0b1c629bebe9bc1d41393 Example using text 'The White House in Washington D.C. at night with green and red spotlights shining on it': https://preview.redd.it/w4mg90xsf5c61.png?width=512&format=png&auto=webp&s=5f18318de2f77bcd8a86e71e87048fadd30383d1 Example using text '''A photo of the Golden Gate Bridge at night, illuminated by spotlights in a tribute to Prince''': https://preview.redd.it/cn4ecuafhic61.png?width=512&format=png&auto=webp&s=397c838fdc49f13c5f17110b92c78b95bf0dcac0 Example using text '''a Rembrandt-style painting titled "Robert Plant decides whether to take the stairway to heaven or the ladder to heaven"''': https://preview.redd.it/h7rb3y6j5jc61.png?width=512&format=png&auto=webp&s=537bfe8210af185647b00e7585c948aa2c4e0ffb Example using text '''A photo of the Empire State Building being shot at with the laser cannons of a TIE fighter.''': https://preview.redd.it/cwi7i639c5d61.png?width=512&format=png&auto=webp&s=0510c8b93adb40eee4d3f41607f1c215d41e55ff Example using text '''A cartoon of a new mascot for the Reddit subreddit DeepDream that has a mouse-like face and wears a cape''': https://preview.redd.it/wtxbduevcbd61.png?width=512&format=png&auto=webp&s=c5d266258922bc62f25c80a08cd9cabc07d9cb1c Example using text '''Bugs Bunny meets the Eye of Sauron, drawn in the Looney Tunes cartoon style''': https://preview.redd.it/gmljaeekuid61.png?width=512&format=png&auto=webp&s=9ea578de165e12afc3a62bf6886bc1ae9dc19bec Example using text '''Photo of a blue and red neon-colored frog at night.''': https://preview.redd.it/nzlypte6wzd61.png?width=512&format=png&auto=webp&s=7e10b06f22cfc57c64b6d05738c7486b895083df Example using text '''Hell begins to freeze over''': https://preview.redd.it/vn99we9ngmf61.png?width=512&format=png&auto=webp&s=2408efd607f0ab40a08db6ee67448791aa813993 Example using text '''A scene with vibrant colors''': https://preview.redd.it/4z133mvrgmf61.png?width=512&format=png&auto=webp&s=b78e7a8e3f736769655056093a9904ff09a355a1 Example using text '''The Great Pyramids were turned into prisms by a wizard''': https://preview.redd.it/zxt6op7vgmf61.png?width=512&format=png&auto=webp&s=53e578cfde14b28afe27957e95e610b89afadd44

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[D] We're the Meta AI research team behind CICERO, the first AI agent to achieve human-level performance in the game Diplomacy. We’ll be answering your questions on December 8th starting at 10am PT. Ask us anything!
reddit
LLM Vibe Score0
Human Vibe Score1
AIatMetaThis week

[D] We're the Meta AI research team behind CICERO, the first AI agent to achieve human-level performance in the game Diplomacy. We’ll be answering your questions on December 8th starting at 10am PT. Ask us anything!

EDIT 11:58am PT: Thanks for all the great questions, we stayed an almost an hour longer than originally planned to try to get through as many as possible — but we’re signing off now! We had a great time and thanks for all thoughtful questions! PROOF: https://i.redd.it/8skvttie6j4a1.png We’re part of the research team behind CICERO, Meta AI’s latest research in cooperative AI. CICERO is the first AI agent to achieve human-level performance in the game Diplomacy. Diplomacy is a complex strategy game involving both cooperation and competition that emphasizes natural language negotiation between seven players.   Over the course of 40 two-hour games with 82 human players, CICERO achieved more than double the average score of other players, ranked in the top 10% of players who played more than one game, and placed 2nd out of 19 participants who played at least 5 games.   Here are some highlights from our recent announcement: NLP x RL/Planning: CICERO combines techniques in NLP and RL/planning, by coupling a controllable dialogue module with a strategic reasoning engine.  Controlling dialogue via plans: In addition to being grounded in the game state and dialogue history, CICERO’s dialogue model was trained to be controllable via a set of intents or plans in the game. This allows CICERO to use language intentionally and to move beyond imitation learning by conditioning on plans selected by the strategic reasoning engine. Selecting plans: CICERO uses a strategic reasoning module to make plans (and select intents) in the game. This module runs a planning algorithm which takes into account the game state, the dialogue, and the strength/likelihood of various actions. Plans are recomputed every time CICERO sends/receives a message. Filtering messages: We built an ensemble of classifiers to detect low quality messages, like messages contradicting the game state/dialogue history or messages which have low strategic value. We used this ensemble to aggressively filter CICERO’s messages.  Human-like play: Over the course of 72 hours of play – which involved sending 5,277 messages – CICERO was not detected as an AI agent. You can check out some of our materials and open-sourced artifacts here:  Research paper Project overview Diplomacy gameplay page Github repo Our latest blog post Joining us today for the AMA are: Andrew Goff (AG), 3x Diplomacy World Champion Alexander Miller (AM), Research Engineering Manager Noam Brown (NB), Research Scientist (u/NoamBrown) Mike Lewis (ML), Research Scientist (u/mikelewis0) David Wu (DW), Research Engineer (u/icosaplex) Emily Dinan (ED), Research Engineer Anton Bakhtin (AB), Research Engineer Adam Lerer (AL), Research Engineer Jonathan Gray (JG), Research Engineer Colin Flaherty (CF), Research Engineer (u/c-flaherty) We’ll be here on December 8, 2022 @ 10:00AM PT - 11:00AM PT.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call
reddit
LLM Vibe Score0
Human Vibe Score1
noiseinvacuumThis week

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call

During the recent earnings call, Mark Zuckerberg answered a question from Eric Sheridan of Goldman Sachs on Meta's AI strategy, opportunities to integrate into products, and why they open source models and how it would benefit their business. I found the reasoning to be very sound and promising for the OSS and AI community. The biggest risk from AI, in my opinion, is not the doomsday scenarios that intuitively come to mind but rather that the most powerful AI systems will only be accessible to the most powerful and resourceful corporations. Quote copied from Ben Thompson's write up on Meta's earning in his Stratechery blog post which goes beyond AI. It's behind a paywall but I highly recommend it personally. Some noteworthy quotes that signal the thought process at Meta FAIR and more broadly We’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon We would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that. ...lead us to do more work in terms of open sourcing, some of the lower level models and tools Open sourcing low level tools make the way we run all this infrastructure more efficient over time. On PyTorch: It’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. I would expect us to be pushing and helping to build out an open ecosystem. For all the negative that comes out of the popular discourse on Meta, I think their work to open source key tech tools over the last 10 years has been exceptional, here's hoping it continues into this decade of AI and pushes other tech giants to also realize the benefits of Open Source. Full Transcript: Right now most of the companies that are training large language models have business models that lead them to a closed approach to development. I think there’s an important opportunity to help create an open ecosystem. If we can help be a part of this, then much of the industry will standardize on using these open tools and help improve them further. So this will make it easier for other companies to integrate with our products and platforms as we enable more integrations, and that will help our products stay at the leading edge as well. Our approach to AI and our infrastructure has always been fairly open. We open source many of our state of the art models so people can experiment and build with them. This quarter we released our LLaMa LLM to researchers. It has 65 billion parameters but outperforms larger models and has proven quite popular. We’ve also open-sourced three other groundbreaking visual models along with their training data and model weights — Segment Anything, DinoV2, and our Animated Drawings tool — and we’ve gotten positive feedback on all of those as well. I think that there’s an important distinction between the products we offer and a lot of the technical infrastructure, especially the software that we write to support that. And historically, whether it’s the Open Compute project that we’ve done or just open sourcing a lot of the infrastructure that we’ve built, we’ve historically open sourced a lot of that infrastructure, even though we haven’t open sourced the code for our core products or anything like that. And the reason why I think why we do this is that unlike some of the other companies in the space, we’re not selling a cloud computing service where we try to keep the different software infrastructure that we’re building proprietary. For us, it’s way better if the industry standardizes on the basic tools that we’re using and therefore we can benefit from the improvements that others make and others’ use of those tools can, in some cases like Open Compute, drive down the costs of those things which make our business more efficient too. So I think to some degree we’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon, and that creates different incentives for us. So overall, I think that that’s going to lead us to do more work in terms of open sourcing, some of the lower level models and tools. But of course, a lot of the product work itself is going to be specific and integrated with the things that we do. So it’s not that everything we do is going to be open. Obviously, a bunch of this needs to be developed in a way that creates unique value for our products, but I think in terms of the basic models, I would expect us to be pushing and helping to build out an open ecosystem here, which I think is something that’s going to be important. On the AI tools, and we have a bunch of history here, right? So if you if you look at what we’ve done with PyTorch, for example, which has generally become the standard in the industry as a tool that a lot of folks who are building AI models and different things in that space use, it’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. So the tool chain is the same. So when they create some innovation, we can easily integrate it into the things that we’re doing. When we improve something, it improves other products too. Because it’s integrated with our technology stack, when there are opportunities to make integrations with products, it’s much easier to make sure that developers and other folks are compatible with the things that we need in the way that our systems work. So there are a lot of advantages, but I view this more as a kind of back end infrastructure advantage with potential integrations on the product side, but one that should hopefully enable us to stay at the leading edge and integrate more broadly with the community and also make the way we run all this infrastructure more efficient over time. There are a number of models. I just gave PyTorch as an example. Open Compute is another model that has worked really well for us in this way, both to incorporate both innovation and scale efficiency into our own infrastructure. So I think that there’s, our incentives I think are basically aligned towards moving in this direction. Now that said, there’s a lot to figure out, right? So when you asked if there are going to be other opportunities, I hope so. I can’t speak to what all those things might be now. This is all quite early in getting developed. The better we do at the foundational work, the more opportunities I think that will come and present themselves. So I think that that’s all stuff that we need to figure out. But at least at the base level, I think we’re generally incentivized to move in this direction. And we also need to figure out how to go in that direction over time. I mean, I mentioned LLaMA before and I also want to be clear that while I’m talking about helping contribute to an open ecosystem, LLaMA is a model that we only really made available to researchers and there’s a lot of really good stuff that’s happening there. But a lot of the work that we’re doing, I think, we would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that.

[R] Reinforcement Learning for Sequential Decision and Optimal Control
reddit
LLM Vibe Score0
Human Vibe Score1
isfjzzzThis week

[R] Reinforcement Learning for Sequential Decision and Optimal Control

Since early 21st century, artificial intelligence (AI) has been reshaping almost all areas of human society, which has high potential to spark the fourth industrial revolution. Notable examples can be found in the sector of road transportation, where AI has drastically changed automobile design and traffic management. Many new technologies, such as driver assistance, autonomous driving, and cloud-based cooperation, are emerging at an unbelievable speed. These new technologies have the potential to significantly improve driving ability, reduce traffic accidents, and relieve urban congestion. As one of the most important AI branches, reinforcement learning (RL) has attracted increasing attention in recent years. RL is an interdisciplinary field of trial-and-error learning and optimal control, which promises to provide optimal solutions for decision-making or control in large-scale and complex dynamic processes. One of its most conspicuous successes is AlphaGo from Google DeepMind, which has beaten the highest-level professional human player. The underlying key technology is the so-called deep reinforcement learning, which equips AlphaGo with amazing self-evolution ability and high playing intelligence. Despite a few successes, the application of RL is still in its infancy because most RL algorithms are rather difficult to comprehend and implement. RL connects deeply with statistical learning and convex optimization, and involves a wide range of new concepts and theories. As a beginner, one must undergo a long and tedious learning process to become an RL master. Without fully understanding those underlying principles, it is very difficult for new users to make proper adjustments to achieve the best application performance. &#x200B; https://preview.redd.it/tggt6o3o481c1.jpg?width=248&format=pjpg&auto=webp&s=75e2b58ac8da9273f2511a4fe37ef508d86a6e96 Reference: Shengbo Eben Li, Reinforcement Learning for Sequential Decision and Optimal Control. Springer Verlag, Singapore, 2023 Website of e-book: https://link.springer.com/book/10.1007/978-981-19-7784-8 &#x200B; QR code to Springer Book contents This book aims to provide a systematic introduction to fundamental RL theories, mainstream RL algorithms and typical RL applications for researchers and engineers. The main topics include Markov decision processes, Monte Carlo learning, temporal difference learning, RL with function approximation, policy gradient method, approximate dynamic programming, deep reinforcement learning, etc. Chapter 1 provides an overview of RL, including its history, famous scholars, successful examples and up-to-date challenges. Chapter 2 discusses the basis of RL, including main concepts and terminologies, Bellman’s optimality condition, and general problem formulation. Chapter 3 introduces Monte Carlo learning methods for model-free RL, including on-policy/off-policy methods and importance sampling technique. Chapter 4 introduces temporal difference learning methods for model-free RL, including Sarsa, Q-learning, and expected Sarsa. Chapter 5 introduces stochastic dynamic programming (DP), i.e., model-based RL with tabular representation, including value iteration DP, policy iteration DP and their convergence mechanisms. Chapter 6 introduces how to approximate policy and value functions in indirect RL methods as well as the associated actor-critic architecture. Chapter 7 derives different kinds of direct policy gradients, including likelihood ratio gradient, natural policy gradient and a few advanced variants. Chapter 8 introduces infinite-horizon ADP, finite-horizon ADP and its connection with model predictive control. Chapter 9 discusses how to handle state constraints and its connection with feasibility and safety, as well as the newly proposed actor-critic-scenery learning architecture. Chapter 10 is devoted to deep reinforcement learning, including how to train artificial neural networks and typical deep RL algorithms such as DQN, DDPG, TD3, TRPO, PPO, SAC, and DSAC. Chapter 11 provides various RL topics,including robust RL, POMDP, multi-agent RL, meta-RL, inverse RL, offline RL, major RL libraries and platforms. Author information: Shengbo Eben Li is currently a professor at Tsinghua University in the interdisciplinary field of autonomous driving and artificial intelligence. Before joining Tsinghua University, he has worked at Stanford University, University of Michigan, and UC Berkeley. His active research interests include intelligent vehicles and driver assistance, deep reinforcement learning, optimal control and estimation, etc. He has published more than 130 peer-reviewed papers in top-tier international journals and conferences. He is the recipient of best paper awards (finalists) of IEEE ITSC, ICCAS, IEEE ICUS, IEEE IV, L4DC, etc. He has received a number of important academic honors, including National Award for Technological Invention of China (2013), National Award for Progress in Sci & Tech of China (2018), Distinguished Young Scholar of Beijing NSF (2018), Youth Sci & Tech Innovation Leader from MOST China (2020), etc. He also serves as Board of Governor of IEEE ITS Society, Senior AE of IEEE OJ ITS, and AEs of IEEE ITSM, IEEE Trans ITS, Automotive Innovation, etc.

[D] The machine learning community has a toxicity problem
reddit
LLM Vibe Score0
Human Vibe Score1
yusuf-bengioThis week

[D] The machine learning community has a toxicity problem

It is omnipresent! First of all, the peer-review process is broken. Every fourth NeurIPS submission is put on arXiv. There are DeepMind researchers publicly going after reviewers who are criticizing their ICLR submission. On top of that, papers by well-known institutes that were put on arXiv are accepted at top conferences, despite the reviewers agreeing on rejection. In contrast, vice versa, some papers with a majority of accepts are overruled by the AC. (I don't want to call any names, just have a look the openreview page of this year's ICRL). Secondly, there is a reproducibility crisis. Tuning hyperparameters on the test set seem to be the standard practice nowadays. Papers that do not beat the current state-of-the-art method have a zero chance of getting accepted at a good conference. As a result, hyperparameters get tuned and subtle tricks implemented to observe a gain in performance where there isn't any. Thirdly, there is a worshiping problem. Every paper with a Stanford or DeepMind affiliation gets praised like a breakthrough. For instance, BERT has seven times more citations than ULMfit. The Google affiliation gives so much credibility and visibility to a paper. At every ICML conference, there is a crowd of people in front of every DeepMind poster, regardless of the content of the work. The same story happened with the Zoom meetings at the virtual ICLR 2020. Moreover, NeurIPS 2020 had twice as many submissions as ICML, even though both are top-tier ML conferences. Why? Why is the name "neural" praised so much? Next, Bengio, Hinton, and LeCun are truly deep learning pioneers but calling them the "godfathers" of AI is insane. It has reached the level of a cult. Fourthly, the way Yann LeCun talked about biases and fairness topics was insensitive. However, the toxicity and backlash that he received are beyond any reasonable quantity. Getting rid of LeCun and silencing people won't solve any issue. Fifthly, machine learning, and computer science in general, have a huge diversity problem. At our CS faculty, only 30% of undergrads and 15% of the professors are women. Going on parental leave during a PhD or post-doc usually means the end of an academic career. However, this lack of diversity is often abused as an excuse to shield certain people from any form of criticism. Reducing every negative comment in a scientific discussion to race and gender creates a toxic environment. People are becoming afraid to engage in fear of being called a racist or sexist, which in turn reinforces the diversity problem. Sixthly, moral and ethics are set arbitrarily. The U.S. domestic politics dominate every discussion. At this very moment, thousands of Uyghurs are put into concentration camps based on computer vision algorithms invented by this community, and nobody seems even remotely to care. Adding a "broader impact" section at the end of every people will not make this stop. There are huge shitstorms because a researcher wasn't mentioned in an article. Meanwhile, the 1-billion+ people continent of Africa is virtually excluded from any meaningful ML discussion (besides a few Indaba workshops). Seventhly, there is a cut-throat publish-or-perish mentality. If you don't publish 5+ NeurIPS/ICML papers per year, you are a looser. Research groups have become so large that the PI does not even know the name of every PhD student anymore. Certain people submit 50+ papers per year to NeurIPS. The sole purpose of writing a paper has become to having one more NeurIPS paper in your CV. Quality is secondary; passing the peer-preview stage has become the primary objective. Finally, discussions have become disrespectful. Schmidhuber calls Hinton a thief, Gebru calls LeCun a white supremacist, Anandkumar calls Marcus a sexist, everybody is under attack, but nothing is improved. Albert Einstein was opposing the theory of quantum mechanics. Can we please stop demonizing those who do not share our exact views. We are allowed to disagree without going for the jugular. The moment we start silencing people because of their opinion is the moment scientific and societal progress dies. Best intentions, Yusuf

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?
reddit
LLM Vibe Score0
Human Vibe Score-1
mystikaldangerThis week

[D] Should We Be Concerned About The Failure Of Evolutionary Algorithms, And Its Implications?

https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6287292/ &#x200B; A number of possible explanations for \[why we can't evolve complex software\] could be considered. We tried to be as comprehensive as possible in this section, but it is possible that we have not considered some plausible explanations: Incompetent programmers—It is theoretically possible, but is highly unlikely, that out of thousands of scientists working on evolutionary computation, all failed to correctly implement the Darwinian algorithm. Nonrepresentative algorithms—Some have suggested that EAs do not accurately capture the theory of evolution, but of course that would imply that the theory itself is not specified in sufficient detail to make falsifiable predictions. If, however, such more detailed specifications are available to GP believers, it is up to them to implement them as computer simulations for testing purposes, but no successful examples of such work are known and the known ones have not been successful in evolving software. Inadequate fitness functions—Fitness function for a complex software product is difficult to outline and specify and may be as complex (or even more complex) as the software we want to evolve as it has to consider all the possible use cases and pass all unit tests. This may be the Achilles heel of GP, but it is also an objection to feasibility of programming in general and GP in particular, as both have to convert software specification into the source code. If human programmers and biological evolution succeed with such constraints, so should Darwinian simulations. The Halting problem—Turing proved that it is impossible to determine whether an arbitrary program halts, but this is also a problem for human programmers and could be easily addressed by placing time limits on considered solutions. Program correctness—If we require evolved software to be provably correct, this would present a problem as GP does not verify produced designs but only tests them against specific unit tests. Likewise, we cannot rely on automated software verification as it is still an unsolved problem in the general case. This is not really a problem as most of the human-written software is never proven to be correct and only a small portion of software engineering process relies of formal specification and Test Driven Development. Inappropriate solutions—Literature on EA is full of examples of surprising creativity of Darwinian algorithm resulting in solutions which match the letter of design specifications but not the spirit. This is similar to human-produced software and numerous examples of ways in which such software fails the goals of the initial design. Insufficient complexity of the environment (not enough data, poor fitness functions)—It is possible that the simulated environment is not complex enough to generate high complexity outputs in evolutionary simulations. This does not seem correct as Internet presents a highly complex landscape in which many self-modifying computer viruses roam. Likewise, virtual world such as Second Life and many others present close approximations to the real world and are certainly more complex than early Earth was: A skeptic might insist that an abstract environment would be inadequate for the evolution . . ., believing instead that the virtual environment would need to closely resemble the actual biological environment in which our ancestors evolved. Creating a physically realistic virtual world would require a far greater investment of computational resources than the simulation of a simple toy world or abstract problem domain (whereas evolution had access to a physically realistic real world “for free”). In the limiting case, if complete microphysical accuracy were insisted upon, the computational requirements would balloon to utterly infeasible proportions. Requiring more realistic environmental conditions may result in an increase in necessary computational resources, a problem addressed in the next bullet. Insufficient resources (compute, memory)—From the history of computer science, we know of many situations (speech recognition, NN training), where we had a correct algorithm but insufficient computational resources to run it to success. It is possible that we simply do not have hardware powerful enough to emulate evolution. We will address this possibility in section “Computational Complexity of Biological Evolution and Available Compute.” Software design is not amenable to evolutionary methods—Space of software designs may be discrete with no continuous path via incremental fitness to the desired solutions. This is possible, but this implies that original goals of GP are unattainable and misguided. In addition, because a clear mapping exists between solutions to problems and animals as solutions to environmental problems, this would also imply that current explanation for the origin of the species is incorrect. Darwinian algorithm is incomplete or wrong—Finally, we have to consider the possibility that the inspiration behind evolutionary computation, the Darwinian algorithm itself is wrong or at least partially incomplete. If that was true, computer simulations of such algorithm would fail to produce results comparable with observations we see in nature and a search for an alternative algorithm would need to take place. This would be an extraordinary claim and would require that we discard all the other possible explanations from this list. We challenge EA community to prove us wrong by producing an experiment, which evolves nontrivial software from scratch and without human help. That would be the only way in which our findings could be shown to be incorrect. Perhaps, reframing the problem in terms of maximizing negentropy of digital organisms, as suggested by Schrödinger, Michaelian, and Ulanowicz and Hannon, with respect to negative energy being a fundamental property of all life-forms may produce better results. On a positive side, the fact that it seems impossible to evolve complex software implies that we are unlikely to be able to evolve highly sophisticated artificially intelligent agents, which may present significant risk to our safety and security. Just imagine what would have happened, if the very first time we ran a simulation of evolution on a computer, it produced a superintelligent agent. Yampolskiy has shown that programming as a problem is AI-complete; if GP can solve programming that would imply that GP = AGI (artificial general intelligence), but we see no experimental evidence for such claim. In fact, it is more likely that once we have AGI, it could be used to create an intelligent fitness function for GP and so evolve software. Genetic programming will not be the cause of AI, but a product of it. However, neuroevolution methods for optimizing deep learning architectures and parameters remain a strong possibility for creation of AGI.

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!
reddit
LLM Vibe Score0
Human Vibe Score-0.5
regalalgorithmThis week

[N] Last Week in AI News Digest 08/15-08/21: detecting hate speech, dogfight simulation, disaster-response, and more!

Hi there, we at Skynet Today produce a weekly newsletter summarizing each week's major AI news, which seems like it'd be of interest to this subreddit. Here's what's in our latest one: Facebook’s AI for detecting hate speech is facing its biggest challenge yet Facebook has made significant progress recently to proactively take down content that violate its community standards. For example, in the second quarter of 2020, Facebook took down 104.6 million pieces of content. While reviews are typically performed by a vast workforce of human moderators, AI-powered tools have enabled Facebook to do this work at a greater scale for textual content. However, there’s a long way to go for these systems to match or exceed the capabilities of human moderators. This is because a large proportion of hate speech and misinformation is in the form of images and memes, and reasoning about the context and language-image interplay is an extremely difficult challenge for AI. Given Facebook’s scale and the speed at which some use it to spread hate, incite violence, and share lies with millions, Facebook will have to keep running to catch up. AI Slays Top F-16 Pilot In DARPA Dogfight Simulation The Defense Advanced Research Project Agency (DARPA) recently hosted a simulated F16 dogfight competition, with different AI bots competing with each other as well as with human pilots. The top AI bot was able to beat a human pilot 5-0 in the simulated contest. DARPA started this program “as a risk-reduction effort \[…\] to flesh out how human and machine pilots share operational control of a fighter jet to maximize its chances of mission success.” Competition runners are broadly optimistic about the demonstration of AI capabilities, even if they are not close to being deployed on a real aircraft. Of concern, the program had little discussion on the ethics of AI military applications, especially with the lethal autonomous weapon systems being considered. News Advances & Business Microsoft, Energy Dept. to Develop Disaster-Response AI Tools \- The U.S. Department of Energy and Microsoft Corp. on Tuesday announced a partnership to develop artificial-intelligence tools aimed at helping first-responders better react to fast-changing natural events, such as floods and wildfires. Coronavirus: Robot CERi is a bilingual Covid-19 expert \- Ceri is bilingual, clued-up on coronavirus and can tell what mood you are in. Ceri also happens to be a robot. Moscow DOH uses AI platform to detect lung cancer symptoms \- Moscow’s department of health is using an artificial intelligence (AI) platform to detect symptoms of lung cancer in CT scans, as part of a project to implement AI technology for radiology. Scientists develop artificial intelligence system for high precision recognition of hand gestures \- The recognition of human hand gestures by AI systems has been a valuable development over the last decade and has been adopted in high-precision surgical robots, health monitoring equipment and in gaming systems. Forget credit cards - now you can pay with your face. Creepy or cool? \- A new way to pay has arrived in Los Angeles: your face. Concerns & Hype The dystopian tech that companies are selling to help schools reopen sooner \- This fall, AI could be watching students social distance and checking their masks. Thousands of schools nationwide will not be reopening this fall. NYPD Used Facial Recognition Technology In Siege Of Black Lives Matter Activist’s Apartment \- The NYPD deployed facial recognition technology in its hunt for a prominent Black Lives Matter activist, whose home was besieged by dozens of officers and police dogs last week, a spokesperson confirmed to Gothamist. Machines can spot mental health issues - if you hand over your personal data \- Digital diagnosis could transform psychiatry by mining your most intimate data for clues. But is the privacy cost worth it? Supporting Black Artists Who Are Examining AI \- Technology has a complicated relationship with racial justice. Smartphones, internet platforms, and other digital tools can be used to document and expose racism. But digital tools can also fuel racism: smart doorbells surveil Black individuals. A-level and GCSE results in England to be based on teacher assessments in U-turn \- All A-level and GCSE results in England will be based on grades assesed by teachers instead of algorithms. Analysis & Policy GPT-3 and The Question of Automation \- Automation is not an all or nothing proposition. An AI model’s automation capability is highly conjoined with the task and application it is used in. An A.I. Movie Service Could One Day Serve You a New Custom Film Every Time \- How long will it be until an A.I. can make an actual feature film on demand? Fairness, evidence, and predictive equality \- How the causal fairness principle relates to predictive equality How robotics and automation could create new jobs in the new normal \- Depending on who you ask, AI and automation will either destroy jobs or create new ones. In reality, a greater push toward automation will probably both kill and create jobs - human workers will become redundant in certain spheres, sure, but many new roles will likely crop up. Expert Opinions & Discussion within the field Too many AI researchers think real-world problems are not relevant \- The community’s hyperfocus on novel methods ignores what’s really important.

[P] Improve AI 8.0: Free Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions
reddit
LLM Vibe Score0
Human Vibe Score1
gogogadgetlegzThis week

[P] Improve AI 8.0: Free Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions

Improve AI 8.0 - Contextual Multi-Armed Bandit Platform for Scoring, Ranking & Decisions Full announcement post at: https://improve.ai/2023/06/08/contextual-bandit.html We’re thrilled to introduce Improve AI 8.0, a modern, free, production-ready contextual multi-armed bandit platform that quickly scores and ranks items using intuitive reward-based training. Multi-armed bandits and contextual bandits are corner-stone machine learning algorithms that power a myriad of applications including recommendation systems, personalization, query re-ranking, automated decisions, and multi-variate optimization. With version 8, we’ve fully delivered on our original vision - providing a high performance, simple to use, low cost contextual multi-armed bandit platform. Key features of v8.0 include: Simplified APIs 90% more memory efficient XGBoost models The reward tracker & trainer is now free for most uses On-device scoring, ranking, and decisions for iOS and Android apps Native Swift SDK that can rank or score any Encodable Ranked Value Encoding* for accurate scoring of String properties Compact hash tables for reduced model sizes when encoding large numbers of string values Balanced exploration vs exploitation using Thompson Sampling Simple APIs With Swift, Python, or Java, create a list of JSON encodable items and simply call Ranker.rank(items). For instance, in an iOS bedtime story app, you may have a list of Story objects: struct Story: Codable { var title: String var author: String var pageCount: Int } To obtain a ranked list of stories, use just one line of code: let rankedStories = try Ranker(modelUrl).rank(stories) The expected best story will be the first element in the ranked list: let bestStory = rankedStories.first Simple Training Easily train your rankers using reinforcement learning. First, track when an item is used: let tracker = RewardTracker("stories", trackUrl) let rewardId = tracker.track(story, from: rankedStories) Later, if a positive outcome occurs, provide a reward: if (purchased) { tracker.addReward(profit, rewardId) } Reinforcement learning uses positive rewards for favorable outcomes (a “carrot”) and negative rewards for undesirable outcomes (a “stick”). By assigning rewards based on business metrics, such as revenue or conversions, the system optimizes these metrics over time. Contextual Ranking & Scoring Improve AI turns XGBoost into a contextual multi-armed bandit, meaning that context is considered when making ranking or scoring decisions. Often, the choice of the best variant depends on the context that the decision is made within. Let’s take the example of greetings for different times of the day: greetings = ["Good Morning", "Good Afternoon", "Good Evening", "Buenos Días", "Buenas Tardes", "Buenas Noches"] rank() also considers the context of each decision. The context can be any JSON-encodable data structure. ranked = ranker.rank(items=greetings, context={ "day_time": 12.0, "language": "en" }) greeting = ranked[0] Trained with appropriate rewards, Improve AI would learn from scratch which greeting is best for each time of day and language. XGBoost Model Improvements Improve AI v8.0 is 90%+ more memory efficient for most use cases. Feature hashing has been replaced with a feature encoding approach that only uses a single feature per item property, substantially improving both training performance as well as ranking / scoring. Ranked Value Encoding Ranked Value Encoding is our novel approach to encoding string values in a manner that is extremely space efficient, accurate, and helps approximate Thompson Sampling for balanced exploration vs exploitation. The concept of Ranked Value Encoding is similar to commonly used Target Value Encoding for encoding string or categorical features. With Target Value Encoding, each string or categorical feature is replaced with the mean of the target values for that string or category. Target Value Encoding tends to provide good results for regression. However, multi-armed bandits are less concerned with the absolute accuracy of the scores and more concerned with the relative scores between items. Since we don’t need the exact target value, we can simply store the relative ranking of the string values, which saves space in the resulting model, increasing performance and lowering distribution costs. Compact String Encoding In conjunction with Ranked Value Encoding, rather than store entire strings, which could be arbitrarily long, Improve AI v8 models only store compact string hashes, resulting in only \~4 bytes per string for typical models. Proven Performance Improve AI is a production ready implementation of a contextual multi-armed bandit algorithm, honed through years of iterative development. By merging Thompson Sampling with XGBoost, it provides a learning system that is both fast and flexible. Thompson Sampling maintains equilibrium between exploring novel possibilities and capitalizing on established options, while XGBoost ensures cost-effective, high-performance training for updated models. Get Started Today Improve AI is available now for Python, Swift, and Java. Check out the Quick-Start Guide for more information. Thank you for your efforts to improve the world a little bit today.

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.
reddit
LLM Vibe Score0
Human Vibe Score0.765
hardmaruThis week

Interview with Juergen Schmidhuber, renowned ‘Father Of Modern AI’, says his life’s work won't lead to dystopia.

Schmidhuber interview expressing his views on the future of AI and AGI. Original source. I think the interview is of interest to r/MachineLearning, and presents an alternate view, compared to other influential leaders in AI. Juergen Schmidhuber, Renowned 'Father Of Modern AI,' Says His Life’s Work Won't Lead To Dystopia May 23, 2023. Contributed by Hessie Jones. Amid the growing concern about the impact of more advanced artificial intelligence (AI) technologies on society, there are many in the technology community who fear the implications of the advancements in Generative AI if they go unchecked. Dr. Juergen Schmidhuber, a renowned scientist, artificial intelligence researcher and widely regarded as one of the pioneers in the field, is more optimistic. He declares that many of those who suddenly warn against the dangers of AI are just seeking publicity, exploiting the media’s obsession with killer robots which has attracted more attention than “good AI” for healthcare etc. The potential to revolutionize various industries and improve our lives is clear, as are the equal dangers if bad actors leverage the technology for personal gain. Are we headed towards a dystopian future, or is there reason to be optimistic? I had a chance to sit down with Dr. Juergen Schmidhuber to understand his perspective on this seemingly fast-moving AI-train that will leap us into the future. As a teenager in the 1970s, Juergen Schmidhuber became fascinated with the idea of creating intelligent machines that could learn and improve on their own, becoming smarter than himself within his lifetime. This would ultimately lead to his groundbreaking work in the field of deep learning. In the 1980s, he studied computer science at the Technical University of Munich (TUM), where he earned his diploma in 1987. His thesis was on the ultimate self-improving machines that, not only, learn through some pre-wired human-designed learning algorithm, but also learn and improve the learning algorithm itself. Decades later, this became a hot topic. He also received his Ph.D. at TUM in 1991 for work that laid some of the foundations of modern AI. Schmidhuber is best known for his contributions to the development of recurrent neural networks (RNNs), the most powerful type of artificial neural network that can process sequential data such as speech and natural language. With his students Sepp Hochreiter, Felix Gers, Alex Graves, Daan Wierstra, and others, he published architectures and training algorithms for the long short-term memory (LSTM), a type of RNN that is widely used in natural language processing, speech recognition, video games, robotics, and other applications. LSTM has become the most cited neural network of the 20th century, and Business Week called it "arguably the most commercial AI achievement." Throughout his career, Schmidhuber has received various awards and accolades for his groundbreaking work. In 2013, he was awarded the Helmholtz Prize, which recognizes significant contributions to the field of machine learning. In 2016, he was awarded the IEEE Neural Network Pioneer Award for "pioneering contributions to deep learning and neural networks." The media have often called him the “father of modern AI,” because the most cited neural networks all build on his lab’s work. He is quick to point out, however, that AI history goes back centuries. Despite his many accomplishments, at the age of 60, he feels mounting time pressure towards building an Artificial General Intelligence within his lifetime and remains committed to pushing the boundaries of AI research and development. He is currently director of the KAUST AI Initiative, scientific director of the Swiss AI Lab IDSIA, and co-founder and chief scientist of AI company NNAISENSE, whose motto is "AI∀" which is a math-inspired way of saying "AI For All." He continues to work on cutting-edge AI technologies and applications to improve human health and extend human lives and make lives easier for everyone. The following interview has been edited for clarity. Jones: Thank you Juergen for joining me. You have signed letters warning about AI weapons. But you didn't sign the recent publication, "Pause Gigantic AI Experiments: An Open Letter"? Is there a reason? Schmidhuber: Thank you Hessie. Glad to speak with you. I have realized that many of those who warn in public against the dangers of AI are just seeking publicity. I don't think the latest letter will have any significant impact because many AI researchers, companies, and governments will ignore it completely. The proposal frequently uses the word "we" and refers to "us," the humans. But as I have pointed out many times in the past, there is no "we" that everyone can identify with. Ask 10 different people, and you will hear 10 different opinions about what is "good." Some of those opinions will be completely incompatible with each other. Don't forget the enormous amount of conflict between the many people. The letter also says, "If such a pause cannot be quickly put in place, governments should intervene and impose a moratorium." The problem is that different governments have ALSO different opinions about what is good for them and for others. Great Power A will say, if we don't do it, Great Power B will, perhaps secretly, and gain an advantage over us. The same is true for Great Powers C and D. Jones: Everyone acknowledges this fear surrounding current generative AI technology. Moreover, the existential threat of this technology has been publicly acknowledged by Sam Altman, CEO of OpenAI himself, calling for AI regulation. From your perspective, is there an existential threat? Schmidhuber: It is true that AI can be weaponized, and I have no doubt that there will be all kinds of AI arms races, but AI does not introduce a new quality of existential threat. The threat coming from AI weapons seems to pale in comparison to the much older threat from nuclear hydrogen bombs that don’t need AI at all. We should be much more afraid of half-century-old tech in the form of H-bomb rockets. The Tsar Bomba of 1961 had almost 15 times more destructive power than all weapons of WW-II combined. Despite the dramatic nuclear disarmament since the 1980s, there are still more than enough nuclear warheads to wipe out human civilization within two hours, without any AI I’m much more worried about that old existential threat than the rather harmless AI weapons. Jones: I realize that while you compare AI to the threat of nuclear bombs, there is a current danger that a current technology can be put in the hands of humans and enable them to “eventually” exact further harms to individuals of group in a very precise way, like targeted drone attacks. You are giving people a toolset that they've never had before, enabling bad actors, as some have pointed out, to be able to do a lot more than previously because they didn't have this technology. Schmidhuber: Now, all that sounds horrible in principle, but our existing laws are sufficient to deal with these new types of weapons enabled by AI. If you kill someone with a gun, you will go to jail. Same if you kill someone with one of these drones. Law enforcement will get better at understanding new threats and new weapons and will respond with better technology to combat these threats. Enabling drones to target persons from a distance in a way that requires some tracking and some intelligence to perform, which has traditionally been performed by skilled humans, to me, it seems is just an improved version of a traditional weapon, like a gun, which is, you know, a little bit smarter than the old guns. But, in principle, all of that is not a new development. For many centuries, we have had the evolution of better weaponry and deadlier poisons and so on, and law enforcement has evolved their policies to react to these threats over time. So, it's not that we suddenly have a new quality of existential threat and it's much more worrisome than what we have had for about six decades. A large nuclear warhead doesn’t need fancy face recognition to kill an individual. No, it simply wipes out an entire city with ten million inhabitants. Jones: The existential threat that’s implied is the extent to which humans have control over this technology. We see some early cases of opportunism which, as you say, tends to get more media attention than positive breakthroughs. But you’re implying that this will all balance out? Schmidhuber: Historically, we have a long tradition of technological breakthroughs that led to advancements in weapons for the purpose of defense but also for protection. From sticks, to rocks, to axes to gunpowder to cannons to rockets… and now to drones… this has had a drastic influence on human history but what has been consistent throughout history is that those who are using technology to achieve their own ends are themselves, facing the same technology because the opposing side is learning to use it against them. And that's what has been repeated in thousands of years of human history and it will continue. I don't see the new AI arms race as something that is remotely as existential a threat as the good old nuclear warheads. You said something important, in that some people prefer to talk about the downsides rather than the benefits of this technology, but that's misleading, because 95% of all AI research and AI development is about making people happier and advancing human life and health. Jones: Let’s touch on some of those beneficial advances in AI research that have been able to radically change present day methods and achieve breakthroughs. Schmidhuber: All right! For example, eleven years ago, our team with my postdoc Dan Ciresan was the first to win a medical imaging competition through deep learning. We analyzed female breast cells with the objective to determine harmless cells vs. those in the pre-cancer stage. Typically, a trained oncologist needs a long time to make these determinations. Our team, who knew nothing about cancer, were able to train an artificial neural network, which was totally dumb in the beginning, on lots of this kind of data. It was able to outperform all the other methods. Today, this is being used not only for breast cancer, but also for radiology and detecting plaque in arteries, and many other things. Some of the neural networks that we have developed in the last 3 decades are now prevalent across thousands of healthcare applications, detecting Diabetes and Covid-19 and what not. This will eventually permeate across all healthcare. The good consequences of this type of AI are much more important than the click-bait new ways of conducting crimes with AI. Jones: Adoption is a product of reinforced outcomes. The massive scale of adoption either leads us to believe that people have been led astray, or conversely, technology is having a positive effect on people’s lives. Schmidhuber: The latter is the likely case. There's intense commercial pressure towards good AI rather than bad AI because companies want to sell you something, and you are going to buy only stuff you think is going to be good for you. So already just through this simple, commercial pressure, you have a tremendous bias towards good AI rather than bad AI. However, doomsday scenarios like in Schwarzenegger movies grab more attention than documentaries on AI that improve people’s lives. Jones: I would argue that people are drawn to good stories – narratives that contain an adversary and struggle, but in the end, have happy endings. And this is consistent with your comment on human nature and how history, despite its tendency for violence and destruction of humanity, somehow tends to correct itself. Let’s take the example of a technology, which you are aware – GANs – General Adversarial Networks, which today has been used in applications for fake news and disinformation. In actuality, the purpose in the invention of GANs was far from what it is used for today. Schmidhuber: Yes, the name GANs was created in 2014 but we had the basic principle already in the early 1990s. More than 30 years ago, I called it artificial curiosity. It's a very simple way of injecting creativity into a little two network system. This creative AI is not just trying to slavishly imitate humans. Rather, it’s inventing its own goals. Let me explain: You have two networks. One network is producing outputs that could be anything, any action. Then the second network is looking at these actions and it’s trying to predict the consequences of these actions. An action could move a robot, then something happens, and the other network is just trying to predict what will happen. Now we can implement artificial curiosity by reducing the prediction error of the second network, which, at the same time, is the reward of the first network. The first network wants to maximize its reward and so it will invent actions that will lead to situations that will surprise the second network, which it has not yet learned to predict well. In the case where the outputs are fake images, the first network will try to generate images that are good enough to fool the second network, which will attempt to predict the reaction of the environment: fake or real image, and it will try to become better at it. The first network will continue to also improve at generating images whose type the second network will not be able to predict. So, they fight each other. The 2nd network will continue to reduce its prediction error, while the 1st network will attempt to maximize it. Through this zero-sum game the first network gets better and better at producing these convincing fake outputs which look almost realistic. So, once you have an interesting set of images by Vincent Van Gogh, you can generate new images that leverage his style, without the original artist having ever produced the artwork himself. Jones: I see how the Van Gogh example can be applied in an education setting and there are countless examples of artists mimicking styles from famous painters but image generation from this instance that can happen within seconds is quite another feat. And you know this is how GANs has been used. What’s more prevalent today is a socialized enablement of generating images or information to intentionally fool people. It also surfaces new harms that deal with the threat to intellectual property and copyright, where laws have yet to account for. And from your perspective this was not the intention when the model was conceived. What was your motivation in your early conception of what is now GANs? Schmidhuber: My old motivation for GANs was actually very important and it was not to create deepfakes or fake news but to enable AIs to be curious and invent their own goals, to make them explore their environment and make them creative. Suppose you have a robot that executes one action, then something happens, then it executes another action, and so on, because it wants to achieve certain goals in the environment. For example, when the battery is low, this will trigger “pain” through hunger sensors, so it wants to go to the charging station, without running into obstacles, which will trigger other pain sensors. It will seek to minimize pain (encoded through numbers). Now the robot has a friend, the second network, which is a world model ––it’s a prediction machine that learns to predict the consequences of the robot’s actions. Once the robot has a good model of the world, it can use it for planning. It can be used as a simulation of the real world. And then it can determine what is a good action sequence. If the robot imagines this sequence of actions, the model will predict a lot of pain, which it wants to avoid. If it plays this alternative action sequence in its mental model of the world, then it will predict a rewarding situation where it’s going to sit on the charging station and its battery is going to load again. So, it'll prefer to execute the latter action sequence. In the beginning, however, the model of the world knows nothing, so how can we motivate the first network to generate experiments that lead to data that helps the world model learn something it didn’t already know? That’s what artificial curiosity is about. The dueling two network systems effectively explore uncharted environments by creating experiments so that over time the curious AI gets a better sense of how the environment works. This can be applied to all kinds of environments, and has medical applications. Jones: Let’s talk about the future. You have said, “Traditional humans won’t play a significant role in spreading intelligence across the universe.” Schmidhuber: Let’s first conceptually separate two types of AIs. The first type of AI are tools directed by humans. They are trained to do specific things like accurately detect diabetes or heart disease and prevent attacks before they happen. In these cases, the goal is coming from the human. More interesting AIs are setting their own goals. They are inventing their own experiments and learning from them. Their horizons expand and eventually they become more and more general problem solvers in the real world. They are not controlled by their parents, but much of what they learn is through self-invented experiments. A robot, for example, is rotating a toy, and as it is doing this, the video coming in through the camera eyes, changes over time and it begins to learn how this video changes and learns how the 3D nature of the toy generates certain videos if you rotate it a certain way, and eventually, how gravity works, and how the physics of the world works. Like a little scientist! And I have predicted for decades that future scaled-up versions of such AI scientists will want to further expand their horizons, and eventually go where most of the physical resources are, to build more and bigger AIs. And of course, almost all of these resources are far away from earth out there in space, which is hostile to humans but friendly to appropriately designed AI-controlled robots and self-replicating robot factories. So here we are not talking any longer about our tiny biosphere; no, we are talking about the much bigger rest of the universe. Within a few tens of billions of years, curious self-improving AIs will colonize the visible cosmos in a way that’s infeasible for humans. Those who don’t won’t have an impact. Sounds like science fiction, but since the 1970s I have been unable to see a plausible alternative to this scenario, except for a global catastrophe such as an all-out nuclear war that stops this development before it takes off. Jones: How long have these AIs, which can set their own goals — how long have they existed? To what extent can they be independent of human interaction? Schmidhuber: Neural networks like that have existed for over 30 years. My first simple adversarial neural network system of this kind is the one from 1990 described above. You don’t need a teacher there; it's just a little agent running around in the world and trying to invent new experiments that surprise its own prediction machine. Once it has figured out certain parts of the world, the agent will become bored and will move on to more exciting experiments. The simple 1990 systems I mentioned have certain limitations, but in the past three decades, we have also built more sophisticated systems that are setting their own goals and such systems I think will be essential for achieving true intelligence. If you are only imitating humans, you will never go beyond them. So, you really must give AIs the freedom to explore previously unexplored regions of the world in a way that no human is really predefining. Jones: Where is this being done today? Schmidhuber: Variants of neural network-based artificial curiosity are used today for agents that learn to play video games in a human-competitive way. We have also started to use them for automatic design of experiments in fields such as materials science. I bet many other fields will be affected by it: chemistry, biology, drug design, you name it. However, at least for now, these artificial scientists, as I like to call them, cannot yet compete with human scientists. I don’t think it’s going to stay this way but, at the moment, it’s still the case. Sure, AI has made a lot of progress. Since 1997, there have been superhuman chess players, and since 2011, through the DanNet of my team, there have been superhuman visual pattern recognizers. But there are other things where humans, at the moment at least, are much better, in particular, science itself. In the lab we have many first examples of self-directed artificial scientists, but they are not yet convincing enough to appear on the radar screen of the public space, which is currently much more fascinated with simpler systems that just imitate humans and write texts based on previously seen human-written documents. Jones: You speak of these numerous instances dating back 30 years of these lab experiments where these self-driven agents are deciding and learning and moving on once they’ve learned. And I assume that that rate of learning becomes even faster over time. What kind of timeframe are we talking about when this eventually is taken outside of the lab and embedded into society? Schmidhuber: This could still take months or even years :-) Anyway, in the not-too-distant future, we will probably see artificial scientists who are good at devising experiments that allow them to discover new, previously unknown physical laws. As always, we are going to profit from the old trend that has held at least since 1941: every decade compute is getting 100 times cheaper. Jones: How does this trend affect modern AI such as ChatGPT? Schmidhuber: Perhaps you know that all the recent famous AI applications such as ChatGPT and similar models are largely based on principles of artificial neural networks invented in the previous millennium. The main reason why they works so well now is the incredible acceleration of compute per dollar. ChatGPT is driven by a neural network called “Transformer” described in 2017 by Google. I am happy about that because a quarter century earlier in 1991 I had a particular Transformer variant which is now called the “Transformer with linearized self-attention”. Back then, not much could be done with it, because the compute cost was a million times higher than today. But today, one can train such models on half the internet and achieve much more interesting results. Jones: And for how long will this acceleration continue? Schmidhuber: There's no reason to believe that in the next 30 years, we won't have another factor of 1 million and that's going to be really significant. In the near future, for the first time we will have many not-so expensive devices that can compute as much as a human brain. The physical limits of computation, however, are much further out so even if the trend of a factor of 100 every decade continues, the physical limits (of 1051 elementary instructions per second and kilogram of matter) won’t be hit until, say, the mid-next century. Even in our current century, however, we’ll probably have many machines that compute more than all 10 billion human brains collectively and you can imagine, everything will change then! Jones: That is the big question. Is everything going to change? If so, what do you say to the next generation of leaders, currently coming out of college and university. So much of this change is already impacting how they study, how they will work, or how the future of work and livelihood is defined. What is their purpose and how do we change our systems so they will adapt to this new version of intelligence? Schmidhuber: For decades, people have asked me questions like that, because you know what I'm saying now, I have basically said since the 1970s, it’s just that today, people are paying more attention because, back then, they thought this was science fiction. They didn't think that I would ever come close to achieving my crazy life goal of building a machine that learns to become smarter than myself such that I can retire. But now many have changed their minds and think it's conceivable. And now I have two daughters, 23 and 25. People ask me: what do I tell them? They know that Daddy always said, “It seems likely that within your lifetimes, you will have new types of intelligence that are probably going to be superior in many ways, and probably all kinds of interesting ways.” How should they prepare for that? And I kept telling them the obvious: Learn how to learn new things! It's not like in the previous millennium where within 20 years someone learned to be a useful member of society, and then took a job for 40 years and performed in this job until she received her pension. Now things are changing much faster and we must learn continuously just to keep up. I also told my girls that no matter how smart AIs are going to get, learn at least the basics of math and physics, because that’s the essence of our universe, and anybody who understands this will have an advantage, and learn all kinds of new things more easily. I also told them that social skills will remain important, because most future jobs for humans will continue to involve interactions with other humans, but I couldn’t teach them anything about that; they know much more about social skills than I do. You touched on the big philosophical question about people’s purpose. Can this be answered without answering the even grander question: What’s the purpose of the entire universe? We don’t know. But what’s happening right now might be connected to the unknown answer. Don’t think of humans as the crown of creation. Instead view human civilization as part of a much grander scheme, an important step (but not the last one) on the path of the universe from very simple initial conditions towards more and more unfathomable complexity. Now it seems ready to take its next step, a step comparable to the invention of life itself over 3.5 billion years ago. Alas, don’t worry, in the end, all will be good! Jones: Let’s get back to this transformation happening right now with OpenAI. There are many questioning the efficacy and accuracy of ChatGPT, and are concerned its release has been premature. In light of the rampant adoption, educators have banned its use over concerns of plagiarism and how it stifles individual development. Should large language models like ChatGPT be used in school? Schmidhuber: When the calculator was first introduced, instructors forbade students from using it in school. Today, the consensus is that kids should learn the basic methods of arithmetic, but they should also learn to use the “artificial multipliers” aka calculators, even in exams, because laziness and efficiency is a hallmark of intelligence. Any intelligent being wants to minimize its efforts to achieve things. And that's the reason why we have tools, and why our kids are learning to use these tools. The first stone tools were invented maybe 3.5 million years ago; tools just have become more sophisticated over time. In fact, humans have changed in response to the properties of their tools. Our anatomical evolution was shaped by tools such as spears and fire. So, it's going to continue this way. And there is no permanent way of preventing large language models from being used in school. Jones: And when our children, your children graduate, what does their future work look like? Schmidhuber: A single human trying to predict details of how 10 billion people and their machines will evolve in the future is like a single neuron in my brain trying to predict what the entire brain and its tens of billions of neurons will do next year. 40 years ago, before the WWW was created at CERN in Switzerland, who would have predicted all those young people making money as YouTube video bloggers? Nevertheless, let’s make a few limited job-related observations. For a long time, people have thought that desktop jobs may require more intelligence than skills trade or handicraft professions. But now, it turns out that it's much easier to replace certain aspects of desktop jobs than replacing a carpenter, for example. Because everything that works well in AI is happening behind the screen currently, but not so much in the physical world. There are now artificial systems that can read lots of documents and then make really nice summaries of these documents. That is a desktop job. Or you give them a description of an illustration that you want to have for your article and pretty good illustrations are being generated that may need some minimal fine-tuning. But you know, all these desktop jobs are much easier to facilitate than the real tough jobs in the physical world. And it's interesting that the things people thought required intelligence, like playing chess, or writing or summarizing documents, are much easier for machines than they thought. But for things like playing football or soccer, there is no physical robot that can remotely compete with the abilities of a little boy with these skills. So, AI in the physical world, interestingly, is much harder than AI behind the screen in virtual worlds. And it's really exciting, in my opinion, to see that jobs such as plumbers are much more challenging than playing chess or writing another tabloid story. Jones: The way data has been collected in these large language models does not guarantee personal information has not been excluded. Current consent laws already are outdated when it comes to these large language models (LLM). The concern, rightly so, is increasing surveillance and loss of privacy. What is your view on this? Schmidhuber: As I have indicated earlier: are surveillance and loss of privacy inevitable consequences of increasingly complex societies? Super-organisms such as cities and states and companies consist of numerous people, just like people consist of numerous cells. These cells enjoy little privacy. They are constantly monitored by specialized "police cells" and "border guard cells": Are you a cancer cell? Are you an external intruder, a pathogen? Individual cells sacrifice their freedom for the benefits of being part of a multicellular organism. Similarly, for super-organisms such as nations. Over 5000 years ago, writing enabled recorded history and thus became its inaugural and most important invention. Its initial purpose, however, was to facilitate surveillance, to track citizens and their tax payments. The more complex a super-organism, the more comprehensive its collection of information about its constituents. 200 years ago, at least, the parish priest in each village knew everything about all the village people, even about those who did not confess, because they appeared in the confessions of others. Also, everyone soon knew about the stranger who had entered the village, because some occasionally peered out of the window, and what they saw got around. Such control mechanisms were temporarily lost through anonymization in rapidly growing cities but are now returning with the help of new surveillance devices such as smartphones as part of digital nervous systems that tell companies and governments a lot about billions of users. Cameras and drones etc. are becoming increasingly tinier and more ubiquitous. More effective recognition of faces and other detection technology are becoming cheaper and cheaper, and many will use it to identify others anywhere on earth; the big wide world will not offer any more privacy than the local village. Is this good or bad? Some nations may find it easier than others to justify more complex kinds of super-organisms at the expense of the privacy rights of their constituents. Jones: So, there is no way to stop or change this process of collection, or how it continuously informs decisions over time? How do you see governance and rules responding to this, especially amid Italy’s ban on ChatGPT following suspected user data breach and the more recent news about the Meta’s record $1.3billion fine in the company’s handling of user information? Schmidhuber: Data collection has benefits and drawbacks, such as the loss of privacy. How to balance those? I have argued for addressing this through data ownership in data markets. If it is true that data is the new oil, then it should have a price, just like oil. At the moment, the major surveillance platforms such as Meta do not offer users any money for their data and the transitive loss of privacy. In the future, however, we will likely see attempts at creating efficient data markets to figure out the data's true financial value through the interplay between supply and demand. Even some of the sensitive medical data should not be priced by governmental regulators but by patients (and healthy persons) who own it and who may sell or license parts thereof as micro-entrepreneurs in a healthcare data market. Following a previous interview, I gave for one of the largest re-insurance companies , let's look at the different participants in such a data market: patients, hospitals, data companies. (1) Patients with a rare form of cancer can offer more valuable data than patients with a very common form of cancer. (2) Hospitals and their machines are needed to extract the data, e.g., through magnet spin tomography, radiology, evaluations through human doctors, and so on. (3) Companies such as Siemens, Google or IBM would like to buy annotated data to make better artificial neural networks that learn to predict pathologies and diseases and the consequences of therapies. Now the market’s invisible hand will decide about the data’s price through the interplay between demand and supply. On the demand side, you will have several companies offering something for the data, maybe through an app on the smartphone (a bit like a stock market app). On the supply side, each patient in this market should be able to profit from high prices for rare valuable types of data. Likewise, competing data extractors such as hospitals will profit from gaining recognition and trust for extracting data well at a reasonable price. The market will make the whole system efficient through incentives for all who are doing a good job. Soon there will be a flourishing ecosystem of commercial data market advisors and what not, just like the ecosystem surrounding the traditional stock market. The value of the data won’t be determined by governments or ethics committees, but by those who own the data and decide by themselves which parts thereof they want to license to others under certain conditions. At first glance, a market-based system seems to be detrimental to the interest of certain monopolistic companies, as they would have to pay for the data - some would prefer free data and keep their monopoly. However, since every healthy and sick person in the market would suddenly have an incentive to collect and share their data under self-chosen anonymity conditions, there will soon be many more useful data to evaluate all kinds of treatments. On average, people will live longer and healthier, and many companies and the entire healthcare system will benefit. Jones: Finally, what is your view on open source versus the private companies like Google and OpenAI? Is there a danger to supporting these private companies’ large language models versus trying to keep these models open source and transparent, very much like what LAION is doing? Schmidhuber: I signed this open letter by LAION because I strongly favor the open-source movement. And I think it's also something that is going to challenge whatever big tech dominance there might be at the moment. Sure, the best models today are run by big companies with huge budgets for computers, but the exciting fact is that open-source models are not so far behind, some people say maybe six to eight months only. Of course, the private company models are all based on stuff that was created in academia, often in little labs without so much funding, which publish without patenting their results and open source their code and others take it and improved it. Big tech has profited tremendously from academia; their main achievement being that they have scaled up everything greatly, sometimes even failing to credit the original inventors. So, it's very interesting to see that as soon as some big company comes up with a new scaled-up model, lots of students out there are competing, or collaborating, with each other, trying to come up with equal or better performance on smaller networks and smaller machines. And since they are open sourcing, the next guy can have another great idea to improve it, so now there’s tremendous competition also for the big companies. Because of that, and since AI is still getting exponentially cheaper all the time, I don't believe that big tech companies will dominate in the long run. They find it very hard to compete with the enormous open-source movement. As long as you can encourage the open-source community, I think you shouldn't worry too much. Now, of course, you might say if everything is open source, then the bad actors also will more easily have access to these AI tools. And there's truth to that. But as always since the invention of controlled fire, it was good that knowledge about how technology works quickly became public such that everybody could use it. And then, against any bad actor, there's almost immediately a counter actor trying to nullify his efforts. You see, I still believe in our old motto "AI∀" or "AI For All." Jones: Thank you, Juergen for sharing your perspective on this amazing time in history. It’s clear that with new technology, the enormous potential can be matched by disparate and troubling risks which we’ve yet to solve, and even those we have yet to identify. If we are to dispel the fear of a sentient system for which we have no control, humans, alone need to take steps for more responsible development and collaboration to ensure AI technology is used to ultimately benefit society. Humanity will be judged by what we do next.

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call
reddit
LLM Vibe Score0
Human Vibe Score1
noiseinvacuumThis week

[Discussion]: Mark Zuckerberg on Meta's Strategy on Open Source and AI during the earnings call

During the recent earnings call, Mark Zuckerberg answered a question from Eric Sheridan of Goldman Sachs on Meta's AI strategy, opportunities to integrate into products, and why they open source models and how it would benefit their business. I found the reasoning to be very sound and promising for the OSS and AI community. The biggest risk from AI, in my opinion, is not the doomsday scenarios that intuitively come to mind but rather that the most powerful AI systems will only be accessible to the most powerful and resourceful corporations. Quote copied from Ben Thompson's write up on Meta's earning in his Stratechery blog post which goes beyond AI. It's behind a paywall but I highly recommend it personally. Some noteworthy quotes that signal the thought process at Meta FAIR and more broadly We’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon We would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that. ...lead us to do more work in terms of open sourcing, some of the lower level models and tools Open sourcing low level tools make the way we run all this infrastructure more efficient over time. On PyTorch: It’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. I would expect us to be pushing and helping to build out an open ecosystem. For all the negative that comes out of the popular discourse on Meta, I think their work to open source key tech tools over the last 10 years has been exceptional, here's hoping it continues into this decade of AI and pushes other tech giants to also realize the benefits of Open Source. Full Transcript: Right now most of the companies that are training large language models have business models that lead them to a closed approach to development. I think there’s an important opportunity to help create an open ecosystem. If we can help be a part of this, then much of the industry will standardize on using these open tools and help improve them further. So this will make it easier for other companies to integrate with our products and platforms as we enable more integrations, and that will help our products stay at the leading edge as well. Our approach to AI and our infrastructure has always been fairly open. We open source many of our state of the art models so people can experiment and build with them. This quarter we released our LLaMa LLM to researchers. It has 65 billion parameters but outperforms larger models and has proven quite popular. We’ve also open-sourced three other groundbreaking visual models along with their training data and model weights — Segment Anything, DinoV2, and our Animated Drawings tool — and we’ve gotten positive feedback on all of those as well. I think that there’s an important distinction between the products we offer and a lot of the technical infrastructure, especially the software that we write to support that. And historically, whether it’s the Open Compute project that we’ve done or just open sourcing a lot of the infrastructure that we’ve built, we’ve historically open sourced a lot of that infrastructure, even though we haven’t open sourced the code for our core products or anything like that. And the reason why I think why we do this is that unlike some of the other companies in the space, we’re not selling a cloud computing service where we try to keep the different software infrastructure that we’re building proprietary. For us, it’s way better if the industry standardizes on the basic tools that we’re using and therefore we can benefit from the improvements that others make and others’ use of those tools can, in some cases like Open Compute, drive down the costs of those things which make our business more efficient too. So I think to some degree we’re just playing a different game on the infrastructure than companies like Google or Microsoft or Amazon, and that creates different incentives for us. So overall, I think that that’s going to lead us to do more work in terms of open sourcing, some of the lower level models and tools. But of course, a lot of the product work itself is going to be specific and integrated with the things that we do. So it’s not that everything we do is going to be open. Obviously, a bunch of this needs to be developed in a way that creates unique value for our products, but I think in terms of the basic models, I would expect us to be pushing and helping to build out an open ecosystem here, which I think is something that’s going to be important. On the AI tools, and we have a bunch of history here, right? So if you if you look at what we’ve done with PyTorch, for example, which has generally become the standard in the industry as a tool that a lot of folks who are building AI models and different things in that space use, it’s generally been very valuable for us to provide that because now all of the best developers across the industry are using tools that we’re also using internally. So the tool chain is the same. So when they create some innovation, we can easily integrate it into the things that we’re doing. When we improve something, it improves other products too. Because it’s integrated with our technology stack, when there are opportunities to make integrations with products, it’s much easier to make sure that developers and other folks are compatible with the things that we need in the way that our systems work. So there are a lot of advantages, but I view this more as a kind of back end infrastructure advantage with potential integrations on the product side, but one that should hopefully enable us to stay at the leading edge and integrate more broadly with the community and also make the way we run all this infrastructure more efficient over time. There are a number of models. I just gave PyTorch as an example. Open Compute is another model that has worked really well for us in this way, both to incorporate both innovation and scale efficiency into our own infrastructure. So I think that there’s, our incentives I think are basically aligned towards moving in this direction. Now that said, there’s a lot to figure out, right? So when you asked if there are going to be other opportunities, I hope so. I can’t speak to what all those things might be now. This is all quite early in getting developed. The better we do at the foundational work, the more opportunities I think that will come and present themselves. So I think that that’s all stuff that we need to figure out. But at least at the base level, I think we’re generally incentivized to move in this direction. And we also need to figure out how to go in that direction over time. I mean, I mentioned LLaMA before and I also want to be clear that while I’m talking about helping contribute to an open ecosystem, LLaMA is a model that we only really made available to researchers and there’s a lot of really good stuff that’s happening there. But a lot of the work that we’re doing, I think, we would aspire to and hope to make even more open than that. So, we’ll need to figure out a way to do that.

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.
reddit
LLM Vibe Score0
Human Vibe Score0.6
AlexSnakeKingThis week

[Discussion] When ML and Data Science are the death of a good company: A cautionary tale.

TD;LR: At Company A, Team X does advanced analytics using on-prem ERP tools and older programming languages. Their tools work very well and are designed based on very deep business and domain expertise. Team Y is a new and ambitious Data Science team that thinks they can replace Team X's tools with a bunch of R scripts and a custom built ML platform. Their models are simplistic, but more "fashionable" compared to the econometric models used by Team X, and team Y benefits from the ML/DS moniker so leadership is allowing Team Y to start a large scale overhaul of the analytics platform in question. Team Y doesn't have the experience for such a larger scale transformation, and is refusing to collaborate with team X. This project is very likely going to fail, and cause serious harm to the company as a whole financially and from a people perspective. I argue that this is not just because of bad leadership, but also because of various trends and mindsets in the DS community at large. Update (Jump to below the line for the original story): Several people in the comments are pointing out that this just a management failure, not something due to ML/DS, and that you can replace DS with any buzz tech and the story will still be relevant. My response: Of course, any failure at an organization level is ultimately a management failure one way or the other. Moreover, it is also the case that ML/DS when done correctly, will always improve a company's bottom line. There is no scenario where the proper ML solution, delivered at a reasonable cost and in a timely fashion, will somehow hurt the company's bottom line. My point is that in this case management is failing because of certain trends and practices that are specific to the ML/DS community, namely: The idea that DS teams should operate independently of tech and business orgs -- too much autonomy for DS teams The disregard for domain knowledge that seems prevalent nowadays thanks to the ML hype, that DS can be generalists and someone with good enough ML chops can solve any business problem. That wasn't the case when I first left academia for the industry in 2009 (back then nobody would even bother with a phone screen if you didn't have the right domain knowledge). Over reliance on resources who check all the ML hype related boxes (knows Python, R, Tensorflow, Shiny, etc..., has the right Coursera certifications, has blogged on the topic, etc...), but are lacking in depth of experience. DS interviews nowadays all seem to be: Can you tell me what a p-value is? What is elastic net regression? Show me how to fit a model in sklearn? How do you impute NAs in an R dataframe? Any smart person can look those up on Stackoverflow or Cross-Validated,.....Instead teams should be asking stuff like: why does portfolio optimization use QP not LP? How does a forecast influence a customer service level? When should a recommendation engine be content based and when should it use collaborative filtering? etc... (This is a true story, happening to the company I currently work for. Names, domains, algorithms, and roles have been shuffled around to protect my anonymity)  Company A has been around for several decades. It is not the biggest name in its domain, but it is a well respected one. Risk analysis and portfolio optimization have been a core of Company A's business since the 90s. They have a large team of 30 or so analysts who perform those tasks on a daily basis. These analysts use ERP solutions implemented for them by one the big ERP companies (SAP, Teradata, Oracle, JD Edwards,...) or one of the major tech consulting companies (Deloitte, Accenture, PWC, Capgemini, etc...) in collaboration with their own in house engineering team. The tools used are embarrassingly old school: Classic RDBMS running on on-prem servers or maybe even on mainframes, code written in COBOL, Fortran, weird proprietary stuff like ABAP or SPSS.....you get the picture. But the models and analytic functions were pretty sophisticated, and surprisingly cutting edge compared to the published academic literature. Most of all, they fit well with the company's enterprise ecosystem, and were honed based on years of deep domain knowledge.  They have a tech team of several engineers (poached from the aforementioned software and consulting companies) and product managers (who came from the experienced pools of analysts and managers who use the software, or poached from business rivals) maintaining and running this software. Their technology might be old school, but collectively, they know the domain and the company's overall architecture very, very well. They've guided the company through several large scale upgrades and migrations and they have a track record of delivering on time, without too much overhead. The few times they've stumbled, they knew how to pick themselves up very quickly. In fact within their industry niche, they have a reputation for their expertise, and have very good relations with the various vendors they've had to deal with. They were the launching pad of several successful ERP consulting careers.  Interestingly, despite dealing on a daily basis with statistical modeling and optimization algorithms, none of the analysts, engineers, or product managers involved describe themselves as data scientists or machine learning experts. It is mostly a cultural thing: Their expertise predates the Data Science/ML hype that started circa 2010, and they got most of their chops using proprietary enterprise tools instead of the open source tools popular nowadays. A few of them have formal statistical training, but most of them came from engineering or domain backgrounds and learned stats on the fly while doing their job. Call this team "Team X".  Sometime around the mid 2010s, Company A started having some serious anxiety issues: Although still doing very well for a company its size, overall economic and demographic trends were shrinking its customer base, and a couple of so called disruptors came up with a new app and business model that started seriously eating into their revenue. A suitable reaction to appease shareholders and Wall Street was necessary. The company already had a decent website and a pretty snazzy app, what more could be done? Leadership decided that it was high time that AI and ML become a core part of the company's business. An ambitious Manager, with no science or engineering background, but who had very briefly toyed with a recommender system a couple of years back, was chosen to build a data science team, call it team "Y" (he had a bachelor's in history from the local state college and worked for several years in the company's marketing org). Team "Y" consists mostly of internal hires who decided they wanted to be data scientists and completed a Coursera certification or a Galvanize boot camp, before being brought on to the team, along with a few of fresh Ph.D or M.Sc holders who didn't like academia and wanted to try their hand at an industry role. All of them were very bright people, they could write great Medium blog posts and give inspiring TED talks, but collectively they had very little real world industry experience. As is the fashion nowadays, this group was made part of a data science org that reported directly to the CEO and Board, bypassing the CIO and any tech or business VPs, since Company A wanted to claim the monikers "data driven" and "AI powered" in their upcoming shareholder meetings. In 3 or 4 years of existence, team Y produced a few Python and R scripts. Their architectural experience  consisted almost entirely in connecting Flask to S3 buckets or Redshift tables, with a couple of the more resourceful ones learning how to plug their models into Tableau or how to spin up a Kuberneties pod.  But they needn't worry: The aforementioned manager, who was now a director (and was also doing an online Masters to make up for his qualifications gap and bolster his chances of becoming VP soon - at least he now understands what L1 regularization is), was a master at playing corporate politics and self-promotion. No matter how few actionable insights team Y produced or how little code they deployed to production, he always had their back and made sure they had ample funding. In fact he now had grandiose plans for setting up an all-purpose machine learning platform that can be used to solve all of the company's data problems.  A couple of sharp minded members of team Y, upon googling their industry name along with the word "data science", realized that risk analysis was a prime candidate for being solved with Bayesian models, and there was already a nifty R package for doing just that, whose tutorial they went through on R-Bloggers.com. One of them had even submitted a Bayesian classifier Kernel for a competition on Kaggle (he was 203rd on the leaderboard), and was eager to put his new-found expertise to use on a real world problem. They pitched the idea to their director, who saw a perfect use case for his upcoming ML platform. They started work on it immediately, without bothering to check whether anybody at Company A was already doing risk analysis. Since their org was independent, they didn't really need to check with anybody else before they got funding for their initiative. Although it was basically a Naive Bayes classifier, the term ML was added to the project tile, to impress the board.  As they progressed with their work however, tensions started to build. They had asked the data warehousing and CA analytics teams to build pipelines for them, and word eventually got out to team X about their project. Team X was initially thrilled: They offered to collaborate whole heartedly, and would have loved to add an ML based feather to their already impressive cap. The product owners and analysts were totally onboard as well: They saw a chance to get in on the whole Data Science hype that they kept hearing about. But through some weird mix of arrogance and insecurity, team Y refused to collaborate with them or share any of their long term goals with them, even as they went to other parts of the company giving brown bag presentations and tutorials on the new model they created.  Team X got resentful: from what they saw of team Y's model, their approach was hopelessly naive and had little chances of scaling or being sustainable in production, and they knew exactly how to help with that. Deploying the model to production would have taken them a few days, given how comfortable they were with DevOps and continuous delivery (team Y had taken several months to figure out how to deploy a simple R script to production). And despite how old school their own tech was, team X were crafty enough to be able to plug it in to their existing architecture. Moreover, the output of the model was such that it didn't take into account how the business will consume it or how it was going to be fed to downstream systems, and the product owners could have gone a long way in making the model more amenable to adoption by the business stakeholders. But team Y wouldn't listen, and their leads brushed off any attempts at communication, let alone collaboration. The vibe that team Y was giving off was "We are the cutting edge ML team, you guys are the legacy server grunts. We don't need your opinion.", and they seemed to have a complete disregard for domain knowledge, or worse, they thought that all that domain knowledge consisted of was being able to grasp the definitions of a few business metrics.  Team X got frustrated and tried to express their concerns to leadership. But despite owning a vital link in Company A's business process, they were only \~50 people in a large 1000 strong technology and operations org, and they were several layers removed from the C-suite, so it was impossible for them to get their voices heard.  Meanwhile, the unstoppable director was doing what he did best: Playing corporate politics. Despite how little his team had actually delivered, he had convinced the board that all analysis and optimization tasks should now be migrated to his yet to be delivered ML platform. Since most leaders now knew that there was overlap between team Y and team X's objectives, his pitch was no longer that team Y was going to create a new insight, but that they were going to replace (or modernize) the legacy statistics based on-prem tools with more accurate cloud based ML tools. Never mind that there was no support in the academic literature for the idea that Naive Bayes works better than the Econometric approaches used by team X, let alone the additional wacky idea that Bayesian Optimization would definitely outperform the QP solvers that were running in production.  Unbeknownst to team X, the original Bayesian risk analysis project has now grown into a multimillion dollar major overhaul initiative, which included the eventual replacement of all of the tools and functions supported by team X along with the necessary migration to the cloud. The CIO and a couple of business VPs are on now board, and tech leadership is treating it as a done deal. An outside vendor, a startup who nobody had heard of, was contracted to help build the platform, since team Y has no engineering skills. The choice was deliberate, as calling on any of the established consulting or software companies would have eventually led leadership to the conclusion that team X was better suited for a transformation on this scale than team Y.  Team Y has no experience with any major ERP deployments, and no domain knowledge, yet they are being tasked with fundamentally changing the business process that is at the core of Company A's business. Their models actually perform worse than those deployed by team X, and their architecture is hopelessly simplistic, compared to what is necessary for running such a solution in production.  Ironically, using Bayesian thinking and based on all the evidence, the likelihood that team Y succeeds is close to 0%. At best, the project is going to end up being a write off of 50 million dollars or more. Once the !@#$!@hits the fan, a couple of executive heads are going to role, and dozens of people will get laid off. At worst, given how vital risk analysis and portfolio optimization is to Company A's revenue stream, the failure will eventually sink the whole company. It probably won't go bankrupt, but it will lose a significant portion of its business and work force. Failed ERP implementations can and do sink large companies: Just see what happened to National Grid US, SuperValu or Target Canada.  One might argue that this is more about corporate disfunction and bad leadership than about data science and AI. But I disagree. I think the core driver of this debacle is indeed the blind faith in Data Scientists, ML models and the promise of AI, and the overall culture of hype and self promotion that is very common among the ML crowd.  We haven't seen the end of this story: I sincerely hope that this ends well for the sake of my colleagues and all involved. Company A is a good company, and both its customers and its employees deserver better. But the chances of that happening are negligible given all the information available, and this failure will hit my company hard.

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup
reddit
LLM Vibe Score0
Human Vibe Score0.667
milaworldThis week

[N] How Stability AI’s Founder Tanked His Billion-Dollar Startup

forbes article: https://www.forbes.com/sites/kenrickcai/2024/03/29/how-stability-ais-founder-tanked-his-billion-dollar-startup/ archive no paywall: https://archive.is/snbeV How Stability AI’s Founder Tanked His Billion-Dollar Startup Mar 29, 2024 Stability AI founder Emad Mostaque took the stage last week at the Terranea Resort in Palos Verdes, California to roaring applause and an introduction from an AI-generated Aristotle who announced him as “a modern Prometheus” with “the astuteness of Athena and the vision of Daedalus.” “Under his stewardship, AI becomes the Herculean force poised to vanquish the twin serpents of illness and ailment and extend the olive branch of longevity,” the faux Aristotle proclaimed. “I think that’s the best intro I’ve ever had,” Mostaque said. But behind Mostaque's hagiographic introduction lay a grim and fast metastasizing truth. Stability, once one of AI’s buzziest startups, was floundering. It had been running out of money for months and Mostaque had been unable to secure enough additional funding. It had defaulted on payments to Amazon whose cloud service undergirded Stability’s core offerings. The star research team behind its flagship text-to-image generator Stable Diffusion had tendered their resignations just three days before — as Forbes would first report — and other senior leaders had issued him an ultimatum: resign, or we walk too. Still, onstage before a massive audience of peers and acolytes, Mostaque talked a big game. “AI is jet planes for the mind,” he opined. “AI is our collective intelligence. It's the human Colossus.” He claimed a new, faster version of the Stable Diffusion image generator released earlier this month could generate “200 cats with hats per second.” But later, when he was asked about Stability’s financial model, Mostaque fumbled. “I can’t say that publicly,” he replied. “But it’s going well. We’re ahead of forecast.” Four days later, Mostaque stepped down as CEO of Stability, as Forbes first reported. In a post to X, the service formerly known as Twitter, he claimed he’d voluntarily abdicated his role to decentralize “the concentration of power in AI.” But sources told Forbes that was hardly the case. Behind the scenes, Mostaque had fought to maintain his position and control despite mounting pressure externally and internally to step down. Company documents and interviews with 32 current and former employees, investors, collaborators and industry observers suggest his abrupt exit was the result of poor business judgment and wild overspending that undermined confidence in his vision and leadership, and ultimately kneecapped the company. Mostaque, through his attorneys, declined to comment on record on a detailed list of questions about the reporting in this story. But in an email to Forbes earlier this week he broadly disputed the allegations. “Nobody tells you how hard it is to be a CEO and there are better CEOs than me to scale a business,” he said in a statement. “I am not sure anyone else would have been able to build and grow the research team to build the best and most widely used models out there and I’m very proud of the team there. I look forward to moving onto the next problem to handle and hopefully move the needle.” In an emailed statement, Christian Laforte and Shan Shan Wong, the interim co-CEOs who replaced Mostaque, said, "the company remains focused on commercializing its world leading technology” and providing it “to partners across the creative industries." After starting Stability in 2019, Mostaque built the company into an early AI juggernaut by seizing upon a promising research project that would become Stable Diffusion and funding it into a business reality. The ease with which the software generated detailed images from the simplest text prompts immediately captivated the public: 10 million people used it on any given day, the company told Forbes in early 2023. For some true believers, Mostaque was a crucial advocate for open-source AI development in a space dominated by the closed systems of OpenAI, Google and Anthropic. But his startup’s rise to one of the buzziest in generative AI was in part built on a series of exaggerations and misleading claims, as Forbes first reported last year (Mostaque disputed some points at the time). And they continued after he raised $100 million at a $1 billion valuation just days after launching Stable Diffusion in 2022. His failure to deliver on an array of grand promises, like building bespoke AI models for nation states, and his decision to pour tens of millions into research without a sustainable business plan, eroded Stability’s foundations and jeopardized its future. "He was just giving shit away,” one former employee told Forbes. “That man legitimately wanted to transform the world. He actually wanted to train AI models for kids in Malawi. Was it practical? Absolutely not." By October 2023, Stability would have less than $4 million left in the bank, according to an internal memo prepared for a board meeting and reviewed by Forbes. And mounting debt, including months of overdue Amazon Web Services payments, had already left it in the red. To avoid legal penalties for skipping Americans staff’s payroll, the document explained, the London-based startup was considering delaying tax payments to the U.K. government. It was Stability’s armada of GPUs, the wildly powerful and equally expensive chips undergirding AI, that were so taxing the company’s finances. Hosted by AWS, they had long been one of Mostaque’s bragging points; he often touted them as one of the world’s 10 largest supercomputers. They were responsible for helping Stability’s researchers build and maintain one of the top AI image generators, as well as break important new ground on generative audio, video and 3D models. “Undeniably, Stability has continued to ship a lot of models,” said one former employee. “They may not have profited off of it, but the broader ecosystem benefitted in a huge, huge way.” But the costs associated with so much compute were now threatening to sink the company. According to an internal October financial forecast seen by Forbes, Stability was on track to spend $99 million on compute in 2023. It noted as well that Stability was “underpaying AWS bills for July (by $1M)” and “not planning to pay AWS at the end of October for August usage ($7M).” Then there were the September and October bills, plus $1 million owed to Google Cloud and $600,000 to GPU cloud data center CoreWeave. (Amazon, Google and CoreWeave declined to comment.) With an additional $54 million allocated to wages and operating expenses, Stability’s total projected costs for 2023 were $153 million. But according to its October financial report, its projected revenue for the calendar year was just $11 million. Stability was on track to lose more money per month than it made in an entire year. The company’s dire financial position had thoroughly soured Stability’s current investors, including Coatue, which had invested tens of millions in the company during its $101 million funding round in 2022. In the middle of 2023, Mostaque agreed to an independent audit after Coatue raised a series of concerns, according to a source with direct knowledge of the matter. The outcome of the investigation is unclear. Coatue declined to comment. Within a week of an early October board meeting where Mostaque shared that financial forecast, Lightspeed Venture Partners, another major investor, sent a letter to the board urging them to sell the company. The distressing numbers had “severely undermined” the firm’s confidence in Mostaque’s ability to lead the company. “In particular, we are surprised and deeply concerned by a cash position just now disclosed to us that is inconsistent with prior discussions on this topic,” Lightspeed’s general counsel Brett Nissenberg wrote in the letter, a copy of which was viewed by Forbes. “Lightspeed believes that the company is not likely financeable on terms that would assure the company’s long term sound financial position.” (Lightspeed declined a request for comment.) The calls for a sale led Stability to quietly begin looking for a buyer. Bloomberg reported in November that Stability approached AI startups Cohere and Jasper to gauge their interest. Stability denied this, and Jasper CEO Timothy Young did the same when reached for comment by Forbes. A Cohere representative declined to comment. But one prominent AI company confirmed that Mostaque’s representatives had reached out to them to test the waters. Those talks did not advance because “the numbers didn’t add up,” this person, who declined to be named due to the confidential nature of the talks, told Forbes. Stability also tried to court Samsung as a buyer, going so far as to redecorate its office in advance of a planned meeting with the Korean electronics giant. (Samsung said that it invested in Stability in 2023 and that it does not comment on M&A discussions.) Coatue had been calling for Mostaque’s resignation for months, according to a source with direct knowledge. But it and other investors were unable to oust him because he was the company’s majority shareholder. When they tried a different tact by rallying other investors to offer him a juicy equity package to resign, Mostaque refused, said two sources. By October, Coatue and Lightspeed had had enough. Coatue left the board and Lightspeed resigned its observer seat. “Emad infuriated our initial investors so much it’s just making it impossible for us to raise more money under acceptable terms,” one current Stability executive told Forbes. The early months of 2024 saw Stability’s already precarious position eroding further still. Employees were quietly laid off. Three people in a position to know estimated that at least 10% of staff were cut. And cash reserves continued to dwindle. Mostaque mentioned a lifeline at the October board meeting: $95 million in tentative funding from new investors, pending due diligence. But in the end, only a fraction of it was wired, two sources say, much of it from Intel, which Forbes has learned invested $20 million, a fraction of what was reported. (Intel did not return a request for comment by publication time.) Two hours after Forbes broke the news of Mostaque’s plans to step down as CEO, Stability issued a press release confirming his resignation. Chief operating officer Wong and chief technology officer Laforte have taken over in the interim. Mostaque, who said on X that he still owns a majority of the company, also stepped down from the board, which has now initiated a search for a permanent CEO. There is a lot of work to be done to turn things around, and very little time in which to do it. Said the current Stability executive, “There’s still a possibility of a turnaround story, but the odds drop by the day.” In July of 2023, Mostaque still thought he could pull it off. Halfway through the month, he shared a fundraising plan with his lieutenants. It was wildly optimistic, detailing the raise of $500 million in cash and another $750 million in computing facilities from marquee investors like Nvidia, Google, Intel and the World Bank (Nvidia and Google declined comment. Intel did not respond. The World Bank said it did not invest in Stability). In a Slack message reviewed by Forbes, Mostaque said Google was “willing to move fast” and the round was “likely to be oversubscribed.” It wasn’t. Three people with direct knowledge of these fundraising efforts told Forbes that while there was some interest in Stability, talks often stalled when it came time to disclose financials. Two of them noted that earlier in the year, Mostaque had simply stopped engaging with VCs who asked for numbers. Only one firm invested around that time: actor Ashton Kutcher’s Sound Ventures, which invested $35 million in the form of a convertible SAFE note during the second quarter, according to an internal document. (Sound Ventures did not respond to a request for comment.) And though he’d managed to score a meeting with Nvidia and its CEO Jensen Huang, it ended in disaster, according to two sources. “Under Jensen's microscopic questions, Emad just fell apart,” a source in position to know told Forbes. Huang quickly concluded Stability wasn’t ready for an investment from Nvidia, the sources said. Mostaque told Forbes in an email that he had not met with Huang since 2022, except to say “hello and what’s up a few times after.” His July 2023 message references a plan to raise $150 million from Nvidia. (Nvidia declined to comment.) After a June Forbes investigation citing more than 30 sources revealed Mostaque’s history of misleading claims, Mostaque struggled to raise funding, a Stability investor told Forbes. (Mostaque disputed the story at the time and called it "coordinated lies" in his email this week to Forbes). Increasingly, investors scrutinized his assertions and pressed for data. And Young, now the CEO of Jasper, turned down a verbal offer to be Stability’s president after reading the article, according to a source with direct knowledge of the matter. The collapse of the talks aggravated the board and other executives, who had hoped Young would compensate for the sales and business management skills that Mostaque lacked, according to four people in a position to know. (Young declined to comment.) When Stability’s senior leadership convened in London for the CogX conference in September, the financing had still not closed. There, a group of executives confronted Mostaque asking questions about the company’s cash position and runway, according to three people with direct knowledge of the incident. They did not get the clarity they’d hoped for. By October, Mostaque had reduced his fundraising target by more than 80%. The months that followed saw a steady drumbeat of departures — general counsel Adam Avrunin, vice presidents Mike Melnicki, Ed Newton-Rex and Joe Penna, chief people officer Ozden Onder — culminating in the demoralizing March exit of Stable Diffusion’s primary developers Robin Rombach, Andreas Blattmann, Patrick Esser and Dominik Lorenz. Rombach, who led the team, had been angling to leave for months, two sources said, first threatening to resign last summer because of the fundraising failures. Others left over concerns about cash flow, as well as liabilities — including what four people described as Mostaque’s lax approach to ensuring that Stability products could not be used to produce child sexual abuse imagery. “Stability AI is committed to preventing the misuse of AI and prohibits the use of our image models and services for unlawful activity, including attempts to edit or create CSAM,” Ella Irwin, senior vice president of integrity, said in a statement. Newton-Rex told Forbes he resigned because he disagreed with Stability’s position that training AI on copyrighted work without consent is fair use. Melnicki and Penna declined to comment. Avrunin and Onder could not be reached for comment. None of the researchers responded to requests for comment. The Stable Diffusion researchers’ departure as a cohort says a lot about the state of Stability AI. The company’s researchers were widely viewed as its crown jewels, their work subsidized with a firehose of pricey compute power that was even extended to people outside the company. Martino Russi, an artificial intelligence researcher, told Forbes that though he was never formally employed by Stability, the company provided him a “staggering” amount of compute between January and April 2023 to play around with developing an AI video generator that Stability might someday use. “It was Candy Land or Coney Island,” said Russi, who estimates that his experiment, which was ultimately shelved, cost the company $2.5 million. Stable Diffusion was simultaneously Stability’s marquee product and its existential cash crisis. One current employee described it to Forbes as “a giant vacuum that absorbed everything: money, compute, people.” While the software was widely used, with Mostaque claiming downloads reaching into the hundreds of millions, Stability struggled to translate that wild success into revenue. Mostaque knew it could be done — peers at Databricks, Elastic and MongoDB had all turned a free product into a lucrative business — he just couldn’t figure out how. His first attempt was Stability’s API, which allowed paying customers to integrate Stable Diffusion into their own products. In early 2023, a handful of small companies, like art generator app NightCafe and presentation software startup Tome, signed on, according to four people with knowledge of the deals. But Stability’s poor account management services soured many, and in a matter of months NightCafe and Tome canceled their contracts, three people said. NightCafe founder Angus Russell told Forbes that his company switched to a competitor which “offered much cheaper inference costs and a broader service.” Tome did not respond to a request for comment. Meanwhile, Mostaque’s efforts to court larger companies like Samsung and Snapchat were failing, according to five people familiar with the effort. Canva, which was already one of the heaviest users of open-sourced Stable Diffusion, had multiple discussions with Stability, which was angling for a contract it hoped would generate several millions in annual revenue. But the deal never materialized, four sources said. “These three companies wanted and needed us,” one former employee told Forbes. “They would have been the perfect customers.” (Samsung, Snap and Canva declined to comment.) “It’s not that there was not an appetite to pay Stability — there were tons of companies that would have that wanted to,” the former employee said. “There was a huge opportunity and demand, but just a resistance to execution.” Mostaque’s other big idea was to provide governments with bespoke national AI models that would invigorate their economies and citizenry. “Emad envisions a world where AI through 100 national models serves not as a tool of the few, but as a benefactor to all promising to confront great adversaries, cancer, autism, and the sands of time itself,” the AI avatar of Aristotle said in his intro at the conference. Mostaque told several prospective customers that he could deliver such models within 60 days — an untenable timeline, according to two people in position to know. Stability attempted to develop a model for the Singaporean government over the protestation of employees who questioned its technical feasibility, three sources familiar with the effort told Forbes. But it couldn’t pull it off and Singapore never became a customer. (The government of Singapore confirmed it did not enter into a deal with Stability, but declined to answer additional questions.) As Stability careened from one new business idea to another, resources were abruptly reallocated and researchers reassigned. The whiplash shifts in a largely siloed organization demoralized and infuriated employees. “There were ‘urgent’ things, ‘urgent urgent’ things and ‘most urgent,’” one former employee complained. “None of these things seem important if everything is important.” Another former Stability executive was far more pointed in their assessment. “Emad is the most disorganized leader I have ever worked with in my career,” this person told Forbes. “He has no vision, and changes directions every week, often based on what he sees on Twitter.” In a video interview posted shortly before this story was published, Mostaque explained his leadership style: “I'm particularly great at taking creatives, developers, researchers, others, and achieving their full potential in designing systems. But I should not be dealing with, you know, HR and operations and business development and other elements. There are far better people than me to do that.” By December 2023, Stability had partially abandoned its open-source roots and announced that any commercial use of Stable Diffusion would cost customers at least $20 per month (non-commercial and research use of Stable Diffusion would remain free). But privately, Stability was considering a potentially more lucrative source of revenue: reselling the compute it was leasing from providers like AWS, according to six people familiar with the effort. Though it was essentially GPU arbitrage, Stability framed the strategy to investors as a “managed services” offering. Its damning October financial report projected optimistically that such an offering would bring in $139 million in 2024 — 98% of its revenue. Multiple employees at the time told Forbes they feared reselling compute, even if the company called it “managed services,” would violate the terms of Stability’s contract with AWS. Amazon declined to comment. “The line internally was that we are not reselling compute,” one former employee said. “This was some of the dirtiest feeling stuff.” Stability also discussed reselling a cluster of Nvidia A100 chips, leased via CoreWeave, to the venture capital firm Andreessen Horowitz, three sources said. “It was under the guise of managed services, but there wasn’t any management happening,” one of these people told Forbes. Andreessen Horowitz and CoreWeave declined to comment. Stability did not respond to questions about if it plans to continue this strategy now that Mostaque is out of the picture. Regardless, interim co-CEOs Wong and Laforte are on a tight timeline to clean up his mess. Board chairman Jim O’Shaughnessy said in a statement that he was confident the pair “will adeptly steer the company forward in developing and commercializing industry-leading generative AI products.” But burn continues to far outpace revenue. The Financial Times reported Friday that the company made $5.4 million of revenue in February, against $8 million in costs. Several sources said there are ongoing concerns about making payroll for the roughly 150 remaining employees. Leadership roles have gone vacant for months amid the disarray, leaving the company increasingly directionless. Meanwhile, a potentially catastrophic legal threat looms over the company: A trio of copyright infringement lawsuits brought by Getty Images and a group of artists in the U.S. and U.K., who claim Stability illegally used their art and photography to train the AI models powering Stable Diffusion. A London-based court has already rejected the company’s bid to throw out one of the lawsuits on the basis that none of its researchers were based in the U.K. And Stability’s claim that Getty’s Delaware lawsuit should be blocked because it's a U.K.-based company was rejected. (Stability did not respond to questions about the litigation.) AI-related copyright litigation “could go on for years,” according to Eric Goldman, a law professor at Santa Clara University. He told Forbes that though plaintiffs suing AI firms face an uphill battle overcoming the existing legal precedent on copyright infringement, the quantity of arguments available to make are virtually inexhaustible. “Like in military theory, if there’s a gap in your lines, that’s where the enemy pours through — if any one of those arguments succeeds, it could completely change the generative AI environment,” he said. “In some sense, generative AI as an industry has to win everything.” Stability, which had more than $100 million in the bank just a year and a half ago, is in a deep hole. Not only does it need more funding, it needs a viable business model — or a buyer with the vision and chops to make it successful in a fast-moving and highly competitive sector. At an all hands meeting this past Monday, Stability’s new leaders detailed a path forward. One point of emphasis: a plan to better manage resources and expenses, according to one person in attendance. It’s a start, but Mostaque’s meddling has left them with little runway to execute. His resignation, though, has given some employees hope. “A few people are 100% going to reconsider leaving after today,” said one current employee. “And the weird gloomy aura of hearing Emad talking nonsense for an hour is gone.” Shortly before Mostaque resigned, one current Stability executive told Forbes that they were optimistic his departure could make Stability appealing enough to receive a small investment or sale to a friendly party. “There are companies that have raised hundreds of millions of dollars that have much less intrinsic value than Stability,” the person said. “A white knight may still appear.”

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!
reddit
LLM Vibe Score0
Human Vibe Score1
regalalgorithmThis week

[N] Last Week in AI News Digest - Automated chemical synthesis, using heartbeats to detect deepfakes, and more!

Hi there, just sharing the latest edition of our AI news digest newsletter! We're just a couple of AI grad students doing this for fun, so hope the self promotion is not too annoying (also, welcome feedback). See it below, and feel free to subscribe. Mini Briefs Robotics, AI, and Cloud Computing Combine to Supercharge Chemical and Drug Synthesis IBM recently demoed a complex system for chemical testing and drug synthesis. The system has an AI component that predicts the results of chemical reactions, and a fully automated robotic experiment setup that runs chemical tests 24/7. Users can access the remote robotics lab online, and IBM can also install the system on-premise. With these tools working together, IBM is hoping to reduce typical drug discovery and verification time by half. AI researchers use heartbeat detection to identify deepfake videos Researchers from multiple groups are tackling the challenge of detecting deepfake videos by analyzing the apparent heartbeat of the people depicted in the video. This is possible, because a person’s blood flow changes their skin color ever so slightly, and this change is often detectable via a process called photoplethysmography (PPG). Because deepfakes are not currently optimizing to generate realisitic heartbeats, temporal or spatial anomalies in PPG signals allow resesarchers to detect deepfakes with a 97% accuracy. Advances & Business This AI Expert From Senegal Is Helping Showcase Africans In STEM \- Adji Bousso Dieng will be Princeton’s School of Engineering’s first Black female faculty. Google’s AI-powered flood alerts now cover all of India and parts of Bangladesh \- India, the world’s second most populated nation, sees more than 20% of the global flood-related fatalities each year as overrun riverbanks sweep tens of thousands of homes with them. Two years ago, Google volunteered to help. Finding magnetic eruptions in space with an AI assistant \- MMS look for explosive reconnection events as it flies through the magnetopause - the boundary region where Earth’s magnetic butts up against the solar wind that flows throughout the solar system. This know-it-all AI learns by reading the entire web nonstop \- Diffbot is building the biggest-ever knowledge graph by applying image recognition and natural-language processing to billions of web pages. Bosch and Ford will test autonomous parking in Detroit \- Ford, Bosch, and Dan Gilbert’s real estate firm Bedrock today detailed an autonomous parking pilot scheduled to launch in September at The Assembly, a mixed-used building in Detroit’s Corktown neighborhood. Create your own moody quarantine music with Google’s AI \- Lo-Fi Player, the latest project out of Google Magenta, lets you mix tunes with the help of machine learning by interacting with a virtual room. Apple launches AI/ML residency program to attract niche experts \- As Apple’s artificial language and machine learning initiatives continue to expand, its interest in attracting talent has grown - a theme that’s barely under the surface of the company’s occasionally updated Machine Learning Research blog. Dusty Robotics CEO Tessa Lau Discusses Robotics Start-Ups and Autonomous Robots for Construction \- Tessa Lau is Founder/CEO at Dusty Robotics, whose mission is to increase construction industry productivity by introducing robotic automation on the jobsite. Concerns & Hype Google Offers to Help Others With the Tricky Ethics of AI \- Companies pay cloud computing providers like Amazon, Microsoft, and Google big money to avoid operating their own digital infrastructure. The Peace Dividends Of The Autonomous Vehicle Wars \- The rapid growth of the mobile market in the late 2000s and early 2010s led to a burst of technological progress. Ethics must be part of the development process’ \- The increasing use of AI (artificial intelligence) in the development of new medical technologies demands greater attention to ethical aspects. Analysis & Policy China’s new AI trade rules could hamper a TikTok sale \- TikTok’s attempt to sell itself and avert a possible US ban may run into some complications. The Wall Street Journal reports that China has unveiled new restrictions on AI technology exports that could affect TikTok. Podcast Check out our weekly podcast covering these stories! Website | RSS | iTunes | Spotify | YouTube

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

If only someone told me this before my first startup
reddit
LLM Vibe Score0
Human Vibe Score0.625
johnrushxThis week

If only someone told me this before my first startup

If only someone told me this before my first startup: Validate idea first. I wasted a decade building stuff nobody needed. Incubators and VCs served to me as a validation, but I was so wrong. Kill my EGO. It’s not about me, but the user. I must want what the user wants, not what I want. My taste isn't important. The user has expectations, and I must fulfill them. Don’t chaise investors. Chase users, and then investors will be chasing me. I've never had more incoming interest from VC than now when I'm the least interested in them. Never hire managers. Only hire doers until PMF. So many people know how to manage people and so few can actually get sh\*t done barehand. Landing page is the least important thing in a startup. Pick a simple template, edit texts with a no-code website builder in less than an hour and that's it! At the early stage, I win traffic outside of my website, people are already interested, so don't make them search for the signup button among the texts! Focus on conversion optimization only when the traffic is consistent. Keep it to one page. Nobody gonna browse this website. Hire only fullstack devs. There is nothing less productive in this world than a team of developers for an early-stage product. One full stack dev building the whole product. That’s it. Chase global market from day 1. If the product and marketing are good, it will work on the global market too, if it’s bad, it won’t work on the local market too. So better go global from day 1, so that if it works, the upside is 100x bigger. I launched all startups for the Norwegian market, hoping we will scale to international at some point. I wish I launched to international from day 1 as I do now. The size of the market is 10000x bigger. I can validate and grow products in days, not in years as it used to be. Do SEO from day 2. As early as I can. I ignored this for 14 years. It’s my biggest regret. It takes just 5 minutes to get it done on my landing page. I go to Google Keyword Planner, enter a few keywords around my product, sort them by traffic, filter out high competition kws, pick the top 10, and place them natively on my home page and meta tags. Add one blog article every week. Either manually or by paying for an AI blogging tool. Sell features, before building them. Ask existing users if they want this feature. I run DMs with 10-20 users every day, where I chat about all my ideas and features I wanna add. I clearly see what resonates with me most and only go build those. If I don't have followers, try HN, Reddit, or just search on X for posts and ask it in the replies. People are helpful, they will reply if the question is easy to understand. Hire only people I would wanna hug. My cofounder, an old Danish man said this to me in 2015. And it was a big shift. I realized that if I don’t wanna hug the person, it means I dislike them on a chemical/animal level. Even if I can’t say why, but that’s the fact. Sooner or later, we would have a conflict and eventually break up. It takes up to 10 years to build a startup, make sure I do it with people I have this connection with. Invest all money into my startups and friends. Not crypt0, not stockmarket, not properties. I did some math, if I kept investing all my money into all my friends’ startups, that would be about 70 investments. 3 of them turned into unicorns eventually. Even 1 would have made the bank. Since 2022, I have invested all my money into my products, friends, and network. If I don't have friends who do startups, invest it in myself. Post on Twitter daily. I started posting here in March last year. It’s my primary source of new connections and growth. I could have started it earlier, I don't know why I didn't. Don’t work/partner with corporates. Corporations always seem like an amazing opportunity. They’re big and rich, they promise huge stuff, millions of users, etc. But every single time none of this happens. Because I talk to a regular employees there. They waste my time, destroy focus, shift priorities, and eventually bring in no users/money. Don’t get ever distracted by hype e.g. crypt0. I lost 1.5 years of my life this way. I met the worst people along the way. Fricks, scammers, thieves. Some of my close friends turned into thieves along the way, just because it was so common in that space. I wish this didn’t happen to me. I wish I was stronger and stayed on my mission. Don’t build consumer apps. Only b2b. Consumer apps are so hard, like a lottery. It’s just 0.00001% who make it big. The rest don’t. Even if I got many users, then there is a monetization challenge. I’ve spent 4 years in consumer apps and regret it. Don’t hold on bad project for too long, max 1 year. Some projects just don’t work. In most cases, it’s either the idea that’s so wrong that I can’t even pivot it or it’s a team that is good one by one but can’t make it as a team. Don’t drag this out for years. Tech conferences are a waste of time. They cost money, take energy, and time and I never really meet anyone there. Most people there are the “good” employees of corporations who were sent there as a perk for being loyal to the corporation. Very few fellow makers. Scrum is a Scam. For small teams and bootstrapped teams. If I had a team that had to be nagged every morning with questions as if they were children in kindergarten, then things would eventually fail. The only good stuff I managed to do happened with people who were grownups and could manage their stuff on their own. We would just do everything over chat as a sync on goals and plans. Outsource nothing at all until PMF. In a startup, almost everything needs to be done in a slightly different way, more creative, and more integrated into the vision. When outsourcing, the external members get no love and no case for the product. It’s just yet another assignment in their boring job. Instead of coming up with great ideas for my project they will be just focusing on ramping up their skills to get a promotion or a better job offer. Bootstrap. I spent way too much time raising money. I raised more than 10 times, preseed, seeded, and series A. But each time it was a 3-9 month project, meetings every week, and lots of destruction. I could afford to bootstrap, but I still went the VC-funded way, I don’t know why. To be honest, I didn’t know bootstrapping was a thing I could do or anyone does. It may take a decade. When I was 20, I was convinced it takes a few years to build and succeed with a startup. So I kept pushing my plans forward, to do it once I exited. Family, kids. I wish I married earlier. I wish I had kids earlier. No Free Tier. I'd launch a tool with a free tier, and it'd get sign-ups, but very few would convert. I'd treat free sign-ups as KPIs and run on it for years. I'd brag about signups and visitors. I'd even raise VC money with these stats. But eventually, I would fail to reach PMF. Because my main feedback would come from free users and the product turned into a perfect free product. Once I switched to "paid only" until I validated the product, things went really well. Free and paid users often need different products. Don't fall into this trap as I did. Being To Cheap. I always started by checking all competitors and setting the lowest price. I thought this would be one of the key advantages of my product. But no, I was wrong. The audience on $5 and $50 are totally different. $5: pain in the \*ss, never happy, never recommend me to a friend, leave in 4 months. $50: polite, give genuine feedback, happy, share with friends, become my big fan if I solve their request. I will fail. When I started my first startup. I thought if I did everything right, it would work out. But it turned out that almost every startup fails. I wish I knew that and I tried to fail faster, to get to the second iteration, then to the third, and keep going on, until I either find out nothing works or make it work. Use boilerplates. I wasted years of dev time and millions of VC money to pay for basic things. To build yet another sidebar, yet another dashboard, and payment integration... I had too much pride, I couldn't see myself taking someone else code as a basis for my product. I wanted it to be 100% mine, original, from scratch. Because my product seems special to me. Spend more time with Family & Friends. I missed the weddings of all my best friends and family. I was so busy. I thought if I didn't do it on time, the world would end. Looking back today, it was so wrong. I meet my friends and can't share those memories with them, which makes me very sad. I realized now, that spending 10% of my time with family and friends would practically make no negative impact on my startups. Build Products For Audiences I Love. I never thought of this. I'd often build products either for corporates, consumers, or for developers. It turns out I have no love for all 3. But I deeply love indie founders. Because they are risk-takers and partly kids in their hearts. Once I switched the focus to indie makers on my products, my level of joy increased by 100x for me. Ignore Badges and Awards I was chasing those awards just like everyone else. Going to ceremonies, signing up for events and stuff. I've won tons of awards, but none of those were eventually useful to my business. I better focused on my business and users. Write Every Single Day. When I was a kid, I loved writing stories. In school, they would give an assignment, and I'd often write a long story for it, however, the teacher would put an F on it. The reason was simple, I had an issue with the direction of the letters and the sequence of letters in the words. I still have it, it's just the Grammarly app helping me to correct these issues. So the teacher would fail my stories because almost every sentence had a spelling mistake that I couldn't even see. It made me think I'm made at writing. So I stopped, for 15 years. But I kept telling stories all these years. Recently I realized that in any group, the setup ends up turning into me telling stories to everyone. So I tried it all again, here on X 10 months ago. I love it, the process, the feedback from people. I write every day. I wish I had done it all these years. The End. \ this is an updated version of my post on the same topic from 2 months ago. I've edited some of the points and added 9 new ones.* \\ This is not advice, it's my self-reflection that might help you avoid same mistakes if you think those were mistakes

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

Turning a Social Media Agency into $1.5 Million in Revenue
reddit
LLM Vibe Score0
Human Vibe Score1
FounderFolksThis week

Turning a Social Media Agency into $1.5 Million in Revenue

Steffie here from Founder Folks, with a recent interview I did with Jason Yormark from Socialistics. Here is his story how he started and grew his social media agency. Name: Jason Yormark Company: Socialistics Employee Size: 10 Revenue: $1,500,000/year Year Founded: 2018 Website: www.socialistics.com Technology Tools: ClickUp, Slack, KumoSpace, Google Workspace, Shift, Zapier, Klayvio, Zoom, Gusto, Calendly, Pipedrive Introduction: I am the founder of Socialistics (www.socialistics.com), a leading social media agency that helps businesses turn their social media efforts into real measurable results. I am a 20+ year marketing veteran whose prior work has included launching and managing social media efforts for Microsoft Advertising, Office for Mac, the Air Force, and Habitat for Humanity. I have been recognized as a top B2B social media influencer and thought leader on multiple lists and publications including Forbes, ranking #30 on their 2012 list. I've recently published the book Anti-Agency: A Realistic Path to a $1,000,000 Business, and host the Anti Agency podcast where I share stories of doing business differently. You can learn more about me at www.jasonyormark.com. The Inspiration To Become An Entrepreneur: I’ve been involved with social media marketing since 2007, and have pretty much carved my career out of that. It was a natural progression for me to transition into starting a social media agency. From Idea to Reality: For me realistically, I had to side hustle something long enough to build it up to a point that I could take the leap and risks going full time on my own. For these reasons, I built the company and brand on the side putting out content regularly, and taking on side hustle projects to build out my portfolio and reputation. This went on for about 18 months at which point I had reached the breaking point of my frustrations of working for someone else, and felt I was ready to take the leap since I had the wheels in motion. While balancing a full-time job, I made sure not to overdo it. My main focus was on building out the website/brand and putting out content regularly to gain some traction and work towards some search visibility. I only took on 1-2 clients at a time to make sure I could still meet their needs while balancing a full time job. Attracting Customers: Initially I tapped into my existing network to get my first few clients. Then it was a mix of trade shows, networking events, and throwing a bit of money at paid directories and paid media. This is really a long game. You have to plant seeds over time with people and nurture those relationships over time. A combination of being helpful, likable and a good resource for folks will position you to make asks in the future. If people respect and like you, it makes it much easier to approach for opportunities when the time comes. Overcoming Challenges in Starting the Business: Plenty. Learning when to say no, only hiring the very best, and ultimately the realization that owning a marketing agency is going to have hills and valleys no matter what you do. Costs and Revenue: My largest expense by FAR is personnel, comprising between 50-60% of the business’ expenses, and justifiably so. It’s a people business. Our revenue doubled from the years 2018 through 2021, and we’ve seen between 10-20% growth year over year. A Day in the Life: I’ve successfully removed myself from the day to day of the business and that’s by design. I have a tremendous team, and a rock start Director of Operations who runs the agency day to day. It frees me up to pursue other opportunities, and to mentor, speak and write more. It also allows me to evangelize the book I wrote detailing my journey to a $1M business titled: Anti-Agency: A Realistic Path To A $1,000,000 Business (www.antiagencybook.com). Staying Ahead in a Changing Landscape: You really have to stay on top of technology trends. AI is a huge impact on marketing these days, so making sure we are up to speed on that, and not abusing it or relying on it too much. You also have to embrace that technology and not hide the fact that it’s used. Non-marketers still don’t and can’t do the work regardless of how much AI can help, so we just need to be transparent and smart on how we integrate it, but the fact is, technology will never replace creativity. As an agency, it’s imperative that we operationally allow our account managers to have bandwidth to be creative for clients all the time. It’s how we keep clients and buck the trend of companies changing agencies every year or two. The Vision for Socialistics: Continuing to evolve to cater to our clients through learning, education, and staying on top of the latest tools and technologies. Attracting bigger and more exciting clients, and providing life changing employment opportunities.

Raised $450k for my startup, here are the lessons I've learned along the way
reddit
LLM Vibe Score0
Human Vibe Score1
marin_smiljanicThis week

Raised $450k for my startup, here are the lessons I've learned along the way

2021 has been a pretty amazing year for Omnisearch. Having started initial work on Omnisearch at the end of 2020, we entered the new year with a working MVP yet no revenue, no significant partnerships, and no funding. Fast forward to the end of 2021, and we now have fantastic revenue growth, a partnership with a public company, and a far more powerful, complete and polished product. But one milestone really changed Omnisearch’s trajectory: our $450,000 USD pre-seed round by GoAhead Ventures. In this post I want to share the story of how it came about and offer a couple of takeaways to keep in mind when preparing for fundraising. &#x200B; The story Contrary to most advice, my co-founder Matej and I didn’t allocate a specific time to switch to “fundraising mode” but rather talked to investors on an ongoing basis. It was a bit of a distraction from working on the product, but on the positive side we were able to constantly get feedback on the idea, pitch, go-to-market strategy and hiring, as well as hearing investors’ major concerns sooner rather than later. That being said, our six-month long fundraising efforts weren’t yielding results - we talked to about twenty investors, mostly angels or smaller funds, with no success. The feedback was generally of the “too early for us” variety (since we were still pre-revenue), with additional questions about our go-to-market strategy and ideal customer persona. The introduction to our eventual investors, California-based GoAhead Ventures, came through a friend who had pitched them previously. We wrote a simple blurb and sent our pitch deck. We then went through GoAhead’s hyper-efficient screening process, consisting of a 30-minute call, a recorded three-minute pitch, and filling out a simple Google doc. Throughout the whole process, the GoAhead team left an awesome impression thanks to their knowledge of enterprise software and their responsiveness. They ended up investing and the whole deal was closed within two weeks, which is super fast even by Silicon Valley standards. While our fundraising experience is a single data point and your case might be different, here are the key takeaways from our journey. &#x200B; Perseverance wins: Like I said above, we talked to about twenty investors before we closed our round. Getting a series of “no”s sucks, but we took the feedback seriously and tried to prepare better for questions that caught us off guard. But we persevered, keeping in mind that from a bird’s eye perspective it’s an amazing time to be building startups and raising funds. Focus on traction: Sounds pretty obvious, right? The truth is, though, that even a small amount of revenue is infinitely better than none at all. One of the major differences between our eventual successful investor pitch and the earlier ones was that we had actual paying customers, though our MRR was low. This allows you to talk about customers in the present tense, showing there’s actual demand for your product and making the use cases more tangible. And ideally, highlight a couple of customer testimonials to boost your credibility. Have a demo ready: In Omnisearch’s case, the demo was oftentimes the best received part of the pitch or call. We’d show investors the live demo, and for bonus points even asked them to choose a video from YouTube and then try searching through it. This always had a “wow” effect on prospective investors and made the subsequent conversation more exciting and positive. Accelerators: Accelerators like Y Combinator or Techstars can add enormous value to a startup, especially in the early stages. And while it’s a great idea to apply, don’t rely on them too heavily. Applications happen only a few times a year, and you should have a foolproof fundraising plan in case you don’t get in. In our case, we just constantly looked for investors who were interested in our space (defined as enterprise SaaS more broadly), using LinkedIn, AngelList, and intros from our own network. Practice the pitch ad nauseam: Pitching is tough to get right even for seasoned pros, so it pays to practice as often as possible. We took every opportunity to perfect the pitch: attending meetups and giving the thirty-second elevator pitch to other attendees over beer and pizza, participating in startup competitions, going to conferences and exhibiting at our own booth, attending pre-accelerator programs, and pitching to friends who are in the startup world. Show an understanding of the competition: Frankly, this was one of the strongest parts of our pitch and investor conversations. If you’re in a similar space to ours, Gartner Magic Quadrants and Forrester Waves are an awesome resource, as well as sites like AlternativeTo or Capterra and G2. By thoroughly studying these resources we gained a great understanding of the industry landscape and were able to articulate our differentiation more clearly and succinctly. Presenting this visually in a coordinate system or a feature grid is, from our experience, even more effective. Remember it’s just the beginning! Getting your first round of funding is just the beginning of the journey, so it’s important to avoid euphoria and get back to building and selling the product as soon as possible. While securing funding enables you to scale the team, and is a particular relief if the founders had worked without a salary, the end goal is still to build a big, profitable, and overall awesome startup.

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!
reddit
LLM Vibe Score0
Human Vibe Score0.973
lazymentorsThis week

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!

Starting to focus on marketing for your business, You will come across the same tools mentioned over and over by marketers. I would like to mention here tools that you might haven’t seen going viral in the community but actually will help you grow faster and efficiently. Starting off with My favourite Marketing Channel! #Email Marketing For SMBs Convertkit / Mailerlite / Mailchimp - These 3 Platforms are the best options for SMBs and entrepreneurs just starting out with email marketing. All 3 have free plans up to 1,000 subscribers. Scribe - Email Signature Tool, Create Great Email signatures for your emails. Liramail - Most Email marketing platforms don’t offer great email templates. This tool will help you build great email templates with drag and drop. Quick mail Auto-Warmer - Most Businesses at the beginning don’t know what to do when open rate drops. You need to use an email warmer like this to keep it up. #Email Marketing For Big Businesses SendGrid - Overall Email Marketing Tools, this tool is best for brands that have huge email lists and email marketing is the key marketing channel. Braze - This tool is leading in email marketing for large Email senders. When I was working for agencies, this was one of the best email marketing tools I had used. NeoCertified - Protect your emails for spammers and threats. To keep your email list healthy, this is a must have! Sparkloop - Referral Marketing For Email Campaigns. Email can generate great huge amount of referrals for you and Sparkloop makes it easier. #Cold Emails & Lead Generation Hunter - A Great Tool to scrape emails from domain names. The tool comes with a green free plan but Pro plan is worth the amount of features it provides. Icyleads - It’s better than Hunter as it’s heavily focused on the sales and prospecting to help you derive great results from your campaigns. Mailshake - Beginner Friend Cold Email Tool with Great features like email list warming. #Communication Tools Twilio - One do the best customer engagement platform used by Companies like Stripe and mine too. Chatlio - Use Live chat feature on your website with slack integration. My favourite easier to catch up on conversations through slack integration. Intercom - Used by Most Marketers, Industry Leading customer communication platform. Great for beginners! Chatwoot - Another Amazing Communication Tool but the best part is they have a great free plan useful for new businesses. Loom - Communicate with your audience through Videos. Loom is great for SaaS and to show human interaction to close new visitors effectively. #CRM Outseta - This tool provides great CRM and their billing system is better than other tools out their which makes it stands out! Hubspot - I don’t think this tool needs an introduction because Hubspot’s CRM is the best in industry. Salesflare - This CRM is a great alternative to hubspot as it’s beginner friendly and helpful for SMBs. #SEO Tools Ahrefs - One of the best SEO tool in the industry. They also just launched a bunch of free tools to help SEO beginners. Screaming frog - The only website crawler I have used since I bought my first domain. It’s the best! Ubersuggest- The Tool by Neil Patel is the best SEO tool for you. (I’m Joking, it’s the worst) Contentking - This tool is good at Real-time SEO Auditing, they do a lot of Marketing work through Newsletters. If you are subscribed to any SEO newsletter. You may have seen this tool. SEOquake & Semrush - SEOquake is a great tool to conduct on-page analysis, SERP, and much more. Great tool but it’s owned by Semrush. You should go for Semrush because that tool will cover all SEO aspects for you. #Content Marketing Buzzsumo - This tool is great for content research and but you may find the regular emails pretty annoying sometimes. Contentrow - Analyse Your Content and find it’s strength. Highly recommended who are weak at content structuring like me. Grammarly - If you are not a native English speaker like me, you might think you need it or not. You need it for sure for grammar corrections. #Graphic Design Tools Visme - At agencies, Infographics can be more effective than usual postscript. Visme is a graphic design tool focused on infographics and designs related to B2B and B2C. It’s great for agencies! Glorify - A Graphic Design Tool focused on E-commerce, filled with Designs useful for E-commerce store owners. Canva - All-in-one Industry leading Graphic Design Tool that everyone knows and every template is overused now. Adobe Creative Cloud ( previously Sparkpost) - It’s a great alternative to Canva filled with Amazing Stock images to use in your visuals but the only backlash is the exports in this tool are not high quality. Snaps - A Canva Alternative that might not have overused templates for your Social Accounts. #Advertising Tools Plai - It’s a great PPC tool to create Ads for Instagram and Tiktok. Wordstream - It’s an industry leading PPC Tool, great for Ad Grading and auditing. AdEspresso - This Is a tool by Hootsuite. They have a lot of Data sourced at the backend, which helps in Ad optimisation through this tool. That’s the reason I recommend this tool. #Video Editing Tools Veed Studio - I have been using Veed from last year. It’s one of the best Video Marketing Tool Optimized for Instagram & Tiktok. Synthesia - It’s a new AI video generation platform. From last few months, if you have seen marketing agencies including Videos in Emails. The chances are that’s not a Agency member taking but AI generated Human. Motionbox - It’s also a great video editing tool focused on video editing for Digital Marketers. Jitter Video - It’s a great motion design tool. Comes with great templates, the only place where other tools I mentioned lacks. It’s great and beginner friendly. #Copywriting Jasper AI - Google’s John Mueller says AI generated content is banned on Search but I think with Jasper AI you can generate SEO optimised Content but you have to put in some efforts like at least give 30 minutes for editing the Copy by yourself. Copy AI - Another AI tool to help you write better copy. This one is more focused on helping you write copy suitable for Ads and Social media campaigns. Hemingway App - To help you write more clearly and Bold. This tool is better than Grammarly if you look for writing perspective and it’s free. #Social Media Management App I’ve used a Lot of SMM Tools and that’s why going to mention all of them with a short review. Sprout social - The Best with deep insights coverage. Hootsuite - Great Scheduling tool just under sprout social. Later - Heavily Focused on Instagram from beginning and Now Tiktok too. SkedSocial - It’s like a Later alternative with great addition features like link-in-bio. Facebook’s Business Manager- Great but sometimes bugs can make a huge issue for you and customer support is like dead. Tweet Hunter & Hypefury- Both are Twitter Scheduling tools growing very fast on platform and are great for growth. Buffer - It’s a great tool but I haven’t seen any new updates to help with management. Zoho Social - It’s a great SMM tool and if you use other marketing solutions from Zoho. It’s a must have! #Market Research Tool • SparkToro - That’s the only one I have ever used. It’s great for audience research and comes with great customer service. Founded by Rand Fishkin, it’s one of the best research tool. #Influencer Marketing & UGC InfluenceGrid - A free search engine To find Tiktok & Instagram Influencers for your campaigns. Tiktok Creative Center- TikTok’s in-built tool called “Creative Center” is the best to find content trends, audience demographics and much more. Archive - Find Instagram Stories and Posts mentioning Your brands and use them as Ads for your business Marketing. #Landing Page Builders Leadpages - Its a great landing page builder because the integration and drag-and-drop features makes it easier to work with! Cardd co - A Great Landing page builder with easy step up but it lacks the copywriting and tracking features. Instapage - It’s one of the best out and I think the overall product is effective enough to help you stand out with your landing page. Unbounce - It’s a great alternative to Instapage due its well polished landing page templates that might be helpful for you. #Community Building Mighty Networks - A Great Community building platform, and you can also sell courses within the platform. Circle so - A great alternative to Mighty networks focused on Communities specifically. We are currently using for small community Of ours. #Sales Tools Drift - You can get much more out of Drift than just sales tools but The Sales solutions provided in Drift are one of the best. Salesforce - It’s the industry Sales solution provider. A go-to and have various pricing plans making it suitable for majority of SMBs. #Social Proof Tools People don’t have enough time to search across internet to decide to trust you after seeing your Ad first time. That’s what you might be facing too. Here are two tools I absolutely love for social proof! Use Proof - Show Recent Activities occurring on your website and build the trust of your visitors. Testimonial to - Gather Testimonials across Social Media platforms related to your business with this tool. Capture tweets and comments mentioning your brands and mention them. #Analytics Tools Plausible Analytics- A privacy friendly Analytics alternative to Google Analytics if you hate Analytics 4 like me. Mixpanel - Product Analytics and funnel reports better than Google Analytics. #Reddit Marketing Gummysearch- This tool will help To find your target audience on Reddit and interact with them with its help and close your new customers. Howitzer- It’s another pretty similar tool to Gummysearch focused on Reddit cold outreach to get clients and new customers. Both are great but Gummysearch provides better customer support while Howtizer is helpful on a large scale Reddit Marketing. #Text Marketing Klaviyo - It’s an email + SMS marketing tool, it’s taking up space in marketing industry very quickly as an industry leader due to its great integrations but you need to learn the platform usage to maximise the outcome. Cartloop - This tool provides great text marketing solutions with integration with Spotify and other e-commerce marketing tools. Attentive Mobile - This is my favourite Text marketing tool due to the interactive dashboard + they have a library of Text marketing examples to help you out with your campaigns. #Other Tools I have used throughout my journey! Triple Whale - It’s a great E-commerce marketing tools with Triple pixel to help you track your campaigns more efficiently. Fastory - To create well optimized Instagram & Tiktok Stories for your business. Jotform - Online Form Builder with integrations with leading marketing tools. Gated - As an entrepreneur and marketer, you may receive a bunch of unwanted emails. Use Gated to get rid of them and receive useful mails only! ClickUp- The main Tool for Project Management, one of the best and highly recommended. Riverside - Forget Zoom or Google Meet, For your Podcast Interviews and Marketing conferences. You need riverside with great video quality and recording features. Manychat- Automate your Instagram DMs and interact with your followers more efficiently + sell out your products/ services when you are offline. Calendy - To schedule meetings with your ideal clients. ServiceProviderPro - It’s a client portal for SEO & Growing Agencies, very helpful in scaling agencies. SendCheckit - Compare your Email Subject Lines with 100,000+ others in the database for free. Otter AI - Using AI track your meetings more effectively, you can easily edit, annotate and share notes from the meetings. Ryte - Optimise your website User experience with this tool focused on UX aspects + SEO too. PhantomBuster - Scrape LinkedIn Profile and Data from Facebook/LinkedIn groups. I clearly love this tool! #Honourable Mentions Zapier - The Only tool you need to integrate your favourite tool with a new effective tool. Elementor - That’s what I use for web design and it’s great! Marketer Hire - To hire world class marketers to work with you. InShot & Capcut - I create Instagram Reels and TikTok’s and life without these tools isn’t possible. Nira - It’s a great tool to Manage your workspace and this tool has launched many marketing templates in-built helpful for marketers and also entrepreneurs. X - The tool you love that wasn’t mentioned here is valuable and I honour that tool and share that if you would like to! I mean thanks for reading what I have curated all over my life as a marketer. I share 5 Marketing Tools, 5 Marketing Resources and 1 Free Resourceevery week in my newsletter, you can subscribe here to receive that for free. Also, You can read an expanded list of email marketing tools in this Reddit post!

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING
reddit
LLM Vibe Score0
Human Vibe Score1
Millionaire_This week

In 2018, I started an AI chatbot company...today, we have over 4000 paying customers and ChatGPT is changing EVERYTHING

Intro: 5 years ago, my co-founders and I ventured into the space of AI chatbots and started our first truly successful company. Never in a million years did I see myself in this business and we truly stumbled upon the opportunity by chance. Prior to that, we ran a successful lead generation business and questioned whether a simple ai chat product would increase our online conversions. Of the 3 co-founders, I was skeptical that it would, but the data was clear that we had something that really worked. We built a really simple MVP version of the product and gave it to some of our top lead buyers who saw even better conversion improvements on their own websites. In just a matter of weeks, a new business opportunity was born and a major pivot away from our lead generation business started. Our growth story: Startup growth is really interesting and in most cases, founders aren't really educated on what a typical growth curve looks like. While we hear about "hockey stick" growth curves, it's really atypical to actually see or experience this. From my experience, growth curves take place in a "stair curve". For example, you can scrap your way to a $100k run rate without much process or tracking. You can even get to $1 million ARR being super disorganized. As you start going beyond $1M ARR, things start to break and growth can flatten out while you put new processes and systems in place. Eventually you'll get to $2M or 3M with your new strategy and then things start breaking again. I've seen the process repeat itself and as you increase your ARR, the processes and systems become more difficult to work through...mainly because more people get involved and the product becomes more complex. When you do end up cracking the code in each step, the growth accelerates faster and faster before things start to break down and flatten out again. Without getting too much into the numbers, here were some of our initial levers for growth: Our first "stair" step was to leverage our existing customer base from our prior lead generation business. Having prior business relationships and a proven track record made it really simple to have conversations with people who already trusted us to try something new that we had to offer. Stair #2 was to build out a partner channel. Since our chat product involved a web developer or agency installing the chat on client sites, we partnered with these developers and agencies to leverage their already existing customer bases. We essentially piggy-backed off of their relationships and gave them a cut of the revenue. We built an internal partner tracking portal which took 6+ months, but it was well worth it. Stair #3 was our most expensive step, biggest headache, but added the most revenue. After COVID, we had and SDR/Account Executive sales team of roughly 30 people. It added revenue fast, but the payback periods were 12+ months so we had to cut back on this strategy after exhausting our universe of clients. Stair #4 involves a variety of paid advertisement strategies with product changes and the introduction of new onboarding features. We're in the middle of this stair and hope it's multiple years before things breakdown again. Don't give up I know it sounds really cliché, but the #1 indicator of success is doing the really boring stuff day in and day out and making incremental improvements. As the weeks, months, and years pass by, you will slowly gain domain expertise and start to see the gaps in the market that can set you apart from your competition. It's so hard for founders to stay focused and not get distracted so I would say it's equally as important to have co-founders who hold each other accountable on what your collective goals are. How GPT is changing everything I could write pages and pages about how GPT is going to change how the world operates, but I'll keep it specific to our business and chatbots. In 2021, we built an industry specific AI model that did a great job of classifying intents which allowed us to train future actions during a chat. It was a great advancement in our customer's industry at the time. With GPT integrated into our system, that training process that would take an employee hours to do, can be done in 5 minutes. The model is also cheaper than our own and more accurate. Because of these training improvements, we have been able to conduct research that is allowing us to leverage GPT models like no one else in the industry. This is both in the realm of chat and also training during onboarding. I really want to refrain from sharing our company, but if you are interested in seeing a model trained for your specific company or website, just PM me your link and I'll send you a free testing link with a model fully trained for your site to play around with. Where we are headed and the dangers of AI The level of advancement in AI is not terribly dangerous in its current state. I'm sure you've heard it before, but those who leverage the technology today will be the ones who get ahead. In the coming years, AI will inevitably replace a large percentage of human labor. This will be great for overall value creation and productivity for the world, but the argument that humans have always adapted and new jobs will be created is sadly not going to be as relevant in this case. As the possibility of AGI becomes a reality in the coming years or decades, productivity through AI will be off the charts. There is a major risk that human innovation and creative thinking will be completely stalled...human potential as we know it will be capped off and there will need to be major economic reform for displaced workers. This may not happen in the next 5 or 10 years, but you would be naïve not to believe the world we live in today will not be completely different in 20 to 30 years. Using AI to create deepfakes, fake voice agents, scam the unsuspecting, or exploit technical vulnerabilities are just a few other examples I could write about, but don't want to go into to much detail for obvious reasons. Concluding If you found the post interesting or you have any questions, please don't hesitate to ask. I'll do my best to answer whatever questions come from this! &#x200B; \*EDIT: Wasn't expecting this sort of response. I posted this right before I went to sleep so I'll get to responding soon.

Follow Along as I Flip this Website - Case Study
reddit
LLM Vibe Score0
Human Vibe Score1
jshogren10This week

Follow Along as I Flip this Website - Case Study

I am starting a new case study where I will be documenting my attempt to flip a website that I just purchased from Flippa. However, unlike most case studies where people hide certain parts and details from the public I will instead be sharing everything. That means you will know the exact URL of the site that I purchased and I will share everything with you all as I progress.I know that case studies are lot more interesting and you can learn better when you can see real examples of what I am talking about. Enough of the chatting, let's jump straight into this new case study and I will explain what this is all about. Before you get into the case study I want to give you the option of reading this one my website where all of the images can be seen within the post and it is easier to read. I also want to say that I have nothing to sell you or anything close to it. So if you want to read it there you can do so here ##Introductory Video I have put together a video that talks about many of the things that I cover in this article. So if you would rather watch a video you can watch that here - https://www.youtube.com/watch?v=EE3SxtNnqts However, I go into more detail in the actual article FYI. Also, I plan on using Youtube very frequently in this case study so be on the lookout for new videos.There is going to be a video that will accompany every single case study post because I like having it being presented in two different mediums. ##The Website I Just Bought Around a week ago I made a new website purchase from Flippa and you can view the website's Flippa listing here - https://flippa.com/6439965-hvactraining101-com Screenshot of the Homepage - http://imgur.com/T6Iv1QN I paid $1,250 for the site and you will soon see that I got a really good deal. As you might be able to tell from the URL, this site is focused around training and education for becoming a HVAC technician. This is a lucrative niche to be in and Adsense pays very well. I do not have control of the site yet due to the transfer process not being completed. However, I am hoping within a few days everything will be finalized and I will take full control of the site. In the meantime, I figured it would be a good time to put together the introduction post for this new case study! ##Why I Bought this Website Now that you have a general idea of the website that I purchased, I now want to explain the reasoning behind the purchase. There are 3 major reasons for this purchase and I will explain each one of them below. GREAT Price As I mentioned earlier, I bought this website for $1,250. However, that doesn't mean a whole lot unless you know how much the site is making each month. Screenshot of the earnings for the last 12 months - http://imgur.com/NptxCHy Average Monthly Profits: 3 Month = $126 6 Month = $128 12 Month = $229.50 Let's use the 6 month average of $128/month as our baseline average. Since it is making on average $128/month and it was sold for $1,250 then that means I bought this site at a multiple of 9.76x! Most sites in today's market go for 20x-30x multiples. As you can see, I got a great deal on this site. Although the great price was the biggest reason for me buying this site there are other factors that persuaded me as well. You need to remember that just because you can get a website for a good price it doesn't mean it is a good deal. There are other factors that you need to look at as well. Extremely Under Optimized This site is currently being monetized mainly by Adsense and a very small amount from Quinstreet. From my experience with testing and optimizing Adsense layouts for my site in my Website Investing case study I know the common ad layouts that work best for maximizing Adsense revenue. With that being said, I can quickly determine if a website is being under optimized in terms of the ad layout. One of the first things I did when analyzing this site was examine the ad layout it was using. Screenshot of the website with the ad layout the previous owner was using - http://imgur.com/wqleLVA There is only ONE ad per page being used, that's it. Google allows up to 6 total ads to be used per page and you can imagine how much money is being left on the table because of this. I am estimating that I can probably double the earnings for the site practically overnight once I add more ads to the site. Adding more ads in combination with my favorite Adsense plugin, AmpedSense, I will be able to easily boost the earnings for this site quickly. It is also worth mentioning how lucrative this niche is and how much advertisers are willing to spend on a per click basis. The average CPC for the top keywords this site is currently ranking for in Google - http://imgur.com/ifxiy8B Look at those average CPC numbers, they are insanely high! I could be making up to $25 per click for some of those keywords, which is so absurd to me. Combine these extremely high CPC with the fact that the site currently only has one ad per page and you can start to understand just how under optimized this site truly is. I also plan on utilizing other ad networks such as Quinstreet and Campus Explorer more as well. These two networks are targeted at the education niche which works very well with my site. I will be testing to see if these convert better than normal Adsense ads. Goldmine of Untapped Keywords One of the biggest opportunities I see for growing this site is to target local keywords related to HVAC training. As of right now, the site has only scratched the surface when it comes to trying to rank for state/city keywords. Currently there are only two pages on the entire website which go after local keywords, those two pages target Texas and Florida HVAC search terms. These two pages are two of the more popular pages in terms of total amount of traffic. See the screenshot of the Google Analytics - http://imgur.com/NB0xJ4G Two out of the top five most popular pages for the entire website are focused on local search terms. However, these are the ONLY two pages that target local search terms on the whole site! There are 48 other states, although there may not be search volume for all states, and countless cities that are not being targeted. Why do I think this is such a good opportunity? For a few reasons: Local keywords are a lot easier to rank for in Google than more general keywords This site has been able to rank for two states successfully already and it proves it is possible Traffic going to these local pages is WAY more targeted and will convert at a much higher rate, which means more commissions for me There are so many more states and cities that get a good amount of searches that I can target To give you an idea of the type of keywords these local pages rank for, you can see the top keywords that the Florida page is ranking for in Google: Top ranking keywords for the Florida page - http://imgur.com/j7uKzl2 As you can see these keywords don't get a ton of searches each month, but ranking 1st for a keyword getting 90 searches a month is better than being ranked 10th for a keyword getting 1,000 searches a month. I have started to do some keyword research for other states and I am liking what I am finding so far. Keywords that I have found which I will be targeting with future articles - http://imgur.com/8CCCCWU I will go into more detail about my keyword research in future articles, but I wanted to give you an idea of what my strategy will be! I also wanted to share why I am super excited about the future potential to grow this site by targeting local keywords. ##Risks Yes, there are many good things about this website, but there are always risks involved no matter what the investment is. The same thing goes for this site. Below are some of the risks that I currently see. HTML Site This website is a HTML site and I will need to transfer it to Wordpress ASAP. I have been doing some research on this process and it shouldn't be too hard to get this over to Wordpress. In doing so it will make adding content, managing the back end and just about everything else easier. Also, I am hoping that when I transfer it to Wordpress that it will become more optimized for Google which will increase keyword rankings. Declining Earnings Looking at the last 12 months of earnings you will notice a drop off from last year till now. Earnings from the last 12 months - http://imgur.com/WsotZsj In May of 2015 it looks like the site earned right around $500, which is much higher than the $128 that it is earning now. However, the last 7 or so months have been consistent which is a good sign. Even though the earnings are much lower now then they were a year ago it is good to know that this site has the potential to earn $500/month because it has done it before. Slightly Declining Traffic In the last 12 months the site's traffic has declined, however, it looks like it is picking back up. Traffic from the last 12 months - http://imgur.com/aiYZW9W The decline is nothing serious, but there is a drop on traffic. Let's take a look at the complete history of this site's traffic so we can get a better idea of what is going on here: Complete traffic history - http://imgur.com/tYmboVn The above screenshot is from 2012 all the way up to right now. In the grand scheme of things you can see that the traffic is still doing well and it looks like it is on the upswing now. Those three risks mentioned above are the three biggest risks with this site at this point. It is always good to note the risks and do everything you can to prevent them from causing a problem. ##My Growth Strategy Whenever I purchase a new site I always create an outline or plan on how I will grow the site. Right now, I have some basic ideas on how I will grow this site, but as I go on I will continue to change and optimize my strategies to be more effective. Below I have outlined my current plans to grow: Add more Adsense Ads The very first thing I will do once I get control of the site is add more ads per page. I am predicting that by just adding a few more ads per page I will be able to more than likely double the earnings. I will touch on exactly how I will be optimizing the ad layouts in future posts. Test other Ad Networks I will be doing a lot of testing and experimenting when it comes to the ad networks. I plan on trying out Adsense, Media.net, Quinstreet, Campus Explorer and finding the combination of those 4 which produces the most revenue. The Adsense and Media.net ads will perform well on the more general pages while Quinstreet and Campus Explorer ads will be geared towards the local search terms. There will probably be other ad networks I will try out but these are the four which I will be using right away. If you are aware of any other ad networks out there which are geared towards the education niche please let me know in the comments below! Target Local Keywords with new Content I have already touched on this, but I will starting to produce content targeting these local keywords ASAP. The sooner I add the content to the site the sooner it will start to rank and bring in traffic. I will not be writing my own content and instead I will be outsourcing all of it via Upwork. I will show you all how I go about outsourcing content production and you can see my process for doing that. ##Goals for this Website My goal for the website is to have it valued at $10,000+ within 12 months. Let's break down this larger goal into smaller chunks which will make achieving it easier and more attainable. Earnings - $500/month To get the site valued at $10,000 the site will need to be making $500/month using a 20x monthly multiple. Right now, the site is making around $130/month so it has a ways to before it reaches the $500 a month mark. However, after doing some Adsense optimization I think we could push the earnings to around $300/month without much work. From there, it will come down to trying to bring in more traffic! Traffic - 5,000 Visitors per Month Why 5,000 visitors? Because that is how much traffic it is going to take to get to the $500/month goal. Let me explain how I came to this conclusion: The average RPM for this site is currently $50, which means for every 1,000 page views the site earns $50. After I optimize the Adsense layout for the site and add more ads per page I think I will be able to double the RPM to $100. Using the RPM of $100 the site will need to have 5,000 monthly visitors to earn $500. So 5,000 monthly visitors is the traffic goal I have set and aiming for! The site is currently getting around 3,000 visitors per month so I will need to add an extra 2,000 visitors to get to this goal. ##Want to Follow this Case Study? I will be using Youtube a lot in this case study so make sure to follow my Youtube channel here - www.youtube.com/c/joshshogren Other than that, I think that is going to bring us to the end of the introductory post for this new case study. I hope that you enjoyed reading and that you are excited to follow along! If you have any suggestions to make this case study better PLEASE let me know in the comment below. I want to make this case study the best one I have done yet. Talk to you all in the comment section.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

I got fired due to automation — lessons learned. Two-month overview.
reddit
LLM Vibe Score0
Human Vibe Score1
WebsterPepsterThis week

I got fired due to automation — lessons learned. Two-month overview.

UPD: Guys, I'm not promoting myself as some of the redditors decided. That's why to deal with contradictions I'll do next things: make additional post with short review and description of the general tools and processes you could apply. help only those who have already written me. So I won't answer on new offers or DMs. As mentioned, damn robots have taken my job. PRE-HISTORY During Covid times, I found myself without my offline job, and since I was interested in marketing and SMM, I began searching for a job there. Completed free Google and Udemy courses and finally landed my first SMM manager position with a business owner. He had several projects so, finally, I started managing three Twitter accounts, two Facebook accs, two IGs, and one TikTok. I handled posting, content editing and responding routine, while freelancers usually took care of video creation for IG and TT. THE STORY ITSELF Things took a turn for the worse in April when my employer introduced ChatGPT and Midjourney, tools I was already using. The owner insisted on integrating them into the workflow, and my wages took a 20% hit. I thought I could roll with it, but it was just the beginning. By midsummer, the owner implemented second-layer AI tools like Visla, Pictory, and Woxo for video (bye freelancers, lol), as well as TweetHunter, Jasper, and Perplexity for content. Midjourney and Firefly joined for image generation. All together, my paycheck was slashed by 50%. Finally, at the end of October, my boss told me he automated stuff with Zapier, cutting costs that way. Additionally, he adopted MarketOwl, autoposting tool for Twitter, and SocialBee for Facebook. He stated that he didn’t need me, as by now he could manage the social media accounts himself. I feel so pissed then and even thought that there's no point in searching for similar jobs. HOW I SPENT TWO MONTHS Well, for the first two weeks, I did nothing but being miserable, drinking and staring at the wall. My gf said it's unbearable and threatened to leave if I not pull myself together. It was not the final push, but definitely made me rethink things. So I decided to learn more about the capabilities of these automation covers and eventually became an AI adviser for small businesses. It's ironic that now I sometimes earn money advising on how to optimize marketing, possibly contributing to other people's job loss. FINAL THOUGHTS I am fully aware of the instability of such a job and have invested my last savings in taking an online marketing course at Columbia to gain more marketing experience and got something more stable afterwards. Message for mods: I'm not promoting myself or anything mentioned here; just sharing the experience that someone might find helpful.

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)
reddit
LLM Vibe Score0
Human Vibe Score0.882
mr_t_forhireThis week

Started a content marketing agency 8 years ago - $0 to $7,863,052 (2025 update)

Hey friends, My name is Tyler and for the past 8 years, I’ve been documenting my experience building a content marketing agency called Optimist. Year 1 — 0 to $500k ARR Year 2 — $500k to $1MM ARR Year 3 — $1MM ARR to $1.5MM(ish) ARR Year 4 — $3,333,686 Revenue Year 5 — $4,539,659 Revenue Year 6 — $5,974,324 Revenue Year 7 - $6,815,503 Revenue (Edit: Seems like links are banned now. You can check my post history for all of my previous updates with lessons and learnings.) How Optimist Works First, an overview/recap of the Optimist business model: We operate as a “collective” of full time/professional freelancers Everyone aside from me is a contractor Entirely remote/distributed team We pay freelancers a flat fee for most work, working out to roughly $65-100/hour. Clients pay us a flat monthly fee for full-service content marketing (research, strategy, writing, editing, design/photography, reporting and analytics, targeted linkbuilding, and more)\ Packages range in price from \~$10-20k/mo \This is something we are revisiting now* The Financials In 2024, we posted $1,032,035.34 in revenue. This brings our lifetime revenue to $7,863,052. Here’s our monthly revenue from January 2017 to December of 2024. (Edit: Seems like I'm not allowed to link to the chart.) The good news: Revenue is up 23% YoY. EBITDA in Q4 trending up 1-2 points. We hosted our first retreat in 4 years, going to Ireland with about half the team. The bad news: Our revenue is still historically low. At $1MM for the year, we’re down about 33% from our previous years over $1.5MM. Revenue has been rocky. It doesn’t feel like we’ve really “recovered” from the bumps last year. The trend doesn’t really look great. Even though, anecdotally, it feels like we are moving in a good direction. EBITDA is still hovering at around 7%. Would love to get that closer to 20%. (For those who may ask: I’m calculating EBITDA after paying taxes and W2 portion of my income.) — Almost every year, my update starts the same way: This has been a year of growth and change. Both for my business—and me personally. 2024 was no different. I guess that tells you something about entrepreneurship. It’s a lot more like sailing a ship than driving a car. You’re constantly adapting, tides are shifting, and any blip of calm is usually just a moment before the next storm. As with past years, there’s a lot to unpack from the last 12 months. Here we go again. Everything is Burning In the last 2 years, everything has turned upside down in the world of content and SEO. Back in 2020, we made a big decision to re-position the agency. (See post history) We decided to narrow our focus to our most successful, profitable, and consistent segment of clients and re-work our entire operation to focus on serving them. We defined our ICP as: \~Series A ($10mm+ funding) with 6-12 months runway to scale organic as a channel Product-led company with “simple” sales cycle involving fewer stakeholders Demonstrable opportunity to use SEO to drive business growth Our services: Content focused on growing organic search (SEO) Full-service engagements that included research, planning, writing, design, reporting And our engagement structure: Engaged directly with an executive; ownership over strategy and day-to-day execution 1-2 points of contact or stakeholders Strategic partner that drives business growth (not a service vendor who makes content) Most importantly, we decided that we were no longer going to offer a broader range of content that we used to sell. That included everything from thought leadership content to case studies and ebooks. We doubled-down on “SEO content” for product-led SaaS companies. And this worked phenomenally for us. We started bringing on more clients than ever. We developed a lot of internal system and processes that helped us scale and take on more work than we’ve ever had and drive great outcomes for our ideal clients. But in 2023 and 2024, things started going awry. One big change, of course, was the rise of AI. Many companies and executives (and writers) feel that AI can write content just as well as an agency like ours. That made it a lot harder to sell a $10,000 per month engagement when they feel like the bulk of the work could be “done for free.” (Lots of thoughts on this if you want my opinions.) But it wasn’t just that. Google also started tinkering with their algorithm, introducing new features like AI Overviews, and generally changing the rules of the game. This created 3 big shifts in our world: The perceived value of content (especially “SEO content”) dropped dramatically in many people’s minds because of AI’s writing capabilities SEO became less predictable as a source of traffic and revenue It’s harder than ever for startups and smaller companies to rank for valuable keywords (let alone generate any meaningful traffic or revenue from them) The effect? The middle of the content market has hollowed out. People—like us—providing good, human-crafted content aimed on driving SEO growth saw a dramatic decline in demand. We felt it all year. Fewer and fewer leads. The leads we did see usually scoffed at our prices. They were indexing us against the cost of content mills and mass-produced AI articles. It was a time of soul-searching and looking for a way forward. I spent the first half of the year convinced that the only way to survive was to run toward the fire. We have to build our own AI workflows. We have to cut our rates internally. We have to get faster and cheaper to stay competitive with the agencies offering the same number of deliverables for a fraction of our rates. It’s the only way forward. But then I asked myself a question… Is this the game I actually want to play? As an entrepreneur, do I want to run a business where I’m competing mostly on price and efficiency rather than quality and value? Do I want to hop into a race toward cheaper and cheaper content? Do I want to help people chase a dwindling amount of organic traffic that’s shrinking in value? No. That’s not the game I want to play. That’s not a business I want to run. I don’t want to be in the content mill business. So I decided to turn the wheel—again. Repositioning Part II: Electric Boogaloo What do you do when the whole world shifts around you and the things that used to work aren’t working anymore? You pivot. You re-position the business and move in another direction. So that’s what we decided to do. Again. There was only one problem: I honestly wasn’t sure what opportunities existed in the content marketing industry outside of what we were already doing. We lived in a little echo chamber of startups and SEO. It felt like the whole market was on fire and I had fight through the smoke to find an escape hatch. So I started making calls. Good ol’ fashioned market research. I reached out to a few dozen marketing and content leaders at a bunch of different companies. I got on the phone and just asked lots of questions about their content programs, their goals, and their pain points. I wanted to understand what was happening in the market and how we could be valuable. And, luckily, this process really paid off. I learned a lot about the fragmentation happening across content and how views were shifting. I noticed key trends and how our old target market really wasn’t buying what we were selling. Startups and small companies are no longer willing to invest in an agency like ours. If they were doing content and SEO at all, they were focused entirely on using AI to scale output and minimize costs. VC money is still scarce and venture-backed companies are more focused on profitability than pure growth and raising another round. Larger companies (\~500+ employees) are doing more content than ever and drowning in content production. They want to focus on strategy but can barely tread water keeping up with content requests from sales, demand gen, the CEO, and everyone else. Many of the companies still investing in content are looking at channels and formats outside of SEO. Things like thought leadership, data reports, interview-driven content, and more. They see it as a way to stand out from the crowd of “bland SEO content.” Content needs are constantly in flux. They range from data reports and blog posts to product one-pagers. The idea of a fixed-scope retainer is a total mismatch for the needs of most companies. All of this led to the logical conclusion: We were talking to the wrong people about the wrong things\.\ Many companies came to one of two logical conclusions: SEO is a risky bet, so it’s gotta be a moonshot—super-low cost with a possibility for a big upside (i.e., use AI to crank out lots of content. If it works, great. If it doesn’t, then at least we aren’t out much money.) SEO is a risky bet, so we should diversify into other strategies and channels to drive growth (i.e., shift our budget from SEO and keyword-focused content to video, podcasts, thought leadership, social, etc) Unless we were going to lean into AI and dramatically cut our costs and rates, our old buyers weren’t interested. And the segment of the market that needs our help most are looking primarily for production support across a big range of content types. They’re not looking for a team to run a full-blown program focused entirely on SEO. So we had to go back to the drawing board. I’ve written before about our basic approach to repositioning the business. But, ultimately it comes down to identifying our unique strengths as a team and then connecting them to needs in the market. After reviewing the insights from my discussions and taking another hard look at our business and our strengths, I decided on a new direction: Move upmarket: Serve mid-size to enterprise businesses with \~500-5,000 employees instead of startups Focus on content that supports a broader range of business goals instead of solely on SEO and organic growth (e.g., sales, demand gen, brand, etc) Shift back to our broader playbook of content deliverables, including thought leadership, data studies, and more Focus on content execution and production to support an internally-directed content strategy across multiple functions In a way, it’s sort of a reverse-niche move. Rather than zooming in specifically on driving organic growth for startups, we want to be more of an end-to-end content production partner that solves issues of execution and operations for all kinds of content teams. It’s early days, but the response here has been promising. We’ve seen an uptick in leads through Q4. And more companies in our pipeline fit the new ICP. They’re bigger, often have more budget. (But they move more slowly). We should know by the end of the quarter if this maneuver is truly paying off. Hopefully, this will work out. Hopefully our research and strategy are right and we’ll find a soft landing serving a different type of client. If it doesn’t? Then it will be time to make some harder decisions. As I already mentioned, I’m not interested in the race to the bottom of AI content. And if that’s the only game left in town, then it might be time to think hard about a much bigger change. — To be done: Build new content playbooks for expanded deliverables Build new showcase page for expanded deliverables Retooling the Operation It’s easy to say we’re doing something new. It’s a lot harder to actually do it—and do it well. Beyond just changing our positioning, we have to do open-heart surgery on the entire content operation behind the scenes. We need to create new systems that work for a broader range of content types, formats, and goals. Here’s the first rub: All of our workflows are tooled specifically for SEO-focused content. Every template, worksheet, and process that we’ve built and scaled in the last 5 years assumes that the primary goal of every piece of content is SEO. Even something as simple as requiring a target keyword is a blocker in a world where we’re not entirely focused on SEO. This is relatively easy to fix, but it requires several key changes: Update content calendars to make keywords optional Update workflows to determine whether we need an optimization report for each deliverable Next, we need to break down the deliverables into parts rather than a single line item. In our old system, we would plan content as a single row in a Content Calendar spreadsheet. It was a really wide sheet with lots of fields where we’d define the dimensions of each individual article. This was very efficient and simple to follow. But every article had the same overall scope when it came to the workflow. In Asana (our project management tool), all of the steps in the creation were strung together in a single task. We would create a few basic templates for each client, and then each piece would flow through the same steps: Briefing Writing Editing Design etc. If we had anything that didn’t fit into the “standard” workflow, we’d just tag it in the calendar with an unofficial notation \[USING BRACKETS\]. It worked. But it wasn’t ideal. Now we need the steps to be more modular. Imagine, for example, a client asks us to create a mix of deliverables: 1 article with writing + design 1 content brief 1 long-form ebook with an interview + writing + design Each of these would require its own steps and its own workflow. We need to break down the work to accommodate for a wider variety of workflows and variables. This means we need to update the fields and structure of our calendar to accommodate for the new dimensions—while also keeping the planning process simple and manageable. This leads to the next challenge: The number of “products” that we’re offering could be almost infinite. Just looking at the example scope above, you can mix and match all of these different building blocks to create a huge variety of different types of work, each requiring its own workflow. This is part of the reason we pivoted away from this model to focus on a productized, SEO-focused content service back in 2020. Take something as simple as a case study. On the surface, it seems like one deliverable that can be easily scoped and priced, right? Well, unpack what goes into a case study: Is there already source material from the customer or do we need to conduct an interview? How long is it? Is it a short overview case study or a long-form narrative? Does it need images and graphics? How many? Each of these variables opens up 2-3 possibilities. And when you combine them, we end up with something like 10 possible permutations for this single type of deliverable. It gets a bit messy. But not only do we have to figure out how to scope and price all for all of these variables, we also have to figure out how to account for these variables in the execution. We have to specify—for every deliverable—what type it is, how long, which steps are involved and not involved, the timeline for delivery, and all of the other factors. We’re approaching infinite complexity, here. We have to figure out a system that allows for a high level of flexibility to serve the diverse needs of our clients but is also productized enough that we can build workflows, process, and templates to deliver the work. I’ve spent the last few months designing that system. Failed Attempt #1: Ultra-Productization In my first pass, I tried to make it as straight forward as possible. Just sit down, make a list of all of the possible deliverables we could provide and then assign them specific scopes and services. Want a case study? Okay that’ll include an interview, up to 2,000 words of content, and 5 custom graphics. It costs $X. But this solution quickly fell apart when we started testing it against real-world scenarios. What if the client provided the brief instead of us creating one? What if they didn’t want graphics? What if this particular case study really needs to be 3,000 words but all of the others should be 2,000? In order for this system to work, we’d need to individual scope and price all of these permutations of each productized service. Then we’d need to somehow keep track of all of these and make sure that we accurately scope, price, and deliver them across dozens of clients. It’s sort of like a restaurant handling food allergies by creating separate versions of every single dish to account for every individual type of allergy. Most restaurants have figured out that it makes way more sense to have a “standard” and an “allergy-free” version. Then you only need 2 options to cover 100% of the cases. Onto the next option. Failed Attempt #2: Deliverable-Agnostic Services Next, I sat down with my head of Ops, Katy, to try to map it out. We took a big step back and said: Why does the deliverable itself even matter? At the end of the day, what we’re selling is just a few types of work (research, writing, editing, design, etc) that can be packaged up in an infinite number of ways. Rather than try to define deliverables, shouldn’t we leave it open ended for maximum flexibility? From there, we decided to break down everything into ultra-modular building blocks. We started working on this super complex system of modular deliverables where we would have services like writing, design, editing, etc—plus a sliding scale for different scopes like the length of writing or the number of images. In theory, it would allow us to mix and match any combination of services to create custom deliverables for the client. In fact, we wanted the work to be deliverable-agnostic. That way we could mold it to fit any client’s needs and deliver any type of content, regardless of the format or goal. Want a 5,000-word case study with 15 custom graphics? That’ll be $X. Want a 2,000-word blog post with an interview and no visuals? $Y. Just want us to create 10 briefs, you handle the writing, and we do design? It’s $Z. Again, this feels like a reasonable solution. But it quickly spiraled out of amuck. (That’s an Office reference.) For this to work, we need to have incredibly precise scoping process for every single deliverable. Before we can begin work (or even quote a price), we need to know pretty much the exact word count of the final article, for example. In the real world? This almost never happens. The content is as long as the content needs to be. Clients rarely know if the blog post should be 2,000 words or 3,000 words. They just want good content. We have a general ballpark, but we can rarely dial it in within just 1,000 words until we’ve done enough research to create the brief. Plus, from a packaging and pricing perspective, it introduces all kind of weird scenarios where clients will owe exactly $10,321 for this ultra-specific combination of services. We were building an open system that could accommodate any and all types of potential deliverables. On the face that seems great because it makes us incredibly flexible. In reality, the ambiguity actually works against us. It makes it harder for us to communicate to clients clearly about what they’ll get, how much it will cost, and how long it will take. That, of course, also means that it hurts our client relationships. (This actually kind of goes back to my personal learnings, which I’ll mention in a bit. I tend to be a “let’s leave things vague so we don’t have to limit our options” kind of person. But I’m working on fixing this to be more precise, specific, and clear in everything that we do.) Dialing It In: Building a Closed System We were trying to build an open system. We need to build a closed system. We need to force clarity and get specific about what we do, what we don’t do, and how much it all costs. Then we need a system to expand on that closed system—add new types of deliverables, new content playbooks, and new workflows if and when the need arises. With that in mind, we can start by mapping out the key dimensions of any type of deliverable that we would ever want to deliver. These are the universal dimensions that determine the scope, workflow, and price of any deliverable—regardless of the specific type output. Dimensions are: Brief scope Writing + editing scope Design scope Interview scope Revision (rounds) Scope, essentially, just tells us how many words, graphics, interviews, etc are required for the content we’re creating. In our first crack at the system, we got super granular with these scopes. But to help force a more manageable system, we realized that we didn’t need tiny increments for most of this work. Instead, we just need boundaries—you pay $X for up to Y words. We still need some variability around the scope of these articles. Obviously, most clients won’t be willing to pay the same price for a 1,000-word article as a 10,000-word article. But we can be smarter about the realistic break points. We boiled it down to the most common ranges: (Up to) 250 words 1,000 words 3,000 words 6,000 words 10,000 words This gives us a much more manageable number of variables. But we still haven’t exactly closed the system. We need one final dimension: Deliverable type. This tells us what we’re actually building with these building blocks. This is how we’ll put a cap on the potentially infinite number of combinations we could offer. The deliverable type will define what the final product should look like (e.g., blog post, case study, ebook, etc). And it will also give us a way to put standards and expectations around different types of deliverables that we want to offer. Then we can expand on this list of deliverables to offer new services. In the mean time, only the deliverables that we have already defined are, “on the menu,” so to speak. If a client comes to us and asks for something like a podcast summary article (which we don’t currently offer), we’ll have to either say we can’t provide that work or create a new deliverable type and define the dimensions of that specific piece. But here’s the kicker: No matter the deliverable type, it has to still fit within the scopes we’ve already defined. And the pricing will be the same. This means that if you’re looking for our team to write up to 1,000 words of content, it costs the same amount—whether it’s a blog post, an ebook, a LinkedIn post, or anything else. Rather than trying to retool our entire system to offer this new podcast summary article deliverable, we’ll just create the new deliverable type, add it to the list of options, and it’s ready to sell with the pre-defined dimensions we’ve already identified. To do: Update onboarding workflow Update contracts and scope documents Dial in new briefing process Know Thyself For the last year, I’ve been going through personal therapy. (Huge shout out to my wife, Laura, for her support and encouragement throughout the process.) It’s taught me a lot about myself and my tendencies. It’s helped me find some of my weaknesses and think about how I can improve as a person, as a partner, and as an entrepreneur. And it’s forced me to face a lot of hard truths. For example, consider some of the critical decisions I’ve made for my business: Unconventional freelance “collective” model No formal management structure Open-ended retainers with near-infinite flexibility General contracts without defined scope “Take it or leave it” approach to sales and marketing Over the years, I’ve talked about almost everything on this list as a huge advantage. I saw these things as a reflection of how I wanted to do things differently and better than other companies. But now, I see them more as a reflection of my fears and insecurities. Why did I design my business like this? Why do I want so much “flexibility” and why do I want things left open-ended rather than clearly defined? One reason that could clearly explain it: I’m avoidant. If you’re not steeped in the world of therapy, this basically means that my fight or flight response gets turned all the way to “flight.” If I’m unhappy or uncomfortable, my gut reaction is usually to withdraw from the situation. I see commitment and specificity as a prelude to future conflict. And I avoid conflict whenever possible. So I built my business to minimize it. If I don’t have a specific schedule of work that I’m accountable for delivering, then we can fudge the numbers a bit and hope they even out in the end. If I don’t set a specific standard for the length of an article, then I don’t have to let the client know when their request exceeds that limit. Conflict….avoided? Now, that’s not to say that everything I’ve built was wrong or bad. There is a lot of value in having flexibility in your business. For example, I would say that our flexible retainers are, overall, an advantage. Clients have changing needs. Having flexibility to quickly adapt to those needs can be a huge value add. And not everything can be clearly defined upfront (at least not without a massive amount of time and work just to decide how long to write an article). Overly-rigid structures and processes can be just as problematic as loosey-goosey ones. But, on the whole, I realized that my avoidant tendencies and laissez faire approach to management have left a vacuum in many areas. The places where I avoided specificity were often the places where there was the most confusion, uncertainty, and frustration from the team and from clients. People simply didn’t know what to expect or what was expected of them. Ironically, this often creates the conflict I’m trying to avoid. For example, if I don’t give feedback to people on my team, then they feel uneasy about their work. Or they make assumptions about expectations that don’t match what I’m actually expecting. Then the client might get upset, I might get upset, and our team members may be upset. Conflict definitely not avoided. This happens on the client side, too. If we don’t define a specific timeline when something will be delivered, the client might expect it sooner than we can deliver—creating frustration when we don’t meet their expectation. This conflict actually would have been avoided if we set clearer expectations upfront. But we didn’t do that. I didn’t do that. So it’s time to step up and close the gaps. Stepping Up and Closing the Gaps If I’m going to address these gaps and create more clarity and stability, I have to step up. Both personally and professionally. I have to actually face the fear and uncertainty that drives me to be avoidant. And then apply that to my business in meaningful ways that aren’t cop-out ways of kinda-sorta providing structure without really doing it. I’ve gotta be all in. This means: Fill the gaps where I rely on other people to do things that aren’t really their job but I haven’t put someone in place to do it Set and maintain expectations about our internal work processes, policies, and standards Define clear boundaries on things like roles, timelines, budgets, and scopes Now, this isn’t going to happen overnight. And just because I say that I need to step up to close these gaps doesn’t mean that I need to be the one who’s responsible for them (at least not forever). It just means that, as the business leader, I need to make sure the gaps get filled—by me or by someone else who has been specifically charged with owning that part of the operation. So, this is probably my #1 focus over the coming quarter. And it starts by identifying the gaps that exist. Then, step into those gaps myself, pay someone else to fill that role, or figure out how to eliminate the gap another way. This means going all the way back to the most basic decisions in our business. One of the foundational things about Optimist is being a “different kind” of agency. I always wanted to build something that solved for the bureaucracy, hierarchy, and siloed structure of agencies. If a client has feedback, they should be able to talk directly to the person doing the work rather than going through 3 layers of account management and creative directors. So I tried to be clever. I tried to design all kinds of systems and processes that eliminated these middle rungs. (In retrospect, what I was actually doing was designing a system that played into my avoidant tendencies and made it easy to abdicate responsibility for lots of things.) Since we didn’t want to create hierarchy, we never implemented things like Junior and Senior roles. We never hired someone to manage or direct the individual creatives. We didn’t have Directors or VPs. (Hell, we barely had a project manager for the first several years of existence.) This aversion to hierarchy aligned with our values around elevating ownership and collective contribution. I still believe in the value a flat structure. But a flat structure doesn’t eliminate the complexity of a growing business. No one to review writers and give them 1:1 feedback? I guess I’ll just have to do that….when I have some spare time. No Content Director? Okay, well someone needs to manage our content playbooks and roll out new ones. Just add it to my task list. Our flat structure didn’t eliminate the need for these roles. It just eliminated the people to do them. All of those unfilled roles ultimately fell back on me or our ops person, Katy. Of course, this isn’t the first time we’ve recognized this. We’ve known there were growing holes in our business as it’s gotten bigger and more complex. Over the years, we’ve experimented with different ways to solve for it. The Old Solution: Distributed Ops One system we designed was a “distributed ops” framework. Basically, we had one person who was the head of ops (at the time, we considered anything that was non-client-facing to be “ops”). They’d plan and organize all of the various things that needed to happen around Optimist. Then they’d assign out the work to whoever was able to help. We had a whole system for tying this into the our profit share and even gave people “Partner” status based on their contributions to ops. It worked—kinda. One big downfall is that all of the tasks and projects were ad hoc. People would pick up jobs, but they didn’t have much context or expertise to apply. So the output often varied. Since we were trying to maintain a flat structure, there was minimal oversight or management of the work. In other words, we didn’t always get the best results. But, more importantly, we still didn’t close all of the gaps entirely. Because everything was an ad-hoc list of tasks and projects, we never really had the “big picture” view of everything that needed to be done across the business. This also meant we rarely had clarity on what was important, what was trivial, and what was critical. We need a better system. Stop Reinventing the Wheel (And Create a Damn Org Chart) It’s time to get serious about filling the gaps in our business. It can’t be a half-fix or an ad hoc set of projects and tasks. We need clarity on the roles that need to be filled and then fill them. The first step here is to create an org chart. A real one. Map out all of the jobs that need to be done for Optimist to be successful besides just writers and designers. Roles like: Content director Design director SEO manager Reporting Finance Account management Business development Sales Marketing Project management It feels a bit laughable listing all of these roles. Because most are either empty or have my name attached to them. And that’s the problem. I can’t do everything. And all of the empty roles are gaps in our structure—places where people aren’t getting the direction, feedback, or guidance they need to do their best work. Or where things just aren’t being done consistently. Content director, for example, should be responsible for steering the output of our content strategists, writers, and editors. They’re not micromanaging every deliverable. But they give feedback, set overall policy, and help our team identify opportunities to get better. Right now we don’t have anyone in that role. Which means it’s my job—when I have time. Looking at the org chart (a real org chart that I actually built to help with this), it’s plain as day how many roles look like this. Even if we aren’t going to implement a traditional agency structure and a strict hierarchy, we still need to address these gaps. And the only way for that to happen is face the reality and then create a plan to close the gaps. Now that we have a list of theoretical roles, we need to clearly define the responsibilities and boundaries of those roles to make sure they cover everything that actually needs to happen. Then we can begin the process of delegating, assigning, hiring, and otherwise addressing each one. So that’s what I need to do. To be done: Create job descriptions for all of the roles we need to fill Hire Biz Dev role Hire Account Lead role(s) Hire Head of Content Playing Offense As we move into Q1 of 2025 and I reflect on the tumultuous few years we’ve had, one thought keeps running through my head. We need to play offense. Most of the last 1-2 years was reacting to changes that were happening around us. Trying to make sense and chart a new path forward. Reeling. But what I really want—as a person and as an entrepreneur—is to be proactive. I want to think and plan ahead. Figure out where we want to go before we’re forced to change course by something that’s out of our control. So my overarching focus for Q1 is playing offense. Thinking longer term. Getting ahead of the daily deluge and creating space to be more proactive, innovative, and forward thinking. To do: Pilot new content formats Audit and update our own content strategy Improve feedback workflows Build out long-term roadmap for 1-2 years for Optimist Final Note on Follow-Through and Cadence In my reflection this year, one of the things I’ve realized is how helpful these posts are for me. I process by writing. So I actually end up making a lot of decisions and seeing things more clearly each time I sit down to reflect and write my yearly recap. It also gives me a space to hold myself accountable for the things I said I would do. So, I’m doing two things a bit differently from here on out. First: I’m identifying clear action items that I’m holding myself accountable for getting done in the next 3 months (listed in the above sections). In each future update, I’ll do an accounting of what I got done and what wasn’t finished (and why). Second: I’m going to start writing shorter quarterly updates. This will gives me more chances each year to reflect, process, and make decisions. Plus it gives me a shorter feedback loop for the action items that I identified above. (See—playing offense.) — Okay friends, enemies, and frenemies. This is my first update for 2025. Glad to share with y’all. And thanks to everyone who’s read, commented, reached out, and shared their own experiences over the years. We are all the accumulation of our connections and our experiences. As always, I will pop in to respond to comments and answer questions. Feel free to share your thoughts, questions, and general disdain down below. Cheers, Tyler

 I just sold my startup for $200,000 after 11 months. AMA
reddit
LLM Vibe Score0
Human Vibe Score0
jeannenThis week

I just sold my startup for $200,000 after 11 months. AMA

Last August, I was looking for a startup idea I could grow and made a MVP in a week then launched it. I received the $200,000 wire from the buyer a couple of days ago I found tons of useful info online for free, so I hope this can be my way of giving back :) Here is some background: Idea I got the idea when trying to write a tweet using Google Doc's transcription tool, which was terrible. I was pretty sure I wasn't the only one too lazy to type, I made my own solution using AI to transcribe and reformat voice notes into any kind of content. I called it Talknotes, mainly because it was the only domain available lol Validation: My rule is to only reinvest what the project generates. After listing on startup directories and posting on Twitter, I generated $700 in 10 days. It wasn't much, but enough to show interest and keep me motivated. I added user-requested features, but the launch effect wore off, and daily revenues dropped to $0 after a few weeks. I almost gave up, but friends encouraged me to continue. In October, I launched on ProductHunt and it blew up. It became Product of the Day and reached $1500 MRR thanks to media coverage. I initially built everything using vanilla JS/CSS/HTML + Node for backend. But it's pretty limited for apps with lots of interactivity so, I rebuilt the app using Nuxt.js to make it easier to ship new features. Then, I launched ads on Facebook and I implemented a feedback loop: Get new users Learn about them through onboarding Make more ads based on onboarding data This doubled MRR in about 2 months. Burnout and Sale: In May, I had a bad burnout after emergency bug fixes. This made it hard to work on the app after. At this point MRR was around $7000 and total revenues around $70,0000 I listed it on Acquire.com for $200,000, a very good price for the buyer considering revenues and growth. I could've gotten $300,000 with buyer financing or earn-outs, but I wanted cash, $200,000 today is better than $300,000 in a year. Everything was smooth until we tried using Escrow, which almost fucked up the deal (details here). Long story short, had to threaten them to make a sponsored post on Twitter explaining what they did + legal action. They sent the refund the very next day, and we completed the transfer directly. Now, this isn't an overnight success. It's the result of 7 years of grind. I launched over 40 projects since I started, and most of them failed. I often worked 100 hours per week, and I rarely go out or meet many people. It's not for everyone, but I'm fine with it With the profit from the app + sale, and other projects, I have close to 1/3 of a million dollar. I could retire in Asia if I wanted Just mind blowing to think I wrote funny characters in a code editor and sold it for the price of a house lol Edit 1: A few people got confused. I said it's 7 years of grind and most of my projects failed, not that I was not making money. I also said I OFTEN worked 100h/week, not every week :) Since I learned to code 2 years ago I've made close to $400k from my app's profit + exit (this one + another one for $65k last year). And before that I was making money as a marketing freelancer. Also, I dropped after high-school, so, I had to learn everything from scratch, it takes time! Edit 2: Lots of people asked how/where I learned to code in 2 months. I wrote a blog/journal about it back then with links to resources, you can find it here if you're interested

Seeking co-founder to build LinkedIn’s biggest rival(curated version)
reddit
LLM Vibe Score0
Human Vibe Score1
ItzdreeThis week

Seeking co-founder to build LinkedIn’s biggest rival(curated version)

How do you connect with likeminded people? You see the polished wins everywhere, but what about the messy drafts , the awkward pitches and the moments you’re not sure you’ve got it right? Problem: The whole idea of founding and starting a business can be super intimidating for some people, specially those who don’t know any founders personally, those who don’t have a large network, those who don’t have rich parents with large networks, those not inserted in an entrepreneurial culture like in the US for example (which is my case). Sometimes all you need is the right support network, and too see others do what you want, to know that it’s possible! Everyone has an “ultimate guide” to make 7 figures or build a business on YouTube but NO ONE shoes you the HOW, just the results… I’ve tried joining founder communities, LinkedIn ,Reddit … you name it. Most of these founder communities are inaccessible for regular people and often ask for you to have an already existing business with a min ARR… or their simply geography based and if you’re not in a certain area you can’t really participate… As of LinkedIn… full of empty AI generated posts about how some random dude raised $10m in 7 days. Okay Jonathan, but what about the HOW?? How did you write your first pitch? How many rejection calls did you get? What is an MVP? There simply isn’t a platform out there to document your founding journey and find inspiration within a community of people who are doing the same as you. What better way to feel motivated then to see someone actually document their process? Solution: I’m working on building a social media platform for aspiring/founders to connect through the RAW, UNFILTERED process of turning ideas into reality in REAL time. It’s all gonna be around the “building in public “ concept and content creation. Picture an instagram/tiktok profile where instead of seeing someone’s dog you see them documenting their founding process—from the moment they had the idea, to the moment they launched, you’re going to see the successes, the fails, the rejected calls, all documented through actual content and not some AI generated LinkedIn post. Imagine if you wanted to learn about how Steve Jobs started Apple , you could simply go through his profile on this app—exactly. To make sure all interactions are meaningful people would have to apply. It’s a truly curated community, with REAL people, building REAL things in REAL time, and not just tell us the story of how they did it… Audience: I’m targeting people who have a burning desire of building a business and early stage founders starting their founder journeys, that don’t have a support network and simply don’t know where to start. People who are tired of watching 30 min “ultimate guides “ on how to make it on YouTube from “business gurus” selling courses. People who haven’t reached the min ARR required to join an “exclusive “ founder a community. People who can’t simply just move to the US to get into the “exclusive” YC combinator. People who want to connect with real people building real things and not anonymous people on Reddit, or LinkedIn influencers again trying to promote their services. I believe in the idea because I’m also part of my audience. Have always wanted to start my own thing just never knew how to and where to find a community of likeminded people . I don’t know any founders myself, I come from a non-entrepreneurial society and I’d pay good money to access a community of REAL passionate founders building REAL things, in REAL time. This would be my first ever business, and I want to share my journey building it and hopefully inspire others to just start so I’ve created a mailing list to keep anyone interested in the project updated on my fails , learnings and successes. I’m not worried about “making it” but just “starting” and hopefully reach the right audience and inspire anyone to start whatever they have marinating in their thoughts. If you’re a founder struggling with staying consistent or an aspiring founder with an insane desire of starting and don’t know how to start, I’d love to get your feedback on what’s stopping you, your challenges starting out and what you’d find useful in such platform. And finally would this be something that interests you?? PS: casually looking for a technical co-founder

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

I Watched My Startup Slowly Dying Over Two Years: Mistakes and Lessons Learned
reddit
LLM Vibe Score0
Human Vibe Score0.429
Personal-Expression3This week

I Watched My Startup Slowly Dying Over Two Years: Mistakes and Lessons Learned

If you are tired of reading successful stories, you may want to listen to my almost failure story. Last year in April, I went full-time on my startup. Nearly two years later, I’ve seen my product gradually dying. I want to share some of the key mistakes I made and the lessons I’ve taken from them so you don't have to go through them. Some mistakes were very obvious in hindsight; others, I’m still not sure if they were mistakes or just bad luck. I’d love to hear your thoughts and advice as well. Background I built an English-learning app, with both web and mobile versions. The idea came from recognizing how expensive it is to hire an English tutor in most countries, especially for practicing speaking skills. With the rise of AI, I saw an opportunity in the education space. My target market was Japan, though I later added support for multiple languages and picked up some users from Indonesia and some Latin American countries too. Most of my users came from influencer marketing on Twitter. The MVP for the web version launched in Japan and got great feedback. People were reposting it on Twitter, and growth was at its peak in the first few weeks. After verifying the requirement with the MVP, I decided to focus on the mobile app to boost user retention, but for various reasons, the mobile version didn’t launch until December 2023— 8 months after the web version. Most of this year has been spent iterating on the mobile app, but it didn’t make much of an impact in the end. Key Events and Lessons Learned Here are some takeaways: Find co-founders as committed as you are I started with two co-founders—both were tech people and working Part-Time. After the web version launched, one dropped out due to family issues. Unfortunately, we didn’t set clear rules for equity allocation, so even after leaving, they still retained part of the equity. The other co-founder also effectively dropped out this year, contributing only minor fixes here and there. So If you’re starting a company with co-founders, make sure they’re as committed as you are. Otherwise, you might be better off going solo. I ended up teaching myself programming with AI tools, starting with Flutter and eventually handling both front-end and back-end work using Windsurf. With dev tools getting more advanced, being a solo developer is becoming a more viable option. Also, have crystal-clear rules for equity—especially around what happens if someone leaves. Outsourcing Pitfalls Outsourcing development was one of my biggest mistakes. I initially hired a former colleague from India to build the app. He dragged the project on for two months with endless excuses, and the final output was unusable. Then I hired a company, but they didn’t have enough skilled Flutter developers. The company’s owner scrambled to find people, which led to rushed work and poor-quality code which took a lot of time revising myself. Outsourcing is a minefield. If you must do it, break the project into small tasks, set clear milestones, and review progress frequently. Catching issues early can save you time and money. Otherwise, you’re often better off learning the tools yourself—modern dev tools are surprisingly beginner-friendly. Trust, but Verify I have a bad habit of trusting people too easily. I don’t like spending time double-checking things, so I tend to assume people will do what they say they’ll do. This mindset is dangerous in a startup. For example, if I had set up milestones and regularly verified the progress of my first outsourced project, I would’ve realized something was wrong within two weeks instead of two months. That would’ve saved me a lot of time and frustration. Like what I mentioned above, set up systems to verify their work—milestones, deliverables, etc.—to minimize risk. Avoid red ocean if you are small My team was tiny (or non-existent, depending on how you see it), with no technical edge. Yet, I chose to enter Japan’s English-learning market, which is incredibly competitive. It’s a red ocean, dominated by big players who’ve been in the game for years. Initially, my product’s AI-powered speaking practice and automatic grammar correction stood out, but within months, competitors rolled out similar features. Looking back, I should’ve gone all-in on marketing during the initial hype and focused on rapidly launching the mobile app. But hindsight is 20/20. 'Understanding your user' helps but what if it's not what you want? I thought I was pretty good at collecting user feedback. I added feedback buttons everywhere in the app and made changes based on what users said. But most of these changes were incremental improvements—not the kind of big updates that spark excitement. Also, my primary users were from Japan and Indonesia, but I’m neither Japanese nor Indonesian. That made it hard to connect with users on social media in an authentic way. And in my opinion, AI translations can only go so far—they lack the human touch and cultural nuance that builds trust. But honestly I'm not sure if the thought is correct to assume that they will not get touched if they recognize you are a foreigner...... Many of my Japanese users were working professionals preparing for the TOEIC exam. I didn’t design any features specifically for that; instead, I aimed to build a general-purpose English-learning tool since I dream to expand it to other markets someday. While there’s nothing wrong with this idealistic approach, it didn’t give users enough reasons to pay for the app. Should You Go Full-Time? From what I read, a lot of successful indie developers started part-time, building traction before quitting their jobs. But for me, I jumped straight into full-time mode, which worked for my lifestyle but might’ve hurt my productivity. I value work-life balance and refused to sacrifice everything for the startup. The reason I chose to leave the corp is I want to escape the 996 toxic working environment in China's internet companies. So even during my most stressful periods, I made time to watch TV with my partner and take weekends off. Anyways, if you’re also building something or thinking about starting a business, I hope my story helps. If I have other thoughts later, I will add them too. Appreciate any advice.

101 best SEO tips to help you drive traffic in 2k21
reddit
LLM Vibe Score0
Human Vibe Score0.543
DrJigsawThis week

101 best SEO tips to help you drive traffic in 2k21

Hey guys! I don't have to tell you how SEO can be good for your business - you can drive leads to your SaaS on autopilot, drive traffic to your store/gym/bar/whatever, etc. The thing with SEO, though, is that most SEO tips on the internet are just not that good. Most of the said tips: Are way too simple & basic (“add meta descriptions to your images”*) Are not impactful. Sure, adding that meta tag to an image is important, but that’s not what’s going to drive traffic to your website Don’t talk much about SEO strategy (which is ultimately the most important thing for SEO). Sure, on-page SEO is great, but you sure as hell won't drive much traffic if you can't hire the right writers to scale your content. And to drive serious SEO traffic, you'll need a LOT more than that. Over the past few years, my and my co-founder have helped grow websites to over 200k+ monthly traffic (check out our older Reddit post if you want to learn more about us, our process, and what we do), and we compiled all our most important SEO tips and tricks, as well as case studies, research, and experiments from the web, into this article. Hope you like it ;) If you think we missed something super important, let us know and we'll add it to the list. And btw, we also published this article on our own blog with images, smart filters, and all that good stuff. If you want to check it out, click here. That said, grab some coffee (or beer) & let's dive in - this is going to be a long one. SEO Strategy Tips Tip #1. A Lot of SEO Tips On The Internet Are NOT Necessarily Factual A lot of the SEO content you’ll read on the internet will be based on personal experiences and hearsay. Unfortunately, Google is a bit vague about SEO advice, so you have to rely more on experiments conducted by SEO pros in the community. So, sometimes, a lot of this information is questionable, wrong, or simply based on inaccurate data.  What we’re getting at here is, whenever you hear some new SEO advice, take it with a grain of salt. Google it to double-check other sources, and really understand what this SEO advice is based on (instead of just taking it at face value). Tip #2. SEO Takes Time - Get Used to It Any way you spin it, SEO takes time.  It can take around 6 months to 2 years (depending on the competition in your niche) before you start seeing some serious results.  So, don’t get disappointed if you don’t see any results within 3 months of publishing content. Tip #3. SEO Isn’t The Best Channel for Everyone That said, if you need results for your business tomorrow, you might want to reconsider SEO altogether.  If you just started your business, for example, and are trying to get to break-even ASAP, SEO is a bad idea - you’ll quit before you even start seeing any results.  If that’s the case, focus on other marketing channels that can have faster results like content marketing, PPC, outreach, etc. Tip #4. Use PPC to Validate Keywords Not sure if SEO is right for your business? Do this: set up Google Search ads for the most high-intent keywords in your niche. See how well the traffic converts and then decide if it’s worthwhile to focus on SEO (and rank on these keywords organically). Tip #5. Use GSC to See If SEO Is Working While it takes a while to see SEO results, it IS possible to see if you’re going in the right direction. On a monthly basis, you can use Search Console to check if your articles are indexed by Google and if their average position is improving over time. Tip #6. Publish a TON of Content The more content you publish on your blog, the better. We recommend a minimum of 10,000 words per month and optimally 20,000 - 30,000 (especially if your website is fresh). If an agency offers you the typical “4 500-word articles per month” deal, stay away. No one’s ever gotten results in SEO with short, once-per-week articles. Tip #7. Upgrade Your Writers Got a writer that’s performing well? Hire them as an editor and get them to oversee content operations / edit other writers’ content. Then, upgrade your best editor to Head of Content and get them to manage the entire editor / writer ops. Tip #8. Use Backlink Data to Prioritize Content When doing keyword research, gather the backlink data of the top 3 ranking articles and add it to your sheet. Then, use this data to help you prioritize which keywords to focus on first. We usually prioritize keywords that have lower competition, high traffic, and a medium to high buyer intent. Tip #9. Conduct In-Depth Keyword Research Make your initial keyword research as comprehensive as possible. This will give you a much more realistic view of your niche and allow you to prioritize content the right way. We usually aim for 100 to 300 keywords (depending on the niche) for the initial keyword research when we start working with a client. Tip #10. Start With Competitive Analysis Start every keyword research with competitive analysis. Extract the keywords your top 3 competitors are ranking on.  Then, use them as inspiration and build upon it. Use tools like UberSuggest to help generate new keyword ideas. Tip #11. Get SEMrush of Ahrefs You NEED SEMrush or Ahrefs, there’s no doubt about it. While they might seem expensive at a glance (99 USD per month billed annually), they’re going to save you a lot of manpower doing menial SEO tasks. Tip #12. Don’t Overdo It With SEO Tools Don’t overdo it with SEO tools. There are hundreds of those out there, and if you’re the type that’s into SaaS, you might be tempted to play around with dozens at a time. And yes, to be fair, most of these tools ARE helpful one way or another. To effectively do organic SEO, though, you don’t really need that many tools. In most cases, you just need the following: SEMrush/Ahrefs Screaming Frog RankMath/Yoast SEO Whichever outreach tool you prefer (our favorite is snov.io). Tip #13. Try Some of the Optional Tools In addition to the tools we mentioned before, you can also try the following 2 which are pretty useful & popular in the SEO community: Surfer SEO - helps with on-page SEO and creating content briefs for writers. ClusterAI - tool that helps simplify keyword research & save time. Tip #14. Constantly Source Writers Want to take your content production to the next level? You’ll need to hire more writers.  There is, however, one thing that makes this really, really difficult: 95 - 99% of writers applying for your gigs won’t be relevant. Up to 80% will be awful at writing, and the remainder just won’t be relevant for your niche. So, in order to scale your writing team, we recommend sourcing constantly, and not just once every few months. Tip #15. Create a Process for Writer Filtering As we just mentioned, when sourcing writers, you’ll be getting a ton of applicants, but most won’t be qualified. Fun fact \- every single time we post a job ad on ProBlogger, we get around 300 - 500 applications (most of which are totally not relevant). Trust us, you don’t want to spend your time going through such a huge list and checking out the writer samples. So, instead, we recommend you do this: Hire a virtual assistant to own the process of evaluating and short-listing writers. Create a process for evaluating writers. We recommend evaluating writers by: Level of English. If their samples aren’t fluent, they’re not relevant. Quality of Samples. Are the samples engaging / long-form content, or are they boring 500-word copy-pastes? Technical Knowledge. Has the writer written about a hard-to-explain topic before? Anyone can write about simple topics like traveling - you want to look for someone who knows how to research a new topic and explain it in a simple and easy to read way. If someone’s written about how to create a perfect cover letter, they can probably write about traveling, but the opposite isn’t true. The VA constantly evaluates new applicants and forwards the relevant ones to the editor. The editor goes through the short-listed writers and gives them trial tasks and hires the ones that perform well. Tip #16. Use The Right Websites to Source Writers “Is UpWork any good?” This question pops up on social media time and time again. If you ask us, no, UpWork is not good at all. Of course, there are qualified writers there (just like anywhere else), but from our experience, those writers are few and far in-between. Instead, here are some of our favorite ways to source writers: Cult of Copy Job Board ProBlogger Headhunting on LinkedIn If you really want to use UpWork, use it for headhunting (instead of posting a job ad) Tip #17. Hire Writers the Right Way If you want to seriously scale your content production, hire your writers full-time. This (especially) makes sense if you’re a content marketing agency that creates a TON of content for clients all the time. If you’re doing SEO just for your own blog, though, it usually makes more sense to use freelancers. Tip #18. Topic Authority Matters Google keeps your website's authoritativeness in mind. Meaning, if you have 100 articles on digital marketing, you’re probably more of an authority on the topic than someone that has just 10. Hence, Google is a lot more likely to reward you with better rankings. This is also partially why content volume really matters: the more frequently you publish content, the sooner Google will view you as an authority. Tip #19. Focus on One Niche at a Time Let’s say your blog covers the following topics: sales, accounting, and business management.  You’re more likely to rank if you have 30 articles on a single topic (e.g. accounting) than if you have 10 articles on each. So, we recommend you double-down on one niche instead of spreading your content team thin with different topics. Tip #20. Don’t Fret on the Details While technical SEO is important, you shouldn’t get too hung up on it.  Sure, there are thousands of technical tips you can find on the internet, and most of them DO matter. The truth, though, is that Google won’t punish you just because your website doesn’t load in 3 milliseconds or there’s a meta description missing on a single page. Especially if you have SEO fundamentals done right: Get your website to run as fast as possible. Create a ton of good SEO content. Get backlinks for your website on a regular basis. You’ll still rank, even if your website isn’t 100% optimized. Tip #21. Do Yourself a Favor and Hire a VA There are a TON of boring SEO tasks that your team should really not be wasting time with. So, hire a full-time VA to help with all that. Some tasks you want to outsource include gathering contacts to reach out to for link-building, uploading articles on WordPress, etc. Tip #22. Google Isn’t Everything While Google IS the dominant search engine in most parts of the world, there ARE countries with other popular search engines.  If you want to improve your SEO in China, for example, you should be more concerned with ranking on Baidu. Targeting Russia? Focus on Yandex. Tip #23. No, Voice Search is Still Not Relevant Voice search is not and will not be relevant (no matter what sensationalist articles might say). It’s just too impractical for most search queries to use voice (as opposed to traditional search). Tip #24. SEO Is Not Dead SEO is not dead and will still be relevant decades down the line. Every year, there’s a sensationalist article talking about this.  Ignore those. Tip #25. Doing Local SEO? Focus on Service Pages If you’re doing local SEO, focus on creating service-based landing pages instead of content.  E.g. if you’re an accounting firm based in Boston, you can make a landing page about /accounting-firm-boston/, /tax-accounting-boston/, /cpa-boston/, and so on. Thing is, you don’t really need to rank on global search terms - you just won’t get leads from there. Even if you ranked on the term “financial accounting,” it wouldn’t really matter for your bottom line that much. Tip #26. Learn More on Local SEO Speaking of local SEO, we definitely don’t do the topic justice in this guide. There’s a lot more you need to know to do local SEO effectively and some of it goes against the general SEO advice we talk about in this article (e.g. you don't necessarily need blog content for local SEO). We're going to publish an article on that soon enough, so if you want to check it out, DM me and I'll hit you up when it's up. Tip #27. Avoid Vanity Metrics Don’t get side-tracked by vanity metrics.  At the end of the day, you should care about how your traffic impacts your bottom line. Fat graphs and lots of traffic are nice and all, but none of it matters if the traffic doesn’t have the right search intent to convert to your product/service. Tip #28. Struggling With SEO? Hire an Expert Failing to make SEO work for your business? When in doubt, hire an organic SEO consultant or an SEO agency.  The #1 benefit of hiring an SEO agency or consultant is that they’ve been there and done that - more than once. They might be able to catch issues an inexperienced SEO can’t. Tip #29. Engage With the Community Need a couple of SEO questions answered?  SEO pros are super helpful & easy to reach! Join these Facebook groups and ask your question - you’ll get about a dozen helpful answers! SEO Signals Lab SEO & Content Marketing The Proper SEO Group. Tip #30. Stay Up to Date With SEO Trends SEO is always changing - Google is constantly pumping out new updates that have a significant impact on how the game is played.  Make sure to stay up to date with the latest SEO trends and Google updates by following the Google Search Central blog. Tip #31. Increase Organic CTR With PPC Want to get the most out of your rankings? Run PPC ads for your best keywords. Googlers who first see your ad are more likely to click your organic listing. Content & On-Page SEO Tips Tip #32. Create 50% Longer Content On average, we recommend you create an article that’s around 50% longer than the best article ranking on the keyword.  One small exception, though, is if you’re in a super competitive niche and all top-ranking articles are already as comprehensive as they can be. For example, in the VPN niche, all articles ranking for the keyword “best VPN” are around 10,000 - 11,000 words long. And that’s the optimal word count - even if you go beyond, you won’t be able to deliver that much value for the reader to make it worth the effort of creating the content. Tip #33. Longer Is Not Always Better Sometimes, a short-form article can get the job done much better.  For example, let’s say you’re targeting the keyword “how to tie a tie.”  The reader expects a short and simple guide, something under 500 words, and not “The Ultimate Guide to Tie Tying for 2021 \[11 Best Tips and Tricks\]” Tip #34. SEO is Not Just About Written Content Written content is not always best. Sometimes, videos can perform significantly better. E.g. If the Googler is looking to learn how to get a deadlift form right, they’re most likely going to be looking for a video. Tip #35. Don’t Forget to Follow Basic Optimization Tips For all your web pages (articles included), follow basic SEO optimization tips. E.g. include the keyword in the URL, use the right headings etc.  Just use RankMath or YoastSEO for this and you’re in the clear! Tip #36. Hire Specialized Writers When hiring content writers, try to look for ones that specialize in creating SEO content.  There are a LOT of writers on the internet, plenty of which are really good.  However, if they haven’t written SEO content before, chances are, they won’t do that good of a job. Tip #37. Use Content Outlines Speaking of writers - when working with writers, create a content outline that summarizes what the article should be about and what kind of topics it needs to cover instead of giving them a keyword and asking them to “knock themselves out.”   This makes it a lot more likely for the writer to create something that ranks. When creating content outlines, we recommend you include the following information: Target keyword Related keywords that should be mentioned in the article Article structure - which headings should the writer use? In what order? Article title Tip #38. Find Writers With Niche Knowledge Try to find a SEO content writer with some experience or past knowledge about your niche. Otherwise, they’re going to take around a month or two to become an expert. Alternatively, if you’re having difficulty finding a writer with niche knowledge, try to find someone with experience in technical or hard to explain topics. Writers who’ve written about cybersecurity in the past, for example, are a lot more likely to successfully cover other complicated topics (as opposed to, for example, a food or travel blogger). Tip #39. Keep Your Audience’s Knowledge in Mind When creating SEO content, always keep your audience’s knowledge in mind. If you’re writing about advanced finance, for example, you don’t need to teach your reader what an income statement is. If you’re writing about income statements, on the other hand, you’d want to start from the very barebone basics. Tip #40. Write for Your Audience If your readers are suit-and-tie lawyers, they’re going to expect professionally written content. 20-something hipsters? You can get away with throwing a Rick and Morty reference here and there. Tip #41. Use Grammarly Trust us, it’ll seriously make your life easier! Keep in mind, though, that the app is not a replacement for a professional editor. Tip #42. Use Hemingway Online content should be very easy to read & follow for everyone, whether they’re a senior profession with a Ph.D. or a college kid looking to learn a new topic. As such, your content should be written in a simple manner - and that’s where Hemingway comes in. It helps you keep your blog content simple. Tip #43. Create Compelling Headlines Want to drive clicks to your articles? You’ll need compelling headlines. Compare the two headlines below; which one would you click? 101 Productivity Tips \[To Get Things Done in 2021\] VS Productivity Tips Guide Exactly! To create clickable headlines, we recommend you include the following elements: Keyword Numbers Results Year (If Relevant) Tip #44. Nail Your Blog Content Formatting Format your blog posts well and avoid overly long walls of text. There’s a reason Backlinko content is so popular - it’s extremely easy to read and follow. Tip #45. Use Relevant Images In Your SEO Content Key here - relevant. Don’t just spray random stock photos of “office people smiling” around your posts; no one likes those.  Instead, add graphs, charts, screenshots, quote blocks, CSS boxes, and other engaging elements. Tip #46. Implement the Skyscraper Technique (The Right Way) Want to implement Backlinko’s skyscraper technique?  Keep this in mind before you do: not all content is meant to be promoted.  Pick a topic that fits the following criteria if you want the internet to care: It’s on an important topic. “Mega-Guide to SaaS Marketing” is good, “top 5 benefits of SaaS marketing” is not. You’re creating something significantly better than the original material. The internet is filled with mediocre content - strive to do better. Tip #47. Get The URL Slug Right for Seasonal Content If you want to rank on a seasonal keyword with one piece of content (e.g. you want to rank on “saas trends 2020, 2021, etc.”), don’t mention the year in the URL slug - keep it /saas-trends/ and just change the headline every year instead.  If you want to rank with separate articles, on the other hand (e.g. you publish a new trends report every year), include the year in the URL. Tip #48. Avoid content cannibalization.  Meaning, don’t write 2+ articles on one topic. This will confuse Google on which article it should rank. Tip #49. Don’t Overdo Outbound Links Don’t include too many outbound links in your content. Yes, including sources is good, but there is such a thing as overdoing it.  If your 1,000 word article has 20 outbound links, Google might consider it as spam (even if all those links are relevant). Tip #50. Consider “People Also Ask” To get the most out of SERP, you want to grab as many spots on the search result as possible, and this includes “people also ask (PAA):” Make a list of the topic’s PAA questions and ensure that your article answers them.  If you can’t fit the questions & answers within the article, though, you can also add an FAQ section at the end where you directly pose these questions and provide the answers. Tip #51. Optimize For Google Snippet Optimize your content for the Google Snippet. Check what’s currently ranking as the snippet. Then, try to do something similar (or even better) in terms of content and formatting. Tip #52. Get Inspired by Viral Content Want to create content that gets insane shares & links?  Reverse-engineer what has worked in the past. Look up content in your niche that went viral on Reddit, Hacker News, Facebook groups, Buzzsumo, etc. and create something similar, but significantly better. Tip #53. Avoid AI Content Tools No, robots can’t write SEO content.  If you’ve seen any of those “AI generated content tools,” you should know to stay away. The only thing those tools are (currently) good for is creating news content. Tip #54. Avoid Bad Content You will never, ever, ever rank with one 500-word article per week.  There are some SEO agencies (even the more reputable ones) that offer this as part of their service. Trust us, this is a waste of time. Tip #55. Update Your Content Regularly Check your top-performing articles annually and see if there’s anything you can do to improve them.  When most companies finally get the #1 ranking for a keyword, they leave the article alone and never touch it again… ...Until they get outranked, of course, by someone who one-upped their original article. Want to prevent this from happening? Analyze your top-performing content once a year and improve it when possible. Tip #56. Experiment With CTR Do your articles have low CTR? Experiment with different headlines and see if you can improve it.  Keep in mind, though, that what a “good CTR” is really depends on the keyword.  In some cases, the first ranking will drive 50% of the traffic. In others, it’s going to be less than 15%. Link-Building Tips Tip #57. Yes, Links Matter. Here’s What You Need to Know “Do I need backlinks to rank?” is probably one of the most common SEO questions.  The answer to the question (alongside all other SEO-related questions) is that it depends on the niche.  If your competitors don’t have a lot of backlinks, chances are, you can rank solely by creating superior content. If you’re in an extremely competitive niche (e.g. VPN, insurance, etc.), though, everyone has amazing, quality content - that’s just the baseline.  What sets top-ranking content apart from the rest is backlinks. Tip #58. Sometimes, You’ll Have to Pay For Links Unfortunately, in some niches, paying for links is unavoidable - e.g. gambling, CBD, and others. In such cases, you either need a hefty link-building budget, or a very creative link-building campaign (create a viral infographic, news-worthy story based on interesting data, etc.). Tip #59. Build Relationships, Not Links The very best link-building is actually relationship building.  Make a list of websites in your niche and build a relationship with them - don’t just spam them with the standard “hey, I have this amazing article, can you link to it?”.  If you spam, you risk ruining your reputation (and this is going to make further outreach much harder). Tip #60. Stick With The Classics At the end of the day, the most effective link-building tactics are the most straightforward ones:  Direct Outreach Broken Link-Building Guest Posting Skyscraper Technique Creating Viral Content Guestposting With Infographics Tip #61. Give, Don’t Just Take! If you’re doing link-building outreach, don’t just ask for links - give something in return.  This will significantly improve the reply rate from your outreach email. If you own a SaaS tool, for example, you can offer the bloggers you’re reaching out to free access to your software. Or, alternatively, if you’re doing a lot of guest posting, you can offer the website owner a link from the guest post in exchange for the link to your website. Tip #62. Avoid Link Resellers That guy DMing you on LinkedIn, trying to sell you links from a Google Sheet?  Don’t fall for it - most of those links are PBNs and are likely to backfire on you. Tip #63. Avoid Fiverr Like The Plague Speaking of spammy links, don’t touch anything that’s sold on Fiverr - pretty much all of the links there are useless. Tip #64. Focus on Quality Links Not all links are created equal. A link is of higher quality if it’s linked from a page that: Is NOT a PBN. Doesn’t have a lot of outbound links. If the page links to 20 other websites, each of them gets less link juice. Has a lot of (quality) backlinks. Is part of a website with a high domain authority. Is about a topic relevant to the page it’s linking to. If your article about pets has a link from an accounting blog, Google will consider it a bit suspicious. Tip #65. Data-Backed Content Just Works Data-backed content can get insane results for link-building.  For example, OKCupid used to publish interesting data & research based on how people interacted with their platform and it never failed to go viral. Each of their reports ended up being covered by dozens of news media (which got them a ton of easy links). Tip #66. Be Creative - SEO Is Marketing, After All Be novel & creative with your link-building initiatives.  Here’s the thing: the very best link-builders are not going to write about the tactics they’re using.  If they did, you’d see half the internet using the exact same tactic as them in less than a week! Which, as you can guess, would make the tactic cliche and significantly less effective. In order to get superior results with your link-building, you’ll need to be creative - think about how you can make your outreach different from what everyone does. Experiment it, measure it, and improve it till it works! Tip #67. Try HARO HARO, or Help a Reporter Out, is a platform that matches journalists with sources. You get an email every day with journalists looking for experts in specific niches, and if you pitch them right, they might feature you in their article or link to your website. Tip #68. No-Follow Links Aren’t That Bad Contrary to what you might’ve heard, no-follow links are not useless. Google uses no-follow as more of a suggestion than anything else.  There have been case studies that prove Google can disregard the no-follow tag and still reward you with increased rankings. Tip #69. Start Fresh With an Expired Domain Starting a new website? It might make sense to buy an expired one with existing backlinks (that’s in a similar niche as yours). The right domain can give you a serious boost to how fast you can rank. Tip #70. Don’t Overspend on Useless Links “Rel=sponsored” links don’t pass pagerank and hence, won’t help increase your website rankings.  So, avoid buying links from media websites like Forbes, Entrepreneur, etc. Tip #71. Promote Your Content Other than link-building, focus on organic content promotion. For example, you can repost your content on Facebook groups, LinkedIn, Reddit, etc. and focus on driving traffic.  This will actually lead to you getting links, too. We got around 95 backlinks to our SEO case study article just because of our successful content promotion. Tons of people saw the article on the net, liked it, and linked to it from their website. Tip #72. Do Expert Roundups Want to build relationships with influencers in your niche, but don’t know where to start?  Create an expert roundup article. If you’re in the sales niche, for example, you can write about Top 21 Sales Influencers in 2021 and reach out to the said influencers letting them know that they got featured. Trust us, they’ll love you for this! Tip #73. .Edu Links are Overhyped .edu links are overrated. According to John Mueller, .edu domains tend to have a ton of outbound links, and as such, Google ignores a big chunk of them. Tip #74. Build Relationships With Your Customers Little-known link-building hack: if you’re a SaaS company doing SEO, you can build relationships with your customers (the ones that are in the same topical niche as you are) and help each other build links! Tip #75. Reciprocal Links Aren’t That Bad Reciprocal links are not nearly as bad as Google makes them out to be. Sure, they can be bad at scale (if trading links is all you’re doing). Exchanging a link or two with another website / blog, though, is completely harmless in 99% of cases. Tip #76. Don’t Overspam Don’t do outreach for every single post you publish - just the big ones.  Most people already don’t care about your outreach email. Chances are, they’re going to care even less if you’re asking them to link to this new amazing article you wrote (which is about the top 5 benefits of adopting a puppy). Technical SEO Tips Tip #77. Use PageSpeed Insights If your website is extremely slow, it’s definitely going to impact your rankings. Use PageSpeed Insights to see how your website is currently performing. Tip #78. Load Speed Matters While load speed doesn’t impact rankings directly, it DOES impact your user experience. Chances are, if your page takes 5 seconds to load, but your competition’s loads instantly, the average Googler will drop off and pick them over you. Tip #79. Stick to a Low Crawl Depth Crawl depth of any page on your website should be lower than 4 (meaning, any given page should be possible to reach in no more than 3 clicks from the homepage).  Tip #80. Use Next-Gen Image Formats Next-gen image formats such as JPEG 2000, JPEG XR, and WebP can be compressed a lot better than PNG or JPG. So, when possible, use next-get formats for images on your website. Tip #81. De-Index Irrelevant Pages Hide the pages you don’t want Google to index (e.g: non-public, or unimportant pages) via your Robots.txt. If you’re a SaaS, for example, this would include most of your in-app pages or your internal knowledge base pages. Tip #82. Make Your Website Mobile-Friendly Make sure that your website is mobile-friendly. Google uses “mobile-first indexing.” Meaning, unless you have a working mobile version of your website, your rankings will seriously suffer. Tip #83. Lazy-Load Images Lazy-load your images. If your pages contain a lot of images, you MUST activate lazy-loading. This allows images that are below the screen, to be loaded only once the visitor scrolls down enough to see the image. Tip #84. Enable Gzip Compression Enable Gzip compression to allow your HTML, CSS and JS files to load faster. Tip #85. Clean Up Your Code If your website loads slowly because you have 100+ external javascript files and stylesheets being requested from the server, you can try minifying, aggregating, and inlining some of those files. Tip 86. Use Rel-Canonical Have duplicate content on your website? Use rel-canonical to show Google which version is the original (and should be prioritized for search results). Tip #87. Install an SSL Certificate Not only does an SSL certificate help keep your website safe, but it’s also a direct ranking factor. Google prioritizes websites that have SSL certificates over the ones that don’t. Tip #88. Use Correct Anchor Texts for Internal Links When linking to an internal page, mention the keyword you’re trying to rank for on that page in the anchor text. This helps Google understand that the page is, indeed, about the keyword you’re associating it with. Tip #89. Use GSC to Make Sure Your Content is Interlinked Internal links can have a serious impact on your rankings. So, make sure that all your blog posts (especially the new ones) are properly linked to/from your past content.  You can check how many links any given page has via Google Search Console. Tip #90. Bounce rate is NOT a Google ranking factor. Meaning, you can still rank high-up even with a high bounce rate. Tip #91. Don’t Fret About a High Bounce Rate Speaking of the bounce rate, you’ll see that some of your web pages have a higher-than-average bounce rate (70%+).  While this can sometimes be a cause for alarm, it’s not necessarily so. Sometimes, the search intent behind a given keyword means that you WILL have a high bounce rate even if your article is the most amazing thing ever.  E.g. if it’s a recipe page, the reader gets the recipe and bounces off (since they don’t need anything else). Tip #92. Google Will Ignore Your Meta Description More often than not, Google won’t use the meta description you provide - that’s normal. It will, instead, automatically pick a part of the text that it thinks is most relevant and use it as a meta description. Despite this, you should always add a meta description to all pages. Tip #93. Disavow Spammy & PBN Links Keep track of your backlinks and disavow anything that’s obviously spammy or PBNy. In most cases, Google will ignore these links anyway. However, you never know when a competitor is deliberately targeting you with too many spammy or PBN links (which might put you at risk for being penalized). Tip #94. Use The Correct Redirect  When permanently migrating your pages, use 301 redirect to pass on the link juice from the old page to the new one. If the redirect is temporary, use a 302 redirect instead. Tip #95. When A/B Testing, Do This A/B testing two pages? Use rel-canonical to show Google which page is the original. Tip #96. Avoid Amp DON’T use Amp.  Unless you’re a media company, Amp will negatively impact your website. Tip #97. Get Your URL Slugs Right Keep your blog URLs short and to-the-point. Good Example: apollodigital.io/blog/seo-case-study Bad Example: apollodigital.io/blog/seo-case-study-2021-0-to-200,000/ Tip #98. Avoid Dates in URLs An outdated date in your URL can hurt your CTR. Readers are more likely to click / read articles published recently than the ones written years back. Tip #99. Social Signals Matter Social signals impact your Google rankings, just not in the way you think. No, your number of shares and likes does NOT impact your ranking at all.  However, if your article goes viral and people use Google to find your article, click it, and read it, then yes, it will impact your rankings.  E.g. you read our SaaS marketing guide on Facebook, then look up “SaaS marketing” on Google, click it, and read it from there. Tip #100. Audit Your Website Frequently Every other month, crawl your website with ScreamingFrog and see if you have any broken links, 404s, etc. Tip #101. Use WordPress Not sure which CMS platform to use?  99% of the time, you’re better off with WordPress.  It has a TON of plugins that will make your life easier.  Want a drag & drop builder? Use Elementor. Wix, SiteGround and similar drag & drops are bad for SEO. Tip #102. Check Rankings the Right Way When checking on how well a post is ranking on Google Search Console, make sure to check Page AND Query to get the accurate number.  If you check just the page, it’s going to give you the average ranking on all keywords the page is ranking for (which is almost always going to be useless data). Conclusion Aaand that's about it - thanks for the read! Now, let's circle back to Tip #1 for a sec. Remember when we said a big chunk of what you read on SEO is based on personal experiences, experiments, and the like? Well, the tips we've mentioned are part of OUR experience. Chances are, you've done something that might be different (or completely goes against) our advice in this article. If that's the case, we'd love it if you let us know down in the comments. If you mention something extra-spicy, we'll even include it in this article.

I sold my AI tool for $35,000
reddit
LLM Vibe Score0
Human Vibe Score1
marclouvThis week

I sold my AI tool for $35,000

Hey Entrepreneurs, Marc here. Last month I wrote here about how sold a habit tracker for $10,000 in October. Earlier this month, I got $35,000 in my bank account after selling a landing page maker with AI. Here's the story: &#x200B; April 2023: Just like everyone, I get massive FOMO with AI. I played with GPT and decided to build a landing page generator with AI: Input text and the AI prefills a template with copy and AI-generated images. I'm working on it with a good friend of mine named Martin. May: The product is called LandingAI. It's an MVP but we launched and made \~$8,000. Unfortunately, Martin and I had different visions for the project so we forked. &#x200B; June: LandingAI is the name of a big corp (bummer) so I rebranded it to MakeLanding. I ditch 90% of the code because users want a very different product: So here I am, building an entire website builder powered with AI... &#x200B; July: I launched again, but made a BIG mistake: I swapped the one-time payment for a monthly subscription and got $20 MRR for 15k visitors... If you can avoid subscriptions, do it New pricing means new positioning—users compared the app to Framer & Webflow August: I removed the subscription and sales came back: \~$7,000 in 3 months. But I realized this was going nowhere... September: I don't use the product The market is gigantic and crowded As a solopreneur, nothing is more important for me than building cool stuff for people I care about. And I didn't really care about this big market so... October: I called my friend Dan and he said: SELL. He was right. I bought my shares of LandingAI from Martin and listed MakeLanding on Acquire: Asking $38,000 for $14,000 TTM (3x profit) Within hours, I received dozens of NDAs and a buyer started the process 🤯 After a few weeks of NDA, LOI, Escrow, etc. the buyer sent the money but... Only a fraction of the transaction. Then he ghosted me. So I canceled the transition. Back to Acquire... Luckily, in 24 hours I got another buyer! &#x200B; November: Within weeks, the money was in my bank account. The buyer and I never called, just a few messages. It's mind-blowing. &#x200B; My takeaways: Don't build AI products just because Don't go on a massive market you don't care Sell if you don't know how to grow the product It's my 3rd acquisition this year. I love the freedom of build, sell, repeat.

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!
reddit
LLM Vibe Score0
Human Vibe Score0.973
lazymentorsThis week

[Ultimate List] A list of Marketing Tools That I’ve tested over the years and found helpful to do better marketing with less work. More than 50 Tools To Help you with Marketing, Copywriting & Sales!

Starting to focus on marketing for your business, You will come across the same tools mentioned over and over by marketers. I would like to mention here tools that you might haven’t seen going viral in the community but actually will help you grow faster and efficiently. Starting off with My favourite Marketing Channel! #Email Marketing For SMBs Convertkit / Mailerlite / Mailchimp - These 3 Platforms are the best options for SMBs and entrepreneurs just starting out with email marketing. All 3 have free plans up to 1,000 subscribers. Scribe - Email Signature Tool, Create Great Email signatures for your emails. Liramail - Most Email marketing platforms don’t offer great email templates. This tool will help you build great email templates with drag and drop. Quick mail Auto-Warmer - Most Businesses at the beginning don’t know what to do when open rate drops. You need to use an email warmer like this to keep it up. #Email Marketing For Big Businesses SendGrid - Overall Email Marketing Tools, this tool is best for brands that have huge email lists and email marketing is the key marketing channel. Braze - This tool is leading in email marketing for large Email senders. When I was working for agencies, this was one of the best email marketing tools I had used. NeoCertified - Protect your emails for spammers and threats. To keep your email list healthy, this is a must have! Sparkloop - Referral Marketing For Email Campaigns. Email can generate great huge amount of referrals for you and Sparkloop makes it easier. #Cold Emails & Lead Generation Hunter - A Great Tool to scrape emails from domain names. The tool comes with a green free plan but Pro plan is worth the amount of features it provides. Icyleads - It’s better than Hunter as it’s heavily focused on the sales and prospecting to help you derive great results from your campaigns. Mailshake - Beginner Friend Cold Email Tool with Great features like email list warming. #Communication Tools Twilio - One do the best customer engagement platform used by Companies like Stripe and mine too. Chatlio - Use Live chat feature on your website with slack integration. My favourite easier to catch up on conversations through slack integration. Intercom - Used by Most Marketers, Industry Leading customer communication platform. Great for beginners! Chatwoot - Another Amazing Communication Tool but the best part is they have a great free plan useful for new businesses. Loom - Communicate with your audience through Videos. Loom is great for SaaS and to show human interaction to close new visitors effectively. #CRM Outseta - This tool provides great CRM and their billing system is better than other tools out their which makes it stands out! Hubspot - I don’t think this tool needs an introduction because Hubspot’s CRM is the best in industry. Salesflare - This CRM is a great alternative to hubspot as it’s beginner friendly and helpful for SMBs. #SEO Tools Ahrefs - One of the best SEO tool in the industry. They also just launched a bunch of free tools to help SEO beginners. Screaming frog - The only website crawler I have used since I bought my first domain. It’s the best! Ubersuggest- The Tool by Neil Patel is the best SEO tool for you. (I’m Joking, it’s the worst) Contentking - This tool is good at Real-time SEO Auditing, they do a lot of Marketing work through Newsletters. If you are subscribed to any SEO newsletter. You may have seen this tool. SEOquake & Semrush - SEOquake is a great tool to conduct on-page analysis, SERP, and much more. Great tool but it’s owned by Semrush. You should go for Semrush because that tool will cover all SEO aspects for you. #Content Marketing Buzzsumo - This tool is great for content research and but you may find the regular emails pretty annoying sometimes. Contentrow - Analyse Your Content and find it’s strength. Highly recommended who are weak at content structuring like me. Grammarly - If you are not a native English speaker like me, you might think you need it or not. You need it for sure for grammar corrections. #Graphic Design Tools Visme - At agencies, Infographics can be more effective than usual postscript. Visme is a graphic design tool focused on infographics and designs related to B2B and B2C. It’s great for agencies! Glorify - A Graphic Design Tool focused on E-commerce, filled with Designs useful for E-commerce store owners. Canva - All-in-one Industry leading Graphic Design Tool that everyone knows and every template is overused now. Adobe Creative Cloud ( previously Sparkpost) - It’s a great alternative to Canva filled with Amazing Stock images to use in your visuals but the only backlash is the exports in this tool are not high quality. Snaps - A Canva Alternative that might not have overused templates for your Social Accounts. #Advertising Tools Plai - It’s a great PPC tool to create Ads for Instagram and Tiktok. Wordstream - It’s an industry leading PPC Tool, great for Ad Grading and auditing. AdEspresso - This Is a tool by Hootsuite. They have a lot of Data sourced at the backend, which helps in Ad optimisation through this tool. That’s the reason I recommend this tool. #Video Editing Tools Veed Studio - I have been using Veed from last year. It’s one of the best Video Marketing Tool Optimized for Instagram & Tiktok. Synthesia - It’s a new AI video generation platform. From last few months, if you have seen marketing agencies including Videos in Emails. The chances are that’s not a Agency member taking but AI generated Human. Motionbox - It’s also a great video editing tool focused on video editing for Digital Marketers. Jitter Video - It’s a great motion design tool. Comes with great templates, the only place where other tools I mentioned lacks. It’s great and beginner friendly. #Copywriting Jasper AI - Google’s John Mueller says AI generated content is banned on Search but I think with Jasper AI you can generate SEO optimised Content but you have to put in some efforts like at least give 30 minutes for editing the Copy by yourself. Copy AI - Another AI tool to help you write better copy. This one is more focused on helping you write copy suitable for Ads and Social media campaigns. Hemingway App - To help you write more clearly and Bold. This tool is better than Grammarly if you look for writing perspective and it’s free. #Social Media Management App I’ve used a Lot of SMM Tools and that’s why going to mention all of them with a short review. Sprout social - The Best with deep insights coverage. Hootsuite - Great Scheduling tool just under sprout social. Later - Heavily Focused on Instagram from beginning and Now Tiktok too. SkedSocial - It’s like a Later alternative with great addition features like link-in-bio. Facebook’s Business Manager- Great but sometimes bugs can make a huge issue for you and customer support is like dead. Tweet Hunter & Hypefury- Both are Twitter Scheduling tools growing very fast on platform and are great for growth. Buffer - It’s a great tool but I haven’t seen any new updates to help with management. Zoho Social - It’s a great SMM tool and if you use other marketing solutions from Zoho. It’s a must have! #Market Research Tool • SparkToro - That’s the only one I have ever used. It’s great for audience research and comes with great customer service. Founded by Rand Fishkin, it’s one of the best research tool. #Influencer Marketing & UGC InfluenceGrid - A free search engine To find Tiktok & Instagram Influencers for your campaigns. Tiktok Creative Center- TikTok’s in-built tool called “Creative Center” is the best to find content trends, audience demographics and much more. Archive - Find Instagram Stories and Posts mentioning Your brands and use them as Ads for your business Marketing. #Landing Page Builders Leadpages - Its a great landing page builder because the integration and drag-and-drop features makes it easier to work with! Cardd co - A Great Landing page builder with easy step up but it lacks the copywriting and tracking features. Instapage - It’s one of the best out and I think the overall product is effective enough to help you stand out with your landing page. Unbounce - It’s a great alternative to Instapage due its well polished landing page templates that might be helpful for you. #Community Building Mighty Networks - A Great Community building platform, and you can also sell courses within the platform. Circle so - A great alternative to Mighty networks focused on Communities specifically. We are currently using for small community Of ours. #Sales Tools Drift - You can get much more out of Drift than just sales tools but The Sales solutions provided in Drift are one of the best. Salesforce - It’s the industry Sales solution provider. A go-to and have various pricing plans making it suitable for majority of SMBs. #Social Proof Tools People don’t have enough time to search across internet to decide to trust you after seeing your Ad first time. That’s what you might be facing too. Here are two tools I absolutely love for social proof! Use Proof - Show Recent Activities occurring on your website and build the trust of your visitors. Testimonial to - Gather Testimonials across Social Media platforms related to your business with this tool. Capture tweets and comments mentioning your brands and mention them. #Analytics Tools Plausible Analytics- A privacy friendly Analytics alternative to Google Analytics if you hate Analytics 4 like me. Mixpanel - Product Analytics and funnel reports better than Google Analytics. #Reddit Marketing Gummysearch- This tool will help To find your target audience on Reddit and interact with them with its help and close your new customers. Howitzer- It’s another pretty similar tool to Gummysearch focused on Reddit cold outreach to get clients and new customers. Both are great but Gummysearch provides better customer support while Howtizer is helpful on a large scale Reddit Marketing. #Text Marketing Klaviyo - It’s an email + SMS marketing tool, it’s taking up space in marketing industry very quickly as an industry leader due to its great integrations but you need to learn the platform usage to maximise the outcome. Cartloop - This tool provides great text marketing solutions with integration with Spotify and other e-commerce marketing tools. Attentive Mobile - This is my favourite Text marketing tool due to the interactive dashboard + they have a library of Text marketing examples to help you out with your campaigns. #Other Tools I have used throughout my journey! Triple Whale - It’s a great E-commerce marketing tools with Triple pixel to help you track your campaigns more efficiently. Fastory - To create well optimized Instagram & Tiktok Stories for your business. Jotform - Online Form Builder with integrations with leading marketing tools. Gated - As an entrepreneur and marketer, you may receive a bunch of unwanted emails. Use Gated to get rid of them and receive useful mails only! ClickUp- The main Tool for Project Management, one of the best and highly recommended. Riverside - Forget Zoom or Google Meet, For your Podcast Interviews and Marketing conferences. You need riverside with great video quality and recording features. Manychat- Automate your Instagram DMs and interact with your followers more efficiently + sell out your products/ services when you are offline. Calendy - To schedule meetings with your ideal clients. ServiceProviderPro - It’s a client portal for SEO & Growing Agencies, very helpful in scaling agencies. SendCheckit - Compare your Email Subject Lines with 100,000+ others in the database for free. Otter AI - Using AI track your meetings more effectively, you can easily edit, annotate and share notes from the meetings. Ryte - Optimise your website User experience with this tool focused on UX aspects + SEO too. PhantomBuster - Scrape LinkedIn Profile and Data from Facebook/LinkedIn groups. I clearly love this tool! #Honourable Mentions Zapier - The Only tool you need to integrate your favourite tool with a new effective tool. Elementor - That’s what I use for web design and it’s great! Marketer Hire - To hire world class marketers to work with you. InShot & Capcut - I create Instagram Reels and TikTok’s and life without these tools isn’t possible. Nira - It’s a great tool to Manage your workspace and this tool has launched many marketing templates in-built helpful for marketers and also entrepreneurs. X - The tool you love that wasn’t mentioned here is valuable and I honour that tool and share that if you would like to! I mean thanks for reading what I have curated all over my life as a marketer. I share 5 Marketing Tools, 5 Marketing Resources and 1 Free Resourceevery week in my newsletter, you can subscribe here to receive that for free. Also, You can read an expanded list of email marketing tools in this Reddit post!

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing
reddit
LLM Vibe Score0
Human Vibe Score0.857
adamkstinsonThis week

AI Content Campaign Got 4M impressions, Thousands of Website Views, Hundreds of Customers for About $100 — This is the future of marketing

Alright. So, a few months ago I tested a marketing strategy for a client that I’ve sense dedicated my life to developing on. The Idea was to take the clients Pillar content (their YouTube videos) and use AI to rewrite the content for all the viable earned media channels (mainly Reddit). The campaign itself was moderately successful. To be specific, after one month it became their 2nd cheapest customer acquisition cost (behind their organic YouTube content). But there is a lot to be done to improve the concept. I will say, having been in growth marketing for a decade, I felt like I had hit something big with the concept. I’m going to detail how I built that AI system, and what worked well and what didn’t here. Hopefully you guys will let me know what you think and whether or not there is something here to keep working on. DEFINING THE GOAL Like any good startup, their marketing budget was minimal. They wanted to see results, fast and cheap. Usually, marketers like me hate to be in this situation because getting results usually either takes time or it takes money. But you can get results fast and cheap if you focus on an earned media strategy - basically getting featured in other people’s publication. The thing is these strategies are pretty hard to scale or grow over time. That was a problem for future me though. I looked through their analytics and saw they were getting referral traffic from Reddit - it was their 5th or 6th largest source of traffic - and they weren’t doing any marketing on the platform. It was all digital word of mouth there. It kind of clicked for me there, that Reddit might be the place to start laying the ground work. So with these considerations in mind the goal became pretty clear: Create content for relevant niche communities on Reddit with the intent of essentially increasing brand awareness. Use an AI system to repurpose their YouTube videos to keep the cost of producing unique content for each subreddit really low. THE HIGH-LEVEL STRATEGY I knew that there are huge amounts of potential customers on Reddit (About 12M people in all the relevant communities combined) AND that most marketers have a really tough time with the platform. I also knew that any earned media strategy, Reddit or not, means Click Through Rates on our content would be extremely low. A lot of people see this as a Reddit specific problem because you can’t self-promote on the platform, but really you have to keep self-promotion to a minimum with any and all earned media. This basically meant we had to get a lot of impressions to make up for it. The thing about Reddit is if your post absolutely crushes it, it can get millions of views. But crushing it is very specific to what the expectations are of that particular subreddit. So we needed to make content that was specifically written for that Subreddit. With that I was able to essentially design how this campaign would work: We would put together a list of channels (specifically subreddits to start) that we wanted to create content for. For each channel, we would write a content guideline that details out how to write great content for this subreddit. These assets would be stored in an AirTable base, along with the transcripts of the YouTube videos that were the base of our content. We would write and optimize different AI Prompts that generated different kinds of posts (discussion starters about a stock, 4-5 paragraph stock analysis, Stock update and what it means, etc…) We would build an automation that took the YouTube transcripts, ran each prompt on it, and then edited each result to match the channel writing guidelines. And then we would find a very contextual way to leave a breadcrumb back to the client. Always as part of the story of the content. At least, this is how I originally thought things would go. CHOOSING THE RIGHT SUBREDDITS Picking the right communities was vital. Here’s the basic rubric we used to pick and prioritize them: • Relevance: We needed communities interested in stock analysis, personal finance, or investing. • Subreddit Size vs. Engagement: Large subreddits offer more potential impressions but can be less focused. Smaller subreddits often have higher engagement rates. • Content Feasibility: We had to ensure we could consistently create high-value posts for each chosen subreddit. We started with about 40 possibilities, then narrowed it down to four or five that consistently delivered upvotes and user signups. CREATING CHANNEL-SPECIFIC GUIDES By the end, creating channel specific writing guidelines looked like a genius decision. Here’s how we approached it and used AI to get it done quickly: Grabbed Top Posts: We filtered the subreddit’s top posts (change filter to “Top” and then “All Time”) of all time to see the kinds of content that performed best Compiled The Relevant Posts: We took the most relevant posts to what we were trying to do and put them all on one document (basically created one document per subreddit that just had the top 10 posts in that subreddit). Had AI Create Writing Guideline Based On Posts: For each channel, we fed the document with the 10 posts with the instructions “Create a writing guideline for this subreddit based on these high performing posts. I had to do some editing on each guideline but this worked pretty well and saved a lot of time. Each subreddit got a custom guideline, and we put these inside the “Channels” table of the AirTable base we were developing with these assets. BUILDING THE AI PROMPTS THAT GENERATED CONTENT Alright this is probably the most important section so I’ll be detailed. Essentially, we took all the assets we developed up until this point, and used them to create unique posts for each channel. This mean each AI prompt was about 2,000 words of context and produced about a 500-word draft. There was a table in our AirTable where we stored the prompts, as I alluded to earlier. And these were basically the instructions for each prompt. More specifically, they detailed out our expectations for the post. In other words, there were different kinds of posts that performed well on each channel. For example, you can write a post that’s a list of resources (5 tools we used to…), or a how to guide (How we built…), etc.. Those weren’t the specific ones we used, but just wanted to really explain what I meant there. That actual automation that generated the content worked as follows: New source content (YouTube video transcript) was added to the Source Content table. This triggered the Automation. The automation grabbed all the prompts in the prompt table. For each prompt in the prompt table, we sent a prompt to OpenAI (gpt-4o) that contained first the prompt and also the source content. Then, for each channel that content prompt could be used on, we sent another prompt to OpenAI that revised the result of the first prompt based on the specific channel guidelines. The output of that prompt was added to the Content table in AirTable. To be clear, our AirTable had 4 tables: Content Channels Prompts Source Content The Source Content, Prompts, and Channel Guidelines were all used in the prompt that generated content. And the output was put in the Content table. Each time the automation ran, the Source Content was turned into about 20 unique posts, each one a specific post type generated for a specific channel. In other words, we were create a ton of content. EDITING & REFINING CONTENT The AI drafts were never perfect. Getting them Reddit-ready took editing and revising The main things I had to go in and edit for were: • Tone Adjustments: We removed excessively cliche language. The AI would say silly things like “Hello fellow redditors!” which sound stupid. • Fact-Checking: Financial data can be tricky. We discovered AI often confused figures, so we fact check all stock related metrics. Probably something like 30-40% error rate here. Because the draft generation was automated, that made the editing and getting publish ready the human bottleneck. In other words, after creating the system I spent basically all my time reviewing the content. There were small things I could do to make this more efficient, but not too much. The bigger the model we used, the less editing the content needed. THE “BREADCRUMB” PROMOTION STRATEGY No where in my prompt to the AI did I mention that we were doing any marketing. I just wanted the AI to focus on creating content that would do well on the channel. So in the editing process I had to find a way to promote the client. I called it a breadcrumb strategy once and that stuck. Basically, the idea was to never overtly promote anything. Instead find a way to leave a breadcrumb that leads back to the client, and let the really interested people follow the trail. Note: this is supposed to be how we do all content marketing. Some examples of how we did this were: Shared Visuals with a Subtle Watermark: Because our client’s product offered stock data, we’d often include a chart or graph showing a company’s financial metric with the client’s branding in the corner. Added Supporting Data from Client’s Website: If we mentioned something like a company’s cash flow statement, we could link to that company’s cash flow statement on the client’s website. It worked only because there was a lot of data on the client’s website that wasn’t gated. These tactics were really specific to the client. Which is should be. For other companies I would rethink what tactics I use here. THE RESULTS I’m pretty happy with the results • Impressions: – Early on posts averaged \~30,000 apiece, but after about a month of optimization, we hit \~70,000 impressions average. Over about two months, we reached 4 million total impressions. • Signups: – In their signups process there was one of those “Where did you find us?” questions and the amount of people who put Reddit jumped into the few hundred a month. Precise tracking of this is impossible. • Cost Efficiency (This is based on what I charged, and not the actual cost of running the campaign which is about $100/mo): – CPM (cost per thousand impressions) was about $0.08, which is far better than most paid channels. – Cost per free user: \~$8-10. After about a 10% conversion rate to a paid plan, our cost per paying user was $80–$100—well below the client’s previous $300–$400. HIGHLIGHTS: WHAT WORKED Subreddit-Specific Content: – Tailoring each post’s format and length to the audience norms boosted engagement. Worked out really well. 1 post got over 1M views alone. We regularly had posts that had hundreds of thousands. Breadcrumbs: – We never had anyone call us out for promoting. And really we weren’t. Our first priority was writing content that would crush on that subreddit. Using the Founder’s Existing Material: – The YouTube transcripts grounded the AI’s content in content we already made. This was really why we were able to produce so much content. CHALLENGES: WHAT DIDN’T WORK AI is still off: – Maybe it’s expecting too much, but still I wish the AI had done a better job. I editing a lot of content. Human oversight was critical. Scheduling all the content was a pain: – Recently I automated this pretty well. But at first I was scheduling everything manually and scheduling a hundred or so posts was a hassle. Getting Data and Analytics: – Not only did we have not very good traffic data, but the data from reddit had to be collected manually. Will probably automate this in the future. COST & TIME INVESTMENT Setup: The setup originally took me a couple weeks. I’ve since figured out how to do much faster (about 1 week). AirTable Setup here was easy and the tools costs $24/mo so not bad. ChatGPT costs were pretty cheap. Less than $75 per month. I’ve sense switched to using o1 which is much more expensive but saves me a lot of editing time Human Editing: Because this is the human part of the process and everything else was automated it mean by default all my time was spent editing content. Still this was a lot better than creating content from scratch probably by a factor of 5 or 10. The main expense was paying an editor (or using your own time) to refine posts. Worth it? Yes even with the editing time I was able to generate way more content that I would have otherwise. LESSONS & ACTIONABLE TAKEAWAYS Reddit as a Growth Channel: – If you genuinely respect each subreddit’s culture, you can achieve massive reach on a tight budget. AI + Human Collaboration: – AI excels at first drafts, but human expertise is non-negotiable for polishing and ensuring factual integrity. Soft Promotion Wins: – The “breadcrumb” approach paid off. It might feel like too light a touch, but is crucial for Reddit communities. Create once, repurpose as many times as possible: – If you have blog posts, videos, podcasts, or transcripts, feed them into AI to keep your message accurate and brand-consistent. CONCLUSION & NEXT STEPS If you try a similar approach: • Begin with smaller tests in a few niches to learn what resonates. • Create a clear “channel guide” for each community. • Carefully fact-check AI-generated posts. • Keep brand mentions low-key until you’ve established credibility.

I’ve professionalized the family business. Now I feel stuck
reddit
LLM Vibe Score0
Human Vibe Score1
2LobstersThis week

I’ve professionalized the family business. Now I feel stuck

I wrote the post below in my own words and then sent to ChatGPT for refinement/clarity. So if it reads like AI, it's because it is, but it's conveying the message from my own words a bit better than my original with a few of my own lines written back in. Hope that's not an issue here. I’m 33, married with two young kids. I have a bachelor’s from a well-regarded public university (though in an underwhelming field—economics adjacent). I used that degree to land a job at a mid-sized distribution company (\~$1B annual revenue), where I rose quickly to a project management role and performed well. In 2018, after four years there, I returned to my family's $3M/yr residential service and repair plumbing business. I saw my father withdrawing from leadership, responsibilities being handed to underqualified middle managers, and overall employee morale declining. I’d worked in the business from a young age, had all the necessary licenses, and earned a degree of respect from the team—not just as “the boss’s kid,” but as someone who had done the work. I spent my first year back in the field, knocking off the rust. From there, I started chipping away at process issues and inefficiencies, without any formal title. In 2020, I became General Manager. Since then, we’ve grown to over $5M in revenue, improved profitability, and automated many of the old pain points. The business runs much smoother and requires less day-to-day oversight from me. That said—I’m running out of motivation. I have no equity in the business. And realistically, I won’t for a long time. The family dynamic is... complicated. There are relatives collecting large salaries despite zero involvement in the business. Profits that should fuel growth get drained, and we can’t make real accountability stick because we rely too heavily on high-producing employees—even when they underperform in every other respect. I want to be clear—this isn’t a sob story. I know how lucky I am. The business supports my family, and for that I’m grateful. But I’ve gone from showing up every day with fresh ideas and energy to slowly becoming the guy who upholds the status quo. I’ve hit most of the goals I set for myself, but I’m stagnating—and that scares me. The safe move is to keep riding this out. My wife also works and has strong earning potential. We’re financially secure, and with two small kids, I’m not eager to gamble that away. But I’m too young to coast for the next decade while I wait for a possible ownership shakeup. At this point, the job isn’t mentally stimulating. One hour I’m building dynamic pricing models; the next, I’m literally dealing with whether a plumber is wiping his ass properly because I've had multiple complaints about his aroma. I enjoy the challenging, high-level work—marketing, systems, strategy—but I’m worn down by the drama, the legacy egos I can’t fire, and the petty dysfunction I’m forced to manage. I'm working on building a middle management gap, but there's something lost in not being as hands-on in a small business like this. I fear that by isolating myself from the bullshit, I'll also be isolating myself from some of the crucial day-to-day that keep us who we are. Hope that makes sense. (To be fair, most of our team is great. We have an outstanding market reputation and loyal employees—but the garbage still hits my desk when it shows up.) I’ve toyed with starting a complementary business or launching a consulting gig for similar-sized companies outside our market. I’ve taken some Udemy and Maven Analytics courses (digital marketing, advanced Excel/Power BI, etc.) to keep learning, but I rarely get to apply that knowledge here. So here I am. Is this burnout? A premature midlife crisis? A motivation slump? I’m not sure what I’m looking for—but if you’ve been here, or have any hard-earned advice, I’d be grateful to hear it.

I spent 18 hours every week tracking marketing trends and latest news. Here are my predictions for 2024
reddit
LLM Vibe Score0
Human Vibe Score0.778
lazymentorsThis week

I spent 18 hours every week tracking marketing trends and latest news. Here are my predictions for 2024

1/ Securing Digital Footprint becomes #1 Priority For Chronically Online Users, Protecting their digital footprint will become one of the main things. We saw influencers getting cancelled over Old Content and Brands used Old Travis Kelce Tweets, we saw what could happen without digital footprint protection. Online Engagement Precautions will be taken again with Twitter & IG showing your usernames above ‘Algorithm Suggested Content’. What you like is more visible to other people in UI Design of these apps, another reason behind why Digital Footprint preservation will matter a lot in 2024. This will impact likes to viewership ratio on your organic and paid content. &#x200B; 2/  TikTok wants Long Videos with Storytelling As I was writing this report, TikTok also released their What’s Next 2024 Report. It focuses heavily on how the audiences on the app demand better storytelling and from the examples in the report, you can judge what TikTok wants. They also rolled out a 30-minute video upload limit. Engaging Content over 1-Minute Mark to keep the audiences longer on the app. I highlighted in the first trend, every social media platform wants the same thing, more time spent. 3/ Use of Shop the Look While Streaming Netflix or Amazon Prime. This year’s one of the most successful TV series, The Bear caused Men to go mad for the T-Shirt worn by Jeremy Allen White in the show. Showing us how TV Shows influence or encourage us to dress in a particular way. It’s nothing new, TV Shows like Friends & Gossip Girl influenced all demographics when they came out. But now, Streamings Services such as Roku & Amazon enable consumers to shop the look while watching the TV Shows. Many Brands will jump on these opportunities in upcoming months. 4/ Brands in Comments & Memes are the new norm By Summer 2024, Most Online Users & Creators will no longer feel too excited or answered when they see your brand in the comments. Why? It’s becoming too common for Brands to show in comments under viral content about them. Or Brands being funny with Internet Culture Trends is known to most users. The Saturation of Every Brand being funny and being present leads to increased competition of levitating the content quality. &#x200B; 5/ Marketers decrease their focus on Traffic & Views With AI recommendations taking over, The Structure of content distributing on social media is changing, the same goes for SEO. Conversational AIs are changing how web traffic is distributed to publishers. An Increased focus on managing the conversion rate and landing page relevancy will be the main focus. 6/ OOH is kind of making a comeback. First, US OOH Ads Industry grew 1.1% in Q3 2023. Second, Outfront Media reported slight revenue increase in Q3 as Billboard Ad Revenue grew in Q3. Many Brands in UK are also aligning more toward traditional media Channels. With Burger King in UK focusing on only OOH for Christmas this year and Fashion Brands like SSENSE launching Billboards as Branding Play. 7/ Rise of Curation Continues This Year, we witnessed success of Pinterest Shuffles App, Gen-Z loved it. Similar Success with formats like IG photo dump & TikTok ‘My Fav Finds’ Carousels being the center of Gen-Z Content. Just look at this recent trend and tell me Curation isn’t personal to Online Teens. Spotify won with their idea of curating Songs with Astrology-type signs. The Fashion Products with Curated Emojis and Stickers on them, that scrappy curated approach is predicted to grow in 2024, data from Pinterest. 8/ Use of AI to Trace Consumers in the wild This year we saw a huge trend of people using Image/ face recognition tools to find or dig dirt about famous people. The biggest example was Dillion Dannis exposing Multiple images of Logan Paul’s girlfriend using AI tools. (Which was Obviously bad) But next year, I believe with better rules, big brands like Adidas or Nike will be able to find worldwide micro-influencers & Online Consumers seen wearing adidas. And partnering with them on a large scale through automated outreach. 9/ More Cartoons than Influencer-Brand Products. All the Cartoon shows are seeing huge rise on IG and TikTok, Shaun the sheep is viral, Snoopy was big this year, Sesame Street’s TikTok is working. Aussie Show Bluey is making a huge spark in the US. More Brand collaborations are on the road. Why? Cartoons have built a very consistent identity and they have social channels. I know many see Cartoons as Kids Content but on social, looking at TikTok Account of Sesame Street & Snoopy. Last month, Powerpuff Girls launched a collaboration with Nike. &#x200B; 10/ The Best Trend to get people off social media &#x200B; Try to get people off the social media apps, build your own loops. You can’t rely on social and you clearly shouldn’t burn out trying to win on social and streaming with Paid Ads or without them. This matters a lot because data shares most of your customers buy from you once or twice a year. And then they interact with your content, how bad will you feel if the only thing they remember as your content is being on TikTok. Nothing about your brand. 11/ The Internet Aesthetic will Die for Cafes & Restaurants When I wrote my post about Instagram Marketing, I mentioned this issue of Every Account looking the same. In reality, It isn’t limited to IG Feeds, This Creator points out the same Problem, mentioning the aesthetic Standards from Internet are changing how new businesses approach their whole business. More Content from Cafes & Restaurants need to be around their people and neighbourhood. 12/ Echo Chambers & Sonic Influence All Podcasts are Echo Chambers because if people wanted a new perspective in form of value. We would have chosen debates, but we chose Podcasts to find new value while being in comfort. People are now looking for more value in comfort than ever, Podcasts will continue to rise. 13/ Clever AI Integration to Better Customer Journeys in B2B & B2C Marketing Agencies can provide clever solutions to B2B Companies, and help them overcome the tag of Boring Ads only. How? Ogilvy India created an AI Ad Campaign for Cadbury, allowing SMBs to have the Bollywood Actor endorse them. They used the AI voice generation allowing businesses to alter the voice and have Shah Rukh Khan endorse their shop. A similar approach was taken by IPG India, An AI Ad with Shah Rukh Khan allowing everyone to add their face in the Branded Content. &#x200B; If I sounded like an Old head in this report or I missed on some elements like Programmatic Advertising and PPC. I will try to include better analysis and new content about future trends. You can find the post shared with examples & research, linked here.

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.
reddit
LLM Vibe Score0
Human Vibe Score1
CALLIRDAN90This week

Only 2 months of cash in the Bank for my business but was able to save it with the help of AI.

Hi there! I’m excited to share something very personal with you. We needed to book at least 2 appointments per day in the next 60 days, or my business would fail. We were already trying two acquisition channels, LinkedIn and email. The problem with these channels was that the positive response rate was very low in both. So I decided to focus on LinkedIn and get the attention of the lead by sending videos directly to them via LinkedIn messages. (You can send videos to your connections on LinkedIn if you use your cell phone.) This wasn’t new, but I added a small twist to get the lead’s attention. All the covers of the videos had a picture of me holding a sign with the person’s name and an interesting phrase. This showed some okay results, but the rest of the video was not personalized. Only the picture on the cover was. I even developed a Chrome extension for this because I thought this would be the answer and that I would book tons of appointments.  But after more trial and outreach, my leads responded, telling me that because the video itself wasn’t personalized for them, they felt like I didn’t put enough effort in, so they would not book a call with me. So after investing time and effort into my “new bright idea” and getting developers to make the Chrome extension, I was back to square one with no results. A few weeks went by, and after researching online, I found an online course from a guy who promised to teach me how to book 30+ appointments per month, guaranteed (at the time, I was making 2 or 3 appointments per week, maximum). He promised that I would only pay if he actually booked appointments for me and even offered to give me money if his course didn’t work for me. I never paid attention to internet gurus, but the offer was actually not bad, so I looked into this guy’s website. I found out he had hundreds of reviews from people who had taken his course and were talking amazing things about it. The more I read, the more excited I got. I booked a call that day and talked to a salesperson. The call was very short, and he promised I would get at least 2 appointments per day, easily. He seemed a bit cocky and told me that I just needed to trust him and the 100+ reviews from people who had taken the course. He didn’t share details, a proposal, or anything. I asked the price, and he told me it was close to $10k. (Not kidding, this was the price.) Then he told me that I would make the money back in no time with the clients I would get following his course, and that if it didn’t work, he would give me the money back. But I needed to follow everything the course said for at least 6 months. I had never paid $10k for anything in my life; it was extremely expensive for me. Also, my salary from my business was not in dollars but in a currency that was worth much less than the dollar. I continued to research more and more, but no other course was close to the number of reviews and promises that this guy had. I got desperate and told myself that I would bet everything on this course. If it worked for so many others, surely it would work for me. I got a loan from the bank and paid for the course. You might read this and think it was the most stupid thing ever, but the reality is that after 2 months in the course (I did the course as fast as I could), I learned a lot. The course was not bad; it was very extensive—probably more than 200 hours or so—and they taught a lot of things. I don’t think it was worth $10k for me, but I can see how for other people it might be worth that. Now, to the question you’re all thinking: did it get me the 2 appointments I needed per day? The answer is no. Here’s the thing: most of the techniques they taught were innovative and disruptive, but the focus was always on personalization, and they didn’t teach any way to automate the personalization. (I think, at the time they made the course, the tools didn’t exist yet.) So they taught how to do everything manually, and it took a lot—a lot of time and effort. And most annoyingly: an incredible amount of time doing operational things. I did get 2 appointments on some days, but it wasn’t consistent, and I didn’t have the time to spend 14 hours a day doing everything manually or the money to hire someone to do this for me. (I needed to also spend time delivering our service to our current clients; otherwise, they would leave.) I told them this, and they were very reasonable. After some negotiation, they gave me part of the money back. (To be fair, there was a lot of value in the course, so asking for the full $10k back would have been excessive because, in the end, it really taught me a lot of things I didn’t know.) So in the end, I spent $10k and 200+ hours on an online course, spent time and effort developing a Chrome extension, and was still not able to hit the meetings I needed. Money in the business was running out, and I needed to do something fast, or I was doomed. After investing time and effort in tools, research, and spending $10k and over 200 hours on a course that didn’t deliver the consistent results I needed, I was at a crossroads. My businesses were running out of money, and I knew I needed to find a solution quickly, or everything I had worked for would collapse. It was during this time of desperation that I started exploring other options. One night, while scrolling through the internet, I stumbled upon a 2024 article about how AI was being used to revolutionize various industries. It wasn’t directly related to appointment booking, but it sparked an idea in my mind. What if I could use AI to automate the personalization process that I had learned in the course? It seemed like a long shot, but I had nothing to lose. I started researching AI tools and technologies—YouTube videos, podcasts, pretty much everything related to AI—desperate to find something that could help me scale my outreach without investing too much time, while still maintaining the personalization that was so important. After a lot of trial and error, I found a few tools that showed promise. All of these tools were extremely new. Some of them had just launched the versions I needed just weeks ago. I can say I researched and tested more than 50 AI startups, experimenting with them, testing different approaches, checking prices (the problem was that most of them were cheap but became very expensive when applying the volume I needed to get results), and gradually refining my process. It wasn’t an overnight success, but for the first time, I felt like I was onto something that could truly work. The idea of combining AI personalization with volume was something new, and it gave me hope that I could finally book the meetings I needed without burning out. One day, I sent a video of myself talking—completely AI-generated—to my family chat group and waited for their response. None of them noticed it wasn’t actually me. At that moment, I said to myself: “Okay, I am ready to test this in the real world and see if it works.” Like everything in life, focus is key. As I mentioned earlier, we were already trying outbound strategies on LinkedIn and email, but I decided to narrow my focus to LinkedIn and specifically to video outreach. My goal was to stand out from the crowd, where most people were using text or sending generic videos. I knew that if my videos were 100% personalized, it would make a strong impression on my leads. I focused on two key metrics during my tests: Time spent on manual personalized outreach vs. AI-generated personalized outreach. Positive reply rate for non-personalized manual outreach vs. AI-generated personalized outreach. I ran a test using a sample of 50 one-minute videos sent to 50 leads, and here are the results: Time Spent to Make the Videos: Manual Process: It took me up to 10 hours to create and send 50 personalized videos. This included looking good on camera, brushing my hair, choosing appropriate clothing, ensuring proper lighting, not messing up the script, using a camera holder, recharging the phone, pausing to drink water, avoiding external sounds, being in an appropriate room, downloading the videos, deleting the videos that were not good, and sending the final ones. On average, it took me at least 12.5 minutes per one-minute video. AI Process: With AI, it took me just 32 seconds to create the exact same one-minute personalized video—without saying a word or recording a second of footage. In total, I could make and send the same 50 personalized videos in just 27 minutes. Result: The AI process was 24 times faster. Completely crazy! Positive Reply Rate: Non-Personalized Script (Manual): Using a good script without personalization (no name, job title, city, company, etc.) resulted in a positive reply rate of 4-6% on LinkedIn, including follow-ups. Personalized Script (AI): Using the same script but adding personalized details like the lead's name, company, city, and job title resulted in a positive reply rate of 15-20%, including follow-ups. Result: AI personalization led to 3x (three times) more replies. The best part was the responses. Almost everyone who replied thanked me for taking the time to research them, congratulated me on my speech, and appreciated the personalization and eloquence of my message.  These metrics were a complete breakthrough for me. I researched online to see if anyone else had done something similar, but I couldn’t find anything close. After achieving these metrics, booking the two appointments I desperately needed became easy. In fact, in the last 10 weeks, I’ve been able to consistently book 3-4 appointments per day. This success allowed me to train someone in my company to handle the process, freeing me up to focus on other aspects of the business and ultimately saving it. With the AI appointment machine we built, I even have free time now—time that I’ve been using to develop a methodology and tech tools that I now teach to others. I named the methodology Clip2Lead as a reference to the first Chrome extension I developed that didn’t work but ended up being the first step toward everything that followed. I’ve condensed everything I learned and throughout my experiences into a simple and short FREE training where I cover the entire AI appointment booking process. This includes how to find leads, create scripts, set up follow-up sequences, generate AI videos, clone your voice, compare non-AI metrics with AI metrics, and even navigate AI safety controls. I also offer Chrome extensions that helped me automate the process even further, so you can spend your time closing deals or focusing on other acquisition channels, while your AI machine for booking appointments runs with minimal effort from you. If you’re interested please get in touch with me and thank you for taking the time to read my personal story.

This founder was about to shut down his business and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What other businesses can scale like this?
reddit
LLM Vibe Score0
Human Vibe Score1
CountryPitifulThis week

This founder was about to shut down his business and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What other businesses can scale like this?

I heard that Jasper scaled to $45m ARR in 12 months...with a team of 8. For context, they are one of the fastest-growing companies ever. Grew from $0 to $45m ARR in 12 months (then raised $125m at a $1.5b valuation). As a fellow founder, their story is really inspiring to me (curious about what others think): In December 2020, Dave Rogenmoser and his co-founders were on the brink of shutting down their business. They'd spent 3+ years building a conversion optimization software called Proof...and it was flatlining. A few weeks prior they had to make the painful decision to let go of half their team. Competition and churn had completely eroded growth. Things were painful. 8 years of work left them with a string of startups that never quite made it: 2 failed software businesses (couldn't make money*) A SMB marketing agency (maxed out at $25k/mo*) An online course company (hard to get big*) The Pivot: In January 2021, they had an idea to use Chat GPT-3, the generative AI model released 6 months earlier, to write high-converting Facebook ads. Within 30 days, they launched the business. With the skeleton crew remaining from the last startup, they scaled the business to $45m ARR and 70,000+ customers without hiring a single new person. Soon after, they raised $125m at a $1.5b valuation. Dave Rogenmoser, CEO at Jasper, had some great one-liners in a few podcasts I listened to on the business. Here are some of his learnings: Right Skill, Wrong Vehicle: He spent 8 years building marketing businesses which gave this team the knowledge and confidence to spend $1m/mo on sales and marketing to scale the business to $45m ARR in year 1. Launch Fast & Iterate Quickly: The team agreed that if the business didn't work in 30 days, they'd shut it down. Dave says, "If you have been working on a problem for more than 18 months and haven't found Product market fit (PMF), odds are you won't...Make the hard pivot."* Ride A Big Wave: Generative AI technology is a new technology that is changing the way we work. But it's not just text. It's images, voice, etc. Identify new customer segments (e.g., Municipalities, Banks, Lawyers, etc.), learn their problems, and apply this novel technology to solve them. What other businesses have you seen scale like this? I've never seen a SaaS business grow that fast. I meet interesting founders 2x per week and share the learnings here.

I’m building a “DesignPickle” for all things Funnels. Would love your feedback...
reddit
LLM Vibe Score0
Human Vibe Score0.846
Gluteous_MaximusThis week

I’m building a “DesignPickle” for all things Funnels. Would love your feedback...

Hey Entrepreneurs, Early next year I’m rolling out a productized service business along the lines of Design Pickle, but instead of design assets, we create on-demand marketing assets: Things like landing pages, lead magnets, email campaigns, etc. This is NOT an agency with client engagements, etc.  It is an on-demand, menu-item style fulfillment platform where we do a few predefined things really, really well, and as much as possible try to reduce the complexity (and required customer inputs) so that creating your next killer Funnel is as easy as ordering dinner on Skip the Dishes. Below I’ve laid out our current thinking (we’re still distilling this into a deck), just so you have the full context.  And at the end, I pose 5 feedback questions. So if this “deck” seems interesting to you, then I’d love to get your feedback at the end 🙂 Thanks! And here goes... \--- The current elevator pitch:  We will research your business, your market and your competitors to develop a killer Lead Magnet, Landing Page, Ad Creatives and a 30-Day Email Drip campaign designed to turn your traffic into a rabid, lifelong buyer tribe (that you can email for years... like having your own, on-demand cash printer).  The overall thesis:  While AI is getting continually better at creating things like one-off graphics, article content, and so on - we do not think it can deeply understand market psychology, what keeps your customers up at night, or the underlying emotions that drive purchase decisions at the individual level, for your specific offer(s). Moreover, it’s also this psychological aspect of marketing where most businesses simply do not have the talent, resources or frankly the experience to create high-performing funnels themselves, regardless of how much "automation" they might have at their fingertips. And that’s because this is where you need to know who your customer really is, and what they’re actually buying (hint: not your features). Few marketers focus on these fundamentals, let alone understand the selling process. This is also why tools like ClickFunnels, HighLevel, LeadPages, etc. while very helpful, can only help with the logistics of selling. It’s still on each business to figure out how to actually tell their story, capture demand, and sell effectively. This is why a productized service that nails market research, competitor analysis & world-class copywriting that can actually turn cold traffic into lifelong customers is going to be a no-brainer for a business that’s currently struggling to actually get a steady flow of online sales. This is not something we see AI replacing effectively, any time soon. Current gaps & unknowns:  At a top level, I’m not overly worried about validation or viability; there are several existing competitors, and obviously the automation platforms have substantial customer bases (ClickFunnels etc). There will be a certain cohort that will want experts to do the actual thinking for them, storytelling, etc. Even if it’s a relatively small cohort, given the CLTV of a service like this, it still makes for a decent sized business. But where I’m less confident is in who our ideal customer actually is... Yes, basically every direct-response internet business needs an effective funnel that can sell. Whether you’re an Enterprise SaaS platform or a solopreneur launching your first $39 ebook, you will benefit from a killer funnel. As a “DesignPickle” type service though, here’s the challenges I see with each core customer category... B2B SaaS: While sales decisions are still emotional, it’s more about account-based considerations; people usually aren’t spending their own money, so it’s more about not looking stupid vs. gaining some benefit. Harder to systemize. Very high stakes. Consumer / SMB SaaS: While I think in general these are ideal customers, there will be resistance to leaning in hard on personality (and personal brand); founders usually want to sell at some point, so if they become the face of the platform, then boosting performance with a high-personality funnel might ironically make it a harder business to sell. SaaS founders are also generally very technical and stereotypically avoid marketing like the plague. Ecommerce: Most DTC brands think of funnels as an extension of their FB ad campaigns; few see their customers as a long-term audience that can become a significant asset. However, certain lifestyle / luxury brands might differ. Online Courses / Coaches: Of all the customer profiles, this group probably has the most appreciation for the effectiveness of marketing psychology, copywriting, etc. and would get the value prop quickly. The problem is that most won’t have the budget or traction to outsource asset creation. This is the “poorest” segment of the market. Service Businesses: Agencies, consultancies, and so on would greatly benefit from having a strong personal brand + storytelling premise (funnel). However, they’re also the worst offenders when it comes to never practicing what they preach / do for others. Client work soaks up all their resources. Local & Brick/Mortar: Generally speaking most local businesses are going to have smaller audiences (email lists under 2K subs), where funnel ops might have limited value long-term due to a lack of scale. And for larger B&M brands with franchises across various locations, you get into stakeholder friction; messaging usually gets watered down to basic corporate-speak as a result. Now, to be clear, I still see a ton of opportunity in each of those main customer categories as well, but I like to be clear-eyed about the overall resistance each niche will have - mainly because this helps to refine messaging to an ideal customer profile within them. In this case though, so far, nothing’s really jumping out at me as a clear “winner” at a category level. So far, what I’m thinking is our ICP might be situational / conditional. For example: A business has a funnel / is invested in the process, but it’s not working yet A business sees their competitor killing it with a funnel, and they’re ultra motivated to do it even better A business has one funnel that’s working awesome, and everything else they try sucks (so they can’t scale / expand) Etc. Basically, our most ideal customer might be ANY type of business who gets it, who’s tried to do this themselves, and now needs the pros to come in and fix things. \--- This is where your feedback would be incredibly valuable... First, if you’ve made it all the way down to this point - thanks for enduring my rambling mess above! But I did think the context might be helpful. Based on our overall biz plan & go-to-market considerations discussed above, if you run a business (or work with one) that might benefit from something like this, I’d love to ask a few questions... What is the nature of your business? (What do you sell)? What do you find hardest about selling to your online audience? Have you built a funnel in the past / are you running one currently? If not, what’s stopping you from building a high-performing funnel? If you had a “magic marketing lamp” where a genie could create ONE amazing marketing asset for you (eg. a killer landing page, video ad, launch strategy, etc), but you could only use it ONCE, what would you have the genie do for you? Please reply below as a comment, or DM me if you’d prefer to keep answers anonymous.  Thanks so much And again, apologies for the novel... Cheers

This founder was about to shut down his startup and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What else have you seen grow that fast?
reddit
LLM Vibe Score0
Human Vibe Score1
CountryPitifulThis week

This founder was about to shut down his startup and open a restaurant. He pivoted the business and grew it to $45m ARR in 12 months. What else have you seen grow that fast?

I heard that Jasper scaled to $45m ARR in 12 months...with a team of 8. For context, they are one of the fastest-growing companies ever. Grew from $0 to $45m ARR in 12 months (then raised $125m at a $1.5b valuation). As a fellow founder, their story is really inspiring to me (curious about what others think): In December 2020, Dave Rogenmoser and his co-founders were on the brink of shutting down their business. They'd spent 3+ years building a conversion optimization software called Proof...and it was flatlining. A few weeks prior they had to make the painful decision to let go of half their team. Competition and churn had completely eroded growth. Things were painful. 8 years of work left them with a string of startups that never quite made it: 2 failed software businesses (couldn't make money*) A SMB marketing agency (maxed out at $25k/mo*) An online course company (hard to get big*) The Pivot: In January 2021, they had an idea to use Chat GPT-3, the generative AI model released 6 months earlier, to write high-converting Facebook ads. Within 30 days, they launched the business. With the skeleton crew remaining from the last startup, they scaled the business to $45m ARR and 70,000+ customers without hiring a single new person. Soon after, they raised $125m at a $1.5b valuation. Dave Rogenmoser, CEO at Jasper, had some great one-liners in a few podcasts I listened to on the business. Here are some of his learnings: Right Skill, Wrong Vehicle: He spent 8 years building marketing businesses which gave this team the knowledge and confidence to spend $1m/mo on sales and marketing to scale the business to $45m ARR in year 1. Launch Fast & Iterate Quickly: The team agreed that if the business didn't work in 30 days, they'd shut it down. Dave says, "If you have been working on a problem for more than 18 months and haven't found Product market fit (PMF), odds are you won't...Make the hard pivot."* Ride A Big Wave: Generative AI technology is a new technology that is changing the way we work. But it's not just text. It's images, voice, etc. Identify new customer segments (e.g., Municipalities, Banks, Lawyers, etc.), learn their problems, and apply this novel technology to solve them. What other businesses have you seen scale like this? I've never seen a SaaS business grow that fast. I meet interesting founders 2x per week and share the learnings here.

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

Need help with the growth I couldn't handle
reddit
LLM Vibe Score0
Human Vibe Score1
luxendaryThis week

Need help with the growth I couldn't handle

Calling all innovators, dreamers, and disruptors! &#x200B; We're pioneering a new frontier in the world of manufacturing with our vision: "Text to Product". I'm seeking individuals passionate about AI, manufacturing, efficiency and automation. While we can't promise immediate financial rewards, we're offering equity in a venture that's setting out to redefine the way things are made and sold. If the prospect of revolutionizing the future of humanity excites you, we'd love to hear from you. &#x200B; &#x200B; P.S. I realized that I can't always use "brute force" for solving problems, so seeking "the right connections" (seasoned entrepreneurs, advisors). Here's the TLDR version of my story: Started a company with ex-boss, bought him out, grinded for 2 years, found a way to 1000x the orders.* Went full speed for a month, got overwhelmed, barely kept up with half the demand (with that production process).* Focused on this one "platform", shipped hundreds of thousands of units in one holiday season.* Next quarter "the platform" returned about 85% of products as "overstock", demanded money back, made legal threats.* I told them that I will go to court and they stopped bothering me.* Then Covid + Nasty divorce which made me put a pause to regroup.* 2 years later, with 2x the production capacity and after relocating to a friendlier state (from NYC to MIA) I'm ready to relaunch (with a clear head, knowledge of fast growth and what to avoid).*

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.
reddit
LLM Vibe Score0
Human Vibe Score1
Alarmed-Argument-605This week

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.

Apart from the story I need a suggestion and validation here. It's a bit long, skip to tl;dr if you couldn't handle length. A few days ago, I saw a person on Reddit sharing his struggles that, Even after generating a lot of leads from ads of Meta and Google (even with lowest cpc cpa cpl), he was not able to convert them into sales. Out of curiosity I dm'ed him with all fancy services that I offer and expressed that as a agency I would work with him for monthly recurring fee. He suggested for one time consulting fee, I agreed. It was literally a eye opener for me. This guy is in coaching business offering courses for people. His niche was too vague. Courses were on mindset coaching, confidence and public speaking coaching, right attitude coaching, manifestation coaching and all crap shits related to this. At first I thought he was not getting sales because who will pay for all this craps. I openly discussed with him that he has to change what he offers because, if I saw this ad I wouldn't buy this for sure. He then showed me how much money people offering similar service are making . I was literally taken back. He was part of a influencer group (the main guy who encourages these guys to start coaching business, looks like some mlm shit) where people post their succes stories. Literally lot of guys were making above 150k and 200k per month. Even with very basic landing page and average offer They are still winning. Here's where it gets interesting. I tried to clone everything that the top people in this industry are doing in marketing from end to end.( like the same landing page, bonus offers around 50k, exclusive community, free 1 on 1 calls for twice a month).Nothing worked for a month and later surprisingly even the sales started dropping a bit more. I got really confused here. So to do a discovery I went and purchased the competitor course and Man I was literally taken back. Like he has automated everything from end to end. You click the ad, see vsl, you have to fill a form and join a free Skool community where he gives away free stuffs and post success stories of people who took the course. Now every part of this journey you will get a follow up mail and follow up sms. Like after filling the form. after that now if you join and don't purchase the course you will be pampered with email and sms filled with success stories. For sure anybody will be tempted to buy the course. Here is the key take away. He was able to make more sales because he was very successful in nurturing the leads with follow ups after follow ups. Even after you purchased his course he is making passive income from 1 on calls and bonus live webinars. So follow ups will be for 1 on 1 calls and webinars after the course is over. Core point is our guy even after spending 2 to 3k per month on ads was not able to bring huge sales like competitors because he failed the nuture them. Even after making the same offers and the same patterns of marketing as competitors, the sales declined because people thought this is some spam that everyone is doing because the template of the ads was very professional and similar. suprising one is people fall for basic templates thinking it's a underrated one. so what we did here is we integrated a few softwares into one and set up all same webinars, automated email and sms follow ups, ad managers for stats, launched him a free LMS platform where without any additional fees so he can uploaded unlimited courses, skool like community and add product's like Shopify ( he was selling few merchandise with his brand name on) where he can add unlimited products with connection to all payment gateway, integrated with crm with unlimited contacts, workflow and lead nurturing with calender syncing for 1 on 1 calls. But these are a bit old school, what we did was even better. integrated a conversational ai with all of his sales platforms and gave a nocode automation builder with ai for the workflow. we also set him up with a ai voice agent that's automatically calls and markets for his course and also replies for queries when called. we also set up him a dedicated afflitate manager portal with automated commissions. Though he didn't cross 100k Mark, He did a great number after this. He was struggling with 6k sales, now he has reached somewhere mid of 45k to 50k mrr. Max he hit was 61.8k. I see this a big difference.So one small thing, nurturing the lead can bring you immense sales. To set up all of this it costs around 1.2k monthly for me with all the bills. ( I know there are few free for Individual user platforms out there, but It gets very costly when you switch to their premium plans. with heavy volumes you would require more than premium they offer.) I offered him like 3k per month to work as a agency for him who takes care of all these stuffs. He declined and offered for one time set up fee stating that he will pay 1.2k directly. The one time fee was also a bit low, though I agreed since this was a learning for me. what happened next after that is, he referred me to a few other people in the same niche. But the problem is they are not interested in spending 1 to 2 k in bills for software. They requested that if, will I be able to provide the saas alone for less than 500 dollars with one time set up fee. I haven't responded yet since I have to take an enterprise plan for all the software used and pay full advance price for billings. Then to break even that I have to make minimum 50 or odd users for that. let's grantly say 100 users with all other future costs. So here's what I'm planning to do. I'm planning to offer this as saas for let's say 239 dollars per month. with may or may not one time set up fee. ( I checked the entire internet, there is no single person offering at this price point for unlimited. Also one can easily start their marketing agency with this.) The suggestion and validation that I need here is 1.are you going through the same struggles or faced these struggles? would you be interested to buy at 239 dollars per month? let's say you're from a different niche, Did the features I told were okay for you or you need something specific for your industry that you will be interested in buying? please answer in comments and if you will purchase for this price let me know in comments/dms. I will take that into account and if the response rate is above 100 queries, then will integrate this and sell for that price. (ps: If you see this post on similar subs, please bear cause I'm trying to get suggestions from different POV) tl;dr - lead nurturing can massively boost sales *I made a software integration for a client for a 1.2k per month billing and here I want to know if more than 100 people are interested so that I will make this into my own saas and sell it for like a cheap price of 239 dollars per month TIA.

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰
reddit
LLM Vibe Score0
Human Vibe Score1
benfromwhereThis week

From Setbacks to $20K Profit: My AI Influencer Earnings Breakdown (Jan 2025) 💰

(Monthly income breakdown is in the end) 📌 Introduction Hey everyone! 👋 Before I dive into this month’s breakdown, I just want to be upfront—English isn’t my first language, so I’ve used ChatGPT to refine this post for better readability. That said, everything here is 100% real—my personal experiences, struggles, and earnings as someone running a full-time AI influencer business. Since I get a lot of DMs asking about my AI models, here are their Instagram links: 📷 Emma – https://www.instagram.com/emmalauireal 📷 Jade – https://www.instagram.com/jadelaui (jadecasual is the second account) Also, if you’ve been wondering about the community I run, where I teach others how to build AI influencers from scratch, here’s the link (I got approval from mods for this link): 🔗 AI Winners Now, let’s get into what happened this month. 🚀 \------- First, a huge thank you! 🎉 Three months ago, I shared my journey of building an AI influencer business, and I was blown away by the response. That post got 263K+ views and was shared over 2.7K times—way more than I ever expected. If you’re new here or want to check out the full story of how I started, you can read it here: 🔗 Click Here (Reddit link) \------- 🔹 What I Did in January After the holiday rush in December, I knew January would be a slow month—people had already spent most of their money at the end of the year. So instead of pushing harder on monetization, I shifted my focus to tech development and optimization. Flux Character Loras: I spent a lot of time refining and testing different Flux-based character Loras for my models. This is still a work in progress, but the goal is to improve long-term consistency and make my workflow even more efficient. NSFW Content Expansion: On Emma’s side, I expanded her content library using a real model body double, making her content look more organic and natural. Jade, however, remains 100% AI-generated, keeping her workflow entirely digital. Social Media Wipeout (Thanks, VA 🙃): I had handed off both Twitter accounts to a virtual assistant to help with engagement and DMs. Big mistake. He ended up spamming DMs, which got both accounts banned—Emma (80K followers) and Jade (20K followers). 🤦‍♂️ Right now, I’m rebuilding Emma’s account from scratch and taking a much more cautious approach. Jade’s account is still offline for now. New Platform: Threads – I hadn’t touched Threads before, but since engagement on Instagram can be unpredictable, I decided to start accounts for both models. So far, they’re performing well, and I’ll continue experimenting. Launched AI Winners Community: After getting flooded with DMs (both here and on Instagram), I realized there was a massive demand for structured learning around AI influencers. So, I launched AI Winners, a paid community where I break down everything I’ve learned. It’s still early, but I see it turning into a solid, long-term community. Investment & Acquisition Talks: I’m still evaluating potential investors and acquisition offers for my AI models. There’s growing interest in buying or investing in Emma & Jade, so I’ve been having conversations to explore different options. Overall, January was about tech, rebuilding, and long-term planning—not immediate revenue. But that’s what keeps this business sustainable. 🚀 \------- ⚠️ Biggest Challenges This Month Lost Both Twitter Accounts (Massive Traffic Hit) 🚨 The biggest blow this month was losing my models’ Twitter accounts. Twitter was responsible for about 40% of my total traffic, meaning both free and paid subs took a direct hit. While Emma’s revenue took a slight dip, Jade’s income dropped significantly—partly due to the account loss and partly because January is naturally slow. (Full revenue breakdown at the end of the post.) Jade’s Instagram Tanked (Possible Shadow Ban?) 🤔 Jade’s Instagram completely lost momentum in early January. Engagement and reach dropped by over 80%, and I still haven’t figured out why. It feels like a shadow ban, but I have no clear confirmation. To counter this, I launched a second backup account, and things are starting to recover. \------- 🚀 Potential Improvements & What’s Next Locking in a Stable Workflow 🔄 Right now, Emma & Jade’s workflow is still evolving, but I’m aiming to fully stabilize it. As I’m writing this, content is generating on my second monitor—a sign that I’m close to achieving full automation without compromising quality. Boosting Jade’s Fanvue Revenue 💰 Jade’s income took a hit this month, and it’s 100% a traffic issue. The solution? More content, more reach. I’ll be increasing social media output to drive consistent traffic back to Fanvue and restore her earnings. Patreon is Done. All Focus on Fanvue 🚫 I shut down both Emma & Jade’s Patreon accounts. The goal is not to split revenue—I want everything funneled into Fanvue for higher engagement and bigger paydays. \------- 💰 January 2025 Earnings Breakdown Despite January being one of the slowest months for online creators, Emma and Jade still brought in over $29K in revenue, with a net profit exceeding $20K after all expenses. Emma Laui generated $20,206.77, with around $6,000 in expenses (chatter payments, NSFW designer fees, and other operational costs). Jade Laui earned $8,939.05, with $2,000 in expenses. Considering Twitter account losses, Instagram setbacks, and the usual January spending slump, this is still a solid outcome. The focus now is on scaling traffic and maximizing Fanvue revenue heading into February. 🚀🔥 That’s the full breakdown for January! If you have questions, feel free to drop a comment, and I’ll answer when I can. Happy to help, just like others helped me when I was starting out! 🚀🔥

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey
reddit
LLM Vibe Score0
Human Vibe Score0.778
benfromwhereThis week

How I Made $250.000+ in a Year: A Case Study of My AI Influencer Journey

Update on February 22th: I changed my AI influencer's names because it caused some problems on my business. One year, two AI-powered influencers, and $250K in revenue. Sounds unreal? It’s not. Today, I’m pulling back the curtain on the strategies, tools, and hard-won lessons that took me from concept to a six-figure success story in the AI influencer space. Hey, I'm Ben—a 32-year-old designer who spent the past year navigating the world of AI influencers. Let me clear up any confusion right from the start: I’m not here to sell you anything. This is purely a case study to share what worked, what didn’t, and what I’ve learned along the way. I’ll also make sure to answer all your questions in the comments for free whenever I can, so don’t hesitate to ask. Links to Past Topics: If you're curious about some of the groundwork I covered, check out a few of my earlier posts here: How I Make $10,000 Monthly | AI Influencer Management How I Earned $7000+ in 15 Days | AI Influencer Business Update These earlier posts cover a lot of the backstory, so feel free to explore them before diving into this one. So if you're ready, here is the full story: \---- The idea of creating an AI influencer was one of those “what if” moments that wouldn’t leave my mind. At first, it sounded futuristic—even a bit too ambitious. It all started when I stumbled upon an AI influencer on Instagram with the handle AnnaMaes2000. Her content blew me away—the quality, the detail, and just how real everything looked. I was instantly hooked and ended up going through every post, just trying to figure out how she was pulling this off. That’s when I knew I had to learn how this was done. The next step? YouTube. I dived into videos on Stable Diffusion, soaking up everything I could about creating AI-generated images. Those tutorials taught me the basics and got me up to speed. Then, I created my first AI influencer, let's call her Mel for now. Right after that, to complete the storyline and boost engagement, I introduced Mel's “mother,” Jess. Adding Jess gave the whole project depth and a narrative that drew people in, creating a unique family dynamic that instantly elevated traffic and interest. After thousands of bad photos, hundreds of deleted posts, and months of trial and error, you can now see the quality that defines my current accounts. Here’s a rundown of the tools and checkpoints I’ve used from day one, in order: Fooocus on RunDiffusion — Juggernaut V8 Fooocus on RunDiffusion — Juggernaut V9 Fooocus on PC (locally) — Juggernaut V9 Fooocus on PC (locally) —Lyuyang Mix + Juggernaut V9 Flux on PC (couple of photos only since it's so slow even on RTX 4090) Flux on Fal.ai. \---- There’s no magic Instagram hack that guarantees success, despite what everyone thinks and keeps asking me. Quality content, consistent uploads, and solid craftsmanship are what actually help your photos hit trends and show up on the Explore page. Unlike 95% of low-quality AI accounts out there, I don’t rely on faceswap videos, spam Reels, or go around liking comments on other accounts. My approach is fully organic, focused solely on creating my own unique content. By following Instagram's guidelines to the letter, I've managed to direct some of Mel and Jess' fans over to Patreon and Fanvue. There, for a small subscription fee, fans can access exclusive lingerie content. For those looking for more, higher-tier subscriptions give access to even more premium content. Some possible questions and their answers: No, you can't share hardcore NSFW content on Patreon. You can do that on Fanvue. Yes, you can create AI creators on Fanvue — OnlyFans doesn't allow it. Yes, you can use your own ID to get KYC. Yes, we're telling both Mel and Jess is (or use) AI to generate content. And yes, some people leave and some people still have fun with chatting, having a good time and get perfect content for their needs. And yes, we have a chatter team to work on these accounts. \---- This journey wasn’t all smooth sailing. I faced unexpected roadblocks, like platform restrictions that limited certain types of content, and managing fan expectations was more challenging than anticipated. Staying within guidelines while keeping fans engaged required constant adaptation. These hurdles forced me to get creative, adjust my approach, and learn fast. Once I saw Mel and Jess gaining traction, I knew it was time to scale up. Expanding meant finding new ways to keep content fresh, creating deeper narratives, and considering how to bring even more followers into the fold. My focus turned to building a sustainable model that could grow without sacrificing quality or authenticity. If you’re thinking about diving into AI content creation, here’s my advice: patience, consistency, and a focus on quality are key. Don’t cut corners or rely on quick-fix hacks. Invest time in learning the right tools, creating engaging stories, and building an audience that values what you bring to the table. This approach took me from zero to six figures, and it’s what makes the journey worth it. \---- And finally, here’s the income breakdown that everyone’s curious about: Mel on Fanvue: $82,331.58 (Gross earnings because we have chatter cuts like 15%) Mel on Patreon: $50,865.98 (Net earnings) Jess on Fanvue: $89,068.26 (Gross earnings because we have chatter cuts like 15%) Jess on Patreon: $39,040.70 And thanks to Reddit and my old posts, I got a perfect investor like after 5 months, so this is a "payback" for that. Like I said, I'll answer every question in the comments — take care and let me know.

Looking for a Business Partner for an AI Stock recommendation SaaS
reddit
LLM Vibe Score0
Human Vibe Score1
armaan-devThis week

Looking for a Business Partner for an AI Stock recommendation SaaS

Hey everyone, I’m a 15-year-old full-stack developer, currently building StockWise, a startup focused on AI-driven stock market insights and analytics. I can handle all engineering, backend, frontend, and AI-related work—but I need a business partner who can take care of the marketing, sales, and user acquisition side of things. So this SaaS is currently in development. Also this I believe this can be both b2c and b2b. Like for b2c - it's the website included, with the recommendations, for individual users, for b2b - we can provide API's. Here is the classic workflow : \-> You can give your preferences, such as your monthly investment capital, if you're expecting short term or long term, and also if there are any specific areas you are more interested like AI, hydrogen fuel related, ev, compaines. \-> Then with this data, we recommend you stocks to buy, analyzing your preferences, looking at market, researching, looking into company's stock history, background, product \-> You will also have a chatbot like interface you can talk to about anything, and it will be personalized \-> Also you can add your portfolio here, and you can get insights based on the market data \-> Also there can be a weekly newsletter, too, if you subscribe to it. I'm much more of a builder, likes to build stuff, is good at it, but not good at the business side of things, that's why I'm really looking for a business partner. If you’re interested in joining as a co-founder or business partner, drop a comment or DM me!, Thanks a lot, Armaan

I had over 1000 visitors in 24h thanks to a post on HN and generated 0$ revenue but here is what I learned:
reddit
LLM Vibe Score0
Human Vibe Score1
sow4codeThis week

I had over 1000 visitors in 24h thanks to a post on HN and generated 0$ revenue but here is what I learned:

I litteraly just have 39 followers ont Twitter, I don't have an audience at all and a vice that entrepreneurs and indie hackers often fall into is looking at others who have an audience and to start hating it and telling themselves that even if their products are crap they will still have traffic on their site given their number of subscribers and their audiences. This thought is just a limiting thought because. Yes, obviously it's easier for the person who already has an andience to bring traffic to their site and acquire these first users but these people have to work to build this audience, it wasn't easy, it required a lot of effort but we quickly forget that when we don't even have a tenth of what this person has and despite this facility it's not an excuse to fill up and abandon your project, telling yourself that no one will ever see my product if I don't already have a built audience. That's not an excuse ! I am proof of this on a small scale, yesterday I launched my new product (EduHunt, a site that helps you find the most relevant educational content that you are looking for to avoid paying for online courses that are worth a fortune but to be honest in the end it was rubbish, the idea seemed good but the market is what it is and there is NO need for a site like that, I still learn lessons from it, failure is necessary to succeed ! ). So I launched EduHunt on Hacker News and on Reddit but Reddit didn't bring me much in the end. 1 hour after the launch I had around fifty visitors and 3 registered (trial period), I told myself that it was going to continue like this and I hoped to have 200 visitors at the end of the day no more. I can't tell you what a surprise it was when I opened Vercel and saw 800 visitors for 50 online as I looked, I went crazy lol. My post on Hacker News "exploded", I had more than 400 people who had just come from Hacker News and other sites linked to Hacker News, I told myself that it was finally the right one but reality quickly caught up with me , I went to see my post and this is the kind of comment I had ( Above the text ) As you see, my product sucks and it's not the end of the world, I learn a lot of lessons from it, I failed in the design of the product in directly reflecting what the idea of the product is (most of the comments do not really target my basic idea, I wanted to create a site to help search for educational content on YouTube with filters that are not in the usual YouTube search and this in text format analyzed by AI, I was told that I monetize free videos, I do not appropriate the videos that I put on my site and that you have to pay to have access, what is monetized here is the means of 'access to the content, not the content itself, but yes I failed in this and in many others of this project but I come out better) Despite this, I attracted more than 1000 visitors to my site in less than 24 hours with a simple post on Hacker News, a good title, a sincere story to go with it and that was it, I have no audience nothing at all. If the product had been much better who knows where I would be today. All this to say and remind you that there are no excuses to hide behind, building an audience requires hard work and takes time ! But just because you don't have one doesn't mean you can never bring traffic to your site. Be honest in what you do, learn from your mistakes, repeat and you should find your happiness.

SUPIR
github
LLM Vibe Score0.599
Human Vibe Score0.8316614420062696
Fanghua-YuMar 28, 2025

SUPIR

(CVPR2024) Scaling Up to Excellence: Practicing Model Scaling for Photo-Realistic Image Restoration In the Wild [Paper] &emsp; [Project Page] &emsp; [[Online App]](https://supir.suppixel.ai/home) Fanghua, Yu, Jinjin Gu, Zheyuan Li, Jinfan Hu, Xiangtao Kong, Xintao Wang, Jingwen He, Yu Qiao, Chao Dong Shenzhen Institute of Advanced Technology; Shanghai AI Laboratory; University of Sydney; The Hong Kong Polytechnic University; ARC Lab, Tencent PCG; The Chinese University of Hong Kong 🚀 We're thrilled to announce the official launch of SupPixel AI! Experience the next level of image processing and upscaling with our cutting-edge AI technology based on SUPIR. Explore now at suppixel.ai. 🔧 Dependencies and Installation Clone repo Install dependent packages Download Checkpoints For users who can connect to huggingface, please setting LLAVACLIPPATH, SDXLCLIP1PATH, SDXLCLIP2CKPTPTH in CKPTPTH.py as None. These CLIPs will be downloaded automatically. Dependent Models SDXL CLIP Encoder-1 SDXL CLIP Encoder-2 SDXL base 1.00.9vae LLaVA CLIP LLaVA v1.5 13B (optional) Juggernaut-XLv9RunDiffusionPhotov2 Replacement of SDXL base 1.0_0.9vae for Photo Realistic (optional) JuggernautRunDiffusionPhoto2Lightning4Steps Distilling model used in SUPIRv0Juggernautv9_lightning.yaml Models we provided: SUPIR-v0Q: Baidu Netdisk, Google Drive Default training settings with paper. High generalization and high image quality in most cases. SUPIR-v0F: Baidu Netdisk, Google Drive Training with light degradation settings. Stage1 encoder of SUPIR-v0F remains more details when facing light degradations. Edit Custom Path for Checkpoints ⚡ Quick Inference Val Dataset RealPhoto60: Baidu Netdisk, Google Drive Usage of SUPIR Python Script Gradio Demo Online App We've just launched SupPixel AI, an easy-to-use tool designed to help with high-quality image processing and upscaling. It builds on SUPIR. Whether you’re into photography, digital art, or just love playing around with image enhancement, we’d love for you to check it out.~ BibTeX @misc{yu2024scaling, title={Scaling Up to Excellence: Practicing Model Scaling for Photo-Realistic Image Restoration In the Wild}, author={Fanghua Yu and Jinjin Gu and Zheyuan Li and Jinfan Hu and Xiangtao Kong and Xintao Wang and Jingwen He and Yu Qiao and Chao Dong}, year={2024}, eprint={2401.13627}, archivePrefix={arXiv}, primaryClass={cs.CV} } 📧 Contact If you have any question, please email fanghuayu96@gmail.com or jinjin.gu@suppixel.ai. Non-Commercial Use Only Declaration The SUPIR ("Software") is made available for use, reproduction, and distribution strictly for non-commercial purposes. For the purposes of this declaration, "non-commercial" is defined as not primarily intended for or directed towards commercial advantage or monetary compensation. By using, reproducing, or distributing the Software, you agree to abide by this restriction and not to use the Software for any commercial purposes without obtaining prior written permission from Dr. Jinjin Gu. This declaration does not in any way limit the rights under any open source license that may apply to the Software; it solely adds a condition that the Software shall not be used for commercial purposes. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. For inquiries or to obtain permission for commercial use, please contact Dr. Jinjin Gu (jinjin.gu@suppixel.ai).

activepieces
github
LLM Vibe Score0.66
Human Vibe Score1
activepiecesMar 28, 2025

activepieces

An open source replacement for Zapier Documentation 🌪️ Create a Piece 🖉 Deploy 🔥 Join Discord 🤯 Welcome to Activepieces Your friendliest open source all-in-one automation tool, designed to be extensible through a type-safe pieces framework written in Typescript. 🔥 Why Activepieces is Different: 💖 Loved by Everyone: Intuitive interface and great experience for both technical and non-technical users with a quick learning curve. 🌐 Open Ecosystem: All pieces are open source and available on npmjs.com, 60% of the pieces are contributed by the community. 🛠️ Pieces are written in Typescript: Pieces are npm packages in TypeScript, offering full customization with the best developer experience, including hot reloading for local piece development on your machine. 😎 🤖 AI-Ready: Native AI pieces let you experiment with various providers, or create your own agents using our AI SDK, and there is Copilot to help you build flows inside the builder. 🏢 Enterprise-Ready: Developers set up the tools, and anyone in the organization can use the no-code builder. Full customization from branding to control. 🔒 Secure by Design: Self-hosted and network-gapped for maximum security and control over your data. 🧠 Human in the Loop: Delay execution for a period of time or require approval. These are just pieces built on top of the piece framework, and you can build many pieces like that. 🎨 💻 Human Input Interfaces: Built-in support for human input triggers like "Chat Interface" 💬 and "Form Interface" 📝 🛠️ Builder Features: [x] Loops [x] Branches [x] Auto Retries [x] HTTP [x] Code with NPM [x] ASK AI in Code Piece (Non technical user can clean data without knowing to code) [x] Flows are fully versioned. [x] Languages Translations [x] Customizable Templates [X] 200+ Pieces, check https://www.activepieces.com/pieces We release updates frequently. Check the product changelog for the latest features. 🔌 Create Your Own Piece Activepieces supports integrations with Google Sheets, OpenAI, Discord, RSS, and over 200 other services. Check out the full list of supported integrations, which is constantly expanding thanks to our community's contributions. As an open ecosystem, all integration source code is accessible in our repository. These integrations are versioned and published directly to npmjs.com upon contribution. You can easily create your own integration using our TypeScript framework. For detailed instructions, please refer to our Contributor's Guide. License Activepieces' Community Edition is released as open source under the MIT license and enterprise features are released under Commercial License Read more about the feature comparison here https://www.activepieces.com/docs/about/editions 💭 Join Our Community 🌐 Contributions We welcome contributions big or small and in different directions. The best way to do this is to check this document and we are always up to talk on our Discord Server. 📚 Translations Not into coding but still interested in contributing? Come join our Discord and visit https://www.activepieces.com/docs/about/i18n for more information. !fr translation].data.translationProgress&url=https%3A%2F%2Fbadges.awesome-crowdin.com%2Fstats-16093902-626364-update.json) !it translation].data.translationProgress&url=https%3A%2F%2Fbadges.awesome-crowdin.com%2Fstats-16093902-626364-update.json) !de translation].data.translationProgress&url=https%3A%2F%2Fbadges.awesome-crowdin.com%2Fstats-16093902-626364-update.json) !ja translation].data.translationProgress&url=https%3A%2F%2Fbadges.awesome-crowdin.com%2Fstats-16093902-626364-update.json) !pt-BR translation].data.translationProgress&url=https%3A%2F%2Fbadges.awesome-crowdin.com%2Fstats-16093902-626364-update.json) 🦫 Contributors ShahedAlMashni🔌 AbdulTheActivePiecer🚧 Khaled Mashaly🚧 Mohammed Abu Aboud🚧 Abdulrahman Zeineddin🔌 ahmad jaber🔌 ashrafsamhouri🔌 Mohammad Abu Musa📆 Mukewa Wekalao🔌 Osama Abdallah Essa Haikal🔌 Arman🛡️ Oskar Krämer📖 Thibaut Patel🤔 🔌 Applesaucesomer🤔 crazyTweek🤔 Muhammad Tabaza🔌 Shay Punter📖 🔌 abaza738🔌 Jona Boeddinghaus🔌 fomojola💻 Alexander Storozhevsky💻 J0LGER🛡️ Patrick Veverka🐛 Berk Sümbül📖 Willian Guedes🔌 Abdullah Ranginwala💻 Dennis Tychsen🔌 MyWay🔌 Bibhuti Bhusan Panda🔌 Tarun Samanta🐛 Herman Kudria🔌 [NULL] Dev🔌 Jan Bebendorf🔌 Nilesh🔌 Vraj Gohil🔌 BastienMe🔌 Stephen Foskett📖 Nathan📖 Marcin Natanek🔌 Mark van Bellen🔌 Olivier Guzzi🔌 Osama Zakarneh🔌 phestvik🤔 Rajdeep Pal📖 Camilo Usuga🔌 Kishan Parmar📖 🔌 BBND🔌 Haseeb Rehman🔌 Rita Gorokhod🔌 Fábio Ferreira🔌 Florin Buffet📖 Drew Lewis🔌 Benjamin André-Micolon🔌 Denis Gurskij🔌 Nefer Lopez📖 fardeenpanjwani-codeglo📖 Landon Moir🔌 Diego Nijboer🔌 Tân Một Nắng🔌 Gavin Foley📖 Dennis Trautwein🐛 Andrew Rosenblatt🐛 rika🔌 Cyril Selasi🔌 Franck Nijimbere🔌 Aleksandr Denisov🔌 Reuben Swartz📖 joselupianez🔌 Awais Manzoor🐛 💻 Andrei🐛 derbbre📖 Maor Rozenfeld💻 Michael Huynh📖 Filip Dunđer💻 Don Thorp📖 Joe Workman🔌 Aykut Akgün💻 Yann Petitjean🔌 🐛 pfernandez98🔌 Daniel O.🔌 Meng-Yuan Huang📖 Leyla🐛 i-nithin🔌 la3rence🔌 Dennis Rongo🐛 🔌 Kartik Mehta📖 💻 Zakher Masri📖 💻 AbdullahBitar🔌 Mario Meyer🔌 Karim Khaleel🔌 CPonchet🐛 Olivier Sambourg🔌 Ahmad(Ed)🔌 leenmashni🔌 M Abdul Rauf📖 Vincent Barrier🔌 John💻 🔌 Joost de Valk🔌 MJ🔌 ShravanShenoy💻 Jon Kristian📖 cr0fters🐛 Bibek Timsina🐛 Viktor Szépe💻 Rendy Tan📖 🔌 Islam Abdelfattah🐛 Yoonjae Choi💻 Javier HM🔌 Mohamed Hassan🐛 Christian Schab🔌 Pratik Kinage🔌 Abdelrahman Mostafa 🔌 Hamza Zagha🐛 Lasse Schuirmann🔌 Cyril Duchon-Doris🔌 Javiink🔌 Harshit Harchani🔌 MrAkber📖 marek-slavicek🔌 hugh-codes🔌 Alex Lewis🐛 Yuanlin Lin📖 Ala Shiban📖 hamsh💻 Anne Mariel Catapang🔌 Carlo Gino Catapang🔌 Aditya Rathore🔌 coderbob2🔌 Ramy Gamal🔌 Alexandru-Dan Pop💻 Frank Micheal 🔌 Emmanuel Ferdman📖 Sany A🔌 Niels Swimberghe🐛 lostinbug🔌 gushkool🔌 Omar Sayed🔌 rSnapkoOpenOps🐛 ahronshor🔌 Cezar🐛 Shawn Lim🔌 Shawn Lim🔌 pavloDeshko🐛 abc💻 manoj kumar d🔌 Feli🔌 Miguel🔌 Instasent DEV🔌 Matthieu Lombard🔌 beyondlevi🔌 Rafal Zawadzki🔌 Simon Courtois🔌 alegria-solutions🔌 D-Rowe-FS🔌 张晟杰🔌 Ashot🔌 Amr Abu Aza🔌 John Goodliff🔌 Diwash Dev🔌 André🔌 Lou &#124; Digital Marketing🔌 Maarten Coppens🔌 Mahmoud Hamed🔌 Theo Dammaretz🔌 s31w4n📖 Abdul Rahman🔌 Kent Smith🔌 Arvind Ramesh💻 valentin-mourtialon🔌 psgpsg16🔌 Mariia Shyn🔌 Joshua Heslin🔌 Ahmad🔌 you💻 Daniel Poon💻 Kévin Yu🔌 노영은🔌 reemayoush🔌 Brice🛡️ Mg Wunna🔌 This project follows the all-contributors specification. Contributions of any kind are welcome!

anything-llm
github
LLM Vibe Score0.572
Human Vibe Score0.4703504093656464
Mintplex-LabsMar 28, 2025

anything-llm

AnythingLLM: The all-in-one AI app you were looking for. Chat with your docs, use AI Agents, hyper-configurable, multi-user, & no frustrating set up required. | | Docs | Hosted Instance English · 简体中文 · 日本語 👉 AnythingLLM for desktop (Mac, Windows, & Linux)! Download Now A full-stack application that enables you to turn any document, resource, or piece of content into context that any LLM can use as references during chatting. This application allows you to pick and choose which LLM or Vector Database you want to use as well as supporting multi-user management and permissions. !Chatting Watch the demo! Product Overview AnythingLLM is a full-stack application where you can use commercial off-the-shelf LLMs or popular open source LLMs and vectorDB solutions to build a private ChatGPT with no compromises that you can run locally as well as host remotely and be able to chat intelligently with any documents you provide it. AnythingLLM divides your documents into objects called workspaces. A Workspace functions a lot like a thread, but with the addition of containerization of your documents. Workspaces can share documents, but they do not talk to each other so you can keep your context for each workspace clean. Cool features of AnythingLLM 🆕 Custom AI Agents 🆕 No-code AI Agent builder 🖼️ Multi-modal support (both closed and open-source LLMs!) 👤 Multi-user instance support and permissioning Docker version only 🦾 Agents inside your workspace (browse the web, etc) 💬 Custom Embeddable Chat widget for your website Docker version only 📖 Multiple document type support (PDF, TXT, DOCX, etc) Simple chat UI with Drag-n-Drop funcitonality and clear citations. 100% Cloud deployment ready. Works with all popular closed and open-source LLM providers. Built-in cost & time-saving measures for managing very large documents compared to any other chat UI. Full Developer API for custom integrations! Much more...install and find out! Supported LLMs, Embedder Models, Speech models, and Vector Databases Large Language Models (LLMs): Any open-source llama.cpp compatible model OpenAI OpenAI (Generic) Azure OpenAI AWS Bedrock Anthropic NVIDIA NIM (chat models) Google Gemini Pro Hugging Face (chat models) Ollama (chat models) LM Studio (all models) LocalAi (all models) Together AI (chat models) Fireworks AI (chat models) Perplexity (chat models) OpenRouter (chat models) DeepSeek (chat models) Mistral Groq Cohere KoboldCPP LiteLLM Text Generation Web UI Apipie xAI Novita AI (chat models) PPIO Embedder models: AnythingLLM Native Embedder (default) OpenAI Azure OpenAI LocalAi (all) Ollama (all) LM Studio (all) Cohere Audio Transcription models: AnythingLLM Built-in (default) OpenAI TTS (text-to-speech) support: Native Browser Built-in (default) PiperTTSLocal - runs in browser OpenAI TTS ElevenLabs Any OpenAI Compatible TTS service. STT (speech-to-text) support: Native Browser Built-in (default) Vector Databases: LanceDB (default) Astra DB Pinecone Chroma Weaviate Qdrant Milvus Zilliz Technical Overview This monorepo consists of three main sections: frontend: A viteJS + React frontend that you can run to easily create and manage all your content the LLM can use. server: A NodeJS express server to handle all the interactions and do all the vectorDB management and LLM interactions. collector: NodeJS express server that process and parses documents from the UI. docker: Docker instructions and build process + information for building from source. embed: Submodule for generation & creation of the web embed widget. browser-extension: Submodule for the chrome browser extension. 🛳 Self Hosting Mintplex Labs & the community maintain a number of deployment methods, scripts, and templates that you can use to run AnythingLLM locally. Refer to the table below to read how to deploy on your preferred environment or to automatically deploy. | Docker | AWS | GCP | Digital Ocean | Render.com | |----------------------------------------|----|-----|---------------|------------| | [![Deploy on Docker][docker-btn]][docker-deploy] | [![Deploy on AWS][aws-btn]][aws-deploy] | [![Deploy on GCP][gcp-btn]][gcp-deploy] | [![Deploy on DigitalOcean][do-btn]][do-deploy] | [![Deploy on Render.com][render-btn]][render-deploy] | | Railway | RepoCloud | Elestio | | --- | --- | --- | | [![Deploy on Railway][railway-btn]][railway-deploy] | [![Deploy on RepoCloud][repocloud-btn]][repocloud-deploy] | [![Deploy on Elestio][elestio-btn]][elestio-deploy] | or set up a production AnythingLLM instance without Docker → How to setup for development yarn setup To fill in the required .env files you'll need in each of the application sections (from root of repo). Go fill those out before proceeding. Ensure server/.env.development is filled or else things won't work right. yarn dev:server To boot the server locally (from root of repo). yarn dev:frontend To boot the frontend locally (from root of repo). yarn dev:collector To then run the document collector (from root of repo). Learn about documents Learn about vector caching External Apps & Integrations These are apps that are not maintained by Mintplex Labs, but are compatible with AnythingLLM. A listing here is not an endorsement. Midori AI Subsystem Manager - A streamlined and efficient way to deploy AI systems using Docker container technology. Coolify - Deploy AnythingLLM with a single click. GPTLocalhost for Microsoft Word - A local Word Add-in for you to use AnythingLLM in Microsoft Word. Telemetry & Privacy AnythingLLM by Mintplex Labs Inc contains a telemetry feature that collects anonymous usage information. More about Telemetry & Privacy for AnythingLLM Why? We use this information to help us understand how AnythingLLM is used, to help us prioritize work on new features and bug fixes, and to help us improve AnythingLLM's performance and stability. Opting out Set DISABLE_TELEMETRY in your server or docker .env settings to "true" to opt out of telemetry. You can also do this in-app by going to the sidebar > Privacy and disabling telemetry. What do you explicitly track? We will only track usage details that help us make product and roadmap decisions, specifically: Type of your installation (Docker or Desktop) When a document is added or removed. No information about the document. Just that the event occurred. This gives us an idea of use. Type of vector database in use. Let's us know which vector database provider is the most used to prioritize changes when updates arrive for that provider. Type of LLM in use. Let's us know the most popular choice and prioritize changes when updates arrive for that provider. Chat is sent. This is the most regular "event" and gives us an idea of the daily-activity of this project across all installations. Again, only the event is sent - we have no information on the nature or content of the chat itself. You can verify these claims by finding all locations Telemetry.sendTelemetry is called. Additionally these events are written to the output log so you can also see the specific data which was sent - if enabled. No IP or other identifying information is collected. The Telemetry provider is PostHog - an open-source telemetry collection service. View all telemetry events in source code 👋 Contributing create issue create PR with branch name format of - LGTM from core-team 🌟 Contributors 🔗 More Products [VectorAdmin][vector-admin]: An all-in-one GUI & tool-suite for managing vector databases. [OpenAI Assistant Swarm][assistant-swarm]: Turn your entire library of OpenAI assistants into one single army commanded from a single agent. [![][back-to-top]](#readme-top) Copyright © 2025 [Mintplex Labs][profile-link]. This project is MIT licensed. [back-to-top]: https://img.shields.io/badge/-BACKTOTOP-222628?style=flat-square [profile-link]: https://github.com/mintplex-labs [vector-admin]: https://github.com/mintplex-labs/vector-admin [assistant-swarm]: https://github.com/Mintplex-Labs/openai-assistant-swarm [docker-btn]: ./images/deployBtns/docker.png [docker-deploy]: ./docker/HOWTOUSE_DOCKER.md [aws-btn]: ./images/deployBtns/aws.png [aws-deploy]: ./cloud-deployments/aws/cloudformation/DEPLOY.md [gcp-btn]: https://deploy.cloud.run/button.svg [gcp-deploy]: ./cloud-deployments/gcp/deployment/DEPLOY.md [do-btn]: https://www.deploytodo.com/do-btn-blue.svg [do-deploy]: ./cloud-deployments/digitalocean/terraform/DEPLOY.md [render-btn]: https://render.com/images/deploy-to-render-button.svg [render-deploy]: https://render.com/deploy?repo=https://github.com/Mintplex-Labs/anything-llm&branch=render [render-btn]: https://render.com/images/deploy-to-render-button.svg [render-deploy]: https://render.com/deploy?repo=https://github.com/Mintplex-Labs/anything-llm&branch=render [railway-btn]: https://railway.app/button.svg [railway-deploy]: https://railway.app/template/HNSCS1?referralCode=WFgJkn [repocloud-btn]: https://d16t0pc4846x52.cloudfront.net/deploylobe.svg [repocloud-deploy]: https://repocloud.io/details/?app_id=276 [elestio-btn]: https://elest.io/images/logos/deploy-to-elestio-btn.png [elestio-deploy]: https://elest.io/open-source/anythingllm

xpert
github
LLM Vibe Score0.457
Human Vibe Score0.0831216059433162
xpert-aiMar 28, 2025

xpert

English | 中文 [uri_license]: https://www.gnu.org/licenses/agpl-3.0.html [urilicenseimage]: https://img.shields.io/badge/License-AGPL%20v3-blue.svg Xpert Cloud · Self-hosting · Documentation · Enterprise inquiry Open-Source AI Platform for Enterprise Data Analysis, Indicator Management and Agents Orchestration Xpert AI is an open-source enterprise-level AI system that perfectly integrates two major platforms: agent orchestration and data analysis. 💡 What's New Agent and Workflow Hybrid Architecture In today's rapidly evolving AI landscape, enterprises face a critical dilemma: how to balance the creativity of LLMs with the stability of processes? While purely agent-based architectures offer flexibility, they are difficult to control; traditional workflows, though reliable, lack adaptability. The Agent and Workflow Hybrid Architecture of the Xpert AI platform is designed to resolve this conflict — it allows AI to possess "free will" while adhering to "rules and order." !agent-workflow-hybrid-architecture Blog - Agent and Workflow Hybrid Architecture Agent Orchestration Platform By coordinating the collaboration of multiple agents, Xpert completes complex tasks. Xpert integrates different types of AI agents through an efficient management mechanism, utilizing their capabilities to solve multidimensional problems. Xpert Agents Data Analysis Platform An agile data analysis platform based on cloud computing for multidimensional modeling, indicator management, and BI display. It supports connecting to various data sources, achieving efficient and flexible data analysis and visualization, and provides multiple intelligent analysis functions and tools to help enterprises quickly and accurately discover business value and make operational decisions. ChatBI ChatBI is an innovative feature we are introducing, combining chat functionality with business intelligence (BI) analysis capabilities. It offers users a more intuitive and convenient data analysis experience through natural language interaction. ChatBI_Demo.mp4 🚀 Quick Start Before installing Xpert, make sure your machine meets the following minimum system requirements: CPU >= 2 Core RAM >= 4 GiB Node.js (ESM and CommonJS) - 18.x, 19.x, 20.x, 22.x The easiest way to start the Xpert server is through docker compose. Before running Xpert with the following commands, make sure that Docker and Docker Compose are installed on your machine: After running, you can access the Xpert dashboard in your browser at http://localhost/onboarding and start the initialization process. Please check our Wiki - Development to get started quickly. 🎯 Mission Empowering enterprises with intelligent collaboration and data-driven insights through innovative AI orchestration and agile analytics. 🌼 Screenshots Show / Hide Screenshots Pareto analysis open in new tab !Pareto analysis Screenshot Product profit analysis open in new tab !Product profit analysis Screenshot Reseller analysis open in new tab !Reseller analysis Screenshot Bigview dashboard open in new tab !Bigview dashboard Screenshot Indicator application open in new tab !Indicator application Screenshot Indicator mobile app open in new tab !Indicator mobile app Screenshot 💻 Demo, Downloads, Testing and Production Demo Xpert AI Platform Demo at . Notes: You can generate samples data in the home dashbaord page. Production (SaaS) Xpert AI Platform SaaS is available at . Note: it's currently in Alpha version / in testing mode, please use it with caution! 🧱 Technology Stack and Requirements TypeScript language NodeJs / NestJs Nx Angular RxJS TypeORM Langchain ECharts Java Mondrian For Production, we recommend: PostgreSQL PM2 See also README.md and CREDITS.md files in relevant folders for lists of libraries and software included in the Platform, information about licenses, and other details 📄 Documentation Please refer to our official Platform Documentation and to our Wiki (WIP). 💌 Contact Us For business inquiries: Xpert AI Platform @ Twitter 🛡️ License We support the open-source community. This software is available under the following licenses: Xpert AI Platform Community Edition Xpert AI Platform Small Business Xpert AI Platform Enterprise Please see LICENSE for more information on licenses. 💪 Thanks to our Contributors Contributors Please give us :star: on Github, it helps! You are more than welcome to submit feature requests in the Xpert AI repo Pull requests are always welcome! Please base pull requests against the develop branch and follow the contributing guide.

GenAI_Agents
github
LLM Vibe Score0.563
Human Vibe Score0.24210481455988786
NirDiamantMar 28, 2025

GenAI_Agents

🌟 Support This Project: Your sponsorship fuels innovation in GenAI agent development. Become a sponsor to help maintain and expand this valuable resource! GenAI Agents: Comprehensive Repository for Development and Implementation 🚀 Welcome to one of the most extensive and dynamic collections of Generative AI (GenAI) agent tutorials and implementations available today. This repository serves as a comprehensive resource for learning, building, and sharing GenAI agents, ranging from simple conversational bots to complex, multi-agent systems. 📫 Stay Updated! 🚀Cutting-edgeUpdates 💡ExpertInsights 🎯Top 0.1%Content Join over 15,000 of AI enthusiasts getting unique cutting-edge insights and free tutorials! Plus, subscribers get exclusive early access and special 33% discounts to my book and the upcoming RAG Techniques course! Introduction Generative AI agents are at the forefront of artificial intelligence, revolutionizing the way we interact with and leverage AI technologies. This repository is designed to guide you through the development journey, from basic agent implementations to advanced, cutting-edge systems. 📚 Learn to Build Your First AI Agent Your First AI Agent: Simpler Than You Think This detailed blog post complements the repository by providing a complete A-Z walkthrough with in-depth explanations of core concepts, step-by-step implementation, and the theory behind AI agents. It's designed to be incredibly simple to follow while covering everything you need to know to build your first working agent from scratch. 💡 Plus: Subscribe to the newsletter for exclusive early access to tutorials and special discounts on upcoming courses and books! Our goal is to provide a valuable resource for everyone - from beginners taking their first steps in AI to seasoned practitioners pushing the boundaries of what's possible. By offering a range of examples from foundational to complex, we aim to facilitate learning, experimentation, and innovation in the rapidly evolving field of GenAI agents. Furthermore, this repository serves as a platform for showcasing innovative agent creations. Whether you've developed a novel agent architecture or found an innovative application for existing techniques, we encourage you to share your work with the community. Related Projects 📚 Dive into my comprehensive guide on RAG techniques to learn about integrating external knowledge into AI systems, enhancing their capabilities with up-to-date and relevant information retrieval. 🖋️ Explore my Prompt Engineering Techniques guide for an extensive collection of prompting strategies, from fundamental concepts to advanced methods, improving your ability to communicate effectively with AI language models. A Community-Driven Knowledge Hub This repository grows stronger with your contributions! Join our vibrant Discord community — the central hub for shaping and advancing this project together 🤝 GenAI Agents Discord Community Whether you're a novice eager to learn or an expert ready to share your knowledge, your insights can shape the future of GenAI agents. Join us to propose ideas, get feedback, and collaborate on innovative implementations. For contribution guidelines, please refer to our CONTRIBUTING.md file. Let's advance GenAI agent technology together! 🔗 For discussions on GenAI, agents, or to explore knowledge-sharing opportunities, feel free to connect on LinkedIn. Key Features 🎓 Learn to build GenAI agents from beginner to advanced levels 🧠 Explore a wide range of agent architectures and applications 📚 Step-by-step tutorials and comprehensive documentation 🛠️ Practical, ready-to-use agent implementations 🌟 Regular updates with the latest advancements in GenAI 🤝 Share your own agent creations with the community GenAI Agent Implementations Explore our extensive list of GenAI agent implementations, sorted by categories: 🌱 Beginner-Friendly Agents Simple Conversational Agent LangChain PydanticAI Overview 🔎 A context-aware conversational AI maintains information across interactions, enabling more natural dialogues. Implementation 🛠️ Integrates a language model, prompt template, and history manager to generate contextual responses and track conversation sessions. Simple Question Answering Agent Overview 🔎 Answering (QA) agent using LangChain and OpenAI's language model understands user queries and provides relevant, concise answers. Implementation 🛠️ Combines OpenAI's GPT model, a prompt template, and an LLMChain to process user questions and generate AI-driven responses in a streamlined manner. Simple Data Analysis Agent LangChain PydanticAI Overview 🔎 An AI-powered data analysis agent interprets and answers questions about datasets using natural language, combining language models with data manipulation tools for intuitive data exploration. Implementation 🛠️ Integrates a language model, data manipulation framework, and agent framework to process natural language queries and perform data analysis on a synthetic dataset, enabling accessible insights for non-technical users. 🔧 Framework Tutorial: LangGraph Introduction to LangGraph: Building Modular AI Workflows Overview 🔎 This tutorial introduces LangGraph, a powerful framework for creating modular, graph-based AI workflows. Learn how to leverage LangGraph to build more complex and flexible AI agents that can handle multi-step processes efficiently. Implementation 🛠️ Step-by-step guide on using LangGraph to create a StateGraph workflow. The tutorial covers key concepts such as state management, node creation, and graph compilation. It demonstrates these principles by constructing a simple text analysis pipeline, serving as a foundation for more advanced agent architectures. Additional Resources 📚 Blog Post 🎓 Educational and Research Agents ATLAS: Academic Task and Learning Agent System Overview 🔎 ATLAS demonstrates how to build an intelligent multi-agent system that transforms academic support through AI-powered assistance. The system leverages LangGraph's workflow framework to coordinate multiple specialized agents that provide personalized academic planning, note-taking, and advisory support. Implementation 🛠️ Implements a state-managed multi-agent architecture using four specialized agents (Coordinator, Planner, Notewriter, and Advisor) working in concert through LangGraph's workflow framework. The system features sophisticated workflows for profile analysis and academic support, with continuous adaptation based on student performance and feedback. Additional Resources 📚 YouTube Explanation Blog Post Scientific Paper Agent - Literature Review Overview 🔎 An intelligent research assistant that helps users navigate, understand, and analyze scientific literature through an orchestrated workflow. The system combines academic APIs with sophisticated paper processing techniques to automate literature review tasks, enabling researchers to efficiently extract insights from academic papers while maintaining research rigor and quality control. Implementation 🛠️ Leverages LangGraph to create a five-node workflow system including decision making, planning, tool execution, and quality validation nodes. The system integrates the CORE API for paper access, PDFplumber for document processing, and advanced language models for analysis. Key features include a retry mechanism for robust paper downloads, structured data handling through Pydantic models, and quality-focused improvement cycles with human-in-the-loop validation options. Additional Resources 📚 YouTube Explanation Blog Post Chiron - A Feynman-Enhanced Learning Agent Overview 🔎 An adaptive learning agent that guides users through educational content using a structured checkpoint system and Feynman-style teaching. The system processes learning materials (either user-provided or web-retrieved), verifies understanding through interactive checkpoints, and provides simplified explanations when needed, creating a personalized learning experience that mimics one-on-one tutoring. Implementation 🛠️ Uses LangGraph to orchestrate a learning workflow that includes checkpoint definition, context building, understanding verification, and Feynman teaching nodes. The system integrates web search for dynamic content retrieval, employs semantic chunking for context processing, and manages embeddings for relevant information retrieval. Key features include a 70% understanding threshold for progression, interactive human-in-the-loop validation, and structured output through Pydantic models for consistent data handling. Additional Resources 📚 YouTube Explanation 💼 Business and Professional Agents Customer Support Agent (LangGraph) Overview 🔎 An intelligent customer support agent using LangGraph categorizes queries, analyzes sentiment, and provides appropriate responses or escalates issues. Implementation 🛠️ Utilizes LangGraph to create a workflow combining state management, query categorization, sentiment analysis, and response generation. Essay Grading Agent (LangGraph) Overview 🔎 An automated essay grading system using LangGraph and an LLM model evaluates essays based on relevance, grammar, structure, and depth of analysis. Implementation 🛠️ Utilizes a state graph to define the grading workflow, incorporating separate grading functions for each criterion. Travel Planning Agent (LangGraph) Overview 🔎 A Travel Planner using LangGraph demonstrates how to build a stateful, multi-step conversational AI application that collects user input and generates personalized travel itineraries. Implementation 🛠️ Utilizes StateGraph to define the application flow, incorporates custom PlannerState for process management. GenAI Career Assistant Agent Overview 🔎 The GenAI Career Assistant demonstrates how to create a multi-agent system that provides personalized guidance for careers in Generative AI. Using LangGraph and Gemini LLM, the system delivers customized learning paths, resume assistance, interview preparation, and job search support. Implementation 🛠️ Leverages a multi-agent architecture using LangGraph to coordinate specialized agents (Learning, Resume, Interview, Job Search) through TypedDict-based state management. The system employs sophisticated query categorization and routing while integrating with external tools like DuckDuckGo for job searches and dynamic content generation. Additional Resources 📚 YouTube Explanation Project Manager Assistant Agent Overview 🔎 An AI agent designed to assist in project management tasks by automating the process of creating actionable tasks from project descriptions, identifying dependencies, scheduling work, and assigning tasks to team members based on expertise. The system includes risk assessment and self-reflection capabilities to optimize project plans through multiple iterations, aiming to minimize overall project risk. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized nodes including task generation, dependency mapping, scheduling, allocation, and risk assessment. Each node uses GPT-4o-mini for structured outputs following Pydantic models. The system implements a feedback loop for self-improvement, where risk scores trigger reflection cycles that generate insights to optimize the project plan. Visualization tools display Gantt charts of the generated schedules across iterations. Additional Resources 📚 YouTube Explanation Contract Analysis Assistant (ClauseAI) Overview 🔎 ClauseAI demonstrates how to build an AI-powered contract analysis system using a multi-agent approach. The system employs specialized AI agents for different aspects of contract review, from clause analysis to compliance checking, and leverages LangGraph for workflow orchestration and Pinecone for efficient clause retrieval and comparison. Implementation 🛠️ Implements a sophisticated state-based workflow using LangGraph to coordinate multiple AI agents through contract analysis stages. The system features Pydantic models for data validation, vector storage with Pinecone for clause comparison, and LLM-based analysis for generating comprehensive contract reports. The implementation includes parallel processing capabilities and customizable report generation based on user requirements. Additional Resources 📚 YouTube Explanation E2E Testing Agent Overview 🔎 The E2E Testing Agent demonstrates how to build an AI-powered system that converts natural language test instructions into executable end-to-end web tests. Using LangGraph for workflow orchestration and Playwright for browser automation, the system enables users to specify test cases in plain English while handling the complexity of test generation and execution. Implementation 🛠️ Implements a structured workflow using LangGraph to coordinate test generation, validation, and execution. The system features TypedDict state management, integration with Playwright for browser automation, and LLM-based code generation for converting natural language instructions into executable test scripts. The implementation includes DOM state analysis, error handling, and comprehensive test reporting. Additional Resources 📚 YouTube Explanation 🎨 Creative and Content Generation Agents GIF Animation Generator Agent (LangGraph) Overview 🔎 A GIF animation generator that integrates LangGraph for workflow management, GPT-4 for text generation, and DALL-E for image creation, producing custom animations from user prompts. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that generates character descriptions, plots, and image prompts using GPT-4, creates images with DALL-E 3, and assembles them into GIFs using PIL. Employs asynchronous programming for efficient parallel processing. TTS Poem Generator Agent (LangGraph) Overview 🔎 An advanced text-to-speech (TTS) agent using LangGraph and OpenAI's APIs classifies input text, processes it based on content type, and generates corresponding speech output. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that classifies input text using GPT models, applies content-specific processing, and converts the processed text to speech using OpenAI's TTS API. The system adapts its output based on the identified content type (general, poem, news, or joke). Music Compositor Agent (LangGraph) Overview 🔎 An AI Music Compositor using LangGraph and OpenAI's language models generates custom musical compositions based on user input. The system processes the input through specialized components, each contributing to the final musical piece, which is then converted to a playable MIDI file. Implementation 🛠️ LangGraph orchestrates a workflow that transforms user input into a musical composition, using ChatOpenAI (GPT-4) to generate melody, harmony, and rhythm, which are then style-adapted. The final AI-generated composition is converted to a MIDI file using music21 and can be played back using pygame. Content Intelligence: Multi-Platform Content Generation Agent Overview 🔎 Content Intelligence demonstrates how to build an advanced content generation system that transforms input text into platform-optimized content across multiple social media channels. The system employs LangGraph for workflow orchestration to analyze content, conduct research, and generate tailored content while maintaining brand consistency across different platforms. Implementation 🛠️ Implements a sophisticated workflow using LangGraph to coordinate multiple specialized nodes (Summary, Research, Platform-Specific) through the content generation process. The system features TypedDict and Pydantic models for state management, integration with Tavily Search for research enhancement, and platform-specific content generation using GPT-4. The implementation includes parallel processing for multiple platforms and customizable content templates. Additional Resources 📚 YouTube Explanation Business Meme Generator Using LangGraph and Memegen.link Overview 🔎 The Business Meme Generator demonstrates how to create an AI-powered system that generates contextually relevant memes based on company website analysis. Using LangGraph for workflow orchestration, the system combines Groq's Llama model for text analysis and the Memegen.link API to automatically produce brand-aligned memes for digital marketing. Implementation 🛠️ Implements a state-managed workflow using LangGraph to coordinate website content analysis, meme concept generation, and image creation. The system features Pydantic models for data validation, asynchronous processing with aiohttp, and integration with external APIs (Groq, Memegen.link) to create a complete meme generation pipeline with customizable templates. Additional Resources 📚 YouTube Explanation Murder Mystery Game with LLM Agents Overview 🔎 A text-based detective game that utilizes autonomous LLM agents as interactive characters in a procedurally generated murder mystery. Drawing inspiration from the UNBOUNDED paper, the system creates unique scenarios each time, with players taking on the role of Sherlock Holmes to solve the case through character interviews and deductive reasoning. Implementation 🛠️ Leverages two LangGraph workflows - a main game loop for story/character generation and game progression, and a conversation sub-graph for character interactions. The system uses a combination of LLM-powered narrative generation, character AI, and structured game mechanics to create an immersive investigative experience with replayable storylines. Additional Resources 📚 YouTube Explanation 📊 Analysis and Information Processing Agents Memory-Enhanced Conversational Agent Overview 🔎 A memory-enhanced conversational AI agent incorporates short-term and long-term memory systems to maintain context within conversations and across multiple sessions, improving interaction quality and personalization. Implementation 🛠️ Integrates a language model with separate short-term and long-term memory stores, utilizes a prompt template incorporating both memory types, and employs a memory manager for storage and retrieval. The system includes an interaction loop that updates and utilizes memories for each response. Multi-Agent Collaboration System Overview 🔎 A multi-agent collaboration system combining historical research with data analysis, leveraging large language models to simulate specialized agents working together to answer complex historical questions. Implementation 🛠️ Utilizes a base Agent class to create specialized HistoryResearchAgent and DataAnalysisAgent, orchestrated by a HistoryDataCollaborationSystem. The system follows a five-step process: historical context provision, data needs identification, historical data provision, data analysis, and final synthesis. Self-Improving Agent Overview 🔎 A Self-Improving Agent using LangChain engages in conversations, learns from interactions, and continuously improves its performance over time through reflection and adaptation. Implementation 🛠️ Integrates a language model with chat history management, response generation, and a reflection mechanism. The system employs a learning system that incorporates insights from reflection to enhance future performance, creating a continuous improvement loop. Task-Oriented Agent Overview 🔎 A language model application using LangChain that summarizes text and translates the summary to Spanish, combining custom functions, structured tools, and an agent for efficient text processing. Implementation 🛠️ Utilizes custom functions for summarization and translation, wrapped as structured tools. Employs a prompt template to guide the agent, which orchestrates the use of tools. An agent executor manages the process, taking input text and producing both an English summary and its Spanish translation. Internet Search and Summarize Agent Overview 🔎 An intelligent web research assistant that combines web search capabilities with AI-powered summarization, automating the process of gathering information from the internet and distilling it into concise, relevant summaries. Implementation 🛠️ Integrates a web search module using DuckDuckGo's API, a result parser, and a text summarization engine leveraging OpenAI's language models. The system performs site-specific or general searches, extracts relevant content, generates concise summaries, and compiles attributed results for efficient information retrieval and synthesis. Multi agent research team - Autogen Overview 🔎 This technique explores a multi-agent system for collaborative research using the AutoGen library. It employs agents to solve tasks collaboratively, focusing on efficient execution and quality assurance. The system enhances research by distributing tasks among specialized agents. Implementation 🛠️ Agents are configured with specific roles using the GPT-4 model, including admin, developer, planner, executor, and quality assurance. Interaction management ensures orderly communication with defined transitions. Task execution involves collaborative planning, coding, execution, and quality checking, demonstrating a scalable framework for various domains. Additional Resources 📚 comprehensive solution with UI Blogpost Sales Call Analyzer Overview 🔎 An intelligent system that automates the analysis of sales call recordings by combining audio transcription with advanced natural language processing. The analyzer transcribes audio using OpenAI's Whisper, processes the text using NLP techniques, and generates comprehensive reports including sentiment analysis, key phrases, pain points, and actionable recommendations to improve sales performance. Implementation 🛠️ Utilizes multiple components in a structured workflow: OpenAI Whisper for audio transcription, CrewAI for task automation and agent management, and LangChain for orchestrating the analysis pipeline. The system processes audio through a series of steps from transcription to detailed analysis, leveraging custom agents and tasks to generate structured JSON reports containing insights about customer sentiment, sales opportunities, and recommended improvements. Additional Resources 📚 YouTube Explanation Weather Emergency & Response System Overview 🔎 A comprehensive system demonstrating two agent graph implementations for weather emergency response: a real-time graph processing live weather data, and a hybrid graph combining real and simulated data for testing high-severity scenarios. The system handles complete workflow from data gathering through emergency plan generation, with automated notifications and human verification steps. Implementation 🛠️ Utilizes LangGraph for orchestrating complex workflows with state management, integrating OpenWeatherMap API for real-time data, and Gemini for analysis and response generation. The system incorporates email notifications, social media monitoring simulation, and severity-based routing with configurable human verification for low/medium severity events. Additional Resources 📚 YouTube Explanation Self-Healing Codebase System Overview 🔎 An intelligent system that automatically detects, diagnoses, and fixes runtime code errors using LangGraph workflow orchestration and ChromaDB vector storage. The system maintains a memory of encountered bugs and their fixes through vector embeddings, enabling pattern recognition for similar errors across the codebase. Implementation 🛠️ Utilizes a state-based graph workflow that processes function definitions and runtime arguments through specialized nodes for error detection, code analysis, and fix generation. Incorporates ChromaDB for vector-based storage of bug patterns and fixes, with automated search and retrieval capabilities for similar error patterns, while maintaining code execution safety through structured validation steps. Additional Resources 📚 YouTube Explanation DataScribe: AI-Powered Schema Explorer Overview 🔎 An intelligent agent system that enables intuitive exploration and querying of relational databases through natural language interactions. The system utilizes a fleet of specialized agents, coordinated by a stateful Supervisor, to handle schema discovery, query planning, and data analysis tasks while maintaining contextual understanding through vector-based relationship graphs. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-agent workflow including discovery, inference, and planning agents, with NetworkX for relationship graph visualization and management. The system incorporates dynamic state management through TypedDict classes, maintains database context between sessions using a db_graph attribute, and includes safety measures to prevent unauthorized database modifications. Memory-Enhanced Email Agent (LangGraph & LangMem) Overview 🔎 An intelligent email assistant that combines three types of memory (semantic, episodic, and procedural) to create a system that improves over time. The agent can triage incoming emails, draft contextually appropriate responses using stored knowledge, and enhance its performance based on user feedback. Implementation 🛠️ Leverages LangGraph for workflow orchestration and LangMem for sophisticated memory management across multiple memory types. The system implements a triage workflow with memory-enhanced decision making, specialized tools for email composition and calendar management, and a self-improvement mechanism that updates its own prompts based on feedback and past performance. Additional Resources 📚 Blog Post 📰 News and Information Agents News TL;DR using LangGraph Overview 🔎 A news summarization system that generates concise TL;DR summaries of current events based on user queries. The system leverages large language models for decision making and summarization while integrating with news APIs to access up-to-date content, allowing users to quickly catch up on topics of interest through generated bullet-point summaries. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow combining multiple components: GPT-4o-mini for generating search terms and article summaries, NewsAPI for retrieving article metadata, BeautifulSoup for web scraping article content, and Asyncio for concurrent processing. The system follows a structured pipeline from query processing through article selection and summarization, managing the flow between components to produce relevant TL;DRs of current news articles. Additional Resources 📚 YouTube Explanation Blog Post AInsight: AI/ML Weekly News Reporter Overview 🔎 AInsight demonstrates how to build an intelligent news aggregation and summarization system using a multi-agent architecture. The system employs three specialized agents (NewsSearcher, Summarizer, Publisher) to automatically collect, process and summarize AI/ML news for general audiences through LangGraph-based workflow orchestration. Implementation 🛠️ Implements a state-managed multi-agent system using LangGraph to coordinate the news collection (Tavily API), technical content summarization (GPT-4), and report generation processes. The system features modular architecture with TypedDict-based state management, external API integration, and markdown report generation with customizable templates. Additional Resources 📚 YouTube Explanation Journalism-Focused AI Assistant Overview 🔎 A specialized AI assistant that helps journalists tackle modern journalistic challenges like misinformation, bias, and information overload. The system integrates fact-checking, tone analysis, summarization, and grammar review tools to enhance the accuracy and efficiency of journalistic work while maintaining ethical reporting standards. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized components including language models for analysis and generation, web search integration via DuckDuckGo's API, document parsing tools like PyMuPDFLoader and WebBaseLoader, text splitting with RecursiveCharacterTextSplitter, and structured JSON outputs. Each component works together through a unified workflow to analyze content, verify facts, detect bias, extract quotes, and generate comprehensive reports. Blog Writer (Open AI Swarm) Overview 🔎 A multi-agent system for collaborative blog post creation using OpenAI's Swarm package. It leverages specialized agents to perform research, planning, writing, and editing tasks efficiently. Implementation 🛠️ Utilizes OpenAI's Swarm Package to manage agent interactions. Includes an admin, researcher, planner, writer, and editor, each with specific roles. The system follows a structured workflow: topic setting, outlining, research, drafting, and editing. This approach enhances content creation through task distribution, specialization, and collaborative problem-solving. Additional Resources 📚 Swarm Repo Podcast Internet Search and Generate Agent 🎙️ Overview 🔎 A two step agent that first searches the internet for a given topic and then generates a podcast on the topic found. The search step uses a search agent and search function to find the most relevant information. The second step uses a podcast generation agent and generation function to create a podcast on the topic found. Implementation 🛠️ Utilizes LangGraph to orchestrate a two-step workflow. The first step involves a search agent and function to gather information from the internet. The second step uses a podcast generation agent and function to create a podcast based on the gathered information. 🛍️ Shopping and Product Analysis Agents ShopGenie - Redefining Online Shopping Customer Experience Overview 🔎 An AI-powered shopping assistant that helps customers make informed purchasing decisions even without domain expertise. The system analyzes product information from multiple sources, compares specifications and reviews, identifies the best option based on user needs, and delivers recommendations through email with supporting video reviews, creating a comprehensive shopping experience. Implementation 🛠️ Uses LangGraph to orchestrate a workflow combining Tavily for web search, Llama-3.1-70B for structured data analysis and product comparison, and YouTube API for review video retrieval. The system processes search results through multiple nodes including schema mapping, product comparison, review identification, and email generation. Key features include structured Pydantic models for consistent data handling, retry mechanisms for robust API interactions, and email delivery through SMTP for sharing recommendations. Additional Resources 📚 YouTube Explanation Car Buyer AI Agent Overview 🔎 The Smart Product Buyer AI Agent demonstrates how to build an intelligent system that assists users in making informed purchasing decisions. Using LangGraph and LLM-based intelligence, the system processes user requirements, scrapes product listings from websites like AutoTrader, and provides detailed analysis and recommendations for car purchases. Implementation 🛠️ Implements a state-based workflow using LangGraph to coordinate user interaction, web scraping, and decision support. The system features TypedDict state management, async web scraping with Playwright, and integrates with external APIs for comprehensive product analysis. The implementation includes a Gradio interface for real-time chat interaction and modular scraper architecture for easy extension to additional product categories. Additional Resources 📚 YouTube Explanation 🎯 Task Management and Productivity Agents Taskifier - Intelligent Task Allocation & Management Overview 🔎 An intelligent task management system that analyzes user work styles and creates personalized task breakdown strategies, born from the observation that procrastination often stems from task ambiguity among students and early-career professionals. The system evaluates historical work patterns, gathers relevant task information through web search, and generates customized step-by-step approaches to optimize productivity and reduce workflow paralysis. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-step workflow including work style analysis, information gathering via Tavily API, and customized plan generation. The system maintains state through the process, integrating historical work pattern data with fresh task research to output detailed, personalized task execution plans aligned with the user's natural working style. Additional Resources 📚 YouTube Explanation Grocery Management Agents System Overview 🔎 A multi-agent system built with CrewAI that automates grocery management tasks including receipt interpretation, expiration date tracking, inventory management, and recipe recommendations. The system uses specialized agents to extract data from receipts, estimate product shelf life, track consumption, and suggest recipes to minimize food waste. Implementation 🛠️ Implements four specialized agents using CrewAI - a Receipt Interpreter that extracts item details from receipts, an Expiration Date Estimator that determines shelf life using online sources, a Grocery Tracker that maintains inventory based on consumption, and a Recipe Recommender that suggests meals using available ingredients. Each agent has specific tools and tasks orchestrated through a crew workflow. Additional Resources 📚 YouTube Explanation 🔍 Quality Assurance and Testing Agents LangGraph-Based Systems Inspector Overview 🔎 A comprehensive testing and validation tool for LangGraph-based applications that automatically analyzes system architecture, generates test cases, and identifies potential vulnerabilities through multi-agent inspection. The inspector employs specialized AI testers to evaluate different aspects of the system, from basic functionality to security concerns and edge cases. Implementation 🛠️ Integrates LangGraph for workflow orchestration, multiple LLM-powered testing agents, and a structured evaluation pipeline that includes static analysis, test case generation, and results verification. The system uses Pydantic for data validation, NetworkX for graph representation, and implements a modular architecture that allows for parallel test execution and comprehensive result analysis. Additional Resources 📚 YouTube Explanation Blog Post EU Green Deal FAQ Bot Overview 🔎 The EU Green Deal FAQ Bot demonstrates how to build a RAG-based AI agent that helps businesses understand EU green deal policies. The system processes complex regulatory documents into manageable chunks and provides instant, accurate answers to common questions about environmental compliance, emissions reporting, and waste management requirements. Implementation 🛠️ Implements a sophisticated RAG pipeline using FAISS vectorstore for document storage, semantic chunking for preprocessing, and multiple specialized agents (Retriever, Summarizer, Evaluator) for query processing. The system features query rephrasing for improved accuracy, cross-reference with gold Q&A datasets for answer validation, and comprehensive evaluation metrics to ensure response quality and relevance. Additional Resources 📚 YouTube Explanation Systematic Review Automation System + Paper Draft Creation Overview 🔎 A comprehensive system for automating academic systematic reviews using a directed graph architecture and LangChain components. The system generates complete, publication-ready systematic review papers, automatically processing everything from literature search through final draft generation with multiple revision cycles. Implementation 🛠️ Utilizes a state-based graph workflow that handles paper search and selection (up to 3 papers), PDF processing, and generates a complete academic paper with all standard sections (abstract, introduction, methods, results, conclusions, references). The system incorporates multiple revision cycles with automated critique and improvement phases, all orchestrated through LangGraph state management. Additional Resources 📚 YouTube Explanation 🌟 Special Advanced Technique 🌟 Sophisticated Controllable Agent for Complex RAG Tasks 🤖 Overview 🔎 An advanced RAG solution designed to tackle complex questions that simple semantic similarity-based retrieval cannot solve. This approach uses a sophisticated deterministic graph as the "brain" 🧠 of a highly controllable autonomous agent, capable of answering non-trivial questions from your own data. Implementation 🛠️ • Implement a multi-step process involving question anonymization, high-level planning, task breakdown, adaptive information retrieval and question answering, continuous re-planning, and rigorous answer verification to ensure grounded and accurate responses. Getting Started To begin exploring and building GenAI agents: Clone this repository: Navigate to the technique you're interested in: Follow the detailed implementation guide in each technique's notebook. Contributing We welcome contributions from the community! If you have a new technique or improvement to suggest: Fork the repository Create your feature branch: git checkout -b feature/AmazingFeature Commit your changes: git commit -m 'Add some AmazingFeature' Push to the branch: git push origin feature/AmazingFeature Open a pull request Contributors License This project is licensed under a custom non-commercial license - see the LICENSE file for details. ⭐️ If you find this repository helpful, please consider giving it a star! Keywords: GenAI, Generative AI, Agents, NLP, AI, Machine Learning, Natural Language Processing, LLM, Conversational AI, Task-Oriented AI

openkore
github
LLM Vibe Score0.567
Human Vibe Score0.2670720058425842
OpenKoreMar 28, 2025

openkore

!logo !Language !Stars !Fork !Watch !Issues !Pull Requests !Contributors !GithubWorkflowstatus !GithubWorkflowCI OpenKore is a custom client and intelligent automated assistant for Ragnarok Online. It is a free, open source and cross-platform program (Linux, Windows and MacOS are supported). Prerequisites To run OpenKore you will need: Read the Requirements page on our wiki Quickstart Download OpenKore and extract it. Alternatively, you could press the Windows Key + R, type in `cmd` & enter. Run the following command in the cmd to clone. Note: Git required. Configure OpenKore: documentation. Run openkore.pl (You can run start.exe or wxstart.exe if you use Windows). F.A.Q. (Frequently Asked Questions) Have a problem? Update your openkore or download a new one. Still having problems? Search in Wiki. Search in Forum. Search in Github issues. Cant find what you need? / Do not understand? Ask in IRC Channel. Is it a problem in Openkore? Read things to know before reporting. Things to know Make sure you've read FAQ especially to run latest commit on master branch & checking existed issue for your request. Please post in English. Please use the issue template. Please include informations about your server & any changes you did in your configuration. Briefly explain what happened, take a screenhot & include the error message (If available). Please be advised any developers here are doing this on their free time. Please give some time for anyone to respond. Status of botting on Official Servers | Server | Description | Protection | Status | Supporter | | --- | --- | --- | --- | --- | | aRO | Asia RO | CheatDefender | Not working | N/A | | bRO | Brazil RO | EAC | Not working | N/A | | cRO | China RO | nProtect | Botable | N/A | | euRO | Europe RO | Frost Security | Not working | N/A | | euRO-Prime | Europe RO (Prime) | Frost Security | Not working | N/A | | iRO Renewal | International RO | EAC | Not working | N/A | | idRO | Indonesia RO | EAC | Not Working | N/A | | idRO-Retro | Indonesia RO (Retro) | Delphine | Not Working | N/A | | jRO | Japan RO | nProtect | Need Verification | N/A | | kRO | Korea RO | nProtect | Botable | N/A | | kRO-Zero | Korea RO (Zero) | nProtect | Botable | N/A | | ruRO-Prime | Russia RO (Prime) | Frost Security | Not Working | ya4ept | | tRO | Thailand RO | EAC | Not Working | N/A | | tRO-Classic | Thailand RO (Classic) | EAC | Not Working | N/A | | twRO | Taiwan RO | CheatDefender | Not Working | N/A | | vRO | Vietnam RO | nProtect | Not Working | N/A | Contributing OpenKore is developed by a team located around the world. Check out the documentation and if necessary, submit a pull request. Contacts OpenKore Wiki OpenKore forum IRC Channel Connect IRC with Kiwiirc Brazilian Community Russian Community Warning Other communities or websites are not affiliated to openkore.com Other Links Openkore History Legacy Changelog Openkore RoadMap Feature Requests and TODO Wiki and Feature Requests GitHub License This software is open source, licensed under the GNU General Public License, version 2. Basically, this means that you're free to use and allowed to modify and distribute this software. However, if you distribute modified versions, you MUST also distribute the source code. See http://www.gnu.org/licenses/gpl.html for the full license.

AITreasureBox
github
LLM Vibe Score0.447
Human Vibe Score0.1014145151561518
superiorluMar 28, 2025

AITreasureBox

AI TreasureBox English | 中文 Collect practical AI repos, tools, websites, papers and tutorials on AI. Translated from ChatGPT, picture from Midjourney. Catalog Repos Tools Websites Report&Paper Tutorials Repos updated repos and stars every 2 hours and re-ranking automatically. | No. | Repos | Description | | ----:|:-----------------------------------------|:------------------------------------------------------------------------------------------------------| | 1|🔥codecrafters-io/build-your-own-x !2025-03-28364681428|Master programming by recreating your favorite technologies from scratch.| | 2|sindresorhus/awesome !2025-03-28353614145|😎 Awesome lists about all kinds of interesting topics| | 3|public-apis/public-apis !2025-03-28334299125|A collective list of free APIs| | 4|kamranahmedse/developer-roadmap !2025-03-2831269540|Interactive roadmaps, guides and other educational content to help developers grow in their careers.| | 5|vinta/awesome-python !2025-03-28238581114|A curated list of awesome Python frameworks, libraries, software and resources| | 6|practical-tutorials/project-based-learning !2025-03-28222661124|Curated list of project-based tutorials| | 7|tensorflow/tensorflow !2025-03-281888714|An Open Source Machine Learning Framework for Everyone| | 8|Significant-Gravitas/AutoGPT !2025-03-2817391338|An experimental open-source attempt to make GPT-4 fully autonomous.| | 9|jackfrued/Python-100-Days !2025-03-2816305141|Python - 100天从新手到大师| | 10|AUTOMATIC1111/stable-diffusion-webui !2025-03-2815011553|Stable Diffusion web UI| | 11|huggingface/transformers !2025-03-2814207850|🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.| | 12|ollama/ollama !2025-03-28135166151|Get up and running with Llama 2, Mistral, Gemma, and other large language models.| | 13|f/awesome-chatgpt-prompts !2025-03-2812212738 |This repo includes ChatGPT prompt curation to use ChatGPT better.| | 14|justjavac/free-programming-books-zhCN !2025-03-2811316119|📚 免费的计算机编程类中文书籍,欢迎投稿| | 15|krahets/hello-algo !2025-03-2811107930|《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing| | 16|yt-dlp/yt-dlp !2025-03-28105801114|A feature-rich command-line audio/video downloader| | 17|langchain-ai/langchain !2025-03-2810449479|⚡ Building applications with LLMs through composability ⚡| | 18|goldbergyoni/nodebestpractices !2025-03-281021629|✅ The Node.js best practices list (July 2024)| | 19|puppeteer/puppeteer !2025-03-289018212|JavaScript API for Chrome and Firefox| | 20|pytorch/pytorch !2025-03-288833938|Tensors and Dynamic neural networks in Python with strong GPU acceleration| | 21|neovim/neovim !2025-03-288781482|Vim-fork focused on extensibility and usability| | 22|🔥🔥langgenius/dify !2025-03-2887342639 |One API for plugins and datasets, one interface for prompt engineering and visual operation, all for creating powerful AI applications.| | 23|mtdvio/every-programmer-should-know !2025-03-28867069|A collection of (mostly) technical things every software developer should know about| | 24|open-webui/open-webui !2025-03-2886025159|User-friendly WebUI for LLMs (Formerly Ollama WebUI)| | 25|ChatGPTNextWeb/NextChat !2025-03-288231521|✨ Light and Fast AI Assistant. Support: Web | | 26|supabase/supabase !2025-03-287990956|The open source Firebase alternative.| | 27|openai/whisper !2025-03-287905542|Robust Speech Recognition via Large-Scale Weak Supervision| | 28|home-assistant/core !2025-03-287773219|🏡 Open source home automation that puts local control and privacy first.| | 29|tensorflow/models !2025-03-28774694|Models and examples built with TensorFlow| | 30| ggerganov/llama.cpp !2025-03-287731836 | Port of Facebook's LLaMA model in C/C++ | | 31|3b1b/manim !2025-03-287641918|Animation engine for explanatory math videos| | 32|microsoft/generative-ai-for-beginners !2025-03-287623860|12 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/| | 33|nomic-ai/gpt4all !2025-03-28729285 |gpt4all: an ecosystem of open-source chatbots trained on a massive collection of clean assistant data including code, stories and dialogue| | 34|comfyanonymous/ComfyUI !2025-03-2872635111|The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.| | 35|bregman-arie/devops-exercises !2025-03-2872225209|Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions| | 36|elastic/elasticsearch !2025-03-28721419|Free and Open, Distributed, RESTful Search Engine| | 37|🔥n8n-io/n8n !2025-03-2872093495|Free and source-available fair-code licensed workflow automation tool. Easily automate tasks across different services.| | 38|fighting41love/funNLP !2025-03-287200422|The Most Powerful NLP-Weapon Arsenal| | 39|hoppscotch/hoppscotch !2025-03-287060134|Open source API development ecosystem - https://hoppscotch.io (open-source alternative to Postman, Insomnia)| | 40|abi/screenshot-to-code !2025-03-286932817|Drop in a screenshot and convert it to clean HTML/Tailwind/JS code| | 41|binary-husky/gptacademic !2025-03-28680374|Academic Optimization of GPT| | 42|d2l-ai/d2l-zh !2025-03-286774142|Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries| | 43|josephmisiti/awesome-machine-learning !2025-03-286739215|A curated list of awesome Machine Learning frameworks, libraries and software.| | 44|grafana/grafana !2025-03-286725414|The open and composable observability and data visualization platform. Visualize metrics, logs, and traces from multiple sources like Prometheus, Loki, Elasticsearch, InfluxDB, Postgres and many more.| | 45|python/cpython !2025-03-286602218|The Python programming language| | 46|apache/superset !2025-03-286519020|Apache Superset is a Data Visualization and Data Exploration Platform| | 47|xtekky/gpt4free !2025-03-28639391 |decentralizing the Ai Industry, free gpt-4/3.5 scripts through several reverse engineered API's ( poe.com, phind.com, chat.openai.com etc...)| | 48|sherlock-project/sherlock !2025-03-286332536|Hunt down social media accounts by username across social networks| | 49|twitter/the-algorithm !2025-03-28630586 |Source code for Twitter's Recommendation Algorithm| | 50|keras-team/keras !2025-03-28627835|Deep Learning for humans| | 51|openai/openai-cookbook !2025-03-28625136 |Examples and guides for using the OpenAI API| | 52|immich-app/immich !2025-03-286238670|High performance self-hosted photo and video management solution.| | 53|AppFlowy-IO/AppFlowy !2025-03-286173528|Bring projects, wikis, and teams together with AI. AppFlowy is an AI collaborative workspace where you achieve more without losing control of your data. The best open source alternative to Notion.| | 54|scikit-learn/scikit-learn !2025-03-286158212|scikit-learn: machine learning in Python| | 55|binhnguyennus/awesome-scalability !2025-03-286117021|The Patterns of Scalable, Reliable, and Performant Large-Scale Systems| | 56|labmlai/annotateddeeplearningpaperimplementations !2025-03-285951726|🧑‍🏫 59 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠| | 57|OpenInterpreter/open-interpreter !2025-03-285894710|A natural language interface for computers| | 58|lobehub/lobe-chat !2025-03-285832054|🤖 Lobe Chat - an open-source, extensible (Function Calling), high-performance chatbot framework. It supports one-click free deployment of your private ChatGPT/LLM web application.| | 59|meta-llama/llama !2025-03-28579536|Inference code for Llama models| | 60|nuxt/nuxt !2025-03-28566437|The Intuitive Vue Framework.| | 61|imartinez/privateGPT !2025-03-28555192|Interact with your documents using the power of GPT, 100% privately, no data leaks| | 62|Stirling-Tools/Stirling-PDF !2025-03-285500846|#1 Locally hosted web application that allows you to perform various operations on PDF files| | 63|PlexPt/awesome-chatgpt-prompts-zh !2025-03-285459720|ChatGPT Chinese Training Guide. Guidelines for various scenarios. Learn how to make it listen to you| | 64|dair-ai/Prompt-Engineering-Guide !2025-03-285451025 |🐙 Guides, papers, lecture, notebooks and resources for prompt engineering| | 65|ageitgey/facerecognition !2025-03-28544382|The world's simplest facial recognition api for Python and the command line| | 66|CorentinJ/Real-Time-Voice-Cloning !2025-03-285384814|Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 67|geekan/MetaGPT !2025-03-285375376|The Multi-Agent Meta Programming Framework: Given one line Requirement, return PRD, Design, Tasks, Repo | | 68|gpt-engineer-org/gpt-engineer !2025-03-285367419|Specify what you want it to build, the AI asks for clarification, and then builds it.| | 69|lencx/ChatGPT !2025-03-2853653-3|🔮 ChatGPT Desktop Application (Mac, Windows and Linux)| | 70|deepfakes/faceswap !2025-03-28535672|Deepfakes Software For All| | 71|langflow-ai/langflow !2025-03-285319584|Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.| | 72|commaai/openpilot !2025-03-28529759|openpilot is an operating system for robotics. Currently, it upgrades the driver assistance system on 275+ supported cars.| | 73|clash-verge-rev/clash-verge-rev !2025-03-2852848124|Continuation of Clash Verge - A Clash Meta GUI based on Tauri (Windows, MacOS, Linux)| | 74|All-Hands-AI/OpenHands !2025-03-285150675|🙌 OpenHands: Code Less, Make More| | 75|xai-org/grok-1 !2025-03-28502504|Grok open release| | 76|meilisearch/meilisearch !2025-03-284999122|A lightning-fast search API that fits effortlessly into your apps, websites, and workflow| | 77|🔥browser-use/browser-use !2025-03-2849910294|Make websites accessible for AI agents| | 78|jgthms/bulma !2025-03-28496783|Modern CSS framework based on Flexbox| | 79|facebookresearch/segment-anything !2025-03-284947116|The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.| |!green-up-arrow.svg 80|hacksider/Deep-Live-Cam !2025-03-2848612146|real time face swap and one-click video deepfake with only a single image (uncensored)| |!red-down-arrow 81|mlabonne/llm-course !2025-03-284860934|Course with a roadmap and notebooks to get into Large Language Models (LLMs).| | 82|PaddlePaddle/PaddleOCR !2025-03-284785530|Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and deployment among server, mobile, embedded and IoT devices)| | 83|alist-org/alist !2025-03-284732618|🗂️A file list/WebDAV program that supports multiple storages, powered by Gin and Solidjs. / 一个支持多存储的文件列表/WebDAV程序,使用 Gin 和 Solidjs。| | 84|infiniflow/ragflow !2025-03-2847027129|RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.| | 85|Avik-Jain/100-Days-Of-ML-Code !2025-03-284679312|100 Days of ML Coding| | 86|v2ray/v2ray-core !2025-03-28458706|A platform for building proxies to bypass network restrictions.| | 87|hiyouga/LLaMA-Factory !2025-03-284555881|Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)| | 88|Asabeneh/30-Days-Of-Python !2025-03-284544930|30 days of Python programming challenge is a step-by-step guide to learn the Python programming language in 30 days. This challenge may take more than100 days, follow your own pace. These videos may help too: https://www.youtube.com/channel/UC7PNRuno1rzYPb1xLa4yktw| | 89|type-challenges/type-challenges !2025-03-284488511|Collection of TypeScript type challenges with online judge| | 90|lllyasviel/Fooocus !2025-03-284402716|Focus on prompting and generating| | 91|RVC-Boss/GPT-SoVITS !2025-03-284327738|1 min voice data can also be used to train a good TTS model! (few shot voice cloning)| | 92|rasbt/LLMs-from-scratch !2025-03-284320667|Implementing a ChatGPT-like LLM from scratch, step by step| | 93|oobabooga/text-generation-webui !2025-03-284302012 |A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, OPT, and GALACTICA.| | 94|vllm-project/vllm !2025-03-2842982102|A high-throughput and memory-efficient inference and serving engine for LLMs| | 95|dani-garcia/vaultwarden !2025-03-284297121|Unofficial Bitwarden compatible server written in Rust, formerly known as bitwarden_rs| | 96|microsoft/autogen !2025-03-284233049|Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ| | 97|jeecgboot/JeecgBoot !2025-03-284205920|🔥「企业级低代码平台」前后端分离架构SpringBoot 2.x/3.x,SpringCloud,Ant Design&Vue3,Mybatis,Shiro,JWT。强大的代码生成器让前后端代码一键生成,无需写任何代码! 引领新的开发模式OnlineCoding->代码生成->手工MERGE,帮助Java项目解决70%重复工作,让开发更关注业务,既能快速提高效率,帮助公司节省成本,同时又不失灵活性。| | 98|Mintplex-Labs/anything-llm !2025-03-284186955|A full-stack application that turns any documents into an intelligent chatbot with a sleek UI and easier way to manage your workspaces.| | 99|THUDM/ChatGLM-6B !2025-03-28410192 |ChatGLM-6B: An Open Bilingual Dialogue Language Model| | 100|hpcaitech/ColossalAI !2025-03-28406902|Making large AI models cheaper, faster and more accessible| | 101|Stability-AI/stablediffusion !2025-03-28406337|High-Resolution Image Synthesis with Latent Diffusion Models| | 102|mingrammer/diagrams !2025-03-28405063|🎨 Diagram as Code for prototyping cloud system architectures| | 103|Kong/kong !2025-03-28404616|🦍 The Cloud-Native API Gateway and AI Gateway.| | 104|getsentry/sentry !2025-03-284040913|Developer-first error tracking and performance monitoring| | 105| karpathy/nanoGPT !2025-03-284034613 |The simplest, fastest repository for training/finetuning medium-sized GPTs| | 106|fastlane/fastlane !2025-03-2840014-1|🚀 The easiest way to automate building and releasing your iOS and Android apps| | 107|psf/black !2025-03-28399765|The uncompromising Python code formatter| | 108|OpenBB-finance/OpenBBTerminal !2025-03-283972074 |Investment Research for Everyone, Anywhere.| | 109|2dust/v2rayNG !2025-03-283943415|A V2Ray client for Android, support Xray core and v2fly core| | 110|apache/airflow !2025-03-283937314|Apache Airflow - A platform to programmatically author, schedule, and monitor workflows| | 111|KRTirtho/spotube !2025-03-283902746|🎧 Open source Spotify client that doesn't require Premium nor uses Electron! Available for both desktop & mobile!| | 112|coqui-ai/TTS !2025-03-283889719 |🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production| | 113|ggerganov/whisper.cpp !2025-03-283882116|Port of OpenAI's Whisper model in C/C++| | 114|ultralytics/ultralytics !2025-03-283866951|NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite| | 115|typst/typst !2025-03-283863914|A new markup-based typesetting system that is powerful and easy to learn.| | 116|streamlit/streamlit !2025-03-283845828|Streamlit — A faster way to build and share data apps.| | 117|LC044/WeChatMsg !2025-03-283836931|提取微信聊天记录,将其导出成HTML、Word、Excel文档永久保存,对聊天记录进行分析生成年度聊天报告,用聊天数据训练专属于个人的AI聊天助手| | 118|lm-sys/FastChat !2025-03-283822112 |An open platform for training, serving, and evaluating large languages. Release repo for Vicuna and FastChat-T5.| | 119|NaiboWang/EasySpider !2025-03-283819013|A visual no-code/code-free web crawler/spider易采集:一个可视化浏览器自动化测试/数据采集/爬虫软件,可以无代码图形化的设计和执行爬虫任务。别名:ServiceWrapper面向Web应用的智能化服务封装系统。| | 120|microsoft/DeepSpeed !2025-03-283765816 |A deep learning optimization library that makes distributed training and inference easy, efficient, and effective| | 121|QuivrHQ/quivr !2025-03-28376067|Your GenAI Second Brain 🧠 A personal productivity assistant (RAG) ⚡️🤖 Chat with your docs (PDF, CSV, ...) & apps using Langchain, GPT 3.5 / 4 turbo, Private, Anthropic, VertexAI, Ollama, LLMs, that you can share with users ! Local & Private alternative to OpenAI GPTs & ChatGPT powered by retrieval-augmented generation.| | 122|freqtrade/freqtrade !2025-03-283757817 |Free, open source crypto trading bot| | 123|suno-ai/bark !2025-03-28373178 |🔊 Text-Prompted Generative Audio Model| | 124|🔥cline/cline !2025-03-2837307282|Autonomous coding agent right in your IDE, capable of creating/editing files, executing commands, and more with your permission every step of the way.| | 125|LAION-AI/Open-Assistant !2025-03-28372712 |OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.| | 126|penpot/penpot !2025-03-283716217|Penpot: The open-source design tool for design and code collaboration| | 127|gradio-app/gradio !2025-03-283713320|Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!| | 128|FlowiseAI/Flowise !2025-03-283667135 |Drag & drop UI to build your customized LLM flow using LangchainJS| | 129|SimplifyJobs/Summer2025-Internships !2025-03-28366506|Collection of Summer 2025 tech internships!| | 130|TencentARC/GFPGAN !2025-03-28365027 |GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration.| | 131|ray-project/ray !2025-03-283626819|Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a toolkit of libraries (Ray AIR) for accelerating ML workloads.| | 132|babysor/MockingBird !2025-03-28360498|🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time| | 133|unslothai/unsloth !2025-03-283603691|5X faster 50% less memory LLM finetuning| | 134|zhayujie/chatgpt-on-wechat !2025-03-283600124 |Wechat robot based on ChatGPT, which uses OpenAI api and itchat library| | 135|upscayl/upscayl !2025-03-283599824|🆙 Upscayl - Free and Open Source AI Image Upscaler for Linux, MacOS and Windows built with Linux-First philosophy.| | 136|freeCodeCamp/devdocs !2025-03-28359738|API Documentation Browser| | 137|XingangPan/DragGAN !2025-03-28359043 |Code for DragGAN (SIGGRAPH 2023)| | 138|2noise/ChatTTS !2025-03-283543922|ChatTTS is a generative speech model for daily dialogue.| | 139|google-research/google-research !2025-03-28352207 |Google Research| | 140|karanpratapsingh/system-design !2025-03-28351003|Learn how to design systems at scale and prepare for system design interviews| | 141|lapce/lapce !2025-03-28350855|Lightning-fast and Powerful Code Editor written in Rust| | 142| microsoft/TaskMatrix !2025-03-2834500-3 | Talking, Drawing and Editing with Visual Foundation Models| | 143|chatchat-space/Langchain-Chatchat !2025-03-283442020|Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain| | 144|unclecode/crawl4ai !2025-03-283434163|🔥🕷️ Crawl4AI: Open-source LLM Friendly Web Crawler & Scrapper| | 145|Bin-Huang/chatbox !2025-03-283374733 |A desktop app for GPT-4 / GPT-3.5 (OpenAI API) that supports Windows, Mac & Linux| | 146|milvus-io/milvus !2025-03-283366525 |A cloud-native vector database, storage for next generation AI applications| | 147|mendableai/firecrawl !2025-03-2833297128|🔥 Turn entire websites into LLM-ready markdown| | 148|pola-rs/polars !2025-03-283269320|Fast multi-threaded, hybrid-out-of-core query engine focussing on DataFrame front-ends| | 149|Pythagora-io/gpt-pilot !2025-03-28325321|PoC for a scalable dev tool that writes entire apps from scratch while the developer oversees the implementation| | 150|hashicorp/vault !2025-03-28320797|A tool for secrets management, encryption as a service, and privileged access management| | 151|shardeum/shardeum !2025-03-28319580|Shardeum is an EVM based autoscaling blockchain| | 152|Chanzhaoyu/chatgpt-web !2025-03-28319242 |A demonstration website built with Express and Vue3 called ChatGPT| | 153|lllyasviel/ControlNet !2025-03-283186413 |Let us control diffusion models!| | 154|google/jax !2025-03-28317727|Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more| | 155|facebookresearch/detectron2 !2025-03-28315987|Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.| | 156|myshell-ai/OpenVoice !2025-03-28315233|Instant voice cloning by MyShell| | 157|TheAlgorithms/C-Plus-Plus !2025-03-283151411|Collection of various algorithms in mathematics, machine learning, computer science and physics implemented in C++ for educational purposes.| | 158|hiroi-sora/Umi-OCR !2025-03-283138129|OCR图片转文字识别软件,完全离线。截屏/批量导入图片,支持多国语言、合并段落、竖排文字。可排除水印区域,提取干净的文本。基于 PaddleOCR 。| | 159|mudler/LocalAI !2025-03-283127815|🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.| | 160|facebookresearch/fairseq !2025-03-28312124 |Facebook AI Research Sequence-to-Sequence Toolkit written in Python.| | 161|alibaba/nacos !2025-03-28310559|an easy-to-use dynamic service discovery, configuration and service management platform for building cloud native applications.| | 162|yunjey/pytorch-tutorial !2025-03-28310326|PyTorch Tutorial for Deep Learning Researchers| | 163|v2fly/v2ray-core !2025-03-28307448|A platform for building proxies to bypass network restrictions.| | 164|mckaywrigley/chatbot-ui !2025-03-283067714|The open-source AI chat interface for everyone.| | 165|TabbyML/tabby !2025-03-28305949 |Self-hosted AI coding assistant| | 166|deepseek-ai/awesome-deepseek-integration !2025-03-283053193|| | 167|danielmiessler/fabric !2025-03-283028914|fabric is an open-source framework for augmenting humans using AI.| | 168|xinntao/Real-ESRGAN !2025-03-283026623 |Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.| | 169|paul-gauthier/aider !2025-03-283014642|aider is GPT powered coding in your terminal| | 170|tatsu-lab/stanfordalpaca !2025-03-28299022 |Code and documentation to train Stanford's Alpaca models, and generate the data.| | 171|DataTalksClub/data-engineering-zoomcamp !2025-03-282971817|Free Data Engineering course!| | 172|HeyPuter/puter !2025-03-282967014|🌐 The Internet OS! Free, Open-Source, and Self-Hostable.| | 173|mli/paper-reading !2025-03-282962314|Classic Deep Learning and In-Depth Reading of New Papers Paragraph by Paragraph| | 174|linexjlin/GPTs !2025-03-28295568|leaked prompts of GPTs| | 175|s0md3v/roop !2025-03-28295286 |one-click deepfake (face swap)| | 176|JushBJJ/Mr.-Ranedeer-AI-Tutor !2025-03-2829465-1 |A GPT-4 AI Tutor Prompt for customizable personalized learning experiences.| | 177|opendatalab/MinerU !2025-03-282927074|A one-stop, open-source, high-quality data extraction tool, supports PDF/webpage/e-book extraction.一站式开源高质量数据提取工具,支持PDF/网页/多格式电子书提取。| | 178|mouredev/Hello-Python !2025-03-282920720|Curso para aprender el lenguaje de programación Python desde cero y para principiantes. 75 clases, 37 horas en vídeo, código, proyectos y grupo de chat. Fundamentos, frontend, backend, testing, IA...| | 179|Lightning-AI/pytorch-lightning !2025-03-28292039|Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.| | 180|crewAIInc/crewAI !2025-03-282919344|Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.| | 181|facebook/folly !2025-03-282916612|An open-source C++ library developed and used at Facebook.| | 182|google-ai-edge/mediapipe !2025-03-28291519|Cross-platform, customizable ML solutions for live and streaming media.| | 183| getcursor/cursor !2025-03-282892025 | An editor made for programming with AI| | 184|chatanywhere/GPTAPIfree !2025-03-282856424|Free ChatGPT API Key, Free ChatGPT API, supports GPT-4 API (free), ChatGPT offers a free domestic forwarding API that allows direct connections without the need for a proxy. It can be used in conjunction with software/plugins like ChatBox, significantly reducing interface usage costs. Enjoy unlimited and unrestricted chatting within China| | 185|meta-llama/llama3 !2025-03-28285552|The official Meta Llama 3 GitHub site| | 186|tinygrad/tinygrad !2025-03-282845811|You like pytorch? You like micrograd? You love tinygrad! ❤️| | 187|google-research/tuningplaybook !2025-03-282841514|A playbook for systematically maximizing the performance of deep learning models.| | 188|huggingface/diffusers !2025-03-282830222|🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.| | 189|tokio-rs/tokio !2025-03-28282408|A runtime for writing reliable asynchronous applications with Rust. Provides I/O, networking, scheduling, timers, ...| | 190|RVC-Project/Retrieval-based-Voice-Conversion-WebUI !2025-03-282823817|Voice data !2025-03-282822612|Jan is an open source alternative to ChatGPT that runs 100% offline on your computer| | 192|openai/CLIP !2025-03-282814720|CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image| | 193|🔥khoj-ai/khoj !2025-03-2828112313|Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.| | 194| acheong08/ChatGPT !2025-03-2828054-2 | Reverse engineered ChatGPT API | | 195|iperov/DeepFaceLive !2025-03-28279345 |Real-time face swap for PC streaming or video calls| | 196|eugeneyan/applied-ml !2025-03-28278471|📚 Papers & tech blogs by companies sharing their work on data science & machine learning in production.| | 197|XTLS/Xray-core !2025-03-282778213|Xray, Penetrates Everything. Also the best v2ray-core, with XTLS support. Fully compatible configuration.| | 198|feder-cr/JobsApplierAIAgent !2025-03-282776410|AutoJobsApplierAI_Agent aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in an automated and personalized way.| | 199|mindsdb/mindsdb !2025-03-282750631|The platform for customizing AI from enterprise data| | 200|DataExpert-io/data-engineer-handbook !2025-03-282721611|This is a repo with links to everything you'd ever want to learn about data engineering| | 201|exo-explore/exo !2025-03-282721633|Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚| | 202|taichi-dev/taichi !2025-03-2826926-1|Productive, portable, and performant GPU programming in Python.| | 203|mem0ai/mem0 !2025-03-282689134|The memory layer for Personalized AI| | 204|svc-develop-team/so-vits-svc !2025-03-28268096 |SoftVC VITS Singing Voice Conversion| | 205|OpenBMB/ChatDev !2025-03-28265624|Create Customized Software using Natural Language Idea (through Multi-Agent Collaboration)| | 206|roboflow/supervision !2025-03-282632010|We write your reusable computer vision tools. 💜| | 207|drawdb-io/drawdb !2025-03-282626913|Free, simple, and intuitive online database design tool and SQL generator.| | 208|karpathy/llm.c !2025-03-28261633|LLM training in simple, raw C/CUDA| | 209|airbnb/lottie-ios !2025-03-28261431|An iOS library to natively render After Effects vector animations| | 210|openai/openai-python !2025-03-282607713|The OpenAI Python library provides convenient access to the OpenAI API from applications written in the Python language.| | 211|academic/awesome-datascience !2025-03-28259876|📝 An awesome Data Science repository to learn and apply for real world problems.| | 212|harry0703/MoneyPrinterTurbo !2025-03-282576618|Generate short videos with one click using a large model| | 213|gabime/spdlog !2025-03-282571511|Fast C++ logging library.| | 214|ocrmypdf/OCRmyPDF !2025-03-2825674217|OCRmyPDF adds an OCR text layer to scanned PDF files, allowing them to be searched| | 215|Vision-CAIR/MiniGPT-4 !2025-03-28256170 |Enhancing Vision-language Understanding with Advanced Large Language Models| | 216|Stability-AI/generative-models !2025-03-28255936|Generative Models by Stability AI| | 217|DS4SD/docling !2025-03-282555662|Get your docs ready for gen AI| | 218|PostHog/posthog !2025-03-282533227|🦔 PostHog provides open-source product analytics, session recording, feature flagging and A/B testing that you can self-host.| | 219|nrwl/nx !2025-03-282509612|Smart Monorepos · Fast CI| | 220|continuedev/continue !2025-03-282500737|⏩ the open-source copilot chat for software development—bring the power of ChatGPT to VS Code| | 221|opentofu/opentofu !2025-03-28247968|OpenTofu lets you declaratively manage your cloud infrastructure.| | 222|invoke-ai/InvokeAI !2025-03-28247293|InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.| | 223|deepinsight/insightface !2025-03-282471615 |State-of-the-art 2D and 3D Face Analysis Project| | 224|apache/flink !2025-03-28246865|Apache Flink| | 225|ComposioHQ/composio !2025-03-28246436|Composio equips agents with well-crafted tools empowering them to tackle complex tasks| | 226|Genesis-Embodied-AI/Genesis !2025-03-282458314|A generative world for general-purpose robotics & embodied AI learning.| | 227|stretchr/testify !2025-03-28243184|A toolkit with common assertions and mocks that plays nicely with the standard library| | 228| yetone/openai-translator !2025-03-28242921 | Browser extension and cross-platform desktop application for translation based on ChatGPT API | | 229|frappe/erpnext !2025-03-282425211|Free and Open Source Enterprise Resource Planning (ERP)| | 230|songquanpeng/one-api !2025-03-282410034|OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.| | 231| microsoft/JARVIS !2025-03-28240604 | a system to connect LLMs with ML community | | 232|google/flatbuffers !2025-03-28239965|FlatBuffers: Memory Efficient Serialization Library| | 233|microsoft/graphrag !2025-03-282398928|A modular graph-based Retrieval-Augmented Generation (RAG) system| | 234|rancher/rancher !2025-03-28239675|Complete container management platform| | 235|bazelbuild/bazel !2025-03-282384618|a fast, scalable, multi-language and extensible build system| | 236|modularml/mojo !2025-03-28238236 |The Mojo Programming Language| | 237|danny-avila/LibreChat !2025-03-282378753|Enhanced ChatGPT Clone: Features OpenAI, GPT-4 Vision, Bing, Anthropic, OpenRouter, Google Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development| |!green-up-arrow.svg 238|🔥🔥🔥Shubhamsaboo/awesome-llm-apps !2025-03-28237391211|Collection of awesome LLM apps with RAG using OpenAI, Anthropic, Gemini and opensource models.| |!red-down-arrow 239|microsoft/semantic-kernel !2025-03-282373611|Integrate cutting-edge LLM technology quickly and easily into your apps| |!red-down-arrow 240|TheAlgorithms/Rust !2025-03-28236995|All Algorithms implemented in Rust| | 241|stanford-oval/storm !2025-03-28236326|An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.| | 242|openai/gpt-2 !2025-03-28232483|Code for the paper "Language Models are Unsupervised Multitask Learners"| | 243|labring/FastGPT !2025-03-282319445|A platform that uses the OpenAI API to quickly build an AI knowledge base, supporting many-to-many relationships.| | 244|pathwaycom/llm-app !2025-03-2822928-10|Ready-to-run cloud templates for RAG, AI pipelines, and enterprise search with live data. 🐳Docker-friendly.⚡Always in sync with Sharepoint, Google Drive, S3, Kafka, PostgreSQL, real-time data APIs, and more.| | 245|warpdotdev/Warp !2025-03-282286825|Warp is a modern, Rust-based terminal with AI built in so you and your team can build great software, faster.| | 246|🔥agno-agi/agno !2025-03-2822833298|Agno is a lightweight library for building Multimodal Agents. It exposes LLMs as a unified API and gives them superpowers like memory, knowledge, tools and reasoning.| | 247|qdrant/qdrant !2025-03-282275214 |Qdrant - Vector Database for the next generation of AI applications. Also available in the cloud https://cloud.qdrant.io/| | 248|ashishpatel26/500-AI-Machine-learning-Deep-learning-Computer-vision-NLP-Projects-with-code !2025-03-282271815|500 AI Machine learning Deep learning Computer vision NLP Projects with code| | 249|stanfordnlp/dspy !2025-03-282268321|Stanford DSPy: The framework for programming—not prompting—foundation models| | 250|PaddlePaddle/Paddle !2025-03-28226246|PArallel Distributed Deep LEarning: Machine Learning Framework from Industrial Practice (『飞桨』核心框架,深度学习&机器学习高性能单机、分布式训练和跨平台部署)| | 251|zulip/zulip !2025-03-28225464|Zulip server and web application. Open-source team chat that helps teams stay productive and focused.| | 252|Hannibal046/Awesome-LLM !2025-03-282240721|Awesome-LLM: a curated list of Large Language Model| | 253|facefusion/facefusion !2025-03-282218812|Next generation face swapper and enhancer| | 254|Mozilla-Ocho/llamafile !2025-03-28220624|Distribute and run LLMs with a single file.| | 255|yuliskov/SmartTube !2025-03-282201614|SmartTube - an advanced player for set-top boxes and tvs running Android OS| | 256|haotian-liu/LLaVA !2025-03-282201316 |Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.| | 257|ashishps1/awesome-system-design-resources !2025-03-282189367|This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems| | 258|Cinnamon/kotaemon !2025-03-28218248|An open-source RAG-based tool for chatting with your documents.| | 259|CodePhiliaX/Chat2DB !2025-03-282179757|🔥🔥🔥AI-driven database tool and SQL client, The hottest GUI client, supporting MySQL, Oracle, PostgreSQL, DB2, SQL Server, DB2, SQLite, H2, ClickHouse, and more.| | 260|blakeblackshear/frigate !2025-03-282177113|NVR with realtime local object detection for IP cameras| | 261|facebookresearch/audiocraft !2025-03-28217111|Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.| | 262|karpathy/minGPT !2025-03-28216567|A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training| | 263|grpc/grpc-go !2025-03-282159510|The Go language implementation of gRPC. HTTP/2 based RPC| | 264|HumanSignal/label-studio !2025-03-282137618|Label Studio is a multi-type data labeling and annotation tool with standardized output format| | 265|yoheinakajima/babyagi !2025-03-28212764 |uses OpenAI and Pinecone APIs to create, prioritize, and execute tasks, This is a pared-down version of the original Task-Driven Autonomous Agent| | 266|deepseek-ai/DeepSeek-Coder !2025-03-282118210|DeepSeek Coder: Let the Code Write Itself| | 267|BuilderIO/gpt-crawler !2025-03-282118010|Crawl a site to generate knowledge files to create your own custom GPT from a URL| | 268| openai/chatgpt-retrieval-plugin !2025-03-2821152-1 | Plugins are chat extensions designed specifically for language models like ChatGPT, enabling them to access up-to-date information, run computations, or interact with third-party services in response to a user's request.| | 269|microsoft/OmniParser !2025-03-282113123|A simple screen parsing tool towards pure vision based GUI agent| | 270|black-forest-labs/flux !2025-03-282107219|Official inference repo for FLUX.1 models| | 271|ItzCrazyKns/Perplexica !2025-03-282099154|Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI| | 272|microsoft/unilm !2025-03-28209876|Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities| | 273|Sanster/lama-cleaner !2025-03-282077614|Image inpainting tool powered by SOTA AI Model. Remove any unwanted object, defect, people from your pictures or erase and replace(powered by stable diffusion) any thing on your pictures.| | 274|assafelovic/gpt-researcher !2025-03-282057222|GPT based autonomous agent that does online comprehensive research on any given topic| | 275|PromtEngineer/localGPT !2025-03-28204230 |Chat with your documents on your local device using GPT models. No data leaves your device and 100% private.| | 276|elastic/kibana !2025-03-28203482|Your window into the Elastic Stack| | 277|fishaudio/fish-speech !2025-03-282033222|Brand new TTS solution| | 278|mlc-ai/mlc-llm !2025-03-282028110 |Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.| | 279|deepset-ai/haystack !2025-03-282005320|🔍 Haystack is an open source NLP framework to interact with your data using Transformer models and LLMs (GPT-4, ChatGPT and alike). Haystack offers production-ready tools to quickly build complex question answering, semantic search, text generation applications, and more.| | 280|tree-sitter/tree-sitter !2025-03-28200487|An incremental parsing system for programming tools| | 281|Anjok07/ultimatevocalremovergui !2025-03-281999811|GUI for a Vocal Remover that uses Deep Neural Networks.| | 282|guidance-ai/guidance !2025-03-28199622|A guidance language for controlling large language models.| | 283|ml-explore/mlx !2025-03-28199619|MLX: An array framework for Apple silicon| | 284|mlflow/mlflow !2025-03-281995314|Open source platform for the machine learning lifecycle| | 285|ml-tooling/best-of-ml-python !2025-03-28198631|🏆 A ranked list of awesome machine learning Python libraries. Updated weekly.| | 286|BerriAI/litellm !2025-03-281981862|Call all LLM APIs using the OpenAI format. Use Bedrock, Azure, OpenAI, Cohere, Anthropic, Ollama, Sagemaker, HuggingFace, Replicate (100+ LLMs)| | 287|LazyVim/LazyVim !2025-03-281981320|Neovim config for the lazy| | 288|wez/wezterm !2025-03-281976018|A GPU-accelerated cross-platform terminal emulator and multiplexer written by @wez and implemented in Rust| | 289|valkey-io/valkey !2025-03-281970416|A flexible distributed key-value datastore that supports both caching and beyond caching workloads.| | 290|LiLittleCat/awesome-free-chatgpt !2025-03-28196185|🆓免费的 ChatGPT 镜像网站列表,持续更新。List of free ChatGPT mirror sites, continuously updated.| | 291|Byaidu/PDFMathTranslate !2025-03-281947645|PDF scientific paper translation with preserved formats - 基于 AI 完整保留排版的 PDF 文档全文双语翻译,支持 Google/DeepL/Ollama/OpenAI 等服务,提供 CLI/GUI/Docker| | 292|openai/swarm !2025-03-281947111|Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.| | 293|HqWu-HITCS/Awesome-Chinese-LLM !2025-03-281921423|Organizing smaller, cost-effective, privately deployable open-source Chinese language models, including related datasets and tutorials| | 294|stitionai/devika !2025-03-28190903|Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.| | 295|OpenBMB/MiniCPM-o !2025-03-28190887|MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone| | 296|samber/lo !2025-03-281904815|💥 A Lodash-style Go library based on Go 1.18+ Generics (map, filter, contains, find...)| | 297|chroma-core/chroma !2025-03-281895221 |the AI-native open-source embedding database| | 298|DarkFlippers/unleashed-firmware !2025-03-28189278|Flipper Zero Unleashed Firmware| | 299|brave/brave-browser !2025-03-281892710|Brave browser for Android, iOS, Linux, macOS, Windows.| | 300| tloen/alpaca-lora !2025-03-28188641 | Instruct-tune LLaMA on consumer hardware| | 301|VinciGit00/Scrapegraph-ai !2025-03-281884618|Python scraper based on AI| | 302|gitroomhq/postiz-app !2025-03-281879110|📨 Schedule social posts, measure them, exchange with other members and get a lot of help from AI 🚀| | 303|PrefectHQ/prefect !2025-03-281878715|Prefect is a workflow orchestration tool empowering developers to build, observe, and react to data pipelines| | 304|ymcui/Chinese-LLaMA-Alpaca !2025-03-28187723 |Chinese LLaMA & Alpaca LLMs| | 305|kenjihiranabe/The-Art-of-Linear-Algebra !2025-03-28187335|Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"| | 306|joonspk-research/generativeagents !2025-03-28187288|Generative Agents: Interactive Simulacra of Human Behavior| | 307|renovatebot/renovate !2025-03-28186820|Universal dependency update tool that fits into your workflows.| | 308|gventuri/pandas-ai !2025-03-28186109 |Pandas AI is a Python library that integrates generative artificial intelligence capabilities into Pandas, making dataframes conversational| | 309|thingsboard/thingsboard !2025-03-28185184|Open-source IoT Platform - Device management, data collection, processing and visualization.| | 310|ente-io/ente !2025-03-28184722|Fully open source, End to End Encrypted alternative to Google Photos and Apple Photos| | 311|serengil/deepface !2025-03-281840113|A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python| | 312|Raphire/Win11Debloat !2025-03-281840132|A simple, easy to use PowerShell script to remove pre-installed apps from windows, disable telemetry, remove Bing from windows search as well as perform various other changes to declutter and improve your windows experience. This script works for both windows 10 and windows 11.| | 313|Avaiga/taipy !2025-03-28179235|Turns Data and AI algorithms into production-ready web applications in no time.| | 314|microsoft/qlib !2025-03-281784231|Qlib is an AI-oriented quantitative investment platform that aims to realize the potential, empower research, and create value using AI technologies in quantitative investment, from exploring ideas to implementing productions. Qlib supports diverse machine learning modeling paradigms. including supervised learning, market dynamics modeling, and RL.| | 315|CopilotKit/CopilotKit !2025-03-281778571|Build in-app AI chatbots 🤖, and AI-powered Textareas ✨, into react web apps.| | 316|QwenLM/Qwen-7B !2025-03-281766017|The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud.| | 317|w-okada/voice-changer !2025-03-28176078 |リアルタイムボイスチェンジャー Realtime Voice Changer| | 318|rlabbe/Kalman-and-Bayesian-Filters-in-Python !2025-03-281756011|Kalman Filter book using Jupyter Notebook. Focuses on building intuition and experience, not formal proofs. Includes Kalman filters,extended Kalman filters, unscented Kalman filters, particle filters, and more. All exercises include solutions.| | 319|Mikubill/sd-webui-controlnet !2025-03-28174794 |WebUI extension for ControlNet| | 320|jingyaogong/minimind !2025-03-2817380116|「大模型」3小时完全从0训练26M的小参数GPT,个人显卡即可推理训练!| | 321|apify/crawlee !2025-03-28172696|Crawlee—A web scraping and browser automation library for Node.js to build reliable crawlers. In JavaScript and TypeScript. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with Puppeteer, Playwright, Cheerio, JSDOM, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 322|apple/ml-stable-diffusion !2025-03-28172395|Stable Diffusion with Core ML on Apple Silicon| | 323| transitive-bullshit/chatgpt-api !2025-03-28172095 | Node.js client for the official ChatGPT API. | | 324|teableio/teable !2025-03-281719222|✨ The Next Gen Airtable Alternative: No-Code Postgres| | 325| xx025/carrot !2025-03-28170900 | Free ChatGPT Site List | | 326|microsoft/LightGBM !2025-03-28170723|A fast, distributed, high-performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.| | 327|VikParuchuri/surya !2025-03-28169827|Accurate line-level text detection and recognition (OCR) in any language| | 328|deepseek-ai/Janus !2025-03-281692825|Janus-Series: Unified Multimodal Understanding and Generation Models| | 329|ardalis/CleanArchitecture !2025-03-28168823|Clean Architecture Solution Template: A starting point for Clean Architecture with ASP.NET Core| | 330|neondatabase/neon !2025-03-28166466|Neon: Serverless Postgres. We separated storage and compute to offer autoscaling, code-like database branching, and scale to zero.| | 331|kestra-io/kestra !2025-03-281661313|⚡ Workflow Automation Platform. Orchestrate & Schedule code in any language, run anywhere, 500+ plugins. Alternative to Zapier, Rundeck, Camunda, Airflow...| | 332|Dao-AILab/flash-attention !2025-03-281659720|Fast and memory-efficient exact attention| | 333|RPCS3/rpcs3 !2025-03-281655712|PS3 emulator/debugger| | 334|meta-llama/llama-recipes !2025-03-28165486|Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger| | 335|emilwallner/Screenshot-to-code !2025-03-28165180|A neural network that transforms a design mock-up into a static website.| | 336|datawhalechina/llm-cookbook !2025-03-281650922|面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版| | 337|e2b-dev/awesome-ai-agents !2025-03-281643923|A list of AI autonomous agents| | 338|QwenLM/Qwen2.5 !2025-03-281641114|Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.| | 339|dair-ai/ML-YouTube-Courses !2025-03-28164114|📺 Discover the latest machine learning / AI courses on YouTube.| | 340|pybind/pybind11 !2025-03-28163620|Seamless operability between C++11 and Python| | 341|graphdeco-inria/gaussian-splatting !2025-03-281627116|Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"| | 342|meta-llama/codellama !2025-03-28162531|Inference code for CodeLlama models| | 343|TransformerOptimus/SuperAGI !2025-03-28161292 | SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.| | 344|microsoft/onnxruntime !2025-03-28161169|ONNX Runtime: cross-platform, high-performance ML inferencing and training accelerator| | 345|IDEA-Research/Grounded-Segment-Anything !2025-03-281601411 |Marrying Grounding DINO with Segment Anything & Stable Diffusion & BLIP - Automatically Detect, Segment and Generate Anything with Image and Text Inputs| | 346|ddbourgin/numpy-ml !2025-03-28160054|Machine learning, in numpy| | 347|eosphoros-ai/DB-GPT !2025-03-281585225|Revolutionizing Database Interactions with Private LLM Technology| | 348|Stability-AI/StableLM !2025-03-28158310 |Stability AI Language Models| | 349|openai/evals !2025-03-28157935 |Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.| | 350|THUDM/ChatGLM2-6B !2025-03-28157500|ChatGLM2-6B: An Open Bilingual Chat LLM | | 351|sunner/ChatALL !2025-03-28156761 |Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vincuna, Claude, ChatGLM, MOSS, iFlytek Spark, ERNIE and more, discover the best answers| | 352|abseil/abseil-cpp !2025-03-28156656|Abseil Common Libraries (C++)| | 353|NVIDIA/open-gpu-kernel-modules !2025-03-28156531|NVIDIA Linux open GPU kernel module source| | 354|letta-ai/letta !2025-03-281563718|Letta (formerly MemGPT) is a framework for creating LLM services with memory.| | 355|typescript-eslint/typescript-eslint !2025-03-28156211|✨ Monorepo for all the tooling which enables ESLint to support TypeScript| | 356|umijs/umi !2025-03-28156211|A framework in react community ✨| | 357|AI4Finance-Foundation/FinGPT !2025-03-281561215|Data-Centric FinGPT. Open-source for open finance! Revolutionize 🔥 We'll soon release the trained model.| | 358|amplication/amplication !2025-03-28156022|🔥🔥🔥 The Only Production-Ready AI-Powered Backend Code Generation| | 359|KindXiaoming/pykan !2025-03-28155477|Kolmogorov Arnold Networks| | 360|arc53/DocsGPT !2025-03-28154900|GPT-powered chat for documentation, chat with your documents| | 361|influxdata/telegraf !2025-03-28154502|Agent for collecting, processing, aggregating, and writing metrics, logs, and other arbitrary data.| | 362|microsoft/Bringing-Old-Photos-Back-to-Life !2025-03-28154084|Bringing Old Photo Back to Life (CVPR 2020 oral)| | 363|GaiZhenbiao/ChuanhuChatGPT !2025-03-2815394-2|GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.| | 364|Zeyi-Lin/HivisionIDPhotos !2025-03-281529710|⚡️HivisionIDPhotos: a lightweight and efficient AI ID photos tools. 一个轻量级的AI证件照制作算法。| | 365| mayooear/gpt4-pdf-chatbot-langchain !2025-03-281529518 | GPT4 & LangChain Chatbot for large PDF docs | | 366|1Panel-dev/MaxKB !2025-03-2815277148|? Based on LLM large language model knowledge base Q&A system. Ready to use out of the box, supports quick integration into third-party business systems. Officially produced by 1Panel| | 367|ai16z/eliza !2025-03-281526811|Conversational Agent for Twitter and Discord| | 368|apache/arrow !2025-03-28151684|Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing| | 369|princeton-nlp/SWE-agent !2025-03-281516119|SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models| | 370|mlc-ai/web-llm !2025-03-281509311 |Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.| | 371|guillaumekln/faster-whisper !2025-03-281507117 |Faster Whisper transcription with CTranslate2| | 372|overleaf/overleaf !2025-03-28150316|A web-based collaborative LaTeX editor| | 373|triton-lang/triton !2025-03-28150169|Development repository for the Triton language and compiler| | 374|soxoj/maigret !2025-03-281500410|🕵️‍♂️ Collect a dossier on a person by username from thousands of sites| | 375|alibaba/lowcode-engine !2025-03-28149841|An enterprise-class low-code technology stack with scale-out design / 一套面向扩展设计的企业级低代码技术体系| | 376|espressif/esp-idf !2025-03-28148545|Espressif IoT Development Framework. Official development framework for Espressif SoCs.| | 377|pgvector/pgvector !2025-03-281484913|Open-source vector similarity search for Postgres| | 378|datawhalechina/leedl-tutorial !2025-03-28148246|《李宏毅深度学习教程》(李宏毅老师推荐👍),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases| | 379|xcanwin/KeepChatGPT !2025-03-28147972 |Using ChatGPT is more efficient and smoother, perfectly solving ChatGPT network errors. No longer do you need to frequently refresh the webpage, saving over 10 unnecessary steps| | 380|m-bain/whisperX !2025-03-281471313|WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)| | 381|HumanAIGC/AnimateAnyone !2025-03-2814706-1|Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation| |!green-up-arrow.svg 382|naklecha/llama3-from-scratch !2025-03-281469024|llama3 implementation one matrix multiplication at a time| |!red-down-arrow 383| fauxpilot/fauxpilot !2025-03-28146871 | An open-source GitHub Copilot server | | 384|LlamaFamily/Llama-Chinese !2025-03-28145111|Llama Chinese Community, the best Chinese Llama large model, fully open source and commercially available| | 385|BradyFU/Awesome-Multimodal-Large-Language-Models !2025-03-281450121|Latest Papers and Datasets on Multimodal Large Language Models| | 386|vanna-ai/vanna !2025-03-281449819|🤖 Chat with your SQL database 📊. Accurate Text-to-SQL Generation via LLMs using RAG 🔄.| | 387|bleedline/aimoneyhunter !2025-03-28144845|AI Side Hustle Money Mega Collection: Teaching You How to Utilize AI for Various Side Projects to Earn Extra Income.| | 388|stefan-jansen/machine-learning-for-trading !2025-03-28144629|Code for Machine Learning for Algorithmic Trading, 2nd edition.| | 389|state-spaces/mamba !2025-03-28144139|Mamba: Linear-Time Sequence Modeling with Selective State Spaces| | 390|vercel/ai-chatbot !2025-03-281434614|A full-featured, hackable Next.js AI chatbot built by Vercel| | 391|steven-tey/novel !2025-03-281428410|Notion-style WYSIWYG editor with AI-powered autocompletions| | 392|unifyai/ivy !2025-03-281409348|Unified AI| | 393|chidiwilliams/buzz !2025-03-281402411 |Buzz transcribes and translates audio offline on your personal computer. Powered by OpenAI's Whisper.| | 394|lukas-blecher/LaTeX-OCR !2025-03-28139769|pix2tex: Using a ViT to convert images of equations into LaTeX code.| | 395|openai/tiktoken !2025-03-28139599|tiktoken is a fast BPE tokeniser for use with OpenAI's models.| | 396|nocobase/nocobase !2025-03-281391522|NocoBase is a scalability-first, open-source no-code/low-code platform for building business applications and enterprise solutions.| | 397|neonbjb/tortoise-tts !2025-03-28139010 |A multi-voice TTS system trained with an emphasis on quality| | 398|yamadashy/repomix !2025-03-281382036|📦 Repomix (formerly Repopack) is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.| | 399|adobe/react-spectrum !2025-03-28136766|A collection of libraries and tools that help you build adaptive, accessible, and robust user experiences.| | 400|THUDM/ChatGLM3 !2025-03-28136684|ChatGLM3 series: Open Bilingual Chat LLMs | | 401|NVIDIA/NeMo !2025-03-28134837|A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)| | 402|BlinkDL/RWKV-LM !2025-03-28134346 |RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it combines the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.| | 403| fuergaosi233/wechat-chatgpt !2025-03-28133330 | Use ChatGPT On Wechat via wechaty | | 404|udecode/plate !2025-03-28133325|A rich-text editor powered by AI| | 405|xenova/transformers.js !2025-03-281331219|State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!| | 406|stas00/ml-engineering !2025-03-281325615|Machine Learning Engineering Guides and Tools| | 407| wong2/chatgpt-google-extension !2025-03-2813241-1 | A browser extension that enhances search engines with ChatGPT, this repos will not be updated from 2023-02-20| | 408|mrdbourke/pytorch-deep-learning !2025-03-281317520|Materials for the Learn PyTorch for Deep Learning: Zero to Mastery course.| | 409|Koenkk/zigbee2mqtt !2025-03-28131544|Zigbee 🐝 to MQTT bridge 🌉, get rid of your proprietary Zigbee bridges 🔨| | 410|vercel-labs/ai !2025-03-281298528|Build AI-powered applications with React, Svelte, and Vue| | 411|netease-youdao/QAnything !2025-03-28129318|Question and Answer based on Anything.| | 412|huggingface/trl !2025-03-281289622|Train transformer language models with reinforcement learning.| | 413|microsoft/BitNet !2025-03-28128503|Official inference framework for 1-bit LLMs| | 414|mediar-ai/screenpipe !2025-03-281283915|24/7 local AI screen & mic recording. Build AI apps that have the full context. Works with Ollama. Alternative to Rewind.ai. Open. Secure. You own your data. Rust.| | 415|Skyvern-AI/skyvern !2025-03-281277612|Automate browser-based workflows with LLMs and Computer Vision| | 416|pytube/pytube !2025-03-28126591|A lightweight, dependency-free Python library (and command-line utility) for downloading YouTube Videos.| | 417|official-stockfish/Stockfish !2025-03-28126574|UCI chess engine| | 418|sgl-project/sglang !2025-03-281260143|SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with LLMs faster and more controllable.| | 419|plasma-umass/scalene !2025-03-28125535|Scalene: a high-performance, high-precision CPU, GPU, and memory profiler for Python with AI-powered optimization proposals| | 420|danswer-ai/danswer !2025-03-28125503|Ask Questions in natural language and get Answers backed by private sources. Connects to tools like Slack, GitHub, Confluence, etc.| | 421|OpenTalker/SadTalker !2025-03-28125226|[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation| | 422|facebookresearch/AnimatedDrawings !2025-03-28123693 |Code to accompany "A Method for Animating Children's Drawings of the Human Figure"| | 423|activepieces/activepieces !2025-03-28123609|Your friendliest open source all-in-one automation tool ✨ Workflow automation tool 100+ integration / Enterprise automation tool / Zapier Alternative| | 424|ggerganov/ggml !2025-03-28121992 |Tensor library for machine learning| | 425|bytebase/bytebase !2025-03-28121694|World's most advanced database DevOps and CI/CD for Developer, DBA and Platform Engineering teams. The GitLab/GitHub for database DevOps.| | 426| willwulfken/MidJourney-Styles-and-Keywords-Reference !2025-03-28120971 | A reference containing Styles and Keywords that you can use with MidJourney AI| | 427|Huanshere/VideoLingo !2025-03-281207013|Netflix-level subtitle cutting, translation, alignment, and even dubbing - one-click fully automated AI video subtitle team | | 428|OpenLMLab/MOSS !2025-03-28120330 |An open-source tool-augmented conversational language model from Fudan University| | 429|llmware-ai/llmware !2025-03-281200727|Providing enterprise-grade LLM-based development framework, tools, and fine-tuned models.| | 430|PKU-YuanGroup/Open-Sora-Plan !2025-03-28119362|This project aim to reproduce Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.| | 431|ShishirPatil/gorilla !2025-03-28119332 |Gorilla: An API store for LLMs| | 432|NVIDIA/Megatron-LM !2025-03-281192716|Ongoing research training transformer models at scale| | 433|illacloud/illa-builder !2025-03-28119192|Create AI-Driven Apps like Assembling Blocks| | 434|marimo-team/marimo !2025-03-281191521|A reactive notebook for Python — run reproducible experiments, execute as a script, deploy as an app, and version with git.| | 435|smol-ai/developer !2025-03-28119111 | With 100k context windows on the way, it's now feasible for every dev to have their own smol developer| | 436|Lightning-AI/litgpt !2025-03-28118878|Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.| | 437|openai/shap-e !2025-03-28118474 |Generate 3D objects conditioned on text or images| | 438|eugeneyan/open-llms !2025-03-28118451 |A list of open LLMs available for commercial use.| | 439|andrewyng/aisuite !2025-03-28118124|Simple, unified interface to multiple Generative AI providers| | 440|hajimehoshi/ebiten !2025-03-28117816|Ebitengine - A dead simple 2D game engine for Go| | 441|kgrzybek/modular-monolith-with-ddd !2025-03-28117493|Full Modular Monolith application with Domain-Driven Design approach.| | 442|h2oai/h2ogpt !2025-03-2811736-1 |Come join the movement to make the world's best open source GPT led by H2O.ai - 100% private chat and document search, no data leaks, Apache 2.0| | 443|owainlewis/awesome-artificial-intelligence !2025-03-28117332|A curated list of Artificial Intelligence (AI) courses, books, video lectures and papers.| | 444|DataTalksClub/mlops-zoomcamp !2025-03-28116643|Free MLOps course from DataTalks.Club| | 445|Rudrabha/Wav2Lip !2025-03-281163410|This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.| | 446|aishwaryanr/awesome-generative-ai-guide !2025-03-281152810|A one stop repository for generative AI research updates, interview resources, notebooks and much more!| | 447|karpathy/micrograd !2025-03-28115146|A tiny scalar-valued autograd engine and a neural net library on top of it with PyTorch-like API| | 448|InstantID/InstantID !2025-03-28115111|InstantID : Zero-shot Identity-Preserving Generation in Seconds 🔥| | 449|facebookresearch/seamlesscommunication !2025-03-28114434|Foundational Models for State-of-the-Art Speech and Text Translation| | 450|anthropics/anthropic-cookbook !2025-03-281140112|A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.| | 451|mastra-ai/mastra !2025-03-281139240|the TypeScript AI agent framework| | 452|NVIDIA/TensorRT !2025-03-28113864|NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.| | 453|plandex-ai/plandex !2025-03-28113645|An AI coding engine for complex tasks| | 454|RUCAIBox/LLMSurvey !2025-03-28112735 |A collection of papers and resources related to Large Language Models.| | 455|kubeshark/kubeshark !2025-03-28112711|The API traffic analyzer for Kubernetes providing real-time K8s protocol-level visibility, capturing and monitoring all traffic and payloads going in, out and across containers, pods, nodes and clusters. Inspired by Wireshark, purposely built for Kubernetes| | 456|electric-sql/pglite !2025-03-28112617|Lightweight Postgres packaged as WASM into a TypeScript library for the browser, Node.js, Bun and Deno from https://electric-sql.com| | 457|lightaime/camel !2025-03-281124441 |🐫 CAMEL: Communicative Agents for “Mind” Exploration of Large Scale Language Model Society| | 458|huggingface/lerobot !2025-03-281120184|🤗 LeRobot: State-of-the-art Machine Learning for Real-World Robotics in Pytorch| | 459|normal-computing/outlines !2025-03-28111657|Generative Model Programming| | 460|libretro/RetroArch !2025-03-28110701|Cross-platform, sophisticated frontend for the libretro API. Licensed GPLv3.| | 461|THUDM/CogVideo !2025-03-28110599|Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)| | 462|bentoml/OpenLLM !2025-03-28110495|An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease.| | 463|vosen/ZLUDA !2025-03-28110429|CUDA on AMD GPUs| | 464|dair-ai/ML-Papers-of-the-Week !2025-03-28110304 |🔥Highlighting the top ML papers every week.| | 465|WordPress/gutenberg !2025-03-28110212|The Block Editor project for WordPress and beyond. Plugin is available from the official repository.| | 466|microsoft/data-formulator !2025-03-281099827|🪄 Create rich visualizations with AI| | 467|LibreTranslate/LibreTranslate !2025-03-28109887|Free and Open Source Machine Translation API. Self-hosted, offline capable and easy to setup.| | 468|block/goose !2025-03-281097737|an open-source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM| | 469|getumbrel/llama-gpt !2025-03-28109553|A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.| | 470|HigherOrderCO/HVM !2025-03-28109182|A massively parallel, optimal functional runtime in Rust| | 471|databrickslabs/dolly !2025-03-2810812-3 | A large language model trained on the Databricks Machine Learning Platform| | 472|srush/GPU-Puzzles !2025-03-28108014|Solve puzzles. Learn CUDA.| | 473|Z3Prover/z3 !2025-03-28107952|The Z3 Theorem Prover| | 474|UFund-Me/Qbot !2025-03-281079313 |Qbot is an AI-oriented quantitative investment platform, which aims to realize the potential, empower AI technologies in quantitative investment| | 475|langchain-ai/langgraph !2025-03-281077336|| | 476|lz4/lz4 !2025-03-28107647|Extremely Fast Compression algorithm| | 477|magic-research/magic-animate !2025-03-28107160|MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model| | 478|PaperMC/Paper !2025-03-281071410|The most widely used, high performance Minecraft server that aims to fix gameplay and mechanics inconsistencies| | 479|getomni-ai/zerox !2025-03-281071015|Zero shot pdf OCR with gpt-4o-mini| |!green-up-arrow.svg 480|🔥NirDiamant/GenAIAgents !2025-03-2810693318|This repository provides tutorials and implementations for various Generative AI Agent techniques, from basic to advanced. It serves as a comprehensive guide for building intelligent, interactive AI systems.| |!red-down-arrow 481|Unstructured-IO/unstructured !2025-03-28106889|Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.| | 482|apache/thrift !2025-03-28106610|Apache Thrift| | 483| TheR1D/shellgpt !2025-03-28106097 | A command-line productivity tool powered by ChatGPT, will help you accomplish your tasks faster and more efficiently | | 484|TheRamU/Fay !2025-03-281060312 |Fay is a complete open source project that includes Fay controller and numeral models, which can be used in different applications such as virtual hosts, live promotion, numeral human interaction and so on| | 485|zyronon/douyin !2025-03-28105566|Vue3 + Pinia + Vite5 仿抖音,Vue 在移动端的最佳实践 . Imitate TikTok ,Vue Best practices on Mobile| | 486|THU-MIG/yolov10 !2025-03-28105485|YOLOv10: Real-Time End-to-End Object Detection| | 487|idootop/mi-gpt !2025-03-281052522|? Transform XiaoAi speaker into a personal voice assistant with ChatGPT and DouBao integration.| | 488|SakanaAI/AI-Scientist !2025-03-281051310|The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery 🧑‍🔬| | 489|szimek/sharedrop !2025-03-28105101|Easy P2P file transfer powered by WebRTC - inspired by Apple AirDrop| | 490|salesforce/LAVIS !2025-03-28103942 |LAVIS - A One-stop Library for Language-Vision Intelligence| | 491|aws/amazon-sagemaker-examples !2025-03-28103654|Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.| | 492|artidoro/qlora !2025-03-28103402 |QLoRA: Efficient Finetuning of Quantized LLMs| | 493|lllyasviel/stable-diffusion-webui-forge !2025-03-281029314| a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference| | 494|NielsRogge/Transformers-Tutorials !2025-03-28102487|This repository contains demos I made with the Transformers library by HuggingFace.| | 495|kedro-org/kedro !2025-03-28102371|Kedro is a toolbox for production-ready data science. It uses software engineering best practices to help you create data engineering and data science pipelines that are reproducible, maintainable, and modular.| | 496| chathub-dev/chathub !2025-03-28102301 | All-in-one chatbot client | | 497|microsoft/promptflow !2025-03-28101612|Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.| | 498|mistralai/mistral-src !2025-03-28101372|Reference implementation of Mistral AI 7B v0.1 model.| | 499|burn-rs/burn !2025-03-28101183|Burn - A Flexible and Comprehensive Deep Learning Framework in Rust| | 500|AIGC-Audio/AudioGPT !2025-03-28101150 |AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head| | 501|facebookresearch/dinov2 !2025-03-281011210 |PyTorch code and models for the DINOv2 self-supervised learning method.| | 502|RockChinQ/LangBot !2025-03-281008455|😎丰富生态、🧩支持扩展、🦄多模态 - 大模型原生即时通信机器人平台 🤖 | | 503|78/xiaozhi-esp32 !2025-03-281008180|Build your own AI friend| | 504|cumulo-autumn/StreamDiffusion !2025-03-28100761|StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation| | 505|DataTalksClub/machine-learning-zoomcamp !2025-03-28100664|The code from the Machine Learning Bookcamp book and a free course based on the book| | 506|nerfstudio-project/nerfstudio !2025-03-28100343|A collaboration friendly studio for NeRFs| | 507|cupy/cupy !2025-03-28100344|NumPy & SciPy for GPU| | 508|NVIDIA/TensorRT-LLM !2025-03-281000823|TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.| | 509|wasp-lang/open-saas !2025-03-2899665|A free, open-source SaaS app starter for React & Node.js with superpowers. Production-ready. Community-driven.| | 510|huggingface/text-generation-inference !2025-03-2899383|Large Language Model Text Generation Inference| | 511|jxnl/instructor !2025-03-2899224|structured outputs for llms| | 512|GoogleCloudPlatform/generative-ai !2025-03-2899086|Sample code and notebooks for Generative AI on Google Cloud| | 513|manticoresoftware/manticoresearch !2025-03-2898799|Easy to use open source fast database for search | | 514|langfuse/langfuse !2025-03-28985134|🪢 Open source LLM engineering platform. Observability, metrics, evals, prompt management, testing, prompt playground, datasets, LLM evaluations -- 🍊YC W23 🤖 integrate via Typescript, Python / Decorators, OpenAI, Langchain, LlamaIndex, Litellm, Instructor, Mistral, Perplexity, Claude, Gemini, Vertex| | 515|keephq/keep !2025-03-2897949|The open-source alert management and AIOps platform| | 516|sashabaranov/go-openai !2025-03-2897843|OpenAI ChatGPT, GPT-3, GPT-4, DALL·E, Whisper API wrapper for Go| | 517|autowarefoundation/autoware !2025-03-2897766|Autoware - the world's leading open-source software project for autonomous driving| | 518|anthropics/courses !2025-03-2897269|Anthropic's educational courses| | 519|popcorn-official/popcorn-desktop !2025-03-2896853|Popcorn Time is a multi-platform, free software BitTorrent client that includes an integrated media player ( Windows / Mac / Linux ) A Butter-Project Fork| | 520|getmaxun/maxun !2025-03-28968515|🔥 Open-source no-code web data extraction platform. Turn websites to APIs and spreadsheets with no-code robots in minutes! [In Beta]| | 521|wandb/wandb !2025-03-2896763|🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.| | 522|karpathy/minbpe !2025-03-2895353|Minimal, clean, code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.| | 523|bigscience-workshop/petals !2025-03-2895142|🌸 Run large language models at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading| | 524|OthersideAI/self-operating-computer !2025-03-2894931|A framework to enable multimodal models to operate a computer.| | 525|mshumer/gpt-prompt-engineer !2025-03-2894911|| | 526| BloopAI/bloop !2025-03-2894710 | A fast code search engine written in Rust| | 527|BlinkDL/ChatRWKV !2025-03-289467-1 |ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.| | 528|timlrx/tailwind-nextjs-starter-blog !2025-03-2894677|This is a Next.js, Tailwind CSS blogging starter template. Comes out of the box configured with the latest technologies to make technical writing a breeze. Easily configurable and customizable. Perfect as a replacement to existing Jekyll and Hugo individual blogs.| | 529|google/benchmark !2025-03-2893634|A microbenchmark support library| | 530|facebookresearch/nougat !2025-03-2893603|Implementation of Nougat Neural Optical Understanding for Academic Documents| | 531|modelscope/facechain !2025-03-2893536|FaceChain is a deep-learning toolchain for generating your Digital-Twin.| | 532|DrewThomasson/ebook2audiobook !2025-03-2893388|Convert ebooks to audiobooks with chapters and metadata using dynamic AI models and voice cloning. Supports 1,107+ languages!| | 533|RayTracing/raytracing.github.io !2025-03-2893035|Main Web Site (Online Books)| | 534|QwenLM/Qwen2.5-VL !2025-03-28930249|Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.| | 535|WongKinYiu/yolov9 !2025-03-2892201|Implementation of paper - YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information| | 536|alibaba-damo-academy/FunASR !2025-03-28920222|A Fundamental End-to-End Speech Recognition Toolkit and Open Source SOTA Pretrained Models.| | 537|Visualize-ML/Book4Power-of-Matrix !2025-03-2891931|Book4 'Power of Matrix' | | 538|dice2o/BingGPT !2025-03-289185-1 |Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)| | 539|browserbase/stagehand !2025-03-28917621|An AI web browsing framework focused on simplicity and extensibility.| | 540|FlagOpen/FlagEmbedding !2025-03-28914111|Dense Retrieval and Retrieval-augmented LLMs| | 541|Const-me/Whisper !2025-03-2890979|High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model| | 542|lucidrains/denoising-diffusion-pytorch !2025-03-2890942|Implementation of Denoising Diffusion Probabilistic Model in Pytorch| | 543|Chainlit/chainlit !2025-03-28904422|Build Conversational AI in minutes ⚡️| | 544|togethercomputer/OpenChatKit !2025-03-2890160 |OpenChatKit provides a powerful, open-source base to create both specialized and general purpose chatbots for various applications| | 545|Stability-AI/StableStudio !2025-03-2889631 |Community interface for generative AI| | 546|voicepaw/so-vits-svc-fork !2025-03-2889482 |so-vits-svc fork with realtime support, improved interface and more features.| | 547|pymc-devs/pymc !2025-03-2889413|Bayesian Modeling and Probabilistic Programming in Python| | 548|espnet/espnet !2025-03-2889302|End-to-End Speech Processing Toolkit| | 549|kedacore/keda !2025-03-2888991|KEDA is a Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes| | 550|open-mmlab/Amphion !2025-03-28886911|Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.| | 551|gorse-io/gorse !2025-03-2888451|Gorse open source recommender system engine| | 552|adams549659584/go-proxy-bingai !2025-03-288768-1 |A Microsoft New Bing demo site built with Vue3 and Go, providing a consistent UI experience, supporting ChatGPT prompts, and accessible within China| | 553|open-mmlab/mmsegmentation !2025-03-2887513|OpenMMLab Semantic Segmentation Toolbox and Benchmark.| | 554|bytedance/monolith !2025-03-2887223|ByteDance's Recommendation System| | 555|LouisShark/chatgptsystemprompt !2025-03-2887216|store all agent's system prompt| | 556|brexhq/prompt-engineering !2025-03-2887080 |Tips and tricks for working with Large Language Models like OpenAI's GPT-4.| | 557|erincatto/box2d !2025-03-2886841|Box2D is a 2D physics engine for games| | 558|🔥microsoft/ai-agents-for-beginners !2025-03-288669323|10 Lessons to Get Started Building AI Agents| | 559|nashsu/FreeAskInternet !2025-03-2886102|FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, without GPU needed. The user can ask a question and the system will make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results.| | 560|goldmansachs/gs-quant !2025-03-2885981|Python toolkit for quantitative finance| | 561|srbhr/Resume-Matcher !2025-03-2885800|Open Source Free ATS Tool to compare Resumes with Job Descriptions and create a score to rank them.| | 562|facebookresearch/ImageBind !2025-03-2885681 |ImageBind One Embedding Space to Bind Them All| | 563|ashawkey/stable-dreamfusion !2025-03-2885481 |A pytorch implementation of text-to-3D dreamfusion, powered by stable diffusion.| | 564|meetecho/janus-gateway !2025-03-2885232|Janus WebRTC Server| | 565|google/magika !2025-03-2885003|Detect file content types with deep learning| | 566|huggingface/chat-ui !2025-03-2884871 |Open source codebase powering the HuggingChat app| | 567|EleutherAI/lm-evaluation-harness !2025-03-28843012|A framework for few-shot evaluation of autoregressive language models.| | 568|jina-ai/reader !2025-03-2884089|Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/| | 569|microsoft/TypeChat !2025-03-288406-1|TypeChat is a library that makes it easy to build natural language interfaces using types.| | 570|thuml/Time-Series-Library !2025-03-28839715|A Library for Advanced Deep Time Series Models.| | 571|OptimalScale/LMFlow !2025-03-2883882|An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.| | 572|baptisteArno/typebot.io !2025-03-2883845|💬 Typebot is a powerful chatbot builder that you can self-host.| | 573|jzhang38/TinyLlama !2025-03-2883504|The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.| | 574|fishaudio/Bert-VITS2 !2025-03-2883472|vits2 backbone with multilingual-bert| | 575|OpenBMB/XAgent !2025-03-2882683|An Autonomous LLM Agent for Complex Task Solving| | 576|Acly/krita-ai-diffusion !2025-03-2882387|Streamlined interface for generating images with AI in Krita. Inpaint and outpaint with optional text prompt, no tweaking required.| | 577|jasonppy/VoiceCraft !2025-03-2882151|Zero-Shot Speech Editing and Text-to-Speech in the Wild| | 578|SJTU-IPADS/PowerInfer !2025-03-2881693|High-speed Large Language Model Serving on PCs with Consumer-grade GPUs| | 579|modelscope/DiffSynth-Studio !2025-03-28814713|Enjoy the magic of Diffusion models!| | 580|o3de/o3de !2025-03-2881443|Open 3D Engine (O3DE) is an Apache 2.0-licensed multi-platform 3D engine that enables developers and content creators to build AAA games, cinema-quality 3D worlds, and high-fidelity simulations without any fees or commercial obligations.| | 581|zmh-program/chatnio !2025-03-2881325|🚀 Next Generation AI One-Stop Internationalization Solution. 🚀 下一代 AI 一站式 B/C 端解决方案,支持 OpenAI,Midjourney,Claude,讯飞星火,Stable Diffusion,DALL·E,ChatGLM,通义千问,腾讯混元,360 智脑,百川 AI,火山方舟,新必应,Gemini,Moonshot 等模型,支持对话分享,自定义预设,云端同步,模型市场,支持弹性计费和订阅计划模式,支持图片解析,支持联网搜索,支持模型缓存,丰富美观的后台管理与仪表盘数据统计。| | 582|leptonai/searchwithlepton !2025-03-2880632|Building a quick conversation-based search demo with Lepton AI.| | 583|sebastianstarke/AI4Animation !2025-03-2880620|Bringing Characters to Life with Computer Brains in Unity| | 584|wangrongding/wechat-bot !2025-03-2880528|🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...| | 585|openvinotoolkit/openvino !2025-03-2880528|OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference| | 586|steven2358/awesome-generative-ai !2025-03-28802610|A curated list of modern Generative Artificial Intelligence projects and services| | 587|adam-maj/tiny-gpu !2025-03-2880234|A minimal GPU design in Verilog to learn how GPUs work from the ground up| | 588| anse-app/chatgpt-demo !2025-03-2880180 | A demo repo based on OpenAI API (gpt-3.5-turbo) | | 589| acheong08/EdgeGPT !2025-03-288015-1 |Reverse engineered API of Microsoft's Bing Chat | | 590|ai-collection/ai-collection !2025-03-2879994 |The Generative AI Landscape - A Collection of Awesome Generative AI Applications| | 591|GreyDGL/PentestGPT !2025-03-2879953 |A GPT-empowered penetration testing tool| | 592|delta-io/delta !2025-03-2879112|An open-source storage framework that enables building a Lakehouse architecture with compute engines including Spark, PrestoDB, Flink, Trino, and Hive and APIs| | 593|dataelement/bisheng !2025-03-2879085|Bisheng is an open LLM devops platform for next generation AI applications.| | 594|e2b-dev/e2b !2025-03-2878447 |Vercel for AI agents. We help developers to build, deploy, and monitor AI agents. Focusing on specialized AI agents that build software for you - your personal software developers.| | 595|01-ai/Yi !2025-03-2878311|A series of large language models trained from scratch by developers @01-ai| | 596|Plachtaa/VALL-E-X !2025-03-287830-1|An open source implementation of Microsoft's VALL-E X zero-shot TTS model. The demo is available at https://plachtaa.github.io| | 597|abhishekkrthakur/approachingalmost !2025-03-2878204|Approaching (Almost) Any Machine Learning Problem| | 598|pydantic/pydantic-ai !2025-03-28781041|Agent Framework / shim to use Pydantic with LLMs| | 599|rany2/edge-tts !2025-03-2877901|Use Microsoft Edge's online text-to-speech service from Python WITHOUT needing Microsoft Edge or Windows or an API key| | 600|CASIA-IVA-Lab/FastSAM !2025-03-2877881|Fast Segment Anything| | 601|netease-youdao/EmotiVoice !2025-03-2877817|EmotiVoice 😊: a Multi-Voice and Prompt-Controlled TTS Engine| | 602|lllyasviel/IC-Light !2025-03-2877804|More relighting!| | 603|kroma-network/tachyon !2025-03-287774-1|Modular ZK(Zero Knowledge) backend accelerated by GPU| | 604|deep-floyd/IF !2025-03-2877731 |A novel state-of-the-art open-source text-to-image model with a high degree of photorealism and language understanding| | 605|oumi-ai/oumi !2025-03-2877705|Everything you need to build state-of-the-art foundation models, end-to-end.| | 606|reorproject/reor !2025-03-2877681|AI note-taking app that runs models locally.| | 607|lightpanda-io/browser !2025-03-28775813|Lightpanda: the headless browser designed for AI and automation| | 608|xiangsx/gpt4free-ts !2025-03-287755-1|Providing a free OpenAI GPT-4 API ! This is a replication project for the typescript version of xtekky/gpt4free| | 609|IDEA-Research/GroundingDINO !2025-03-28773311|Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"| | 610|bunkerity/bunkerweb !2025-03-2877326|🛡️ Make your web services secure by default !| | 611|vikhyat/moondream !2025-03-2877057|tiny vision language model| | 612|firmai/financial-machine-learning !2025-03-287703-1|A curated list of practical financial machine learning tools and applications.| | 613|n8n-io/self-hosted-ai-starter-kit !2025-03-28765121|The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.| | 614|intel-analytics/ipex-llm !2025-03-2876507|Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.| | 615|jrouwe/JoltPhysics !2025-03-28764510|A multi core friendly rigid body physics and collision detection library. Written in C++. Suitable for games and VR applications. Used by Horizon Forbidden West.| | 616|THUDM/CodeGeeX2 !2025-03-2876270|CodeGeeX2: A More Powerful Multilingual Code Generation Model| | 617|meta-llama/llama-stack !2025-03-2875866|Composable building blocks to build Llama Apps| | 618|sweepai/sweep !2025-03-287530-1|Sweep is an AI junior developer| | 619|lllyasviel/Omost !2025-03-2875301|Your image is almost there!| | 620|ahmedbahaaeldin/From-0-to-Research-Scientist-resources-guide !2025-03-2875050|Detailed and tailored guide for undergraduate students or anybody want to dig deep into the field of AI with solid foundation.| | 621|dair-ai/ML-Papers-Explained !2025-03-2875050|Explanation to key concepts in ML| | 622|zaidmukaddam/scira !2025-03-28750110|Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.| | 623|Portkey-AI/gateway !2025-03-28749416|A Blazing Fast AI Gateway. Route to 100+ LLMs with 1 fast & friendly API.| | 624|web-infra-dev/midscene !2025-03-28748729|An AI-powered automation SDK can control the page, perform assertions, and extract data in JSON format using natural language.| | 625|zilliztech/GPTCache !2025-03-2874801 |GPTCache is a library for creating semantic cache to store responses from LLM queries.| | 626|niedev/RTranslator !2025-03-2874742|RTranslator is the world's first open source real-time translation app.| |!green-up-arrow.svg 627|roboflow/notebooks !2025-03-2874666|Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.| |!red-down-arrow 628|openlm-research/openllama !2025-03-2874652|OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset| | 629|LiheYoung/Depth-Anything !2025-03-2874155|Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data| | 630|enso-org/enso !2025-03-2874040|Hybrid visual and textual functional programming.| | 631|bigcode-project/starcoder !2025-03-287401-1 |Home of StarCoder: fine-tuning & inference!| | 632|git-ecosystem/git-credential-manager !2025-03-2873975|Secure, cross-platform Git credential storage with authentication to GitHub, Azure Repos, and other popular Git hosting services.| | 633|OpenGVLab/InternVL !2025-03-2873634|[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4V. 接近GPT-4V表现的可商用开源模型| | 634|WooooDyy/LLM-Agent-Paper-List !2025-03-2873551|The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.| | 635|lencx/Noi !2025-03-2873157|🦄 AI + Tools + Plugins + Community| | 636|udlbook/udlbook !2025-03-2873075|Understanding Deep Learning - Simon J.D. Prince| | 637|OpenBMB/MiniCPM !2025-03-2872841|MiniCPM-2B: An end-side LLM outperforms Llama2-13B.| | 638|jaywalnut310/vits !2025-03-2872815 |VITS: Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech| | 639|xorbitsai/inference !2025-03-28727528|Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.| | 640|PWhiddy/PokemonRedExperiments !2025-03-2872492|Playing Pokemon Red with Reinforcement Learning| | 641|Canner/WrenAI !2025-03-28723213|🤖 Open-source AI Agent that empowers data-driven teams to chat with their data to generate Text-to-SQL, charts, spreadsheets, reports, and BI. 📈📊📋🧑‍💻| | 642|miurla/morphic !2025-03-2872258|An AI-powered answer engine with a generative UI| | 643|ml-explore/mlx-examples !2025-03-2872168|Examples in the MLX framework| | 644|PKU-YuanGroup/ChatLaw !2025-03-2872010|Chinese Legal Large Model| | 645|NVIDIA/cutlass !2025-03-2871883|CUDA Templates for Linear Algebra Subroutines| | 646|FoundationVision/VAR !2025-03-28717444|[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction"| | 647|ymcui/Chinese-LLaMA-Alpaca-2 !2025-03-2871561|Chinese LLaMA-2 & Alpaca-2 LLMs| | 648|nadermx/backgroundremover !2025-03-2871514 |Background Remover lets you Remove Background from images and video using AI with a simple command line interface that is free and open source.| | 649|onuratakan/gpt-computer-assistant !2025-03-28714514|gpt-4o for windows, macos and ubuntu| | 650|graviraja/MLOps-Basics !2025-03-2871326|| | 651|Future-House/paper-qa !2025-03-287118-1|High accuracy RAG for answering questions from scientific documents with citations| | 652|open-mmlab/mmagic !2025-03-2871102 |OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox| | 653|bhaskatripathi/pdfGPT !2025-03-2870941 |PDF GPT allows you to chat with the contents of your PDF file by using GPT capabilities. The only open source solution to turn your pdf files in a chatbot!| | 654|ollama/ollama-python !2025-03-28709117|Ollama Python library| | 655|facebookresearch/DiT !2025-03-2870376|Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"| | 656|geekyutao/Inpaint-Anything !2025-03-2870262 |Inpaint anything using Segment Anything and inpainting models.| | 657|AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin !2025-03-2870160 |A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using Automatic1111-sd-webui as a backend.| | 658|apple/corenet !2025-03-2869990|CoreNet: A library for training deep neural networks| | 659|openstatusHQ/openstatus !2025-03-2869926|🏓 The open-source synthetic monitoring platform 🏓| | 660|weaviate/Verba !2025-03-2869772|Retrieval Augmented Generation (RAG) chatbot powered by Weaviate| | 661|meshery/meshery !2025-03-2869630|Meshery, the cloud native manager| | 662|OpenTalker/video-retalking !2025-03-2869530|[SIGGRAPH Asia 2022] VideoReTalking: Audio-based Lip Synchronization for Talking Head Video Editing In the Wild| | 663|digitalinnovationone/dio-lab-open-source !2025-03-28689013|Repositório do lab "Contribuindo em um Projeto Open Source no GitHub" da Digital Innovation One.| | 664|jianchang512/ChatTTS-ui !2025-03-2868842|一个简单的本地网页界面,直接使用ChatTTS将文字合成为语音,同时支持对外提供API接口。| | 665|patchy631/ai-engineering-hub !2025-03-28686434|In-depth tutorials on LLMs, RAGs and real-world AI agent applications.| | 666|gunnarmorling/1brc !2025-03-2868512|1️⃣🐝🏎️ The One Billion Row Challenge -- A fun exploration of how quickly 1B rows from a text file can be aggregated with Java| | 667|Azure-Samples/azure-search-openai-demo !2025-03-2868482 |A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.| | 668|mit-han-lab/streaming-llm !2025-03-2868382|Efficient Streaming Language Models with Attention Sinks| | 669|InternLM/InternLM !2025-03-2868352|InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system.| | 670|dependency-check/DependencyCheck !2025-03-2868191|OWASP dependency-check is a software composition analysis utility that detects publicly disclosed vulnerabilities in application dependencies.| | 671|Soulter/AstrBot !2025-03-28678643|✨易上手的多平台 LLM 聊天机器人及开发框架✨。支持 QQ、QQ频道、Telegram、微信平台(Gewechat, 企业微信)、内置 Web Chat,OpenAI GPT、DeepSeek、Ollama、Llama、GLM、Gemini、OneAPI、LLMTuner,支持 LLM Agent 插件开发,可视化面板。一键部署。支持 Dify 工作流、代码执行器、Whisper 语音转文字。| | 672|react-native-webview/react-native-webview !2025-03-2867792|React Native Cross-Platform WebView| | 673|modelscope/agentscope !2025-03-28676916|Start building LLM-empowered multi-agent applications in an easier way.| | 674|mylxsw/aidea !2025-03-2867381|AIdea is a versatile app that supports GPT and domestic large language models,also supports "Stable Diffusion" text-to-image generation, image-to-image generation, SDXL 1.0, super-resolution, and image colorization| | 675|langchain-ai/ollama-deep-researcher !2025-03-28668635|Fully local web research and report writing assistant| | 676|threestudio-project/threestudio !2025-03-2866653|A unified framework for 3D content generation.| | 677|gaomingqi/Track-Anything !2025-03-2866631 |A flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.| | 678|spdustin/ChatGPT-AutoExpert !2025-03-2866570|🚀🧠💬 Supercharged Custom Instructions for ChatGPT (non-coding) and ChatGPT Advanced Data Analysis (coding).| | 679|HariSekhon/DevOps-Bash-tools !2025-03-2866463|1000+ DevOps Bash Scripts - AWS, GCP, Kubernetes, Docker, CI/CD, APIs, SQL, PostgreSQL, MySQL, Hive, Impala, Kafka, Hadoop, Jenkins, GitHub, GitLab, BitBucket, Azure DevOps, TeamCity, Spotify, MP3, LDAP, Code/Build Linting, pkg mgmt for Linux, Mac, Python, Perl, Ruby, NodeJS, Golang, Advanced dotfiles: .bashrc, .vimrc, .gitconfig, .screenrc, tmux..| | 680|modelscope/swift !2025-03-28661530|ms-swift: Use PEFT or Full-parameter to finetune 200+ LLMs or 15+ MLLMs| | 681|langchain-ai/opengpts !2025-03-2866080|This is an open source effort to create a similar experience to OpenAI's GPTs and Assistants API| | 682| yihong0618/xiaogpt !2025-03-2865131 | Play ChatGPT with xiaomi ai speaker | | 683| civitai/civitai !2025-03-2865111 | Build a platform where people can share their stable diffusion models | | 684|KoljaB/RealtimeSTT !2025-03-28649513|A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.| | 685|qunash/chatgpt-advanced !2025-03-2864910 | A browser extension that augments your ChatGPT prompts with web results.| | 686|Licoy/ChatGPT-Midjourney !2025-03-2864850|🎨 Own your own ChatGPT+Midjourney web service with one click| | 687|friuns2/BlackFriday-GPTs-Prompts !2025-03-2864744|List of free GPTs that doesn't require plus subscription| | 688|PixarAnimationStudios/OpenUSD !2025-03-2864700|Universal Scene Description| | 689|linyiLYi/street-fighter-ai !2025-03-2864630 |This is an AI agent for Street Fighter II Champion Edition.| | 690|run-llama/rags !2025-03-2864380|Build ChatGPT over your data, all with natural language| | 691|frdel/agent-zero !2025-03-2864154|Agent Zero AI framework| | 692|microsoft/DeepSpeedExamples !2025-03-2863911 |Example models using DeepSpeed| | 693|k8sgpt-ai/k8sgpt !2025-03-2863882|Giving Kubernetes Superpowers to everyone| | 694|open-metadata/OpenMetadata !2025-03-2863514|OpenMetadata is a unified platform for discovery, observability, and governance powered by a central metadata repository, in-depth lineage, and seamless team collaboration.| | 695|google/gemma.cpp !2025-03-2863163|lightweight, standalone C++ inference engine for Google's Gemma models.| | 696|RayVentura/ShortGPT !2025-03-286314-1|🚀🎬 ShortGPT - An experimental AI framework for automated short/video content creation. Enables creators to rapidly produce, manage, and deliver content using AI and automation.| | 697|openai/consistencymodels !2025-03-2862940 |Official repo for consistency models.| | 698|yangjianxin1/Firefly !2025-03-2862924|Firefly: Chinese conversational large language model (full-scale fine-tuning + QLoRA), supporting fine-tuning of Llma2, Llama, Baichuan, InternLM, Ziya, Bloom, and other large models| | 699|enricoros/big-AGI !2025-03-2862665|Generative AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.| | 700|aptos-labs/aptos-core !2025-03-2862633|Aptos is a layer 1 blockchain built to support the widespread use of blockchain through better technology and user experience.| | 701|wenda-LLM/wenda !2025-03-286262-1 |Wenda: An LLM invocation platform. Its objective is to achieve efficient content generation tailored to specific environments while considering the limited computing resources of individuals and small businesses, as well as knowledge security and privacy concerns| | 702|Project-MONAI/MONAI !2025-03-2862603|AI Toolkit for Healthcare Imaging| | 703|HVision-NKU/StoryDiffusion !2025-03-2862470|Create Magic Story!| | 704|deepseek-ai/DeepSeek-LLM !2025-03-2862463|DeepSeek LLM: Let there be answers| | 705|Tohrusky/Final2x !2025-03-2862393|2^x Image Super-Resolution| | 706|OpenSPG/KAG !2025-03-28619611|KAG is a logical form-guided reasoning and retrieval framework based on OpenSPG engine and LLMs. It is used to build logical reasoning and factual Q&A solutions for professional domain knowledge bases. It can effectively overcome the shortcomings of the traditional RAG vector similarity calculation model.| | 707|Moonvy/OpenPromptStudio !2025-03-2861861 |AIGC Hint Word Visualization Editor| | 708|levihsu/OOTDiffusion !2025-03-2861761|Official implementation of OOTDiffusion| | 709|tmc/langchaingo !2025-03-2861729|LangChain for Go, the easiest way to write LLM-based programs in Go| | 710|vladmandic/automatic !2025-03-2861374|SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models| | 711|clovaai/donut !2025-03-2861231 |Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022| | 712|Shaunwei/RealChar !2025-03-286121-1|🎙️🤖Create, Customize and Talk to your AI Character/Companion in Realtime(All in One Codebase!). Have a natural seamless conversation with AI everywhere(mobile, web and terminal) using LLM OpenAI GPT3.5/4, Anthropic Claude2, Chroma Vector DB, Whisper Speech2Text, ElevenLabs Text2Speech🎙️🤖| | 713|microsoft/TinyTroupe !2025-03-2861142|LLM-powered multiagent persona simulation for imagination enhancement and business insights.| | 714| rustformers/llm !2025-03-2861010 | Run inference for Large Language Models on CPU, with Rust| | 715|firebase/firebase-ios-sdk !2025-03-2860950|Firebase SDK for Apple App Development| | 716|vespa-engine/vespa !2025-03-2860824|The open big data serving engine. https://vespa.ai| | 717|n4ze3m/page-assist !2025-03-28607610|Use your locally running AI models to assist you in your web browsing| | 718|Dooy/chatgpt-web-midjourney-proxy !2025-03-2860646|chatgpt web, midjourney, gpts,tts, whisper 一套ui全搞定| | 719|ethereum-optimism/optimism !2025-03-2860213|Optimism is Ethereum, scaled.| | 720|sczhou/ProPainter !2025-03-2859971|[ICCV 2023] ProPainter: Improving Propagation and Transformer for Video Inpainting| | 721|MineDojo/Voyager !2025-03-2859951 |An Open-Ended Embodied Agent with Large Language Models| | 722|lavague-ai/LaVague !2025-03-2859800|Automate automation with Large Action Model framework| | 723|SevaSk/ecoute !2025-03-2859770 |Ecoute is a live transcription tool that provides real-time transcripts for both the user's microphone input (You) and the user's speakers output (Speaker) in a textbox. It also generates a suggested response using OpenAI's GPT-3.5 for the user to say based on the live transcription of the conversation.| | 724|google/mesop !2025-03-2859661|| | 725|pengxiao-song/LaWGPT !2025-03-2859542 |Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge| | 726|fr0gger/Awesome-GPT-Agents !2025-03-2859434|A curated list of GPT agents for cybersecurity| | 727|google-deepmind/graphcast !2025-03-2859412|| | 728|comet-ml/opik !2025-03-28594126|Open-source end-to-end LLM Development Platform| | 729|SciPhi-AI/R2R !2025-03-28594033|A framework for rapid development and deployment of production-ready RAG systems| | 730|SkalskiP/courses !2025-03-2859272 |This repository is a curated collection of links to various courses and resources about Artificial Intelligence (AI)| | 731|QuivrHQ/MegaParse !2025-03-2859122|File Parser optimised for LLM Ingestion with no loss 🧠 Parse PDFs, Docx, PPTx in a format that is ideal for LLMs.| | 732|pytorch-labs/gpt-fast !2025-03-2858971|Simple and efficient pytorch-native transformer text generation in !2025-03-2858886|Curated list of chatgpt prompts from the top-rated GPTs in the GPTs Store. Prompt Engineering, prompt attack & prompt protect. Advanced Prompt Engineering papers.| | 734|nilsherzig/LLocalSearch !2025-03-2858852|LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.| | 735|kuafuai/DevOpsGPT !2025-03-285874-2|Multi agent system for AI-driven software development. Convert natural language requirements into working software. Supports any development language and extends the existing base code.| | 736|myshell-ai/MeloTTS !2025-03-2858486|High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.| | 737|OpenGVLab/LLaMA-Adapter !2025-03-2858421 |Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters| | 738|volcengine/verl !2025-03-28582563|veRL: Volcano Engine Reinforcement Learning for LLM| | 739|a16z-infra/companion-app !2025-03-2858171|AI companions with memory: a lightweight stack to create and host your own AI companions| | 740|HumanAIGC/OutfitAnyone !2025-03-285816-1|Outfit Anyone: Ultra-high quality virtual try-on for Any Clothing and Any Person| | 741|josStorer/RWKV-Runner !2025-03-2857472|A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.| | 742|648540858/wvp-GB28181-pro !2025-03-2857414|WEB VIDEO PLATFORM是一个基于GB28181-2016标准实现的网络视频平台,支持NAT穿透,支持海康、大华、宇视等品牌的IPC、NVR、DVR接入。支持国标级联,支持rtsp/rtmp等视频流转发到国标平台,支持rtsp/rtmp等推流转发到国标平台。| | 743|ToonCrafter/ToonCrafter !2025-03-2857345|a research paper for generative cartoon interpolation| | 744|PawanOsman/ChatGPT !2025-03-2857191|OpenAI API Free Reverse Proxy| | 745|apache/hudi !2025-03-2857091|Upserts, Deletes And Incremental Processing on Big Data.| | 746| nsarrazin/serge !2025-03-2857081 | A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API| | 747|homanp/superagent !2025-03-2857021|🥷 Superagent - Build, deploy, and manage LLM-powered agents| | 748|ramonvc/freegpt-webui !2025-03-2856910|GPT 3.5/4 with a Chat Web UI. No API key is required.| | 749|baichuan-inc/baichuan-7B !2025-03-2856901|A large-scale 7B pretraining language model developed by BaiChuan-Inc.| | 750|Azure/azure-sdk-for-net !2025-03-2856792|This repository is for active development of the Azure SDK for .NET. For consumers of the SDK we recommend visiting our public developer docs at https://learn.microsoft.com/dotnet/azure/ or our versioned developer docs at https://azure.github.io/azure-sdk-for-net.| | 751|mnotgod96/AppAgent !2025-03-2856643|AppAgent: Multimodal Agents as Smartphone Users, an LLM-based multimodal agent framework designed to operate smartphone apps.| | 752|microsoft/TaskWeaver !2025-03-2856243|A code-first agent framework for seamlessly planning and executing data analytics tasks.| | 753| yetone/bob-plugin-openai-translator !2025-03-285600-1 | A Bob Plugin base ChatGPT API | | 754|PrefectHQ/marvin !2025-03-2855840 |A batteries-included library for building AI-powered software| | 755|microsoft/promptbase !2025-03-2855832|All things prompt engineering| | 756|fullstackhero/dotnet-starter-kit !2025-03-2855560|Production Grade Cloud-Ready .NET 8 Starter Kit (Web API + Blazor Client) with Multitenancy Support, and Clean/Modular Architecture that saves roughly 200+ Development Hours! All Batteries Included.| | 757|deepseek-ai/DeepSeek-Coder-V2 !2025-03-2855435|DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence| | 758|aiwaves-cn/agents !2025-03-2855391|An Open-source Framework for Autonomous Language Agents| | 759|microsoft/Mastering-GitHub-Copilot-for-Paired-Programming !2025-03-2855158|A 6 Lesson course teaching everything you need to know about harnessing GitHub Copilot and an AI Paired Programing resource.| | 760|allenai/OLMo !2025-03-2854506|Modeling, training, eval, and inference code for OLMo| | 761|apify/crawlee-python !2025-03-2854493|Crawlee—A web scraping and browser automation library for Python to build reliable crawlers. Extract data for AI, LLMs, RAG, or GPTs. Download HTML, PDF, JPG, PNG, and other files from websites. Works with BeautifulSoup, Playwright, and raw HTTP. Both headful and headless mode. With proxy rotation.| | 762|k2-fsa/sherpa-onnx !2025-03-28541520|Speech-to-text, text-to-speech, and speaker recongition using next-gen Kaldi with onnxruntime without Internet connection. Support embedded systems, Android, iOS, Raspberry Pi, RISC-V, x86_64 servers, websocket server/client, C/C++, Python, Kotlin, C#, Go, NodeJS, Java, Swift| | 763|TEN-framework/TEN-Agent !2025-03-28541411|TEN Agent is a realtime conversational AI agent powered by TEN. It seamlessly integrates the OpenAI Realtime API, RTC capabilities, and advanced features like weather updates, web search, computer vision, and Retrieval-Augmented Generation (RAG).| | 764|google/gemmapytorch !2025-03-2854010|The official PyTorch implementation of Google's Gemma models| | 765|snakers4/silero-vad !2025-03-2853858|Silero VAD: pre-trained enterprise-grade Voice Activity Detector| | 766|livekit/agents !2025-03-2853836|Build real-time multimodal AI applications 🤖🎙️📹| | 767|pipecat-ai/pipecat !2025-03-28537811|Open Source framework for voice and multimodal conversational AI| | 768|EricLBuehler/mistral.rs !2025-03-28536324|Blazingly fast LLM inference.| | 769|asg017/sqlite-vec !2025-03-28535810|Work-in-progress vector search SQLite extension that runs anywhere.| | 770|albertan017/LLM4Decompile !2025-03-2853563|Reverse Engineering: Decompiling Binary Code with Large Language Models| | 771|Permify/permify !2025-03-2853235|An open-source authorization as a service inspired by Google Zanzibar, designed to build and manage fine-grained and scalable authorization systems for any application.| | 772|imoneoi/openchat !2025-03-2853171|OpenChat: Advancing Open-source Language Models with Imperfect Data| | 773|mosaicml/composer !2025-03-2853140|Train neural networks up to 7x faster| | 774|dsdanielpark/Bard-API !2025-03-285277-1 |The python package that returns a response of Google Bard through API.| | 775|lxfater/inpaint-web !2025-03-2852552|A free and open-source inpainting & image-upscaling tool powered by webgpu and wasm on the browser。| | 776|leanprover/lean4 !2025-03-2852441|Lean 4 programming language and theorem prover| | 777|AILab-CVC/YOLO-World !2025-03-2852415|Real-Time Open-Vocabulary Object Detection| | 778|openchatai/OpenChat !2025-03-2852260 |Run and create custom ChatGPT-like bots with OpenChat, embed and share these bots anywhere, the open-source chatbot console.| | 779|mufeedvh/code2prompt !2025-03-28519414|A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.| | 780|biobootloader/wolverine !2025-03-2851700 |Automatically repair python scripts through GPT-4 to give them regenerative abilities.| | 781|huggingface/parler-tts !2025-03-2851671|Inference and training library for high-quality TTS models.| | 782|Akegarasu/lora-scripts !2025-03-2851308 |LoRA training scripts use kohya-ss's trainer, for diffusion model.| | 783|openchatai/OpenCopilot !2025-03-285128-3|🤖 🔥 Let your users chat with your product features and execute things by text - open source Shopify sidekick| | 784|e2b-dev/fragments !2025-03-2851228|Open-source Next.js template for building apps that are fully generated by AI. By E2B.| | 785|microsoft/SynapseML !2025-03-2851132|Simple and Distributed Machine Learning| | 786|aigc-apps/sd-webui-EasyPhoto !2025-03-285108-1|📷 EasyPhoto | | 787|ChaoningZhang/MobileSAM !2025-03-2850944|This is the official code for Faster Segment Anything (MobileSAM) project that makes SAM lightweight| | 788|huggingface/alignment-handbook !2025-03-2850932|Robust recipes for to align language models with human and AI preferences| | 789|alpkeskin/mosint !2025-03-2850920|An automated e-mail OSINT tool| | 790|TaskingAI/TaskingAI !2025-03-2850891|The open source platform for AI-native application development.| | 791|lipku/metahuman-stream !2025-03-28507615|Real time interactive streaming digital human| | 792|OpenInterpreter/01 !2025-03-2850530|The open-source language model computer| | 793|open-compass/opencompass !2025-03-28505111|OpenCompass is an LLM evaluation platform, supporting a wide range of models (InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.| | 794|xxlong0/Wonder3D !2025-03-2850491|A cross-domain diffusion model for 3D reconstruction from a single image| | 795|pytorch/torchtune !2025-03-2850342|A Native-PyTorch Library for LLM Fine-tuning| | 796|SuperDuperDB/superduperdb !2025-03-2850192|🔮 SuperDuperDB: Bring AI to your database: Integrate, train and manage any AI models and APIs directly with your database and your data.| | 797|WhiskeySockets/Baileys !2025-03-2850057|Lightweight full-featured typescript/javascript WhatsApp Web API| | 798| mpociot/chatgpt-vscode !2025-03-2849890 | A VSCode extension that allows you to use ChatGPT | | 799|OpenGVLab/DragGAN !2025-03-2849880|Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)| | 800|microsoft/LLMLingua !2025-03-2849824|To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.| | 801|Zipstack/unstract !2025-03-2849745|No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents| | 802|OpenBMB/ToolBench !2025-03-2849621|An open platform for training, serving, and evaluating large language model for tool learning.| | 803|Fanghua-Yu/SUPIR !2025-03-2849593|SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild| | 804|GaiaNet-AI/gaianet-node !2025-03-2849360|Install and run your own AI agent service| | 805|qodo-ai/qodo-cover !2025-03-284922-1|Qodo-Cover: An AI-Powered Tool for Automated Test Generation and Code Coverage Enhancement! 💻🤖🧪🐞| | 806|Zejun-Yang/AniPortrait !2025-03-2849042|AniPortrait: Audio-Driven Synthesis of Photorealistic Portrait Animation| | 807|lvwzhen/law-cn-ai !2025-03-2848901 |⚖️ AI Legal Assistant| | 808|developersdigest/llm-answer-engine !2025-03-2848740|Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper| | 809|Plachtaa/VITS-fast-fine-tuning !2025-03-2848640|This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion| | 810|espeak-ng/espeak-ng !2025-03-2848601|eSpeak NG is an open source speech synthesizer that supports more than hundred languages and accents.| | 811|ant-research/CoDeF !2025-03-2848581|[CVPR'24 Highlight] Official PyTorch implementation of CoDeF: Content Deformation Fields for Temporally Consistent Video Processing| | 812|deepseek-ai/DeepSeek-V2 !2025-03-2848512|| | 813|XRPLF/rippled !2025-03-2848210|Decentralized cryptocurrency blockchain daemon implementing the XRP Ledger protocol in C++| | 814|AutoMQ/automq !2025-03-28478721|AutoMQ is a cloud-first alternative to Kafka by decoupling durability to S3 and EBS. 10x cost-effective. Autoscale in seconds. Single-digit ms latency.| | 815|AILab-CVC/VideoCrafter !2025-03-2847800|VideoCrafter1: Open Diffusion Models for High-Quality Video Generation| | 816|nautechsystems/nautilustrader !2025-03-2847702|A high-performance algorithmic trading platform and event-driven backtester| | 817|kyegomez/swarms !2025-03-2847563|The Enterprise-Grade Production-Ready Multi-Agent Orchestration Framework Join our Community: https://discord.com/servers/agora-999382051935506503| | 818|Deci-AI/super-gradients !2025-03-2847310 |Easily train or fine-tune SOTA computer vision models with one open source training library. The home of Yolo-NAS.| | 819|QwenLM/Qwen2.5-Coder !2025-03-2847236|Qwen2.5-Coder is the code version of Qwen2.5, the large language model series developed by Qwen team, Alibaba Cloud.| | 820|SCIR-HI/Huatuo-Llama-Med-Chinese !2025-03-2847191 |Repo for HuaTuo (华驼), Llama-7B tuned with Chinese medical knowledge| | 821|togethercomputer/RedPajama-Data !2025-03-2846841 |code for preparing large datasets for training large language models| | 822|mishushakov/llm-scraper !2025-03-2846704|Turn any webpage into structured data using LLMs| | 823|1rgs/jsonformer !2025-03-2846663 |A Bulletproof Way to Generate Structured JSON from Language Models| | 824|anti-work/shortest !2025-03-2846565|QA via natural language AI tests| | 825|dnhkng/GlaDOS !2025-03-2846510|This is the Personality Core for GLaDOS, the first steps towards a real-life implementation of the AI from the Portal series by Valve.| | 826|Nukem9/dlssg-to-fsr3 !2025-03-2846380|Adds AMD FSR3 Frame Generation to games by replacing Nvidia DLSS-G Frame Generation (nvngx_dlssg).| | 827|BuilderIO/ai-shell !2025-03-2846373 |A CLI that converts natural language to shell commands.| | 828|facebookincubator/AITemplate !2025-03-2846220 |AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.| | 829|terraform-aws-modules/terraform-aws-eks !2025-03-2846030|Terraform module to create AWS Elastic Kubernetes (EKS) resources 🇺🇦| | 830|timescale/pgai !2025-03-2845915|A suite of tools to develop RAG, semantic search, and other AI applications more easily with PostgreSQL| | 831|awslabs/multi-agent-orchestrator !2025-03-2845788|Flexible and powerful framework for managing multiple AI agents and handling complex conversations| | 832|sanchit-gandhi/whisper-jax !2025-03-2845771 |Optimised JAX code for OpenAI's Whisper Model, largely built on the Hugging Face Transformers Whisper implementation| | 833|NVIDIA/NeMo-Guardrails !2025-03-2845755|NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.| | 834|PathOfBuildingCommunity/PathOfBuilding !2025-03-2845480|Offline build planner for Path of Exile.| | 835|UX-Decoder/Segment-Everything-Everywhere-All-At-Once !2025-03-2845412 |Official implementation of the paper "Segment Everything Everywhere All at Once"| | 836|build-trust/ockam !2025-03-2845171|Orchestrate end-to-end encryption, cryptographic identities, mutual authentication, and authorization policies between distributed applications – at massive scale.| | 837|google-research/timesfm !2025-03-2845135|TimesFM (Time Series Foundation Model) is a pretrained time-series foundation model developed by Google Research for time-series forecasting.| | 838|luosiallen/latent-consistency-model !2025-03-2844842|Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference| | 839|NVlabs/neuralangelo !2025-03-2844740|Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)| | 840|kyegomez/tree-of-thoughts !2025-03-2844720 |Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%| | 841|sjvasquez/handwriting-synthesis !2025-03-2844720 |Handwriting Synthesis with RNNs ✏️| | 842| madawei2699/myGPTReader !2025-03-2844420 | A slack bot that can read any webpage, ebook or document and summarize it with chatGPT | | 843|OpenBMB/AgentVerse !2025-03-2844413|🤖 AgentVerse 🪐 provides a flexible framework that simplifies the process of building custom multi-agent environments for large language models (LLMs).| | 844|argmaxinc/WhisperKit !2025-03-2844395|Swift native speech recognition on-device for iOS and macOS applications.| | 845|landing-ai/vision-agent !2025-03-2844346|Vision agent| | 846|InternLM/xtuner !2025-03-2844273|An efficient, flexible and full-featured toolkit for fine-tuning large models (InternLM, Llama, Baichuan, Qwen, ChatGLM)| | 847|google-deepmind/alphageometry !2025-03-284421-1|Solving Olympiad Geometry without Human Demonstrations| | 848|ostris/ai-toolkit !2025-03-2844093|Various AI scripts. Mostly Stable Diffusion stuff.| | 849|LLM-Red-Team/kimi-free-api !2025-03-2844004|🚀 KIMI AI 长文本大模型白嫖服务,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。| | 850|argilla-io/argilla !2025-03-2843991|Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.| | 851|spring-projects/spring-ai !2025-03-28438419|An Application Framework for AI Engineering| | 852|alibaba-damo-academy/FunClip !2025-03-2843555|Open-source, accurate and easy-to-use video clipping tool, LLM based AI clipping intergrated | | 853|yisol/IDM-VTON !2025-03-2843541|IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild| | 854|fchollet/ARC-AGI !2025-03-2843368|The Abstraction and Reasoning Corpus| | 855|MahmoudAshraf97/whisper-diarization !2025-03-2843064|Automatic Speech Recognition with Speaker Diarization based on OpenAI Whisper| | 856|Speykious/cve-rs !2025-03-2843047|Blazingly 🔥 fast 🚀 memory vulnerabilities, written in 100% safe Rust. 🦀| | 857|Blealtan/efficient-kan !2025-03-2842770|An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).| | 858|smol-ai/GodMode !2025-03-284249-1|AI Chat Browser: Fast, Full webapp access to ChatGPT / Claude / Bard / Bing / Llama2! I use this 20 times a day.| | 859|openai/plugins-quickstart !2025-03-284235-4 |Get a ChatGPT plugin up and running in under 5 minutes!| | 860|Doriandarko/maestro !2025-03-2842260|A framework for Claude Opus to intelligently orchestrate subagents.| | 861|philz1337x/clarity-upscaler !2025-03-2842204|Clarity-Upscaler: Reimagined image upscaling for everyone| | 862|facebookresearch/co-tracker !2025-03-2842142|CoTracker is a model for tracking any point (pixel) on a video.| | 863|xlang-ai/OpenAgents !2025-03-2842031|OpenAgents: An Open Platform for Language Agents in the Wild| | 864|alibaba/higress !2025-03-28419514|🤖 AI Gateway | | 865|ray-project/llm-numbers !2025-03-2841920 |Numbers every LLM developer should know| | 866|fudan-generative-vision/champ !2025-03-2841820|Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance| | 867|NVIDIA/garak !2025-03-2841795|the LLM vulnerability scanner| | 868|leetcode-mafia/cheetah !2025-03-2841740 |Whisper & GPT-based app for passing remote SWE interviews| | 869|ragapp/ragapp !2025-03-2841710|The easiest way to use Agentic RAG in any enterprise| | 870|collabora/WhisperSpeech !2025-03-2841692|An Open Source text-to-speech system built by inverting Whisper.| | 871|Facico/Chinese-Vicuna !2025-03-2841520 |Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model| | 872|openai/grok !2025-03-2841381|| | 873|CrazyBoyM/llama3-Chinese-chat !2025-03-2841361|Llama3 Chinese Repository with modified versions, and training and deployment resources| | 874|luban-agi/Awesome-AIGC-Tutorials !2025-03-2841301|Curated tutorials and resources for Large Language Models, AI Painting, and more.| | 875|damo-vilab/AnyDoor !2025-03-2841192|Official implementations for paper: Anydoor: zero-shot object-level image customization| | 876|raspberrypi/pico-sdk !2025-03-2841072|| | 877|mshumer/gpt-llm-trainer !2025-03-284097-1|| | 878|metavoiceio/metavoice-src !2025-03-284076-1|AI for human-level speech intelligence| | 879|intelowlproject/IntelOwl !2025-03-2840763|IntelOwl: manage your Threat Intelligence at scale| | 880|a16z-infra/ai-getting-started !2025-03-2840682|A Javascript AI getting started stack for weekend projects, including image/text models, vector stores, auth, and deployment configs| | 881|MarkFzp/mobile-aloha !2025-03-2840641|Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation| | 882| keijiro/AICommand !2025-03-2840380 | ChatGPT integration with Unity Editor | | 883|Tencent/HunyuanDiT !2025-03-2840214|Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding| | 884|hengyoush/kyanos !2025-03-2840061|Visualize the time packets spend in the kernel, watch & analyze in command line.| | 885|agiresearch/AIOS !2025-03-2840045|AIOS: LLM Agent Operating System| | 886|truefoundry/cognita !2025-03-2839773|RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry| | 887|X-PLUG/MobileAgent !2025-03-2839557|Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception| | 888|jackMort/ChatGPT.nvim !2025-03-2839231|ChatGPT Neovim Plugin: Effortless Natural Language Generation with OpenAI's ChatGPT API| | 889|microsoft/RD-Agent !2025-03-28388422|Research and development (R&D) is crucial for the enhancement of industrial productivity, especially in the AI era, where the core aspects of R&D are mainly focused on data and models. We are committed to automate these high-value generic R&D processes through our open source R&D automation tool RD-Agent, which let AI drive data-driven AI.| | 890|Significant-Gravitas/Auto-GPT-Plugins !2025-03-283882-1 |Plugins for Auto-GPT| | 891|apple/ml-mgie !2025-03-2838770|| | 892|OpenDriveLab/UniAD !2025-03-2838727|[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving| | 893|llSourcell/DoctorGPT !2025-03-2838640|DoctorGPT is an LLM that can pass the US Medical Licensing Exam. It works offline, it's cross-platform, & your health data stays private.| | 894|FlagAI-Open/FlagAI !2025-03-2838601|FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.| | 895|krishnaik06/Roadmap-To-Learn-Generative-AI-In-2024 !2025-03-2838513|Roadmap To Learn Generative AI In 2024| | 896|SysCV/sam-hq !2025-03-2838491|Segment Anything in High Quality| | 897|google/security-research !2025-03-2838420|This project hosts security advisories and their accompanying proof-of-concepts related to research conducted at Google which impact non-Google owned code.| | 898|shroominic/codeinterpreter-api !2025-03-2838330|Open source implementation of the ChatGPT Code Interpreter 👾| | 899|Yonom/assistant-ui !2025-03-2838308|React Components for AI Chat 💬 🚀| | 900|nucleuscloud/neosync !2025-03-2838262|Open source data anonymization and synthetic data orchestration for developers. Create high fidelity synthetic data and sync it across your environments.| | 901|ravenscroftj/turbopilot !2025-03-2838230 |Turbopilot is an open source large-language-model based code completion engine that runs locally on CPU| | 902|NVlabs/Sana !2025-03-28380810|SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer| | 903|huggingface/distil-whisper !2025-03-2838061|Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.| | 904|Codium-ai/AlphaCodium !2025-03-2837971|code generation tool that surpasses most human competitors in CodeContests| | 905|fixie-ai/ultravox !2025-03-2837710|A fast multimodal LLM for real-time voice| | 906|unit-mesh/auto-dev !2025-03-28375715|🧙‍AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀| | 907|Marker-Inc-Korea/AutoRAG !2025-03-2837432|AutoML tool for RAG| | 908|deepseek-ai/DeepSeek-VL !2025-03-283734-1|DeepSeek-VL: Towards Real-World Vision-Language Understanding| | 909|hiyouga/ChatGLM-Efficient-Tuning !2025-03-283692-1|Fine-tuning ChatGLM-6B with PEFT | | 910| Yue-Yang/ChatGPT-Siri !2025-03-2836921 | Shortcuts for Siri using ChatGPT API gpt-3.5-turbo model | | 911|0hq/WebGPT !2025-03-2836901 |Run GPT model on the browser with WebGPU. An implementation of GPT inference in less than ~2000 lines of vanilla Javascript.| | 912|cvg/LightGlue !2025-03-2836903|LightGlue: Local Feature Matching at Light Speed (ICCV 2023)| | 913|deanxv/coze-discord-proxy !2025-03-2836791|代理Discord-Bot对话Coze-Bot,实现API形式请求GPT4对话模型/微调模型| | 914|MervinPraison/PraisonAI !2025-03-2836764|PraisonAI application combines AutoGen and CrewAI or similar frameworks into a low-code solution for building and managing multi-agent LLM systems, focusing on simplicity, customisation, and efficient human-agent collaboration.| | 915|Ironclad/rivet !2025-03-2836345 |The open-source visual AI programming environment and TypeScript library| | 916|BasedHardware/OpenGlass !2025-03-2835851|Turn any glasses into AI-powered smart glasses| | 917|ricklamers/gpt-code-ui !2025-03-2835840 |An open source implementation of OpenAI's ChatGPT Code interpreter| | 918|whoiskatrin/chart-gpt !2025-03-2835830 |AI tool to build charts based on text input| | 919|github/CopilotForXcode !2025-03-2835788|Xcode extension for GitHub Copilot| | 920|hemansnation/God-Level-Data-Science-ML-Full-Stack !2025-03-2835570 |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 921|pytorch/torchchat !2025-03-2835461|Run PyTorch LLMs locally on servers, desktop and mobile| | 922| Kent0n-Li/ChatDoctor !2025-03-2835451 | A Medical Chat Model Fine-tuned on LLaMA Model using Medical Domain Knowledge | | 923|xtekky/chatgpt-clone !2025-03-283519-1 |ChatGPT interface with better UI| | 924|jupyterlab/jupyter-ai !2025-03-2835120|A generative AI extension for JupyterLab| | 925|pytorch/torchtitan !2025-03-2835064|A native PyTorch Library for large model training| | 926|minimaxir/simpleaichat !2025-03-2835031|Python package for easily interfacing with chat apps, with robust features and minimal code complexity.| | 927|srush/Tensor-Puzzles !2025-03-2834930|Solve puzzles. Improve your pytorch.| | 928|Helicone/helicone !2025-03-2834918|🧊 Open source LLM-Observability Platform for Developers. One-line integration for monitoring, metrics, evals, agent tracing, prompt management, playground, etc. Supports OpenAI SDK, Vercel AI SDK, Anthropic SDK, LiteLLM, LLamaIndex, LangChain, and more. 🍓 YC W23| | 929|run-llama/llama-hub !2025-03-2834740|A library of data loaders for LLMs made by the community -- to be used with LlamaIndex and/or LangChain| | 930|NExT-GPT/NExT-GPT !2025-03-2834700|Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model| | 931|souzatharsis/podcastfy !2025-03-2834661|An Open Source Python alternative to NotebookLM's podcast feature: Transforming Multimodal Content into Captivating Multilingual Audio Conversations with GenAI| | 932|Dataherald/dataherald !2025-03-2834450|Interact with your SQL database, Natural Language to SQL using LLMs| | 933|iryna-kondr/scikit-llm !2025-03-2834350 |Seamlessly integrate powerful language models like ChatGPT into scikit-learn for enhanced text analysis tasks.| | 934|Netflix/maestro !2025-03-2834230|Maestro: Netflix’s Workflow Orchestrator| | 935|CanadaHonk/porffor !2025-03-2833560|A from-scratch experimental AOT JS engine, written in JS| | 936|hustvl/Vim !2025-03-2833323|Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model| | 937|pashpashpash/vault-ai !2025-03-2833250 |OP Vault ChatGPT: Give ChatGPT long-term memory using the OP Stack (OpenAI + Pinecone Vector Database). Upload your own custom knowledge base files (PDF, txt, etc) using a simple React frontend.| | 938|tencentmusic/supersonic !2025-03-28330611|SuperSonic is the next-generation BI platform that integrates Chat BI (powered by LLM) and Headless BI (powered by semantic layer) paradigms.| | 939|billmei/every-chatgpt-gui !2025-03-2832981|Every front-end GUI client for ChatGPT| | 940|microsoft/torchgeo !2025-03-2832772|TorchGeo: datasets, samplers, transforms, and pre-trained models for geospatial data| | 941|LLMBook-zh/LLMBook-zh.github.io !2025-03-28326110|《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣| | 942|dvlab-research/MiniGemini !2025-03-2832601|Official implementation for Mini-Gemini| | 943|rashadphz/farfalle !2025-03-2832460|🔍 AI search engine - self-host with local or cloud LLMs| | 944|Luodian/Otter !2025-03-2832450|🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.| | 945|AprilNEA/ChatGPT-Admin-Web !2025-03-2832370 | ChatGPT WebUI with user management and admin dashboard system| | 946|MarkFzp/act-plus-plus !2025-03-2832365|Imitation Learning algorithms with Co-traing for Mobile ALOHA: ACT, Diffusion Policy, VINN| | 947|ethen8181/machine-learning !2025-03-2832310|🌎 machine learning tutorials (mainly in Python3)| | 948|opengeos/segment-geospatial !2025-03-2832312 |A Python package for segmenting geospatial data with the Segment Anything Model (SAM)| | 949|iusztinpaul/hands-on-llms !2025-03-283225-2|🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦𝘢𝘥𝘪𝘯𝘨 𝘮𝘢𝘵𝘦𝘳𝘪𝘢𝘭𝘴| | 950|ToTheBeginning/PuLID !2025-03-2832221|Official code for PuLID: Pure and Lightning ID Customization via Contrastive Alignment| | 951|neo4j-labs/llm-graph-builder !2025-03-2832164|Neo4j graph construction from unstructured data using LLMs| | 952|OpenGVLab/InternGPT !2025-03-2832150 |InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)| | 953|PKU-YuanGroup/Video-LLaVA !2025-03-2832060 |Video-LLaVA: Learning United Visual Representation by Alignment Before Projection| | 954|DataTalksClub/llm-zoomcamp !2025-03-2832030|LLM Zoomcamp - a free online course about building an AI bot that can answer questions about your knowledge base| | 955|gptscript-ai/gptscript !2025-03-2832010|Natural Language Programming| |!green-up-arrow.svg 956|isaac-sim/IsaacLab !2025-03-28320113|Unified framework for robot learning built on NVIDIA Isaac Sim| |!red-down-arrow 957|ai-boost/Awesome-GPTs !2025-03-2832003|Curated list of awesome GPTs 👍.| | 958|huggingface/safetensors !2025-03-2831901|Simple, safe way to store and distribute tensors| | 959|linyiLYi/bilibot !2025-03-2831771|A local chatbot fine-tuned by bilibili user comments.| | 960| project-baize/baize-chatbot !2025-03-283168-1 | Let ChatGPT teach your own chatbot in hours with a single GPU! | | 961|Azure-Samples/cognitive-services-speech-sdk !2025-03-2831280|Sample code for the Microsoft Cognitive Services Speech SDK| | 962|microsoft/Phi-3CookBook !2025-03-2831231|This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.| | 963|neuralmagic/deepsparse !2025-03-2831180|Sparsity-aware deep learning inference runtime for CPUs| | 964|sugarforever/chat-ollama !2025-03-2831000|ChatOllama is an open source chatbot based on LLMs. It supports a wide range of language models, and knowledge base management.| | 965|amazon-science/chronos-forecasting !2025-03-2830974|Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting| | 966|damo-vilab/i2vgen-xl !2025-03-2830902|Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models| | 967|google-deepmind/gemma !2025-03-2830733|Open weights LLM from Google DeepMind.| | 968|iree-org/iree !2025-03-2830733|A retargetable MLIR-based machine learning compiler and runtime toolkit.| | 969|NVlabs/VILA !2025-03-2830724|VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)| | 970|microsoft/torchscale !2025-03-2830661|Foundation Architecture for (M)LLMs| | 971|openai/openai-realtime-console !2025-03-2830656|React app for inspecting, building and debugging with the Realtime API| | 972|daveshap/OpenAIAgentSwarm !2025-03-2830610|HAAS = Hierarchical Autonomous Agent Swarm - "Resistance is futile!"| | 973|microsoft/PromptWizard !2025-03-2830555|Task-Aware Agent-driven Prompt Optimization Framework| | 974|CVI-SZU/Linly !2025-03-2830490 |Chinese-LLaMA basic model; ChatFlow Chinese conversation model; NLP pre-training/command fine-tuning dataset| | 975|cohere-ai/cohere-toolkit !2025-03-2830130|Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.| | 976|adamcohenhillel/ADeus !2025-03-2830131|An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own server. You can then chat with Adeus using the app, and it will have all the right context about what you want to talk about - a truly personalized, personal AI.| | 977|Lightning-AI/LitServe !2025-03-2830132|Lightning-fast serving engine for AI models. Flexible. Easy. Enterprise-scale.| | 978|potpie-ai/potpie !2025-03-2829973|Prompt-To-Agent : Create custom engineering agents for your codebase| | 979|ant-design/x !2025-03-28299529|Craft AI-driven interfaces effortlessly 🤖| | 980|meta-llama/PurpleLlama !2025-03-2829832|Set of tools to assess and improve LLM security.| | 981|williamyang1991/RerenderAVideo !2025-03-2829800|[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation| | 982|baichuan-inc/Baichuan-13B !2025-03-2829790|A 13B large language model developed by Baichuan Intelligent Technology| | 983|Stability-AI/stable-audio-tools !2025-03-2829761|Generative models for conditional audio generation| | 984|li-plus/chatglm.cpp !2025-03-2829720|C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs| | 985|NVIDIA/GenerativeAIExamples !2025-03-2829546|Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.| | 986|Josh-XT/AGiXT !2025-03-2829521 |AGiXT is a dynamic AI Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.| | 987|MrForExample/ComfyUI-3D-Pack !2025-03-2829515|An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)| | 988|olimorris/codecompanion.nvim !2025-03-28295111|✨ AI-powered coding, seamlessly in Neovim. Supports Anthropic, Copilot, Gemini, Ollama, OpenAI and xAI LLMs| | 989|salesforce/CodeT5 !2025-03-282940-1 |Home of CodeT5: Open Code LLMs for Code Understanding and Generation| | 990|facebookresearch/ijepa !2025-03-2829391|Official codebase for I-JEPA, the Image-based Joint-Embedding Predictive Architecture. First outlined in the CVPR paper, "Self-supervised learning from images with a joint-embedding predictive architecture."| | 991|eureka-research/Eureka !2025-03-2829351|Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models"| | 992|NVIDIA/trt-llm-rag-windows !2025-03-282934-1|A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM| | 993|gmpetrov/databerry !2025-03-282930-1|The no-code platform for building custom LLM Agents| | 994|AI4Finance-Foundation/FinRobot !2025-03-28291946|FinRobot: An Open-Source AI Agent Platform for Financial Applications using LLMs 🚀 🚀 🚀| | 995|nus-apr/auto-code-rover !2025-03-2829013|A project structure aware autonomous software engineer aiming for autonomous program improvement| | 996|deepseek-ai/DreamCraft3D !2025-03-2828921|[ICLR 2024] Official implementation of DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior| | 997|mlabonne/llm-datasets !2025-03-2828848|High-quality datasets, tools, and concepts for LLM fine-tuning.| | 998|facebookresearch/jepa !2025-03-2828712|PyTorch code and models for V-JEPA self-supervised learning from video.| | 999|facebookresearch/habitat-sim !2025-03-2828604|A flexible, high-performance 3D simulator for Embodied AI research.| | 1000|xenova/whisper-web !2025-03-2828581|ML-powered speech recognition directly in your browser| | 1001|cvlab-columbia/zero123 !2025-03-2828530|Zero-1-to-3: Zero-shot One Image to 3D Object: https://zero123.cs.columbia.edu/| | 1002|yuruotong1/autoMate !2025-03-28285121|Like Manus, Computer Use Agent(CUA) and Omniparser, we are computer-using agents.AI-driven local automation assistant that uses natural language to make computers work by themselves| | 1003|muellerberndt/mini-agi !2025-03-282845-1 |A minimal generic autonomous agent based on GPT3.5/4. Can analyze stock prices, perform network security tests, create art, and order pizza.| | 1004|allenai/open-instruct !2025-03-2828432|| | 1005|CodingChallengesFYI/SharedSolutions !2025-03-2828360|Publicly shared solutions to Coding Challenges| | 1006|hegelai/prompttools !2025-03-2828220|Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).| | 1007|mazzzystar/Queryable !2025-03-2828222|Run CLIP on iPhone to Search Photos.| | 1008|Doubiiu/DynamiCrafter !2025-03-2828173|DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors| | 1009|SamurAIGPT/privateGPT !2025-03-282805-1 |An app to interact privately with your documents using the power of GPT, 100% privately, no data leaks| | 1010|facebookresearch/Pearl !2025-03-2827951|A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.| | 1011|intuitem/ciso-assistant-community !2025-03-2827954|CISO Assistant is a one-stop-shop for GRC, covering Risk, AppSec and Audit Management and supporting +70 frameworks worldwide with auto-mapping: NIST CSF, ISO 27001, SOC2, CIS, PCI DSS, NIS2, CMMC, PSPF, GDPR, HIPAA, Essential Eight, NYDFS-500, DORA, NIST AI RMF, 800-53, 800-171, CyFun, CJIS, AirCyber, NCSC, ECC, SCF and so much more| | 1012|facebookresearch/audio2photoreal !2025-03-2827840|Code and dataset for photorealistic Codec Avatars driven from audio| | 1013|Azure/azure-rest-api-specs !2025-03-2827770|The source for REST API specifications for Microsoft Azure.| | 1014|SCUTlihaoyu/open-chat-video-editor !2025-03-2827690 |Open source short video automatic generation tool| | 1015|Alpha-VLLM/LLaMA2-Accessory !2025-03-2827642|An Open-source Toolkit for LLM Development| | 1016|johnma2006/mamba-minimal !2025-03-2827601|Simple, minimal implementation of the Mamba SSM in one file of PyTorch.| | 1017|nerfstudio-project/gsplat !2025-03-2827576|CUDA accelerated rasterization of gaussian splatting| | 1018|Physical-Intelligence/openpi !2025-03-28274617|| | 1019|leptonai/leptonai !2025-03-2827246|A Pythonic framework to simplify AI service building| |!green-up-arrow.svg 1020|joanrod/star-vector !2025-03-28271149|StarVector is a foundation model for SVG generation that transforms vectorization into a code generation task. Using a vision-language modeling architecture, StarVector processes both visual and textual inputs to produce high-quality SVG code with remarkable precision.| |!red-down-arrow 1021|jqnatividad/qsv !2025-03-2827092|CSVs sliced, diced & analyzed.| | 1022|FranxYao/chain-of-thought-hub !2025-03-2826991|Benchmarking large language models' complex reasoning ability with chain-of-thought prompting| | 1023|princeton-nlp/SWE-bench !2025-03-2826965|[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?| | 1024|elastic/otel-profiling-agent !2025-03-2826930|The production-scale datacenter profiler| | 1025|src-d/hercules !2025-03-2826900|Gaining advanced insights from Git repository history.| | 1026|lanqian528/chat2api !2025-03-2826695|A service that can convert ChatGPT on the web to OpenAI API format.| | 1027|ishan0102/vimGPT !2025-03-2826681|Browse the web with GPT-4V and Vimium| | 1028|TMElyralab/MuseV !2025-03-2826650|MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising| | 1029|georgia-tech-db/eva !2025-03-2826600 |AI-Relational Database System | | 1030|kubernetes-sigs/controller-runtime !2025-03-2826590|Repo for the controller-runtime subproject of kubebuilder (sig-apimachinery)| | 1031|gptlink/gptlink !2025-03-2826550 |Build your own free commercial ChatGPT environment in 10 minutes. The setup is simple and includes features such as user management, orders, tasks, and payments| | 1032|pytorch/executorch !2025-03-2826534|On-device AI across mobile, embedded and edge for PyTorch| | 1033|NVIDIA/nv-ingest !2025-03-2826290|NVIDIA Ingest is an early access set of microservices for parsing hundreds of thousands of complex, messy unstructured PDFs and other enterprise documents into metadata and text to embed into retrieval systems.| | 1034|SuperTux/supertux !2025-03-2826081|SuperTux source code| | 1035|abi/secret-llama !2025-03-2826050|Fully private LLM chatbot that runs entirely with a browser with no server needed. Supports Mistral and LLama 3.| | 1036|liou666/polyglot !2025-03-2825841 |Desktop AI Language Practice Application| | 1037|janhq/nitro !2025-03-2825821|A fast, lightweight, embeddable inference engine to supercharge your apps with local AI. OpenAI-compatible API| | 1038|deepseek-ai/DeepSeek-Math !2025-03-2825825|DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models| | 1039|anthropics/prompt-eng-interactive-tutorial !2025-03-2825781|Anthropic's Interactive Prompt Engineering Tutorial| | 1040|microsoft/promptbench !2025-03-2825741|A unified evaluation framework for large language models| | 1041|baaivision/Painter !2025-03-2825580 |Painter & SegGPT Series: Vision Foundation Models from BAAI| | 1042|OpenPipe/OpenPipe !2025-03-2825581|Turn expensive prompts into cheap fine-tuned models| | 1043|TracecatHQ/tracecat !2025-03-2825531|😼 The AI-native, open source alternative to Tines / Splunk SOAR.| | 1044|JoshuaC215/agent-service-toolkit !2025-03-2825528|Full toolkit for running an AI agent service built with LangGraph, FastAPI and Streamlit| | 1045|databricks/dbrx !2025-03-2825460|Code examples and resources for DBRX, a large language model developed by Databricks| | 1046|lamini-ai/lamini !2025-03-2825271 |Official repo for Lamini's data generator for generating instructions to train instruction-following LLMs| | 1047|mshumer/gpt-author !2025-03-282510-1|| | 1048|TMElyralab/MusePose !2025-03-2824971|MusePose: a Pose-Driven Image-to-Video Framework for Virtual Human Generation| | 1049|Kludex/fastapi-tips !2025-03-2824974|FastAPI Tips by The FastAPI Expert!| | 1050|openai/simple-evals !2025-03-2824813|| | 1051|iterative/datachain !2025-03-2824732|AI-data warehouse to enrich, transform and analyze data from cloud storages| | 1052|girafe-ai/ml-course !2025-03-2824703|Open Machine Learning course| | 1053|kevmo314/magic-copy !2025-03-2824620 |Magic Copy is a Chrome extension that uses Meta's Segment Anything Model to extract a foreground object from an image and copy it to the clipboard.| | 1054|Eladlev/AutoPrompt !2025-03-2824432|A framework for prompt tuning using Intent-based Prompt Calibration| | 1055|OpenBMB/CPM-Bee !2025-03-282434-1 |A bilingual large-scale model with trillions of parameters| | 1056|IDEA-Research/T-Rex !2025-03-2824310|T-Rex2: Towards Generic Object Detection via Text-Visual Prompt Synergy| | 1057|microsoft/genaiscript !2025-03-2824202|Automatable GenAI Scripting| | 1058|paulpierre/RasaGPT !2025-03-2824090 |💬 RasaGPT is the first headless LLM chatbot platform built on top of Rasa and Langchain. Built w/ Rasa, FastAPI, Langchain, LlamaIndex, SQLModel, pgvector, ngrok, telegram| | 1059|ashishpatel26/LLM-Finetuning !2025-03-2823911|LLM Finetuning with peft| | 1060|SoraWebui/SoraWebui !2025-03-2823570|SoraWebui is an open-source Sora web client, enabling users to easily create videos from text with OpenAI's Sora model.| | 1061|6drf21e/ChatTTScolab !2025-03-2823491|🚀 一键部署(含离线整合包)!基于 ChatTTS ,支持音色抽卡、长音频生成和分角色朗读。简单易用,无需复杂安装。| | 1062|Azure/PyRIT !2025-03-2823343|The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.| | 1063|tencent-ailab/V-Express !2025-03-2823201|V-Express aims to generate a talking head video under the control of a reference image, an audio, and a sequence of V-Kps images.| | 1064|THUDM/CogVLM2 !2025-03-2823170|GPT4V-level open-source multi-modal model based on Llama3-8B| | 1065|dvmazur/mixtral-offloading !2025-03-2823001|Run Mixtral-8x7B models in Colab or consumer desktops| | 1066|semanser/codel !2025-03-2822950|✨ Fully autonomous AI Agent that can perform complicated tasks and projects using terminal, browser, and editor.| | 1067|mshumer/gpt-investor !2025-03-2822590|| | 1068|aixcoder-plugin/aiXcoder-7B !2025-03-2822550|official repository of aiXcoder-7B Code Large Language Model| | 1069|Azure-Samples/graphrag-accelerator !2025-03-2822503|One-click deploy of a Knowledge Graph powered RAG (GraphRAG) in Azure| | 1070|emcf/engshell !2025-03-2821830 |An English-language shell for any OS, powered by LLMs| | 1071|hncboy/chatgpt-web-java !2025-03-2821771|ChatGPT project developed in Java, based on Spring Boot 3 and JDK 17, supports both AccessToken and ApiKey modes| | 1072|openai/consistencydecoder !2025-03-2821692|Consistency Distilled Diff VAE| | 1073|Alpha-VLLM/Lumina-T2X !2025-03-2821681|Lumina-T2X is a unified framework for Text to Any Modality Generation| | 1074|bghira/SimpleTuner !2025-03-2821612|A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.| | 1075|JiauZhang/DragGAN !2025-03-2821530 |Implementation of DragGAN: Interactive Point-based Manipulation on the Generative Image Manifold| | 1076|cgpotts/cs224u !2025-03-2821390|Code for Stanford CS224u| | 1077|PKU-YuanGroup/MoE-LLaVA !2025-03-2821300|Mixture-of-Experts for Large Vision-Language Models| | 1078|darrenburns/elia !2025-03-2820831|A snappy, keyboard-centric terminal user interface for interacting with large language models. Chat with ChatGPT, Claude, Llama 3, Phi 3, Mistral, Gemma and more.| | 1079|ageerle/ruoyi-ai !2025-03-28207898|RuoYi AI 是一个全栈式 AI 开发平台,旨在帮助开发者快速构建和部署个性化的 AI 应用。| | 1080|NVIDIA/gpu-operator !2025-03-2820510|NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes| | 1081|BAAI-Agents/Cradle !2025-03-2820481|The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.| | 1082|microsoft/aici !2025-03-2820080|AICI: Prompts as (Wasm) Programs| | 1083|PRIS-CV/DemoFusion !2025-03-2820040|Let us democratise high-resolution generation! (arXiv 2023)| | 1084|apple/axlearn !2025-03-2820012|An Extensible Deep Learning Library| | 1085|naver/mast3r !2025-03-2819685|Grounding Image Matching in 3D with MASt3R| | 1086|liltom-eth/llama2-webui !2025-03-281958-1|Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.| | 1087|GaParmar/img2img-turbo !2025-03-2819582|One-step image-to-image with Stable Diffusion turbo: sketch2image, day2night, and more| | 1088|Niek/chatgpt-web !2025-03-2819560|ChatGPT web interface using the OpenAI API| | 1089|huggingface/cookbook !2025-03-2819421|Open-source AI cookbook| | 1090|pytorch/ao !2025-03-2819241|PyTorch native quantization and sparsity for training and inference| | 1091|emcie-co/parlant !2025-03-2819053|The behavior guidance framework for customer-facing LLM agents| | 1092|ymcui/Chinese-LLaMA-Alpaca-3 !2025-03-2818980|中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3| | 1093|Nutlope/notesGPT !2025-03-2818811|Record voice notes & transcribe, summarize, and get tasks| | 1094|InstantStyle/InstantStyle !2025-03-2818791|InstantStyle: Free Lunch towards Style-Preserving in Text-to-Image Generation 🔥| | 1095|idaholab/moose !2025-03-2818771|Multiphysics Object Oriented Simulation Environment| | 1096|The-OpenROAD-Project/OpenROAD !2025-03-2818351|OpenROAD's unified application implementing an RTL-to-GDS Flow. Documentation at https://openroad.readthedocs.io/en/latest/| | 1097|alibaba/spring-ai-alibaba !2025-03-281831121|Agentic AI Framework for Java Developers| | 1098|ytongbai/LVM !2025-03-2817990|Sequential Modeling Enables Scalable Learning for Large Vision Models| | 1099|microsoft/sample-app-aoai-chatGPT !2025-03-2817981|[PREVIEW] Sample code for a simple web chat experience targeting chatGPT through AOAI.| | 1100|AI-Citizen/SolidGPT !2025-03-2817830|Chat everything with your code repository, ask repository level code questions, and discuss your requirements. AI Scan and learning your code repository, provide you code repository level answer🧱 🧱| | 1101|YangLing0818/RPG-DiffusionMaster !2025-03-2817784|Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (PRG)| | 1102|kyegomez/BitNet !2025-03-2817710|Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch| | 1103|eloialonso/diamond !2025-03-2817671|DIAMOND (DIffusion As a Model Of eNvironment Dreams) is a reinforcement learning agent trained in a diffusion world model.| | 1104|flowdriveai/flowpilot !2025-03-2817250|flow-pilot is an openpilot based driver assistance system that runs on linux, windows and android powered machines.| | 1105|xlang-ai/OSWorld !2025-03-2817200|OSWorld: Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments| | 1106|linyiLYi/snake-ai !2025-03-2817031|An AI agent that beats the classic game "Snake".| | 1107|baaivision/Emu !2025-03-2816991|Emu Series: Generative Multimodal Models from BAAI| | 1108|kevmo314/scuda !2025-03-2816870|SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.| | 1109|SharifiZarchi/IntroductiontoMachineLearning !2025-03-2816701|دوره‌ی مقدمه‌ای بر یادگیری ماشین، برای دانشجویان| | 1110|google/maxtext !2025-03-2816670|A simple, performant and scalable Jax LLM!| | 1111|ml-explore/mlx-swift-examples !2025-03-2816471|Examples using MLX Swift| | 1112|unitreerobotics/unitreerlgym !2025-03-2816256|| | 1113|collabora/WhisperFusion !2025-03-2815901|WhisperFusion builds upon the capabilities of WhisperLive and WhisperSpeech to provide a seamless conversations with an AI.| | 1114|lichao-sun/Mora !2025-03-2815520|Mora: More like Sora for Generalist Video Generation| | 1115|GoogleCloudPlatform/localllm !2025-03-2815370|Run LLMs locally on Cloud Workstations| | 1116|TencentARC/BrushNet !2025-03-2815330|The official implementation of paper "BrushNet: A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion"| | 1117|ai-christianson/RA.Aid !2025-03-2815288|Develop software autonomously.| | 1118|stephansturges/WALDO !2025-03-2815170|Whereabouts Ascertainment for Low-lying Detectable Objects. The SOTA in FOSS AI for drones!| | 1119|skills/copilot-codespaces-vscode !2025-03-2815112|Develop with AI-powered code suggestions using GitHub Copilot and VS Code| | 1120|andrewnguonly/Lumos !2025-03-2814920|A RAG LLM co-pilot for browsing the web, powered by local LLMs| | 1121|TeamNewPipe/NewPipeExtractor !2025-03-2814811|NewPipe's core library for extracting data from streaming sites| | 1122|mhamilton723/FeatUp !2025-03-2814770|Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024| | 1123|AnswerDotAI/fsdpqlora !2025-03-2814671|Training LLMs with QLoRA + FSDP| | 1124|jgravelle/AutoGroq !2025-03-2814330|| | 1125|OpenGenerativeAI/llm-colosseum !2025-03-2814130|Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM| | 1126|microsoft/vscode-ai-toolkit !2025-03-2814000|| | 1127|McGill-NLP/webllama !2025-03-2813930|Llama-3 agents that can browse the web by following instructions and talking to you| | 1128|lucidrains/self-rewarding-lm-pytorch !2025-03-2813760|Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI| | 1129|ishaan1013/sandbox !2025-03-2813650|A cloud-based code editing environment with an AI copilot and real-time collaboration.| | 1130|goatcorp/Dalamud !2025-03-2813275|FFXIV plugin framework and API| | 1131|Lightning-AI/lightning-thunder !2025-03-2813151|Make PyTorch models Lightning fast! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once.| | 1132|PKU-YuanGroup/MagicTime !2025-03-2813052|MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators| | 1133|SakanaAI/evolutionary-model-merge !2025-03-2813000|Official repository of Evolutionary Optimization of Model Merging Recipes| | 1134|a-real-ai/pywinassistant !2025-03-2812950|The first open source Large Action Model generalist Artificial Narrow Intelligence that controls completely human user interfaces by only using natural language. PyWinAssistant utilizes Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models.| | 1135|TraceMachina/nativelink !2025-03-2812630|NativeLink is an open source high-performance build cache and remote execution server, compatible with Bazel, Buck2, Reclient, and other RBE-compatible build systems. It offers drastically faster builds, reduced test flakiness, and significant infrastructure cost savings.| | 1136|MLSysOps/MLE-agent !2025-03-2812500|🤖 MLE-Agent: Your intelligent companion for seamless AI engineering and research. 🔍 Integrate with arxiv and paper with code to provide better code/research plans 🧰 OpenAI, Ollama, etc supported. 🎆 Code RAG| | 1137|wpilibsuite/allwpilib !2025-03-2811610|Official Repository of WPILibJ and WPILibC| | 1138|elfvingralf/macOSpilot-ai-assistant !2025-03-2811470|Voice + Vision powered AI assistant that answers questions about any application, in context and in audio.| | 1139|langchain-ai/langchain-extract !2025-03-2811210|🦜⛏️ Did you say you like data?| | 1140|FoundationVision/GLEE !2025-03-2811120|【CVPR2024】GLEE: General Object Foundation Model for Images and Videos at Scale| | 1141|Profluent-AI/OpenCRISPR !2025-03-2810990|AI-generated gene editing systems| | 1142|zju3dv/EasyVolcap !2025-03-2810821|[SIGGRAPH Asia 2023 (Technical Communications)] EasyVolcap: Accelerating Neural Volumetric Video Research| | 1143|PaddlePaddle/PaddleHelix !2025-03-2810560|Bio-Computing Platform Featuring Large-Scale Representation Learning and Multi-Task Deep Learning “螺旋桨”生物计算工具集| | 1144|myshell-ai/JetMoE !2025-03-289800|Reaching LLaMA2 Performance with 0.1M Dollars| | 1145|likejazz/llama3.np !2025-03-289770|llama3.np is pure NumPy implementation for Llama 3 model.| | 1146|mustafaaljadery/gemma-2B-10M !2025-03-289500|Gemma 2B with 10M context length using Infini-attention.| | 1147|HITsz-TMG/FilmAgent !2025-03-289382|Resources of our paper "FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces". New versions in the making!| | 1148|aws-samples/amazon-bedrock-samples !2025-03-289362|This repository contains examples for customers to get started using the Amazon Bedrock Service. This contains examples for all available foundational models| | 1149|Akkudoktor-EOS/EOS !2025-03-2893154|This repository features an Energy Optimization System (EOS) that optimizes energy distribution, usage for batteries, heat pumps& household devices. It includes predictive models for electricity prices (planned), load forecasting& dynamic optimization to maximize energy efficiency & minimize costs. Founder Dr. Andreas Schmitz (YouTube @akkudoktor)| Tip: | symbol| rule | | :----| :---- | |🔥 | 256 1k| |!green-up-arrow.svg !red-down-arrow | ranking up / down| |⭐ | on trending page today| [Back to Top] Tools | No. | Tool | Description | | ----:|:----------------------------------------------- |:------------------------------------------------------------------------------------------- | | 1 | ChatGPT | A sibling model to InstructGPT, which is trained to follow instructions in a prompt and provide a detailed response | | 2 | DALL·E 2 | Create original, realistic images and art from a text description | | 3 | Murf AI | AI enabled, real people's voices| | 4 | Midjourney | An independent research lab that produces an artificial intelligence program under the same name that creates images from textual descriptions, used in Discord | 5 | Make-A-Video | Make-A-Video is a state-of-the-art AI system that generates videos from text | | 6 | Creative Reality™ Studio by D-ID| Use generative AI to create future-facing videos| | 7 | chat.D-ID| The First App Enabling Face-to-Face Conversations with ChatGPT| | 8 | Notion AI| Access the limitless power of AI, right inside Notion. Work faster. Write better. Think bigger. | | 9 | Runway| Text to Video with Gen-2 | | 10 | Resemble AI| Resemble’s AI voice generator lets you create human–like voice overs in seconds | | 11 | Cursor| Write, edit, and chat about your code with a powerful AI | | 12 | Hugging Face| Build, train and deploy state of the art models powered by the reference open source in machine learning | | 13 | Claude | A next-generation AI assistant for your tasks, no matter the scale | | 14 | Poe| Poe lets you ask questions, get instant answers, and have back-and-forth conversations with AI. Gives access to GPT-4, gpt-3.5-turbo, Claude from Anthropic, and a variety of other bots| [Back to Top] Websites | No. | WebSite |Description | | ----:|:------------------------------------------ |:---------------------------------------------------------------------------------------- | | 1 | OpenAI | An artificial intelligence research lab | | 2 | Bard | Base Google's LaMDA chatbots and pull from internet | | 3 | ERNIE Bot | Baidu’s new generation knowledge-enhanced large language model is a new member of the Wenxin large model family | | 4 | DALL·E 2 | An AI system that can create realistic images and art from a description in natural language | | 5 | Whisper | A general-purpose speech recognition model | | 6| CivitAI| A platform that makes it easy for people to share and discover resources for creating AI art| | 7|D-ID| D-ID’s Generative AI enables users to transform any picture or video into extraordinary experiences| | 8| Nvidia eDiff-I| Text-to-Image Diffusion Models with Ensemble of Expert Denoisers | | 9| Stability AI| The world's leading open source generative AI company which opened source Stable Diffusion | | 10| Meta AI| Whether it be research, product or infrastructure development, we’re driven to innovate responsibly with AI to benefit the world | | 11| ANTHROPIC| AI research and products that put safety at the frontier | [Back to Top] Reports&Papers | No. | Report&Paper | Description | |:---- |:-------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------- | | 1 | GPT-4 Technical Report | GPT-4 Technical Report | | 2 | mli/paper-reading | Deep learning classics and new papers are read carefully paragraph by paragraph. | | 3 | labmlai/annotateddeeplearningpaperimplementations| A collection of simple PyTorch implementations of neural networks and related algorithms, which are documented with explanations | | 4 | Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models | Talking, Drawing and Editing with Visual Foundation Models | | 5 | OpenAI Research | The latest research report and papers from OpenAI | | 6 | Make-A-Video: Text-to-Video Generation without Text-Video Data|Meta's Text-to-Video Generation| | 7 | eDiff-I: Text-to-Image Diffusion Models with Ensemble of Expert Denoisers| Nvidia eDiff-I - New generation of generative AI content creation tool | | 8 | Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo | 2023 GPT4All Technical Report | | 9 | Segment Anything| Meta Segment Anything | | 10 | LLaMA: Open and Efficient Foundation Language Models| LLaMA: a collection of foundation language models ranging from 7B to 65B parameters| | 11 | papers-we-love/papers-we-love |Papers from the computer science community to read and discuss| | 12 | CVPR 2023 papers |The most exciting and influential CVPR 2023 papers| [Back to Top] Tutorials | No. | Tutorial | Description| |:---- |:---------------------------------------------------------------- | --- | | 1 | Coursera - Machine Learning | The Machine Learning Specialization Course taught by Dr. Andrew Ng| | 2 | microsoft/ML-For-Beginners | 12 weeks, 26 lessons, 52 quizzes, classic Machine Learning for all| | 3 | ChatGPT Prompt Engineering for Developers | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI) will teach how to use a large language model (LLM) to quickly build new and powerful applications | | 4 | Dive into Deep Learning |Targeting Chinese readers, functional and open for discussion. The Chinese and English versions are used for teaching in over 400 universities across more than 60 countries | | 5 | AI Expert Roadmap | Roadmap to becoming an Artificial Intelligence Expert in 2022 | | 6 | Computer Science courses |List of Computer Science courses with video lectures| | 7 | Machine Learning with Python | Machine Learning with Python Certification on freeCodeCamp| | 8 | Building Systems with the ChatGPT API | This short course taught by Isa Fulford (OpenAI) and Andrew Ng (DeepLearning.AI), you will learn how to automate complex workflows using chain calls to a large language model| | 9 | LangChain for LLM Application Development | This short course taught by Harrison Chase (Co-Founder and CEO at LangChain) and Andrew Ng. you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework| | 10 | How Diffusion Models Work | This short course taught by Sharon Zhou (CEO, Co-founder, Lamini). you will gain a deep familiarity with the diffusion process and the models which carry it out. More than simply pulling in a pre-built model or using an API, this course will teach you to build a diffusion model from scratch| | 11 | Free Programming Books For AI |📚 Freely available programming books for AI | | 12 | microsoft/AI-For-Beginners |12 Weeks, 24 Lessons, AI for All!| | 13 | hemansnation/God-Level-Data-Science-ML-Full-Stack |A collection of scientific methods, processes, algorithms, and systems to build stories & models. This roadmap contains 16 Chapters, whether you are a fresher in the field or an experienced professional who wants to transition into Data Science & AI| | 14 | datawhalechina/prompt-engineering-for-developers |Chinese version of Andrew Ng's Big Model Series Courses, including "Prompt Engineering", "Building System", and "LangChain"| | 15 | ossu/computer-science |🎓 Path to a free self-taught education in Computer Science!| | 16 | microsoft/Data-Science-For-Beginners | 10 Weeks, 20 Lessons, Data Science for All! | |17 |jwasham/coding-interview-university !2023-09-29268215336 |A complete computer science study plan to become a software engineer.| [Back to Top] Thanks If this project has been helpful to you in any way, please give it a ⭐️ by clicking on the star.

RD-Agent
github
LLM Vibe Score0.548
Human Vibe Score0.27921589729164453
microsoftMar 28, 2025

RD-Agent

🖥️ Live Demo | 🎥 Demo Video ▶️YouTube | 📖 Documentation | 📃 Papers Data Science Agent Preview Check out our demo video showcasing the current progress of our Data Science Agent under development: https://github.com/user-attachments/assets/3eccbecb-34a4-4c81-bce4-d3f8862f7305 📰 News | 🗞️ News | 📝 Description | | -- | ------ | | Support LiteLLM Backend | We now fully support LiteLLM as a backend for integration with multiple LLM providers. | | More General Data Science Agent | 🚀Coming soon! | | Kaggle Scenario release | We release Kaggle Agent, try the new features! | | Official WeChat group release | We created a WeChat group, welcome to join! (🗪QR Code) | | Official Discord release | We launch our first chatting channel in Discord (🗪) | | First release | RDAgent is released on GitHub | 🌟 Introduction RDAgent aims to automate the most critical and valuable aspects of the industrial R&D process, and we begin with focusing on the data-driven scenarios to streamline the development of models and data. Methodologically, we have identified a framework with two key components: 'R' for proposing new ideas and 'D' for implementing them. We believe that the automatic evolution of R&D will lead to solutions of significant industrial value. R&D is a very general scenario. The advent of RDAgent can be your 💰 Automatic Quant Factory (🎥Demo Video|▶️YouTube) 🤖 Data Mining Agent: Iteratively proposing data & models (🎥Demo Video 1|▶️YouTube) (🎥Demo Video 2|▶️YouTube) and implementing them by gaining knowledge from data. 🦾 Research Copilot: Auto read research papers (🎥Demo Video|▶️YouTube) / financial reports (🎥Demo Video|▶️YouTube) and implement model structures or building datasets. 🤖 Kaggle Agent: Auto Model Tuning and Feature Engineering([🎥Demo Video Coming Soon...]()) and implementing them to achieve more in competitions. ... You can click the links above to view the demo. We're continuously adding more methods and scenarios to the project to enhance your R&D processes and boost productivity. Additionally, you can take a closer look at the examples in our 🖥️ Live Demo. ⚡ Quick start You can try above demos by running the following command: 🐳 Docker installation. Users must ensure Docker is installed before attempting most scenarios. Please refer to the official 🐳Docker page for installation instructions. Ensure the current user can run Docker commands without using sudo. You can verify this by executing docker run hello-world. 🐍 Create a Conda Environment Create a new conda environment with Python (3.10 and 3.11 are well-tested in our CI): Activate the environment: 🛠️ Install the RDAgent You can directly install the RDAgent package from PyPI: 💊 Health check rdagent provides a health check that currently checks two things. whether the docker installation was successful. whether the default port used by the rdagent ui is occupied. ⚙️ Configuration The demos requires following ability: ChatCompletion json_mode embedding query For example: If you are using the OpenAI API, you have to configure your GPT model in the .env file like this. However, not every API services support these features by default. For example: AZURE OpenAI, you have to configure your GPT model in the .env file like this. We now support LiteLLM as a backend for integration with multiple LLM providers. If you use LiteLLM Backend to use models, you can configure as follows: For more configuration information, please refer to the documentation. 🚀 Run the Application The 🖥️ Live Demo is implemented by the following commands(each item represents one demo, you can select the one you prefer): Run the Automated Quantitative Trading & Iterative Factors Evolution: Qlib self-loop factor proposal and implementation application Run the Automated Quantitative Trading & Iterative Model Evolution: Qlib self-loop model proposal and implementation application Run the Automated Medical Prediction Model Evolution: Medical self-loop model proposal and implementation application (1) Apply for an account at PhysioNet. (2) Request access to FIDDLE preprocessed data: FIDDLE Dataset. (3) Place your username and password in .env. Run the Automated Quantitative Trading & Factors Extraction from Financial Reports: Run the Qlib factor extraction and implementation application based on financial reports Run the Automated Model Research & Development Copilot: model extraction and implementation application Run the Automated Kaggle Model Tuning & Feature Engineering: self-loop model proposal and feature engineering implementation application Using sf-crime (San Francisco Crime Classification) as an example. Register and login on the Kaggle website. Configuring the Kaggle API. (1) Click on the avatar (usually in the top right corner of the page) -> Settings -> Create New Token, A file called kaggle.json will be downloaded. (2) Move kaggle.json to ~/.config/kaggle/ (3) Modify the permissions of the kaggle.json file. Reference command: chmod 600 ~/.config/kaggle/kaggle.json Join the competition: Click Join the competition -> I Understand and Accept at the bottom of the competition details page. Description of the above example: Kaggle competition data, contains two parts: competition description file (json file) and competition dataset (zip file). We prepare the competition description file for you, the competition dataset will be downloaded automatically when you run the program, as in the example. If you want to download the competition description file automatically, you need to install chromedriver, The instructions for installing chromedriver can be found in the documentation. The Competition List Available can be found here. 🖥️ Monitor the Application Results You can run the following command for our demo program to see the run logs. Note: Although port 19899 is not commonly used, but before you run this demo, you need to check if port 19899 is occupied. If it is, please change it to another port that is not occupied. You can check if a port is occupied by running the following command. 🏭 Scenarios We have applied RD-Agent to multiple valuable data-driven industrial scenarios. 🎯 Goal: Agent for Data-driven R&D In this project, we are aiming to build an Agent to automate Data-Driven R\&D that can 📄 Read real-world material (reports, papers, etc.) and extract key formulas, descriptions of interested features and models, which are the key components of data-driven R&D . 🛠️ Implement the extracted formulas (e.g., features, factors, and models) in runnable codes. Due to the limited ability of LLM in implementing at once, build an evolving process for the agent to improve performance by learning from feedback and knowledge. 💡 Propose new ideas based on current knowledge and observations. 📈 Scenarios/Demos In the two key areas of data-driven scenarios, model implementation and data building, our system aims to serve two main roles: 🦾Copilot and 🤖Agent. The 🦾Copilot follows human instructions to automate repetitive tasks. The 🤖Agent, being more autonomous, actively proposes ideas for better results in the future. The supported scenarios are listed below: | Scenario/Target | Model Implementation | Data Building | | -- | -- | -- | | 💹 Finance | 🤖 Iteratively Proposing Ideas & Evolving▶️YouTube | 🤖 Iteratively Proposing Ideas & Evolving ▶️YouTube 🦾 Auto reports reading & implementation▶️YouTube | | 🩺 Medical | 🤖 Iteratively Proposing Ideas & Evolving▶️YouTube | - | | 🏭 General | 🦾 Auto paper reading & implementation▶️YouTube 🤖 Auto Kaggle Model Tuning | 🤖Auto Kaggle feature Engineering | RoadMap: Currently, we are working hard to add new features to the Kaggle scenario. Different scenarios vary in entrance and configuration. Please check the detailed setup tutorial in the scenarios documents. Here is a gallery of successful explorations (5 traces showed in 🖥️ Live Demo). You can download and view the execution trace using this command from the documentation. Please refer to 📖readthedocs_scen for more details of the scenarios. ⚙️ Framework Automating the R&D process in data science is a highly valuable yet underexplored area in industry. We propose a framework to push the boundaries of this important research field. The research questions within this framework can be divided into three main categories: | Research Area | Paper/Work List | |--------------------|-----------------| | Benchmark the R&D abilities | Benchmark | | Idea proposal: Explore new ideas or refine existing ones | Research | | Ability to realize ideas: Implement and execute ideas | Development | We believe that the key to delivering high-quality solutions lies in the ability to evolve R&D capabilities. Agents should learn like human experts, continuously improving their R&D skills. More documents can be found in the 📖 readthedocs. 📃 Paper/Work list 📊 Benchmark Towards Data-Centric Automatic R&D !image 🔍 Research In a data mining expert's daily research and development process, they propose a hypothesis (e.g., a model structure like RNN can capture patterns in time-series data), design experiments (e.g., finance data contains time-series and we can verify the hypothesis in this scenario), implement the experiment as code (e.g., Pytorch model structure), and then execute the code to get feedback (e.g., metrics, loss curve, etc.). The experts learn from the feedback and improve in the next iteration. Based on the principles above, we have established a basic method framework that continuously proposes hypotheses, verifies them, and gets feedback from the real-world practice. This is the first scientific research automation framework that supports linking with real-world verification. For more detail, please refer to our 🖥️ Live Demo page. 🛠️ Development Collaborative Evolving Strategy for Automatic Data-Centric Development !image 🤝 Contributing We welcome contributions and suggestions to improve RD-Agent. Please refer to the Contributing Guide for more details on how to contribute. Before submitting a pull request, ensure that your code passes the automatic CI checks. 📝 Guidelines This project welcomes contributions and suggestions. Contributing to this project is straightforward and rewarding. Whether it's solving an issue, addressing a bug, enhancing documentation, or even correcting a typo, every contribution is valuable and helps improve RDAgent. To get started, you can explore the issues list, or search for TODO: comments in the codebase by running the command grep -r "TODO:". Before we released RD-Agent as an open-source project on GitHub, it was an internal project within our group. Unfortunately, the internal commit history was not preserved when we removed some confidential code. As a result, some contributions from our group members, including Haotian Chen, Wenjun Feng, Haoxue Wang, Zeqi Ye, Xinjie Shen, and Jinhui Li, were not included in the public commits. ⚖️ Legal disclaimer The RD-agent is provided “as is”, without warranty of any kind, express or implied, including but not limited to the warranties of merchantability, fitness for a particular purpose and noninfringement. The RD-agent is aimed to facilitate research and development process in the financial industry and not ready-to-use for any financial investment or advice. Users shall independently assess and test the risks of the RD-agent in a specific use scenario, ensure the responsible use of AI technology, including but not limited to developing and integrating risk mitigation measures, and comply with all applicable laws and regulations in all applicable jurisdictions. The RD-agent does not provide financial opinions or reflect the opinions of Microsoft, nor is it designed to replace the role of qualified financial professionals in formulating, assessing, and approving finance products. The inputs and outputs of the RD-agent belong to the users and users shall assume all liability under any theory of liability, whether in contract, torts, regulatory, negligence, products liability, or otherwise, associated with use of the RD-agent and any inputs and outputs thereof.

LLMStack
github
LLM Vibe Score0.535
Human Vibe Score0.022778788676674117
trypromptlyMar 28, 2025

LLMStack

LLMStack is a no-code platform for building generative AI agents, workflows and chatbots, connecting them to your data and business processes. Quickstart | Documentation | Promptly Overview Build tailor-made generative AI agents, applications and chatbots that cater to your unique needs by chaining multiple LLMs. Seamlessly integrate your own data, internal tools and GPT-powered models without any coding experience using LLMStack's no-code builder. Trigger your AI chains from Slack or Discord. Deploy to the cloud or on-premise. !llmstack-quickstart See full demo video here Getting Started Check out our Cloud offering at Promptly or follow the instructions below to deploy LLMStack on your own infrastructure. LLMStack deployment comes with a default admin account whose credentials are admin and promptly. Be sure to change the password from admin panel after logging in. Installation Prerequisites LLMStack depends on a background docker container to run jobs. Make sure you have Docker installed on your machine if want to use jobs. You can follow the instructions here to install Docker. Install LLMStack using pip If you are on windows, please use WSL2 (Windows Subsystem for Linux) to install LLMStack. You can follow the instructions here to install WSL2. Once you are in a WSL2 terminal, you can install LLMStack using the above command. Start LLMStack using the following command: Above commands will install and start LLMStack. It will create .llmstack in your home directory and places the database and config files in it when run for the first time. Once LLMStack is up and running, it should automatically open your browser and point it to localhost:3000. You can add your own keys to providers like OpenAI, Cohere, Stability etc., from Settings page. If you want to provide default keys for all the users of your LLMStack instance, you can add them to the ~/.llmstack/config file. LLMStack: Quickstart video Features 🤖 Agents: Build generative AI agents like AI SDRs, Research Analysts, RPA Automations etc., without writing any code. Connect agents to your internal or external tools, search the web or browse the internet with agents. 🔗 Chain multiple models: LLMStack allows you to chain multiple LLMs together to build complex generative AI applications. 📊 Use generative AI on your Data: Import your data into your accounts and use it in AI chains. LLMStack allows importing various types (CSV, TXT, PDF, DOCX, PPTX etc.,) of data from a variety of sources (gdrive, notion, websites, direct uploads etc.,). Platform will take care of preprocessing and vectorization of your data and store it in the vector database that is provided out of the box. 🛠️ No-code builder: LLMStack comes with a no-code builder that allows you to build AI chains without any coding experience. You can chain multiple LLMs together and connect them to your data and business processes. ☁️ Deploy to the cloud or on-premise: LLMStack can be deployed to the cloud or on-premise. You can deploy it to your own infrastructure or use our cloud offering at Promptly. 🚀 API access: Apps or chatbots built with LLMStack can be accessed via HTTP API. You can also trigger your AI chains from Slack or Discord. 🏢 Multi-tenant: LLMStack is multi-tenant. You can create multiple organizations and add users to them. Users can only access the data and AI chains that belong to their organization. What can you build with LLMStack? Using LLMStack you can build a variety of generative AI applications, chatbots and agents. Here are some examples: 👩🏻‍💼 AI SDRs: You can build AI SDRs (Sales Development Representatives) that can generate personalized emails, LinkedIn messages, cold calls, etc., for your sales team 👩🏻‍💻 Research Analysts: You can build AI Research Analysts that can generate research reports, investment thesis, etc., for your investment team 🤖 RPA Automations: You can build RPA automations that can automate your business processes by generating emails, filling forms, etc., 📝 Text generation: You can build apps that generate product descriptions, blog posts, news articles, tweets, emails, chat messages, etc., by using text generation models and optionally connecting your data. Check out this marketing content generator for example 🤖 Chatbots: You can build chatbots trained on your data powered by ChatGPT like Promptly Help that is embedded on Promptly website 🎨 Multimedia generation: Build complex applications that can generate text, images, videos, audio, etc. from a prompt. This story generator is an example 🗣️ Conversational AI: Build conversational AI systems that can have a conversation with a user. Check out this Harry Potter character chatbot 🔍 Search augmentation: Build search augmentation systems that can augment search results with additional information using APIs. Sharebird uses LLMStack to augment search results with AI generated answer from their content similar to Bing's chatbot 💬 Discord and Slack bots: Apps built on LLMStack can be triggered from Slack or Discord. You can easily connect your AI chains to Slack or Discord from LLMStack's no-code app editor. Check out our Discord server to interact with one such bot. Administration Login to http://localhost:3000/admin using the admin account. You can add users and assign them to organizations in the admin panel. Cloud Offering Check out our cloud offering at Promptly. You can sign up for a free account and start building your own generative AI applications. Documentation Check out our documentation at docs.trypromptly.com/llmstack to learn more about LLMStack. Development Check out our development guide at docs.trypromptly.com/llmstack/development to learn more about how to run and develop LLMStack. Contributing We welcome contributions to LLMStack. Please check out our contributing guide to learn more about how you can contribute to LLMStack.

CrewAI-Studio
github
LLM Vibe Score0.488
Human Vibe Score0.0100269728798468
strnadMar 28, 2025

CrewAI-Studio

CrewAI Studio Welcome to CrewAI Studio! This application provides a user-friendly interface written in Streamlit for interacting with CrewAI, suitable even for those who don't want to write any code. Follow the steps below to install and run the application using Docker/docker-compose or Conda/venv. Features Multi-platform support: Works on Windows, Linux and MacOS. No coding required: User-friendly interface for interacting with CrewAI. Conda and virtual environment support: Choose between Conda and a Python virtual environment for installation. Results history: You can view previous results. Knowledge sources: You can add knowledge sources for your crews CrewAI tools You can use crewai tools to interact with real world. ~~Crewai studio uses a forked version of crewai-tools with some bugfixes and enhancements (https://github.com/strnad/crewAI-tools)~~ (bugfixes already merged to crewai-tools) Custom Tools Custom tools for calling APIs, writing files, enhanced code interpreter, enhanced web scraper... More will be added soon LLM providers supported: Currently OpenAI, Groq, Anthropic, ollama, Grok and LM Studio backends are supported. OpenAI key is probably still needed for embeddings in many tools. Don't forget to load an embedding model when using LM Studio. Single Page app export: Feature to export crew as simple single page streamlit app. Threaded crew run: Crews can run in background and can be stopped. Support CrewAI Studio Your support helps fund the development and growth of our project. Every contribution is greatly appreciated! Donate with Bitcoin Sponsor via GitHub Screenshots Installation Using Virtual Environment For Virtual Environment: Ensure you have Python installed. If you dont have python instaled, you can simply use the conda installer. On Linux or MacOS Clone the repository (or use downloaded ZIP file): Run the installation script: Run the application: On Windows Clone the repository (or use downloaded ZIP file): Run the Conda installation script: Run the application: Using Conda Conda will be installed locally in the project folder. No need for a pre-existing Conda installation. On Linux Clone the repository (or use downloaded ZIP file): Run the Conda installation script: Run the application: On Windows Clone the repository (or use downloaded ZIP file): Run the Conda installation script: Run the application: One-Click Deployment Running with Docker Compose To quickly set up and run CrewAI-Studio using Docker Compose, follow these steps: Prerequisites Ensure Docker and Docker Compose are installed on your system. Steps Clone the repository: Create a .env file for configuration. Edit for your own configuration: Start the application with Docker Compose: Access the application: http://localhost:8501 Configuration Before running the application, ensure you update the .env file with your API keys and other necessary configurations. An example .env file is provided for reference. Troubleshooting In case of problems: Delete the venv/miniconda folder and reinstall crewai-studio. Rename crewai.db (it contains your crews but sometimes new versions can break compatibility). Raise an issue and I will help you. Video tutorial Video tutorial on CrewAI Studio made by Josh Poco Star History

ai-hub-gateway-solution-accelerator
github
LLM Vibe Score0.562
Human Vibe Score0.14530291803566378
Azure-SamplesMar 28, 2025

ai-hub-gateway-solution-accelerator

AI Hub Gateway Landing Zone accelerator The AI Hub Gateway Landing Zone is a solution accelerator that provides a set of guidelines and best practices for implementing a central AI API gateway to empower various line-of-business units in an organization to leverage Azure AI services. !user-story User Story The AI Hub Gateway Landing Zone architecture designed to be a central hub for AI services, providing a single point of entry for AI services, and enabling the organization to manage and govern AI services in a consistent manner. !AI Hub Gateway Landing Zone Key features !ai-hub-gateway-benefits.png Recent release updates: About: here you can see the recent updates to the gateway implementation Now this solution accelerator is updated to be enterprise ready with the following features: Improved OpenAI Usage Ingestion with the ability to ingest usage data from Azure OpenAI API for both streaming and non-streaming requests. Check the guide here Bring your own VNet is now supported with the ability to deploy the AI Hub Gateway Landing Zone in your own VNet. Check the guide here Throttling events monitoring is now supported with the ability to capture and raise too many requests status code as a custom metric in Application Insights. Check the guide here New gpt-4o Global Deployment is now part of the OpenAI resource provisioning Azure OpenAI API spec version was updated to to bring APIs for audio and batch among other advancements (note it is backward compatible with previous versions) AI usage reports enhancements with Cosmos Db now include a container for which include the $ pricing for AI models tokens (sample data can be found here), along with updated PowerBI dashboard design. Private connectivity now can be enabled by setting APIM deployment to External or Internal (require SKU to be either Developer or Premium) and it will provision all included Azure resources like (Azure OpenAI, Cosmos, Event Hub,...) with private endpoints. The AI Hub Gateway Landing Zone provides the following features: Centralized AI API Gateway: A central hub for AI services, providing a single point of entry for AI services that can be shared among multiple use-cases in a secure and governed approach. Seamless integration with Azure AI services: Ability to just update endpoints and keys in existing apps to switch to use AI Hub Gateway. AI routing and orchestration: The AI Hub Gateway Landing Zone provides a mechanism to route and orchestrate AI services, based on priority and target model enabling the organization to manage and govern AI services in a consistent manner. Granular access control: The AI Hub Gateway Landing Zone does not use master keys to access AI services, instead, it uses managed identities to access AI services while consumers can use gateway keys. Private connectivity: The AI Hub Gateway Landing Zone is designed to be deployed in a private network, and it uses private endpoints to access AI services. Capacity management: The AI Hub Gateway Landing Zone provides a mechanism to manage capacity based on requests and tokens. Usage & charge-back: The AI Hub Gateway Landing Zone provides a mechanism to track usage and charge-back to the respective business units with flexible integration with existing charge-back & data platforms. Resilient and scalable: The AI Hub Gateway Landing Zone is designed to be resilient and scalable, and it uses Azure API Management with its zonal redundancy and regional gateways which provides a scalable and resilient solution. Full observability: The AI Hub Gateway Landing Zone provides full observability with Azure Monitor, Application Insights, and Log Analytics with detailed insights into performance, usage, and errors. Hybrid support: The AI Hub Gateway Landing Zone approach the deployment of backends and gateway on Azure, on-premises or other clouds. !one-click-deploy One-click deploy This solution accelerator provides a one-click deploy option to deploy the AI Hub Gateway Landing Zone in your Azure subscription through Azure Developer CLI (azd) or Bicep (IaC). What is being deployed? !Azure components The one-click deploy option will deploy the following components in your Azure subscription: Azure API Management: Azure API Management is a fully managed service that powers most of the GenAI gateway capabilities. Application Insights: Application Insights is an extensible Application Performance Management (APM) service that will provides critical insights on the gateway operational performance. It will also include a dashboard for the key metrics. Event Hub: Event Hub is a fully managed, real-time data ingestion service that’s simple, trusted, and scalable and it is used to stream usage and charge-back data to target data and charge back platforms. Azure OpenAI: 3 instances of Azure OpenAI across 3 regions. Azure OpenAI is a cloud deployment of cutting edge generative models from OpenAI (like ChatGPT, DALL.E and more). Cosmos DB: Azure Cosmos DB is a fully managed NoSQL database for storing usage and charge-back data. Azure Function App: to support real-time event processing service that will be used to process the usage and charge-back data from Event Hub and push it to Cosmos DB. User Managed Identity: A user managed identity to be used by the Azure API Management to access the Azure OpenAI services/Event Hub and another for Azure Stream Analytics to access Event Hub and Cosmos DB. Virtual Network: A virtual network to host the Azure API Management and the other Azure resources. Private Endpoints & Private DNS Zones: Private endpoints for Azure OpenAI, Cosmos DB, Azure Function, Azure Monitor and Event Hub to enable private connectivity. Prerequisites In order to deploy and run this solution accelerator, you'll need Azure Account - If you're new to Azure, get an Azure account for free and you'll get some free Azure credits to get started. Azure subscription with access enabled for the Azure OpenAI service - You can request access. You can also visit the Cognitive Search docs to get some free Azure credits to get you started. Azure account permissions - Your Azure Account must have Microsoft.Authorization/roleAssignments/write permissions, such as User Access Administrator or Owner. For local development, you'll need: Azure CLI - The Azure CLI is a command-line tool that provides a great experience for managing Azure resources. You can install the Azure CLI on your local machine by following the instructions here. Azure Developer CLI (azd) - The Azure Developer CLI is a command-line tool that provides a great experience for deploying Azure resources. You can install the Azure Developer CLI on your local machine by following the instructions here VS Code - Visual Studio Code is a lightweight but powerful source code editor which runs on your desktop and is available for Windows, macOS, and Linux. You can install Visual Studio Code on your local machine by following the instructions here How to deploy? It is recommended to check first the main.bicep file that includes the deployment configuration and parameters. Make sure you have enough OpenAI capacity for gpt-35-turbo and embedding in the selected regions. Currently these are the default values: When you are happy with the configuration, you can deploy the solution using the following command: NOTE: If you faced any deployment errors, try to rerun the command as you might be facing a transient error. After that, you can start using the AI Hub Gateway Landing Zone through the Azure API Management on Azure Portal: !apim-test NOTE: You can use Azure Cloud Shell to run the above command, just clone this repository and run the command from the repo root folder. !docs Supporting documents To dive deeper into the AI Hub Gateway technical mechanics, you can check out the following guides: Architecture guides Architecture deep dive Deployment components API Management configuration OpenAI Usage Ingestion Bring your own Network Onboarding guides OpenAI Onboarding AI Search Onboarding Power BI Dashboard Throttling Events Alerts AI Studio Integration Additional guides End-to-end scenario (Chat with data) Hybrid deployment of AI Hub Gateway Deployment troubleshooting

awesome-ai-in-finance
github
LLM Vibe Score0.58
Human Vibe Score1
georgezouqMar 28, 2025

awesome-ai-in-finance

Awesome AI in Finance There are millions of trades made in the global financial market every day. Data grows very quickly and people are hard to understand. With the power of the latest artificial intelligence research, people analyze & trade automatically and intelligently. This list contains the research, tools and code that people use to beat the market. [中文资源] Contents LLMs Papers Courses & Books Strategies & Research Time Series Data Portfolio Management High Frequency Trading Event Drive Crypto Currencies Strategies Technical Analysis Lottery & Gamble Arbitrage Data Sources Research Tools Trading System TA Lib Exchange API Articles Others LLMs 🌟🌟 MarS - A Financial Market Simulation Engine Powered by Generative Foundation Model. 🌟🌟 Financial Statement Analysis with Large Language Models - GPT-4 can outperform professional financial analysts in predicting future earnings changes, generating useful narrative insights, and resulting in superior trading strategies with higher Sharpe ratios and alphas, thereby suggesting a potential central role for LLMs in financial decision-making. PIXIU - An open-source resource providing a financial large language model, a dataset with 136K instruction samples, and a comprehensive evaluation benchmark. FinGPT - Provides a playground for all people interested in LLMs and NLP in Finance. MACD + RSI + ADX Strategy (ChatGPT-powered) by TradeSmart - Asked ChatGPT on which indicators are the most popular for trading. We used all of the recommendations given. A ChatGPT trading algorithm delivered 500% returns in stock market. My breakdown on what this means for hedge funds and retail investors Use chatgpt to adjust strategy parameters Hands-on LLMs: Train and Deploy a Real-time Financial Advisor - Train and deploy a real-time financial advisor chatbot with Falcon 7B and CometLLM. ChatGPT Strategy by OctoBot - Use ChatGPT to determine which cryptocurrency to trade based on technical indicators. Papers The Theory of Speculation L. Bachelier, 1900 - The influences which determine the movements of the Stock Exchange are. Brownian Motion in the Stock Market Osborne, 1959 - The common-stock prices can be regarded as an ensemble of decisions in statistical equilibrium. An Investigation into the Use of Reinforcement Learning Techniques within the Algorithmic Trading Domain, 2015 A Deep Reinforcement Learning Framework for the Financial Portfolio Management Problem Reinforcement Learning for Trading, 1994 Dragon-Kings, Black Swans and the Prediction of Crises Didier Sornette - The power laws in the distributions of event sizes under a broad range of conditions in a large variety of systems. Financial Trading as a Game: A Deep Reinforcement Learning Approach - Deep reinforcement learning provides a framework toward end-to-end training of such trading agent. Machine Learning for Trading - With an appropriate choice of the reward function, reinforcement learning techniques can successfully handle the risk-averse case. Ten Financial Applications of Machine Learning, 2018 - Slides review few important financial ML applications. FinRL: A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance, 2020 - Introduce a DRL library FinRL that facilitates beginners to expose themselves to quantitative finance and to develop their own stock trading strategies. Deep Reinforcement Learning for Automated Stock Trading: An Ensemble Strategy, 2020 - Propose an ensemble strategy that employs deep reinforcement schemes to learn a stock trading strategy by maximizing investment return. Courses & Books & Blogs 🌟 QuantResearch - Quantitative analysis, strategies and backtests https://letianzj.github.io/ NYU: Overview of Advanced Methods of Reinforcement Learning in Finance Udacity: Artificial Intelligence for Trading AI in Finance - Learn Fintech Online. Advanced-Deep-Trading - Experiments based on "Advances in financial machine learning" book. Advances in Financial Machine Learning - Using advanced ML solutions to overcome real-world investment problems. Build Financial Software with Generative AI - Book about how to build financial software hands-on using generative AI tools like ChatGPT and Copilot. Mastering Python for Finance - Sources codes for: Mastering Python for Finance, Second Edition. MLSys-NYU-2022 - Slides, scripts and materials for the Machine Learning in Finance course at NYU Tandon, 2022. Train and Deploy a Serverless API to predict crypto prices - In this tutorial you won't build an ML system that will make you rich. But you will master the MLOps frameworks and tools you need to build ML systems that, together with tons of experimentation, can take you there. Strategies & Research Time Series Data Price and Volume process with Technology Analysis Indices 🌟🌟 stockpredictionai - A complete process for predicting stock price movements. 🌟 Personae - Implements and environment of Deep Reinforcement Learning & Supervised Learning for Quantitative Trading. 🌟 Ensemble-Strategy - Deep Reinforcement Learning for Automated Stock Trading. FinRL - A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance. AutomatedStockTrading-DeepQ-Learning - Build a Deep Q-learning reinforcement agent model as automated trading robot. tfdeeprltrader - Trading environment(OpenAI Gym) + PPO(TensorForce). trading-gym - Trading agent to train with episode of short term trading itself. trading-rl - Deep Reinforcement Learning for Financial Trading using Price Trailing. deeprltrader - Trading environment(OpenAI Gym) + DDQN (Keras-RL). Quantitative-Trading - Papers and code implementing Quantitative-Trading. gym-trading - Environment for reinforcement-learning algorithmic trading models. zenbrain - A framework for machine-learning bots. DeepLearningNotes - Machine learning in quant analysis. stockmarketreinforcementlearning - Stock market trading OpenAI Gym environment with Deep Reinforcement Learning using Keras. Chaos Genius - ML powered analytics engine for outlier/anomaly detection and root cause analysis.. mlforecast - Scalable machine learning based time series forecasting. Portfolio Management Deep-Reinforcement-Stock-Trading - A light-weight deep reinforcement learning framework for portfolio management. qtrader - Reinforcement Learning for portfolio management. PGPortfolio - A Deep Reinforcement Learning framework for the financial portfolio management problem. DeepDow - Portfolio optimization with deep learning. skfolio - Python library for portfolio optimization built on top of scikit-learn. High Frequency Trading High-Frequency-Trading-Model-with-IB - A high-frequency trading model using Interactive Brokers API with pairs and mean-reversion. 🌟 SGX-Full-OrderBook-Tick-Data-Trading-Strategy - Solutions for high-frequency trading (HFT) strategies using data science approaches (Machine Learning) on Full Orderbook Tick Data. HFTBitcoin - Analysis of High Frequency Trading on Bitcoin exchanges. Event Drive 🌟🌟 stockpredictionai - Complete process for predicting stock price movements. 🌟 trump2cash - A stock trading bot powered by Trump tweets. Crypto Currencies Strategies LSTM-Crypto-Price-Prediction - Predicting price trends in crypto markets using an LSTM-RNN for trading. tforcebtctrader - TensorForce Bitcoin trading bot. Tensorflow-NeuroEvolution-Trading-Bot - A population model that trade cyrpto and breed and mutate iteratively. gekkoga - Genetic algorithm for solving optimization of trading strategies using Gekko. GekkoANNStrategies - ANN trading strategies for the Gekko trading bot. gekko-neuralnet - Neural network strategy for Gekko. bitcoinprediction - Code for "Bitcoin Prediction" by Siraj Raval on YouTube. Technical Analysis quant-trading - Python quantitative trading strategies. Gekko-Bot-Resources - Gekko bot resources. gekkotools - Gekko strategies, tools etc. gekko RSIWR - Gekko RSIWR strategies. gekko HL - Calculate down peak and trade on. EthTradingAlgorithm - Ethereum trading algorithm using Python 3.5 and the library ZipLine. gekkotradingstuff - Awesome crypto currency trading platform. forex.analytics - Node.js native library performing technical analysis over an OHLC dataset with use of genetic algorithmv. BitcoinMACDStrategy - Bitcoin MACD crossover trading strategy backtest. crypto-signal - Automated crypto trading & technical analysis (TA) bot for Bittrex, Binance, GDAX, and more. Gekko-Strategies - Strategies to Gekko trading bot with backtests results and some useful tools. gekko-gannswing - Gann's Swing trade strategy for Gekko trade bot. Lottery & Gamble LotteryPredict - Use LSTM to predict lottery. Arbitrage ArbitrageBot - Arbitrage bot that currently works on bittrex & poloniex. r2 - Automatic arbitrage trading system powered by Node.js + TypeScript. cryptocurrency-arbitrage - A crypto currency arbitrage opportunity calculator. Over 800 currencies and 50 markets. bitcoin-arbitrage - Bitcoin arbitrage opportunity detector. blackbird - Long / short market-neutral strategy. Data Sources Traditional Markets 🌟 Quandl - Get millions of financial and economic dataset from hundreds of publishers via a single free API. yahoo-finance - Python module to get stock data from Yahoo! Finance. Tushare - Crawling historical data of Chinese stocks. Financial Data - Stock Market and Financial Data API. Crypto Currencies CryptoInscriber - A live crypto currency historical trade data blotter. Download live historical trade data from any crypto exchange. Gekko-Datasets - Gekko trading bot dataset dumps. Download and use history files in SQLite format. Research Tools Synthical - AI-powered collaborative environment for Research. 🌟🌟 TensorTrade - Trade efficiently with reinforcement learning. ML-Quant - Quant resources from ArXiv (sanity), SSRN, RePec, Journals, Podcasts, Videos, and Blogs. JAQS - An open source quant strategies research platform. pyfolio - Portfolio and risk analytics in Python. alphalens - Performance analysis of predictive (alpha) stock factors. empyrical - Common financial risk and performance metrics. Used by Zipline and pyfolio. zvt - Zero vector trader. Trading System For Back Test & Live trading Traditional Market System 🌟🌟🌟 OpenBB - AI-powered opensource research and analytics workspace. 🌟🌟 zipline - A python algorithmic trading library. 🌟 TradingView - Get real-time information and market insights. rqalpha - A extendable, replaceable Python algorithmic backtest & trading framework. backtrader - Python backtesting library for trading strategies. kungfu - Kungfu Master trading system. lean - Algorithmic trading engine built for easy strategy research, backtesting and live trading. Combine & Rebuild pylivetrader - Python live trade execution library with zipline interface. CoinMarketCapBacktesting - As backtest frameworks for coin trading strategy. Crypto Currencies zenbot - Command-line crypto currency trading bot using Node.js and MongoDB. bot18 - High-frequency crypto currency trading bot developed by Zenbot. magic8bot - Crypto currency trading bot using Node.js and MongoDB. catalyst - An algorithmic trading library for Crypto-Assets in python. QuantResearchDev - Quant Research dev & Traders open source project. MACD - Zenbot MACD Auto-Trader. abu - A quant trading system base on python. Plugins CoinMarketCapBacktesting - Tests bt and Quantopian Zipline as backtesting frameworks for coin trading strategy. Gekko-BacktestTool - Batch backtest, import and strategy params optimalization for Gekko Trading Bot. TA Lib pandastalib - A Python Pandas implementation of technical analysis indicators. finta - Common financial technical indicators implemented in Python-Pandas (70+ indicators). tulipnode - Official Node.js wrapper for Tulip Indicators. Provides over 100 technical analysis overlay and indicator functions. techan.js - A visual, technical analysis and charting (Candlestick, OHLC, indicators) library built on D3. Exchange API Do it in real world! IbPy - Python API for the Interactive Brokers on-line trading system. HuobiFeeder - Connect HUOBIPRO exchange, get market/historical data for ABAT trading platform backtest analysis and live trading. ctpwrapper - Shanghai future exchange CTP api. PENDAX - Javascript SDK for Trading/Data API and Websockets for cryptocurrency exchanges like FTX, FTXUS, OKX, Bybit, & More Framework tf-quant-finance - High-performance TensorFlow library for quantitative finance. Visualizing playground - Play with neural networks. netron - Visualizer for deep learning and machine learning models. KLineChart - Highly customizable professional lightweight financial charts GYM Environment 🌟 TradingGym - Trading and Backtesting environment for training reinforcement learning agent. TradzQAI - Trading environment for RL agents, backtesting and training. btgym - Scalable, event-driven, deep-learning-friendly backtesting library. Articles The-Economist - The Economist. nyu-mlif-notes - NYU machine learning in finance notes. Using LSTMs to Turn Feelings Into Trades Others zipline-tensorboard - TensorBoard as a Zipline dashboard. gekko-quasar-ui - An UI port for gekko trading bot using Quasar framework. Floom AI gateway and marketplace for developers, enables streamlined integration and least volatile approach of AI features into products Other Resource 🌟🌟🌟 Stock-Prediction-Models - Stock-Prediction-Models, Gathers machine learning and deep learning models for Stock forecasting, included trading bots and simulations. 🌟🌟 Financial Machine Learning - A curated list of practical financial machine learning (FinML) tools and applications. This collection is primarily in Python. 🌟 Awesome-Quant-Machine-Learning-Trading - Quant / Algorithm trading resources with an emphasis on Machine Learning. awesome-quant - A curated list of insanely awesome libraries, packages and resources for Quants (Quantitative Finance). FinancePy - A Python Finance Library that focuses on the pricing and risk-management of Financial Derivatives, including fixed-income, equity, FX and credit derivatives. Explore Finance Service Libraries & Projects - Explore a curated list of Fintech popular & new libraries, top authors, trending project kits, discussions, tutorials & learning resources on kandi.

BERT-pytorch
github
LLM Vibe Score0.514
Human Vibe Score0.16971233963995486
codertimoMar 28, 2025

BERT-pytorch

BERT-pytorch !GitHub issues Pytorch implementation of Google AI's 2018 BERT, with simple annotation BERT 2018 BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding Paper URL : https://arxiv.org/abs/1810.04805 Introduction Google AI's BERT paper shows the amazing result on various NLP task (new 17 NLP tasks SOTA), including outperform the human F1 score on SQuAD v1.1 QA task. This paper proved that Transformer(self-attention) based encoder can be powerfully used as alternative of previous language model with proper language model training method. And more importantly, they showed us that this pre-trained language model can be transfer into any NLP task without making task specific model architecture. This amazing result would be record in NLP history, and I expect many further papers about BERT will be published very soon. This repo is implementation of BERT. Code is very simple and easy to understand fastly. Some of these codes are based on The Annotated Transformer Currently this project is working on progress. And the code is not verified yet. Installation Quickstart NOTICE : Your corpus should be prepared with two sentences in one line with tab(\t) separator Prepare your corpus or tokenized corpus (tokenization is not in package) Building vocab based on your corpus Train your own BERT model Language Model Pre-training In the paper, authors shows the new language model training methods, which are "masked language model" and "predict next sentence". Masked Language Model Original Paper : 3.3.1 Task #1: Masked LM Rules: Randomly 15% of input token will be changed into something, based on under sub-rules Randomly 80% of tokens, gonna be a [MASK] token Randomly 10% of tokens, gonna be a [RANDOM] token(another word) Randomly 10% of tokens, will be remain as same. But need to be predicted. Predict Next Sentence Original Paper : 3.3.2 Task #2: Next Sentence Prediction "Is this sentence can be continuously connected?" understanding the relationship, between two text sentences, which is not directly captured by language modeling Rules: Randomly 50% of next sentence, gonna be continuous sentence. Randomly 50% of next sentence, gonna be unrelated sentence. Author Junseong Kim, Scatter Lab (codertimo@gmail.com / junseong.kim@scatterlab.co.kr) License This project following Apache 2.0 License as written in LICENSE file Copyright 2018 Junseong Kim, Scatter Lab, respective BERT contributors Copyright (c) 2018 Alexander Rush : The Annotated Trasnformer

awesome-quantum-machine-learning
github
LLM Vibe Score0.64
Human Vibe Score1
krishnakumarsekarMar 27, 2025

awesome-quantum-machine-learning

Awesome Quantum Machine Learning A curated list of awesome quantum machine learning algorithms,study materials,libraries and software (by language). Table of Contents INTRODUCTION Why Quantum Machine Learning? BASICS What is Quantum Mechanics? What is Quantum Computing? What is Topological Quantum Computing? Quantum Computing vs Classical Computing QUANTUM COMPUTING Atom Structure Photon wave Electron Fluctuation or spin States SuperPosition SuperPosition specific for machine learning(Quantum Walks) Classical Bit Quantum Bit or Qubit or Qbit Basic Gates in Quantum Computing Quantum Diode Quantum Transistor Quantum Processor Quantum Registery QRAM Quantum Entanglement QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers Tensors Tensors Network Oracle Hadamard transform Hilbert Space eigenvalues and eigenvectors Schr¨odinger Operators Quantum lambda calculus Quantum Amplitute Phase Qubits Encode and Decode convert classical bit to qubit Quantum Dirac and Kets Quantum Complexity Arbitrary State Generation QUANTUM ALGORITHMS Quantum Fourier Transform Variational-Quantum-Eigensolver Grovers Algorithm Shor's algorithm Hamiltonian Oracle Model Bernstein-Vazirani Algorithm Simon’s Algorithm Deutsch-Jozsa Algorithm Gradient Descent Phase Estimation Haar Tansform Quantum Ridgelet Transform Quantum NP Problem QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour Quantum K-Means Quantum Fuzzy C-Means Quantum Support Vector Machine Quantum Genetic Algorithm Quantum Hidden Morkov Models Quantum state classification with Bayesian methods Quantum Ant Colony Optimization Quantum Cellular Automata Quantum Classification using Principle Component Analysis Quantum Inspired Evolutionary Algorithm Quantum Approximate Optimization Algorithm Quantum Elephant Herding Optimization Quantum-behaved Particle Swarm Optimization Quantum Annealing Expectation-Maximization QAUNTUM NEURAL NETWORK Quantum perceptrons Qurons Quantum Auto Encoder Quantum Annealing Photonic Implementation of Quantum Neural Network Quantum Feed Forward Neural Network Quantum Boltzman Neural Network Quantum Neural Net Weight Storage Quantum Upside Down Neural Net Quantum Hamiltonian Neural Net QANN QPN SAL Quantum Hamiltonian Learning Compressed Quantum Hamiltonian Learning QAUNTUM STATISTICAL DATA ANALYSIS Quantum Probability Theory Kolmogorovian Theory Quantum Measurement Problem Intuitionistic Logic Heyting Algebra Quantum Filtering Paradoxes Quantum Stochastic Process Double Negation Quantum Stochastic Calculus Hamiltonian Calculus Quantum Ito's Formula Quantum Stochastic Differential Equations(QSDE) Quantum Stochastic Integration Itō Integral Quasiprobability Distributions Quantum Wiener Processes Quantum Statistical Ensemble Quantum Density Operator or Density Matrix Gibbs Canonical Ensemble Quantum Mean Quantum Variance Envariance Polynomial Optimization Quadratic Unconstrained Binary Optimization Quantum Gradient Descent Quantum Based Newton's Method for Constrained Optimization Quantum Based Newton's Method for UnConstrained Optimization Quantum Ensemble Quantum Topology Quantum Topological Data Analysis Quantum Bayesian Hypothesis Quantum Statistical Decision Theory Quantum Minimax Theorem Quantum Hunt-Stein Theorem Quantum Locally Asymptotic Normality Quantum Ising Model Quantum Metropolis Sampling Quantum Monte Carlo Approximation Quantum Bootstrapping Quantum Bootstrap Aggregation Quantum Decision Tree Classifier Quantum Outlier Detection Cholesky-Decomposition for Quantum Chemistry Quantum Statistical Inference Asymptotic Quantum Statistical Inference Quantum Gaussian Mixture Modal Quantum t-design Quantum Central Limit Theorem Quantum Hypothesis Testing Quantum Chi-squared and Goodness of Fit Testing Quantum Estimation Theory Quantum Way of Linear Regression Asymptotic Properties of Quantum Outlier Detection in Quantum Concepts QAUNTUM ARTIFICIAL INTELLIGENCE Heuristic Quantum Mechanics Consistent Quantum Reasoning Quantum Reinforcement Learning QAUNTUM COMPUTER VISION QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES ALL QUANTUM ALGORITHMS SOURCE CODES , GITHUBS QUANTUM HOT TOPICS Quantum Cognition Quantum Camera Quantum Mathematics Quantum Information Processing Quantum Image Processing Quantum Cryptography Quantum Elastic Search Quantum DNA Computing Adiabetic Quantum Computing Topological Big Data Anlytics using Quantum Hamiltonian Time Based Quantum Computing Deep Quantum Learning Quantum Tunneling Quantum Entanglment Quantum Eigen Spectrum Quantum Dots Quantum elctro dynamics Quantum teleportation Quantum Supremacy Quantum Zeno Effect Quantum Cohomology Quantum Chromodynamics Quantum Darwinism Quantum Coherence Quantum Decoherence Topological Quantum Computing Topological Quantum Field Theory Quantum Knots Topological Entanglment Boson Sampling Quantum Convolutional Code Stabilizer Code Quantum Chaos Quantum Game Theory Quantum Channel Tensor Space Theory Quantum Leap Quantum Mechanics for Time Travel Quantum Secured Block Chain Quantum Internet Quantum Optical Network Quantum Interference Quantum Optical Network Quantum Operating System Electron Fractionalization Flip-Flop Quantum Computer Quantum Information with Gaussian States Quantum Anomaly Detection Distributed Secure Quantum Machine Learning Decentralized Quantum Machine Learning Artificial Agents for Quantum Designs Light Based Quantum Chips for AI Training QUANTUM STATE PREPARATION ALGORITHM FOR MACHINE LEARNING Pure Quantum State Product State Matrix Product State Greenberger–Horne–Zeilinger State W state AKLT model Majumdar–Ghosh Model Multistate Landau–Zener Models Projected entangled-pair States Infinite Projected entangled-pair States Corner Transfer Matrix Method Tensor-entanglement Renormalization Tree Tensor Network for Supervised Learning QUANTUM MACHINE LEARNING VS DEEP LEARNING QUANTUM MEETUPS QUANTUM GOOGLE GROUPS QUANTUM BASED COMPANIES QUANTUM LINKEDLIN QUANTUM BASED DEGREES CONSOLIDATED QUANTUM ML BOOKS CONSOLIDATED QUANTUM ML VIDEOS CONSOLIDATED QUANTUM ML Reserach Papers CONSOLIDATED QUANTUM ML Reserach Scientist RECENT QUANTUM UPDATES FORUM ,PAGES AND NEWSLETTER INTRODUCTION Why Quantum Machine Learning? Machine Learning(ML) is just a term in recent days but the work effort start from 18th century. What is Machine Learning ? , In Simple word the answer is making the computer or application to learn themselves . So its totally related with computing fields like computer science and IT ? ,The answer is not true . ML is a common platform which is mingled in all the aspects of the life from agriculture to mechanics . Computing is a key component to use ML easily and effectively . To be more clear ,Who is the mother of ML ?, As no option Mathematics is the mother of ML . The world tremendous invention complex numbers given birth to this field . Applying mathematics to the real life problem always gives a solution . From Neural Network to the complex DNA is running under some specific mathematical formulas and theorems. As computing technology growing faster and faster mathematics entered into this field and makes the solution via computing to the real world . In the computing technology timeline once a certain achievements reached peoples interested to use advanced mathematical ideas such as complex numbers ,eigen etc and its the kick start for the ML field such as Artificial Neural Network ,DNA Computing etc. Now the main question, why this field is getting boomed now a days ? , From the business perspective , 8-10 Years before during the kick start time for ML ,the big barrier is to merge mathematics into computing field . people knows well in computing has no idea on mathematics and research mathematician has no idea on what is computing . The education as well as the Job Opportunities is like that in that time . Even if a person tried to study both then the business value for making a product be not good. Then the top product companies like Google ,IBM ,Microsoft decided to form a team with mathematician ,a physician and a computer science person to come up with various ideas in this field . Success of this team made some wonderful products and they started by providing cloud services using this product . Now we are in this stage. So what's next ? , As mathematics reached the level of time travel concepts but the computing is still running under classical mechanics . the companies understood, the computing field must have a change from classical to quantum, and they started working on the big Quantum computing field, and the market named this field as Quantum Information Science .The kick start is from Google and IBM with the Quantum Computing processor (D-Wave) for making Quantum Neural Network .The field of Quantum Computer Science and Quantum Information Science will do a big change in AI in the next 10 years. Waiting to see that........... .(google, ibm). References D-Wave - Owner of a quantum processor Google - Quantum AI Lab IBM - Quantum Computer Lab Quora - Question Regarding future of quantum AI NASA - NASA Quantum Works Youtube - Google Video of a Quantum Processor external-link - MIT Review microsoft new product - Newly Launched Microsoft Quantum Language and Development Kit microsoft - Microsoft Quantum Related Works Google2 - Google Quantum Machine Learning Blog BBC - About Google Quantum Supremacy,IBM Quantum Computer and Microsoft Q Google Quantum Supremacy - Latest 2019 Google Quantum Supremacy Achievement IBM Quantum Supremacy - IBM Talk on Quantum Supremacy as a Primer VICE on the fight - IBM Message on Google Quantum Supremacy IBM Zurich Quantum Safe Cryptography - An interesting startup to replace all our Certificate Authority Via Cloud and IBM Q BASICS What is Quantum Mechanics? In a single line study of an electron moved out of the atom then its classical mechanic ,vibrates inside the atom its quantum mechanics WIKIPEDIA - Basic History and outline LIVESCIENCE. - A survey YOUTUBE - Simple Animation Video Explanining Great. What is Quantum Computing? A way of parallel execution of multiple processess in a same time using qubit ,It reduces the computation time and size of the processor probably in neuro size WIKIPEDIA - Basic History and outline WEBOPEDIA. - A survey YOUTUBE - Simple Animation Video Explanining Great. Quantum Computing vs Classical Computing LINK - Basic outline Quantum Computing Atom Structure one line : Electron Orbiting around the nucleous in an eliptical format YOUTUBE - A nice animation video about the basic atom structure Photon Wave one line : Light nornmally called as wave transmitted as photons as similar as atoms in solid particles YOUTUBE - A nice animation video about the basic photon 1 YOUTUBE - A nice animation video about the basic photon 2 Electron Fluctuation or spin one line : When a laser light collide with solid particles the electrons of the atom will get spin between the orbitary layers of the atom ) YOUTUBE - A nice animation video about the basic Electron Spin 1 YOUTUBE - A nice animation video about the basic Electron Spin 2 YOUTUBE - A nice animation video about the basic Electron Spin 3 States one line : Put a point on the spinning electron ,if the point is in the top then state 1 and its in bottom state 0 YOUTUBE - A nice animation video about the Quantum States SuperPosition two line : During the spin of the electron the point may be in the middle of upper and lower position, So an effective decision needs to take on the point location either 0 or 1 . Better option to analyse it along with other electrons using probability and is called superposition YOUTUBE - A nice animation video about the Quantum Superposition SuperPosition specific for machine learning(Quantum Walks) one line : As due to computational complexity ,quantum computing only consider superposition between limited electrons ,In case to merge more than one set quantum walk be the idea YOUTUBE - A nice video about the Quantum Walks Classical Bits one line : If electron moved from one one atom to other ,from ground state to excited state a bit value 1 is used else bit value 0 used Qubit one line : The superposition value of states of a set of electrons is Qubit YOUTUBE - A nice video about the Quantum Bits 1 YOUTUBE - A nice video about the Bits and Qubits 2 Basic Gates in Quantum Computing one line : As like NOT, OR and AND , Basic Gates like NOT, Hadamard gate , SWAP, Phase shift etc can be made with quantum gates YOUTUBE - A nice video about the Quantum Gates Quantum Diode one line : Quantum Diodes using a different idea from normal diode, A bunch of laser photons trigger the electron to spin and the quantum magnetic flux will capture the information YOUTUBE - A nice video about the Quantum Diode Quantum Transistors one line : A transistor default have Source ,drain and gate ,Here source is photon wave ,drain is flux and gate is classical to quantum bits QUORA -Discussion about the Quantum Transistor YOUTUBE - Well Explained Quantum Processor one line : A nano integration circuit performing the quantum gates operation sorrounded by cooling units to reduce the tremendous amount of heat YOUTUBE - Well Explained Quantum Registery QRAM one line : Comapring the normal ram ,its ultrafast and very small in size ,the address location can be access using qubits superposition value ,for a very large memory set coherent superposition(address of address) be used PDF - very Well Explained QUANTUM COMPUTING MACHINE LEARNING BRIDGE Complex Numbers one line : Normally Waves Interference is in n dimensional structure , to find a polynomial equation n order curves ,better option is complex number YOUTUBE - Wonderful Series very super Explained Tensors one line : Vectors have a direction in 2D vector space ,If on a n dimensional vector space ,vectors direction can be specify with the tensor ,The best solution to find the superposition of a n vector electrons spin space is representing vectors as tensors and doing tensor calculus YOUTUBE - Wonderful super Explained tensors basics YOUTUBE - Quantum tensors basics Tensors Network one line : As like connecting multiple vectors ,multple tensors form a network ,solving such a network reduce the complexity of processing qubits YOUTUBE - Tensors Network Some ideas specifically for quantum algorithms QUANTUM MACHINE LEARNING ALGORITHMS Quantum K-Nearest Neighbour info : Here the centroid(euclidean distance) can be detected using the swap gates test between two states of the qubit , As KNN is regerssive loss can be tally using the average PDF1 from Microsoft - Theory Explanation PDF2 - A Good Material to understand the basics Matlab - Yet to come soon Python - Yet to come soon Quantum K-Means info : Two Approaches possible ,1. FFT and iFFT to make an oracle and calculate the means of superposition 2. Adiobtic Hamiltonian generation and solve the hamiltonian to determine the cluster PDF1 - Applying Quantum Kmeans on Images in a nice way PDF2 - Theory PDF3 - Explaining well the K-means clustering using hamiltonian Matlab - Yet to come soon Python - Yet to come soon Quantum Fuzzy C-Means info : As similar to kmeans fcm also using the oracle dialect ,but instead of means,here oracle optimization followed by a rotation gate is giving a good result PDF1 - Theory Matlab - Yet to come soon Python - Yet to come soon Quantum Support Vector Machine info : A little different from above as here kernel preparation is via classical and the whole training be in oracles and oracle will do the classification, As SVM is linear ,An optimal Error(Optimum of the Least Squares Dual Formulation) Based regression is needed to improve the performance PDF1 - Nice Explanation but little hard to understand :) PDF2 - Nice Application of QSVM Matlab - Yet to come soon Python - Yet to come soon Quantum Genetic Algorithm info : One of the best algorithm suited for Quantum Field ,Here the chromosomes act as qubit vectors ,the crossover part carrying by an evaluation and the mutation part carrying by the rotation of gates ![Flow Chart]() PDF1 - Very Beautiful Article , well explained and superp PDF2 - A big theory :) PDF3 - Super Comparison Matlab - Simulation Python1 - Simulation Python2 - Yet to come Quantum Hidden Morkov Models info : As HMM is already state based ,Here the quantum states acts as normal for the markov chain and the shift between states is using quantum operation based on probability distribution ![Flow Chart]() PDF1 - Nice idea and explanation PDF2 - Nice but a different concept little Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum state classification with Bayesian methods info : Quantum Bayesian Network having the same states concept using quantum states,But here the states classification to make the training data as reusable is based on the density of the states(Interference) ![Bayesian Network Sample1]() ![Bayesian Network Sample2]() ![Bayesian Network Sample3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Ant Colony Optimization info : A good algorithm to process multi dimensional equations, ACO is best suited for Sales man issue , QACO is best suited for Sales man in three or more dimension, Here the quantum rotation circuit is doing the peromene update and qubits based colony communicating all around the colony in complex space ![Ant Colony Optimization 1]() PDF1 - Good Concept PDF2 - Good Application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come Quantum Cellular Automata info : One of the very complex algorithm with various types specifically used for polynomial equations and to design the optimistic gates for a problem, Here the lattice is formed using the quatum states and time calculation is based on the change of the state between two qubits ,Best suited for nano electronics ![Quantum Cellular Automata]() Wikipedia - Basic PDF1 - Just to get the keywords PDF2 - Nice Explanation and an easily understandable application Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM NEURAL NETWORK one line : Its really one of the hardest topic , To understand easily ,Normal Neural Network is doing parallel procss ,QNN is doing parallel of parallel processess ,In theory combination of various activation functions is possible in QNN ,In Normal NN more than one activation function reduce the performance and increase the complexity Quantum perceptrons info : Perceptron(layer) is the basic unit in Neural Network ,The quantum version of perceptron must satisfy both linear and non linear problems , Quantum Concepts is combination of linear(calculus of superposition) and nonlinear(State approximation using probability) ,To make a perceptron in quantum world ,Transformation(activation function) of non linearity to certain limit is needed ,which is carrying by phase estimation algorithm ![Quantum Perceptron 3]() PDF1 - Good Theory PDF2 - Good Explanation Matlab - Yet to come Python1 - Yet to come Python2 - Yet to come QAUNTUM STATISTICAL DATA ANALYSIS one line : An under research concept ,It can be seen in multiple ways, one best way if you want to apply n derivative for a problem in current classical theory its difficult to compute as its serialization problem instead if you do parallelization of differentiation you must estimate via probability the value in all flows ,Quantum Probability Helps to achieve this ,as the loss calculation is very less . the other way comparatively booming is Quantum Bayesianism, its a solution to solve most of the uncertainity problem in statistics to combine time and space in highly advanced physical research QUANTUM PROGRAMMING LANGUAGES , TOOLs and SOFTWARES All info : All Programming languages ,softwares and tools in alphabetical order Software - Nice content of all Python library - A python library Matlab based python library - Matlab Python Library Quantum Tensor Network Github - Tensor Network Bayesforge - A Beautiful Amazon Web Service Enabled Framework for Quantum Alogorithms and Data Analytics Rigetti - A best tools repository to use quantum computer in real time Rigetti Forest - An API to connect Quantum Computer quil/pyQuil - A quantum instruction language to use forest framework Grove - Grove is a repository to showcase quantum Fourier transform, phase estimation, the quantum approximate optimization algorithm, and others developed using Forest QISKit - A IBM Kit to access quantum computer and mainly for quantum circuits IBM Bluemix Simulator - A Bluemix Simulator for Quantum Circuits Microsoft Quantum Development Kit - Microsoft Visual Studio Enbaled Kit for Quantum Circuit Creation Microsoft "Q#" - Microsoft Q Sharp a new Programming Language for Quantum Circuit Creation qiskit api python - An API to connect IBM Quantum Computer ,With the generated token its easy to connect ,but very limited utils ,Lot of new utils will come soon Cyclops Tensor Framework - A framework to do tensor network simulations Python ToolKit for chemistry and physics Quantum Algorithm simulations - A New Started Project for simulating molecule and solids Bayesian Based Quatum Projects Repository - A nice repository and the kickstarter of bayesforge Google Fermion Products - A newly launched product specifivally for chemistry simulation Tree Tensor Networks - Interesting Tensor Network in Incubator Deep Tensor Neural Network - Some useful information about Tensor Neural Network in Incubator Generative Tensorial Networks - A startup to apply machine learning via tensor network for drug discovery Google Bristlecone - A new Quantum Processor from Google , Aimed for Future Hardwares with full fledged AI support XANADU - A Light based Quantum Hardware(chips supports) and Software Company Started in Preparation Stage. Soon will be in market fathom computing - A new concept to train the ai in a processor using light and quantum based concepts. soon products will be launch Alibaba Quantum Computing Cloud Service - Cloud Service to access 11 Bit Quantum Computing Processor Atomistic Machine Learning Project - Seems something Interesting with Deep Tensor Network for Quantum Chemistry Applications circQ and Google Works - Google Top Efforts on Tools IBM Safe Cryptography on Cloud - IBM Started and Developing a Quantm Safe Cryptography to replace all our Certificate Authority via Cloud Google Tensor Network Open Source - Google Started the Most Scientist Preferred Way To Use a Quantum Computer Circuit. Tensor Flow Which Makes Easy to Design the Network and Will Leave the Work Effect Of Gates, Processor Preparation and also going to tell the beauty of Maths Google Tensor Network Github - Github Project of Google Tensor Network Quantum Tensorflow - Yet to come soon Quantum Spark - Yet to come soon Quatum Map Reduce - Yet to come soon Quantum Database - Yet to come soon Quantum Server - Yet to come soon Quantum Data Analytics - Yet to come soon QUANTUM HOT TOPICS Deep Quantum Learning why and what is deep learning? In one line , If you know deep learning you can get a good job :) ,Even a different platform undergraduated and graduated person done a master specialization in deep learning can work in this big sector :), Practically speaking machine learning (vector mathematics) , deep learning (vector space(Graphics) mathematics) and big data are the terms created by big companies to make a trend in the market ,but in science and research there is no word such that , Now a days if you ask a junior person working in this big companies ,what is deep learning ,you will get some reply as "doing linear regression with stochastic gradient for a unsupervised data using Convolutional Neural Network :)" ,They knows the words clearly and knows how to do programming using that on a bunch of "relative data" , If you ask them about the FCM , SVM and HMM etc algorithms ,they will simply say these are olden days algorithms , deep learning replaced all :), But actually they dont know from the birth to the till level and the effectiveness of algorithms and mathematics ,How many mathematical theorems in vector, spaces , tensors etc solved to find this "hiding the complexity technology", They did not played with real non relative data like medical images, astro images , geology images etc , finding a relation and features is really complex and looping over n number of images to do pattern matching is a giant work , Now a days the items mentioned as deep learning (= multiple hidden artifical neural network) is not suitable for that why quantum deep learning or deep quantum learning? In the mid of Artificial Neural Network Research people realised at the maximum extreme only certain mathematical operations possible to do with ANN and the aim of this ANN is to achieve parallel execution of many mathematical operations , In artificial Intelligence ,the world intelligence stands for mathematics ,how effective if a probem can be solvable is based on the mathematics logic applying on the problem , more the logic will give more performance(more intelligent), This goal open the gate for quantum artificial neural network, On applying the ideas behind the deep learning to quantum mechanics environment, its possible to apply complex mathematical equations to n number of non relational data to find more features and can improve the performance Quantum Machine Learning vs Deep Learning Its fun to discuss about this , In recent days most of the employees from Product Based Companies Like google,microsoft etc using the word deep learning ,What actually Deep Learning ? and is it a new inventions ? how to learn this ? Is it replacing machine learning ? these question come to the mind of junior research scholars and mid level employees The one answer to all questions is deep learning = parallel "for" loops ,No more than that ,Its an effective way of executing multiple tasks repeatly and to reduce the computation cost, But it introduce a big cap between mathematics and computerscience , How ? All classical algorithms based on serial processing ,Its depends on the feedback of the first loop ,On applying a serial classical algorithm in multiple clusters wont give a good result ,but some light weight parallel classical algorithms(Deep learning) doing the job in multiple clusters and its not suitable for complex problems, What is the solution for then? As in the title Quantum Machine Learning ,The advantage behind is deep learning is doing the batch processing simply on the data ,but quantum machine learning designed to do batch processing as per the algorithm The product companies realised this one and they started migrating to quantum machine learning and executing the classical algorithms on quantum concept gives better result than deep learning algorithms on classical computer and the target to merge both to give very wonderful result References Quora - Good Discussion Quora - The Bridge Discussion Pdf - Nice Discussion Google - Google Research Discussion Microsoft - Microsoft plan to merge both IBM - IBM plan to merge both IBM Project - IBM Project idea MIT and Google - Solutions for all questions QUANTUM MEETUPS Meetup 1 - Quantum Physics Meetup 2 - Quantum Computing London Meetup 3 - Quantum Computing New York Meetup 4 - Quantum Computing Canada Meetup 5 - Quantum Artificial Intelligence Texas Meetup 6 - Genarl Quantum Mechanics , Mathematics New York Meetup 7 - Quantum Computing Mountain View California Meetup 8 - Statistical Analysis New York Meetup 9 - Quantum Mechanics London UK Meetup 10 - Quantum Physics Sydney Australia Meetup 11 - Quantum Physics Berkeley CA Meetup 12 - Quantum Computing London UK Meetup 13 - Quantum Mechanics Carmichael CA Meetup 14 - Maths and Science Group Portland Meetup 15 - Quantum Physics Santa Monica, CA Meetup 16 - Quantum Mechanics London Meetup 17 - Quantum Computing London Meetup 18 - Quantum Meta Physics ,Kansas City , Missouri ,US Meetup 19 - Quantum Mechanics and Physics ,Boston ,Massachusetts ,US Meetup 20 - Quantum Physics and Mechanics ,San Francisco ,California Meetup 21 - Quantum Mechanics ,Langhorne, Pennsylvania Meetup 22 - Quantum Mechanics ,Portland QUANTUM BASED DEGREES Plenty of courses around the world and many Universities Launching it day by day ,Instead of covering only Quantum ML , Covering all Quantum Related topics gives more idea in the order below Available Courses Quantum Mechanics for Science and Engineers Online Standford university - Nice Preparatory Course edx - Quantum Mechanics for Everyone NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum mechanics NPTEL 2 NPTEL 3 NPTEL 4 NPTEL 5 Class Based Course UK Bristol Australia Australian National University Europe Maxs Planks University Quantum Physics Online MIT - Super Explanation and well basics NPTEL - Nice Series of Courses to understand basics and backbone of quantum Physics Class Based Course Europe University of Copenhagen Quantum Chemistry Online NPTEL 1 - Nice Series of Courses to understand basics and backbone of quantum Chemistry NPTEL 2 - Class Based Course Europe UGent Belgium Quantum Computing Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum Computing Class Based Course Canada uwaterloo Singapore National University Singapore USA Berkley China Baidu Quantum Technology Class Based Course Canada uwaterloo Singapore National University Singapore Europe Munich Russia Skoltech Quantum Information Science External Links quantwiki Online MIT - Super Explanation and well basics edx - Nice Explanation NPTEL - Nice Series of Courses to understand basics and backbone of quantum information and computing Class Based Course USA MIT Standford University Joint Center for Quantum Information and Computer Science - University of Maryland Canada Perimeter Institute Singapore National University Singapore Europe ULB Belgium IQOQI Quantum Electronics Online MIT - Wonderful Course NPTEL - Nice Series of Courses to understand basics and backbone of quantum Electronics Class Based Course USA Texas Europe Zurich ICFO Asia Tata Institute Quantum Field Theory Online Standford university - Nice Preparatory Course edx - Some QFT Concepts available Class Based Course UK Imperial Europe Vrije Quantum Computer Science Class Based Course USA Oxford Joint Center for Quantum Information and Computer Science - University of Maryland Quantum Artificial Intelligence and Machine Learning External Links Quora 1 Quora 1 Artificial Agents Research for Quantum Designs Quantum Mathematics Class Based Course USA University of Notre CONSOLIDATED Quantum Research Papers scirate - Plenty of Quantum Research Papers Available Peter Wittek - Famous Researcher for the Quantum Machine Leanrning , Published a book in this topic [Murphy Yuezhen Niu] (https://scholar.google.com/citations?user=0wJPxfkAAAAJ&hl=en) - A good researcher published some nice articles Recent Quantum Updates forum ,pages and newsletter Quantum-Tech - A Beautiful Newsletter Page Publishing Amazing Links facebook Quantum Machine Learning - Running By me . Not that much good :). You can get some ideas Linkedlin Quantum Machine Learning - A nice page running by experts. Can get plenty of ideas FOSDEM 2019 Quantum Talks - A one day talk in fosdem 2019 with more than 10 research topics,tools and ideas FOSDEM 2020 Quantum Talks - Live talk in fosdem 2020 with plenty new research topics,tools and ideas License Dedicated Opensources ![Dedicated Opensources]() Source code of plenty of Algortihms in Image Processing , Data Mining ,etc in Matlab, Python ,Java and VC++ Scripts Good Explanations of Plenty of algorithms with flow chart etc Comparison Matrix of plenty of algorithms Is Quantum Machine Learning Will Reveal the Secret Maths behind Astrology? Awesome Machine Learning and Deep Learning Mathematics is online Published Basic Presentation of the series Quantum Machine Learning Contribution If you think this page might helpful. Please help for World Education Charity or kids who wants to learn

eiten
github
LLM Vibe Score0.549
Human Vibe Score0.754375921646308
tradyticsMar 27, 2025

eiten

Eiten - Algorithmic Investing Strategies for Everyone Eiten is an open source toolkit by Tradytics that implements various statistical and algorithmic investing strategies such as Eigen Portfolios, Minimum Variance Portfolios, Maximum Sharpe Ratio Portfolios, and Genetic Algorithms based Portfolios. It allows you to build your own portfolios with your own set of stocks that can beat the market. The rigorous testing framework included in Eiten enables you to have confidence in your portfolios. If you are looking to discuss these tools in depth and talk about more tools that we are working on, please feel free to join our Discord channel where we have a bunch of more tools too. Files Description | Path | Description | :--- | :---------- | eiten | Main folder. | &boxur; figures | Figures for this github repositories. | &boxur; stocks | Folder to keep your stock lists that you want to use to create your portfolios. | &boxur; strategies | A bunch of strategies implemented in python. | backtester.py | Backtesting module that both backtests and forward tests all portfolios. | data_loader.py | Module for loading data from yahoo finance. | portfolio_manager.py | Main file that takes in a bunch of arguments and generates several portfolios for you. | simulator.py | Simulator that uses historical returns and monte carlo to simulate future prices for the portfolios. | strategy_manager.py | Manages the strategies implemented in the 'strategies' folder. Required Packages You will need to install the following package to train and test the models. Scikit-learn Numpy Tqdm Yfinance Pandas Scipy You can install all packages using the following command. Please note that the script was written using python3. Build your portfolios Let us see how we can use all the strategies given in the toolkit to build our portfolios. The first thing you need to do is modify the stocks.txt file in the stocks folder and add the stocks of your choice. It is recommended to keep the list small i.e anywhere between 5 to 50 stocks should be fine. We have already put a small stocks list containing a bunch of tech stocks like AAPL, MSFT, TSLA etc. Let us build our portfolios now. This is the main command that you need to run. This command will use last 5 years of daily data excluding the last 90 days and build several portfolios for you. Based on those portfolios, it will then test them on the out of sample data of 90 days and show you the performance of each portfolio. Finally, it will also compare the performance with your choice of market index which is QQQ here. Let's dive into each of the parameters in detail. istest: The value determined if the program is going to keep some separate data for future testing. When this is enabled, the value of futurebars should be larger than 5. future_bars: These are the bars that the tool will exclude during portfolio building and will forward test the portfolios on the excluded set. This is also called out of sample data. datagranularityminutes: How much granular data do you want to use to build your portfolios. For long term portfolios, you should use daily data but for short term, you can use hourly or minute level data. The possible values here are 3600, 60, 30, 15, 5, 1. 3600 means daily. historytouse: Whether to use a specific number of historical bars or use everything that we receive from yahoo finance. For minute level data, we only receive up to one month of historical data. For daily, we receive 5 years worth of historical data. If you want to use all available data, the value should be all but if you want to use smaller history, you can set it to an integer value e.g 100 which will only use the last 100 bars to build the portfolios. applynoisefiltering: This uses random matrix theory to filter out the covariance matrix from randomness thus yielding better portfolios. A value of 1 will enable it and 0 will disable it. market_index: Which index do you want to use to compare your portfolios. This should mostly be SPY but since we analyzed tech stocks, we used QQQ. only_long: Whether to use long only portfolio or enable short selling as well. Long only portfolios have shown to have better performance using algorithmic techniques. eigenportfolionumber: Which eigen portfolio to use. Any value between 1-5 should work. The first eigen portfolio (1) represents the market portfolio and should act just like the underlying index such as SPY or QQQ. The second one is orthogonal and uncorrelated to the market and poses the greatest risk and reward. The following ones have reduced risk and reward. Read more on eigen-portfolios. stocksfilepath: File that contains the list of stocks that you want to use to build your portfolio. Some Portfolio Building Examples Here are a few examples for building different types of portfolios. Both long and short portfolios by analyzing last 90 days data and keeping the last 30 days as testing data. This will give us 60 days of portfolio construction data and 30 days of testing. Only long portfolio on 60 minute bars of the last 30 days. No future testing. Compare the results with SPY index instead of QQQ. Do not apply noise filtering on the covariance matrix. Use the first eigen portfolio (market portfolio) and compare with SQQQ, Portfolio Strategies Four different portfolio strategies are currently supported by the toolkit. Eigen Portfolios These portfolios are orthogonal and uncorrelated to the market in general thus yielding high reward and alpha. However, since they are uncorrelated to the market, they can also provide great risk. The first eigen portfolio is considered to be a market portfolio which is often ignored. The second one is uncorrelated to the others and provides the highest risk and reward. As we go down the numbering, the risk as well as the reward are reduced. Minimum Variance Portfolio (MVP) MVP tries to minimize the variance of the portfolio. These portfolios are lowest risk and reward. Maximum Sharpe Ratio Portfolio (MSR) MSR solves an optimization problem that tries to maximize the sharpe ratio of the portfolio. It uses past returns during the optimization process which means if past returns are not the same as future returns, the results can vary in future. Genetic Algorithm (GA) based Portfolio This is our own implementation of a GA based portfolio that again tries to maximize the sharpe ratio but in a slightly more robust way. This usually provides more robust portfolios than the others. When you run the command above, our tool will generate portfolios from all these strategies and give them to you. Let us look at some resulting portfolios. Resulting Portfolios For the purpose these results, we will use the 9 stocks in the stocks/stocks.txt file. When we run the above command, we first get the portfolio weights for all four strategies. For testing purposes, the above command used last five years of daily data up till April 29th. The remaining data for this year was used for forward testing i.e the portfolio strategies had no access to it when building the portfolios. What if my portfolio needs different stocks?: All you need to do is change the stocks in the stocks.txt file and run the tool again. Here is the final command again that we run in order to get our portfolios: Portfolio Weights We can see that the eigen portfolio is giving a large weight to TSLA while the others are dividing their weights more uniformly. An interesting phenomena happening here is the hedging with SQQQ that all the strategies have learned automatically. Every tool is assigning some positive weight to SQQQ while also assigning positive weights to other stocks which indicates that the strategies are automatically trying to hedge the portfolios from risk. Obviously this is not perfect, but just the fact that it's happening is fascinating. Let us look at the backtest results on the last five years prior to April 29, 2020. Backtest Results The backtests look pretty encouraging. The black dotted line is the market index i.e QQQ. Other lines are the strategies. Our custom genetic algorithm implementation seems to have the best backtest results because it's an advanced version of other strategies. The eigen portfolio that weighed TSLA the most have the most volatility but its profits are also very high. Finally, as expected, the MVP has the minimum variance and ultimately the least profits. However, since the variance is extremely low, it is a good portfolio for those who want to stay safe. The most interesting part comes next, let us look at the forward or future test results for these portfolios. Forward Test Results These results are from April 29th, 2020 to September 4th, 2020. The eigen portfolio performed the best but it also had a lot of volatility. Moreover, most of those returns are due to TSLA rocketing in the last few months. After that, our GA algorithm worked quite effectively as it beat the market index. Again, as expected, the MVP had the lowest risk and reward and slowly went up in 4-5 months. This shows the effectiveness and power of these algorithmic portfolio optimization strategies where we've developed different portfolios for different kinds of risk and reward profiles. Conclusion and Discussion We are happy to share this toolkit with the trading community and hope that people will like and contribute to it. As is the case with everything in trading, these strategies are not perfect but they are based on rigorous theory and some great empirical results. Please take care when trading with these strategies and always manage your risk. The above results were not cherry picked but the market has been highly bullish in the last few months which has led to the strong results shown above. We would love for the community to try out different strategies and share them with us. Special Thanks Special thanks to Scott Rome's blog. The eigen portfolios and minimum variance portfolio concepts came from his blog posts. The code for filtering eigen values of the covariance matrix was also mostly obtained from one of his posts. License A product by Tradytics Copyright (c) 2020-present, Tradytics.com

machine-learning-blackjack-solution
github
LLM Vibe Score0.42
Human Vibe Score0.022610872675250356
GregSommervilleMar 27, 2025

machine-learning-blackjack-solution

machine-learning-blackjack-solution Introduction A genetic algorithm is a type of artificial intelligence programming that uses ideas from evolution to solve complex problems. It works by creating a population of (initially random) candidate solutions, then repeatedly selecting pairs of candidates and combining their solutions using a process similar to genetic crossover. Sometimes candidate solutions even go through mutation, just to introduce new possibilities into the population. After a large number of generations, the best solution found up to that point is often the optimal, best solution possible. Genetic algorithms are particularly well-suited for combinatorial problems, where there are huge numbers of potential solutions to a problem. The evolutionary process they go through is, in essence, a search through a huge solution space. A solution space so large that you simply could never use a brute force approach. This project is a demonstration of using a genetic algorithm to find an optimal strategy for playing the casino game Blackjack. Please see this article for a story about how this program was used, and what the results were. The article describes some of the available settings, and shows how different values for those settings affect the final result. The source code is for a Windows application written in Cthat allows you to play with different settings like population size, selection style and mutation rate. Each generation's best solution is displayed, so you can watch the program literally evolve a solution. !blackjack strategy tester screenshot The property grid located at the upper left of the screen is where you adjust settings. There's an informational area below that, and the right side of the screen is the display area for the three tables that represent a strategy for playing Blackjack. The tall table on the left is for hard hands, the table in the upper right is for soft hands, and the table in the lower right is for pairs. We'll talk more about how to interpret this strategy in a bit. The columns along the tops of the three tables are for the dealer upcard. When you play Blackjack the dealer has one of his two cards initially turned face up, and the rank of that card has a big impact on recommended strategy. Notice that the upcard ranks don't include Jack, Queen or King. That's because those cards all count 10, so we group them and the Ten together and simplify the tables. To use the tables, first, determine if you have a pair, soft hand, or hard hand. Then look in the appropriate table, with the correct dealer upcard column. The cell in the table will be "H" when the correct strategy is to hit, "S" when the correct strategy is to stand, "D" for double-down, and (in the pairs table only) "P" for split. A Word About This "Optimal" Strategy Before we go any further, it needs to be stated that this problem of finding an optimal Blackjack strategy has already been solved. Back in the 1960s, a mathematician named Edward O. Thorp authored a book called Beat the Dealer, which included charts showing the optimal "Basic" strategy. That strategy looks like this: !optimal blackjack strategy So we're solving a problem that has already been solved, but that's actually good. That means we can compare our results to the known best solution. For example, if our result strategy tells us to do anything but stand when holding a pair of Tens, Jacks, Queens or Kings, we know there's a problem. There's one other thing to get out of the way before we go any further, and that's the idea of nondeterministic code. That means that if we run the same code twice in a row, we're likely to get two different results. That's something that happens with genetic algorithms due to their inherent randomness. There's no guarantee you'll find the absolute optimal solution, but it is assured that you will find an optimal or near-optimal solution. It's something that isn't typical when writing code, so it takes some adjustment for most programmers. Genetic Algorithms Now let's talk about the details of a genetic algorithm. Fitness Scores First of all, we need a way to evaluate candidates so we can compare them to each other. That means a numeric fitness score, which in this case is quite simple: you simulate playing a certain number of hands using the strategy, and then count the number of chips you have at the end. The big question is, how many hands should we test with? The challenge of trying to test a strategy is that due to the innate randomness of Blackjack, you could use the same strategy ten times and get ten completely different results. Obviously, the more hands you play, the more the randomness gets smoothed out, and the quality of the underlying strategy starts to emerge. If you doubt this, just think about flipping a coin. If you only flip it five times, there's certainly a possibility that it'll come up heads all five times (in fact, that happens just over 3% of the time). However, if you flip it 500 times, there's no way it's going to end up all heads - the odds of it happening are 0.5500, which works out to be roughly once every 3 x 10150 times you try it. After some testing and analysis, it was determined that a minimum of 100,000 hands per test is needed for a reasonable level of accuracy. There's still variance even at that number, but in order to cut the variance in half, you'd need to bump the number of hands to 500,000. One reason this accuracy is important is that in the later generations, the differences between candidates are very small. Evolution has caused the main parts of the strategy to converge on a particular approach, and towards the end all it's doing is refining the minor details. In those cases it's important to accurately determine the difference between two similar candidates. Representation Representation is simply the idea that we need to use a data structure for a candidate solution that can be combined via crossover, and possibly mutated. In this case, that's also quite simple because the way that human beings represent a Blackjack strategy is to use three tables, as we've seen. Representing those in code with three two-dimensional arrays is the obvious approach. Each cell in those three tables will have "Hit", "Stand", "Double-Down", or (only for pairs) "Split". By the way, since there are 160 cells in the hard hands table, and 80 cells in the soft hands table, and 100 cells in the pairs table, we can calculate exactly how many possible distinct strategies there are for Blackjack: 4100 x 380 x 3160 = 5 x 10174 possible Blackjack strategies That's a big number, which is obviously impossible to search using brute force. Genetic algorithms (GAs) are extremely helpful when trying to find an optimal solution from a very large set of possible solutions like this. Blackjack Rules and Strategies The rules of Blackjack are fairly simple. The dealer and the player both are dealt two cards. The player sees both of their cards (they are usually dealt face up), and one of the dealer's cards is dealt face up. Each card has a value - for cards between 2 and 10, the value is the same as the card's rank (so an Eight of Spades counts as 8, for example). All face cards count as 10, and an Ace can either be 1 or 11 (it counts as 11 only when that does not result in a hand that exceeds 21). The suit of a card does not matter. After the cards are dealt, if the player has Blackjack (a total of 21) and the dealer does not, the player is immediately paid 1.5 times their original bet, and a new hand is dealt. If the player has 21 and the dealer does also, then it's a tie and the player gets their original bet back, and a new hand is dealt. If the player wasn't dealt a Blackjack, then play continues with the player deciding whether to Stand (not get any more cards), Hit (receive an additional card), Double-down (place an additional bet, and receive one and only one more card), or, in the case of holding a pair, splitting the hand, which means placing an additional bet and receiving two new cards, so the end result is that the player is now playing two (or, in the case of multiple splits, more than two) hands simultaneously. If the player hits or double-downs and has a resulting hand that exceeds 21, then they lose and play continues with the next hand. If not, then the dealer draws until their hand totals at least 17. If the dealer exceeds 21 at this point, the player receives a payment equal to twice their original bet. If the dealer doesn't exceed 21, then the hands are compared and the player with the highest total that doesn't exceed 21 wins. Because of these rules, certain effective strategies emerge. One common strategy is that if you hold a hard hand with a value of 20, 19 or 18, you should Stand, since you avoid busting by going over 21, and you have a nice hand total that might win in a showdown with the dealer. Another common strategy is to split a pair of Aces, since Aces are so powerful (due to the fact that count as 11 or 1, you can often Hit a hand with a soft Ace with no risk of busting). Likewise, splitting a pair of 8s is a good idea because with a hard total of 16, it's likely you will bust if you take a Hit (since so many cards count as 10). As a human being, all it takes is a little knowledge about the rules in order to construct a strategy. The GA program doesn't have that advantage, and operates completely without any pre-programmed knowledge of Blackjack. It simply uses the relative fitness scores and the mechanism of evolution to find the solution. GA Settings There are many variables or settings for a GA. You can adjust population size, how parent candidates are selected, how the resulting children may be mutated, and several other items. The following sections describe some of these settings: Setting: Selection Style Once we've solved representation and have a fitness function, the next step is to select two candidates for crossover during the process of building a new generation. There are three common styles for selection, and this program supports all of them. First, you can choose Roulette Wheel selection. It's named for a Roulette wheel because you can imagine each candidate's fitness score being a wedge in a pie chart, with a size proportionate to its relative fitness compared to the other candidates. (Of course, this assumes that all fitness scores are positive, which we will talk about shortly). The main benefit of Roulette Wheel selection is that selection is fitness-proportionate. Imagine if you had only three candidates, with fitness scores of 1, 3, and 8. The relative selection probabilities for those candidates will be 1/12, 3/12, and 8/12. The downside of Roulette Wheel selection is that it tends to be somewhat slow in terms of processing. The selection process is done by iterating through the candidates until a particular condition is matched - in other words, O(N) performance. Another potential problem with Roulette Wheel selection is that there may be situations where fitness scores vary widely, to such an extent that only certain candidates have any reasonable chance of being selected. This happens frequently in early generations, since the majority of candidates are mostly random. Although this might sound like a positive (since you ultimately want to select candidates with high fitness scores), it also results in a loss of genetic diversity. In other words, even though a particular candidate may have a low fitness score in an early generation, it may contain elements that are needed to find the ultimate solution in later generations. Ranked Selection is the solution to this problem. Instead of using raw fitness scores during the selection process, the candidates are sorted by fitness, with the worst candidate receiving a score of 0, the second worse receiving 1, and so forth, all the way to the best candidate, which has a score equal to the population size - 1. Ranked Selection is quite slow, since it combines the O(N) performance of Roulette Wheel, with the additional requirement that the candidates be sorted before selection. However, there may be circumstances where it performs better than other selection approaches. Finally, the fastest selection method of all is called Tournament Selection. This method simply selects N random candidates from the current generation, and then uses the one with the best fitness score. A tournament size of 2 means two random candidates are selected, and the best of those two is used. If you have a large tournament size (like 10), then 10 different candidates will be selected, with the best of those being the ultimate selection. That obviously tilts the balance between randomness and quality. Tournament selection works well in most cases, but it does require some experimentation to find the best tourney size. Setting: Elitism Elitism is a technique that helps ensure that the best candidates are always maintained. Since all selection methods are random to some degree, it is possible to completely lose the best candidates from one generation to another. By using Elitism, we automatically advance a certain percentage of the best candidates to the next generation. Elitism does have a negative impact on performance since all of the candidates must be sorted by fitness score. Typically Elitism is done before filling the rest of a new generation with new candidates created by crossover. Crossover Details Once two candidate solutions have been selected, the next step in building a new generation is to combine those two into a single new candidate, hopefully using the best of both parent strategies. There are a number of ways to do crossover, but the method used in this program is quite straightforward - the two fitness scores are compared, and crossover happens in a relatively proportionate way. If one candidate has a fitness of 10, and the other has a fitness of 5, then the one with fitness 10 contributes twice as much to the child as the parent with a fitness of 5. Since the fitness scores in this program are based on how much the strategy would win over thousands of hands, almost all fitness scores will be negative. (This is obviously because the rules are set up so the house always wins.) This makes it difficult to calculate relative fitnesses (how do you compare a positive number with a negative, and find relative proportions?), and also causes problems with selection methods like Roulette Wheel or Ranked. To solve this, we find the lowest fitness score of the generation and add that value to each candidate. This results in an adjusted fitness score of 0 for the very worse candidate, so it never gets selected. Mutation As has been mentioned a few times, maintaining genetic diversity in our population of candidate solutions is a good thing. It helps the GA ultimately find the very best solution, by occasionally altering a candidate in a positive direction. There are two settings for mutation. MutationRate controls what percentage of new candidates have mutation done on them. MutationImpact controls what percentage of their strategy is randomized. Population Size Population size has a significant impact on performance. The smaller the population size, the faster the GA will execute. On the other hand, if the size is too low the population may not have enough genetic diversity to find the ultimate solution. During testing, it looks like 700 to 1000 is a good balance between speed and correctness. Performance Notes This program consumes a lot of processing power. Running tests of hundreds of thousands of hands of Blackjack for hundreds or thousands of candidates consumes a lot of time. It's really imperative to write the code so that it works as efficiently as possible. If your CPU isn't consistently at or above 95% usage, there's still room for improvement. Multi-threading is a natural fit for genetic algorithms because we often want to perform the same action on each candidate. The best example of this is when we calculate fitness scores. This is often an operation that takes quite a bit of time. In our case, we're dealing out 100,000 hands, and each hand has to be played until the end. If we're single-threading that code, it's going to take a long time. Multi-threading is really the way to go. Luckily, there's a ridiculously simple way to efficiently use all of your processors for an operation like this. This code loops over all of the candidates in the currentGeneration list, calls the fitness function and sets the fitness property for each: Regardless of the number of items in the list or the number of processors on your machine, the code will efficiently run the code in a multi-threaded manner, and continue only when all of the threads are complete. One of the side effects of making this code multi-threaded is that all of the code relating to evaluating a candidate must be thread-safe, including any Singleton objects. When making code thread-safe, pay attention that you don't accidentally introduce code that will slow your program down unintentionally, because sometimes it can be quite subtle. Random numbers are central to how genetic algorithms work, so it's critical that they can be used correctly from a multithreaded environment. That means that each random number generator must be separate from the others, and it also means that each must produce a distinct series of random numbers. Random number generators use seed values which are usually time-based, like the number of milliseconds the computer has been turned on. Starting with that seed, subsequent calls will return a series of numbers that look random, but really aren't. If you start with the same seed, you get the same sequence. And that's a problem because if you create multiple random number generator objects in a loop using the default time-based seed, several of them will have the same time-based initial seed value, which will result in the same sequence of "random" numbers. That's a bug, because it can reduce the true randomness of the program a great deal, and that's vital to a genetic algorithm. There are a couple of ways to solve this problem. First, you can make the random object truly a singleton, and restrict access to it by using a Clock statement. The makes all access serialized for any random number need, which reduces performance. Another approach is to make the variable static per thread. By declaring the variable as static and also marking it with the [ThreadStatic] attribute, the .NET runtime allocates one static variable per thread. That eliminates the locking/serialization, but also has performance issues. The approach used in this application is to use a non-default seed value. In this case we call Guid.NewGuid().GetHashCode(), which generates a new, unique GUID, then gets an integer hashcode value that should be unique, depending on how GetHashCode is implemented. While multithreading really helps performance, there are also other things we can do to improve performance. For example, when dealing with large populations, the hundreds or thousands of objects that will be generated each generation can quickly turn into a huge problem related to garbage collection. In the end, the easiest way to solve that is to look through the code and find objects being allocate inside a loop. It's better to declare the variable outside of the loop, and then clear it in the loop, rather than reallocate it. In a program like this one where you could be looping hundreds of thousands of times, this can result in a very significant performance boost. For example, in an early version of this code, a Deck object was created for each hand. Since there are hundreds of candidate solutions running hundreds of thousands of trial hands, this was a huge inefficiency. The code was changed to allocate one deck per test sequence. The deck was shuffled as needed, so it never needs to be reallocated. Beyond the cards in the deck, another object type that was repeatedly created and destroyed were the candidate strategies. To mitigate this problem, a StrategyPool class was created that handles allocation and deallocation. This means that strategy objects are reused, rather than dynamically created when needed. The pool class has to be thread-safe, so it does serialize access to its methods via a Clock statement, but overall using the pool approach produced a good performance increase. Finally, a subtle form of object allocation is conversion. In an early version of the code, a utility card function used Convert.ToInt32(rankEnum). Obviously, the easiest way to convert from an enum to an int is simply to cast it, like (int)rankEnum. But it's hard to know exactly what the difference is between that approach, int.Parse(), int.TryParse(), or Convert.ToInt32(), since they can all be used and are roughly equivalent. Perhaps the compiler was boxing the enum value before passing it to Convert.ToInt32(), because the profiler identified this as a function that had large amounts of thread contention waiting - and the problem got much, much worse as the generations passed. By rewriting the conversion to use a simple cast, the program performance increased threefold (3x). Contributing Please read CONTRIBUTING.md for details on our code of conduct, and the process for submitting pull requests to us. Author Greg Sommerville - Initial work* License This project is licensed under the Apache 2.0 License - see the LICENSE.md file for details

Solana_AIAgent_Trading
github
LLM Vibe Score0.464
Human Vibe Score0.05777682403433476
solagent99Mar 25, 2025

Solana_AIAgent_Trading

Solana AI Agent Trading Tool An open-source trading toolkit for connecting AI agents to Solana protocols. Now, any agent, using any model can autonomously perform 15+ Solana actions: Trade tokens Launch new tokens Lend assets Send compressed airdrops Execute blinks Launch tokens on AMMs And more... 💬 Contact Me If you have any question or something, feel free to reach out me anytime via telegram, discord or twitter. 🌹 You're always welcome 🌹 Telegram: @Leo Replit template created by Arpit Singh 🔧 Core Blockchain Features Token Operations Deploy SPL tokens by Metaplex Transfer assets Balance checks Stake SOL Zk compressed Airdrop by Light Protocol and Helius NFTs on 3.Land Create your own collection NFT creation and automatic listing on 3.land List your NFT for sale in any SPL token NFT Management via Metaplex Collection deployment NFT minting Metadata management Royalty configuration DeFi Integration Jupiter Exchange swaps Launch on Pump via PumpPortal Raydium pool creation (CPMM, CLMM, AMMv4) Orca Whirlpool integration Manifest market creation, and limit orders Meteora Dynamic AMM, DLMM Pool, and Alpha Vault Openbook market creation Register and Resolve SNS Jito Bundles Pyth Price feeds for fetching Asset Prices Register/resolve Alldomains Perpetuals Trading with Adrena Protocol Drift Vaults, Perps, Lending and Borrowing Solana Blinks Lending by Lulo (Best APR for USDC) Send Arcade Games JupSOL staking Solayer SOL (sSOL)staking Non-Financial Actions Gib Work for registering bounties 🤖 AI Integration Features LangChain Integration Ready-to-use LangChain tools for blockchain operations Autonomous agent support with React framework Memory management for persistent interactions Streaming responses for real-time feedback Vercel AI SDK Integration Vercel AI SDK for AI agent integration Framework agnostic support Quick and easy toolkit setup Autonomous Modes Interactive chat mode for guided operations Autonomous mode for independent agent actions Configurable action intervals Built-in error handling and recovery AI Tools DALL-E integration for NFT artwork generation Natural language processing for blockchain commands Price feed integration for market analysis Automated decision-making capabilities 📃 Documentation You can view the full documentation of the kit at docs.solanaagentkit.xyz 📦 Installation Quick Start Usage Examples Deploy a New Token Create NFT Collection on 3Land Create NFT on 3Land When creating an NFT using 3Land's tool, it automatically goes for sale on 3.land website Create NFT Collection Swap Tokens Lend Tokens Stake SOL Stake SOL on Solayer Send an SPL Token Airdrop via ZK Compression Fetch Price Data from Pyth Open PERP Trade Close PERP Trade Close Empty Token Accounts Create a Drift account Create a drift account with an initial token deposit. Create a Drift Vault Create a drift vault. Deposit into a Drift Vault Deposit tokens into a drift vault. Deposit into your Drift account Deposit tokens into your drift account. Derive a Drift Vault address Derive a drift vault address. Do you have a Drift account Check if agent has a drift account. Get Drift account information Get drift account information. Request withdrawal from Drift vault Request withdrawal from drift vault. Carry out a perpetual trade using a Drift vault Open a perpertual trade using a drift vault that is delegated to you. Carry out a perpetual trade using your Drift account Open a perpertual trade using your drift account. Update Drift vault parameters Update drift vault parameters. Withdraw from Drift account Withdraw tokens from your drift account. Borrow from Drift Borrow tokens from drift. Repay Drift loan Repay a loan from drift. Withdraw from Drift vault Withdraw tokens from a drift vault after the redemption period has elapsed. Update the address a Drift vault is delegated to Update the address a drift vault is delegated to. Get Voltr Vault Position Values Get the current position values and total value of assets in a Voltr vault. Deposit into Voltr Strategy Deposit assets into a specific strategy within a Voltr vault. Withdraw from Voltr Strategy Withdraw assets from a specific strategy within a Voltr vault. Get a Solana asset by its ID Get a price inference from Allora Get the price for a given token and timeframe from Allora's API List all topics from Allora Get an inference for an specific topic from Allora Examples LangGraph Multi-Agent System The repository includes an advanced example of building a multi-agent system using LangGraph and Solana Agent Kit. Located in examples/agent-kit-langgraph, this example demonstrates: Multi-agent architecture using LangGraph's StateGraph Specialized agents for different tasks: General purpose agent for basic queries Transfer/Swap agent for transaction operations Read agent for blockchain data queries Manager agent for routing and orchestration Fully typed TypeScript implementation Environment-based configuration Check out the LangGraph example for a complete implementation of an advanced Solana agent system. Dependencies The toolkit relies on several key Solana and Metaplex libraries: @solana/web3.js @solana/spl-token @metaplex-foundation/digital-asset-standard-api @metaplex-foundation/mpl-token-metadata @metaplex-foundation/mpl-core @metaplex-foundation/umi @lightprotocol/compressed-token @lightprotocol/stateless.js Contributing Contributions are welcome! Please feel free to submit a Pull Request. Refer to CONTRIBUTING.md for detailed guidelines on how to contribute to this project. Contributors Star History License Apache-2 License Funding If you wanna give back any tokens or donations to the OSS community -- The Public Solana Agent Kit Treasury Address: Solana Network : EKHTbXpsm6YDgJzMkFxNU1LNXeWcUW7Ezf8mjUNQQ4Pa Security This toolkit handles private keys and transactions. Always ensure you're using it in a secure environment and never share your private keys.

video-killed-the-radio-star
github
LLM Vibe Score0.48
Human Vibe Score0.018384486870142776
dmarxMar 23, 2025

video-killed-the-radio-star

Video Killed The Radio Star Requirements ffmpeg - https://ffmpeg.org/ pytorch - https://pytorch.org/get-started/locally/ vktrs - (this repo) - pip install vktrs[api] stability_sdk api token - https://beta.dreamstudio.ai/ > circular icon in top right > membership > API Key whisper - pip install git+https://github.com/openai/whisper FAQ What is this? TLDR: Automated music video maker, given an mp3 or a youtube URL How does this animation technique work? For each text prompt you provide, the notebook will... Generate an image based on that text prompt (using stable diffusion) Use the generated image as the init_image to recombine with the text prompt to generate variations similar to the first image. This produces a sequence of extremely similar images based on the original text prompt Images are then intelligently reordered to find the smoothest animation sequence of those frames This image sequence is then repeated to pad out the animation duration as needed The technique demonstrated in this notebook was inspired by a video created by Ben Gillin. How are lyrics transcribed? This notebook uses openai's recently released 'whisper' model for performing automatic speech recognition. OpenAI was kind of to offer several different sizes of this model which each have their own pros and cons. This notebook uses the largest whisper model for transcribing the actual lyrics. Additionally, we use the smallest model for performing the lyric segmentation. Neither of these models is perfect, but the results so far seem pretty decent. The first draft of this notebook relied on subtitles from youtube videos to determine timing, which was then aligned with user-provided lyrics. Youtube's automated captions are powerful and I'll update the notebook shortly to leverage those again, but for the time being we're just using whisper for everything and not referencing user-provided captions at all. Something didn't work quite right in the transcription process. How do fix the timing or the actual lyrics? The notebook is divided into several steps. Between each step, a "storyboard" file is updated. If you want to make modifications, you can edit this file directly and those edits should be reflected when you next load the file. Depending on what you changed and what step you run next, your changes may be ignored or even overwritten. Still playing with different solutions here. Can I provide my own images to 'bring to life' and associate with certain lyrics/sequences? Yes, you can! As described above: you just need to modify the storyboard. Will describe this functionality in greater detail after the implementation stabilizes a bit more. This gave me an idea and I'd like to use just a part of your process here. What's the best way to reuse just some of the machinery you've developed here? Most of the functionality in this notebook has been offloaded to library I published to pypi called vktrs. I strongly encourage you to import anything you need from there rather than cutting and pasting function into a notebook. Similarly, if you have ideas for improvements, please don't hesitate to submit a PR! Dev notes

How-to-learn-Deep-Learning
github
LLM Vibe Score0.524
Human Vibe Score0.1392403398579415
emilwallnerMar 23, 2025

How-to-learn-Deep-Learning

Approach A practical, top-down approach, starting with high-level frameworks with a focus on Deep Learning. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Getting started [2 months] There are three main goals to get up to speed with deep learning: 1) Get familiar to the tools you will be working with, e.g. Python, the command line and Jupyter notebooks 2) Get used to the workflow, everything from finding the data to deploying a trained model 3) Building a deep learning mindset, an intuition for how deep learning models behave and how to improve them Spend a week on codecademy.com and learn the python syntax, command line and git. If you don't have any previous programming experience, it's good to spend a few months learning how to program. Otherwise, it's easy to become overwhelmed. Spend one to two weeks using Pandas and Scikit-learn on Kaggle problems using Jupyter Notebook on Colab, e.g. Titanic, House prices, and Iris. This gives you an overview of the machine learning mindset and workflow. Spend one month implementing models on cloud GPUs. Start with FastAI and PyTorch. The FastAI community is the go-to place for people wanting to apply deep learning and share the state of the art techniques. Once you have done this, you will know how to add value with ML. Portfolio [3 - 12 months] Think of your portfolio as evidence to a potential employer that you can provide value for them. When you are looking for your first job, there are four main roles you can apply for Machine Learning Engineering, Applied Machine Learning Researcher / Residencies, Machine Learning Research Scientist, and Software Engineering. A lot of the work related to machine learning is pure software engineering roles (category 4), e.g. scaling infrastructure, but that's out of scope for this article. It's easiest to get a foot in the door if you aim for Machine Learning Engineering roles. There are a magnitude more ML engineering roles compared to category 2 & 3 roles, they require little to no theory, and they are less competitive. Most employers prefer scaling and leveraging stable implementations, often ~1 year old, instead of allocating scarce resources to implement SOTA papers, which are often time-consuming and seldom work well in practice. Once you can cover your bills and have a few years of experience, you are in a better position to learn theory and advance to category 2 & 3 roles. This is especially true if you are self-taught, you often have an edge against an average university graduate. In general, graduates have weak practical skills and strong theory skills. Context You'll have a mix of 3 - 10 technical and non-technical people looking at your portfolio, regardless of their background, you want to spark the following reactions: the applicant has experience tackling our type of problems, the applicant's work is easy to understand and well organized, and the work was without a doubt 100% made by the applicant. Most ML learners end up with the same portfolio as everyone else. Portfolio items include things as MOOC participation, dog/cat classifiers, and implementations on toy datasets such as the titanic and iris datasets. They often indicate that you actively avoid real-world problem-solving, and prefer being in your comfort zone by copy-pasting from tutorials. These portfolio items often signal negative value instead of signaling that you are a high-quality candidate. A unique portfolio item implies that you have tackled a unique problem without a solution, and thus have to engage in the type of problem-solving an employee does daily. A good starting point is to look for portfolio ideas on active Kaggle competitions, and machine learning consulting projects, and demo versions of common production pipelines. Here's a Twitter thread on how to come up with portfolio ideas. Here are rough guidelines to self-assess the strength of your portfolio: Machine learning engineering: Even though ML engineering roles are the most strategic entry point, they are still highly competitive. In general, there are ~50 software engineering roles for every ML role. From the self-learners I know, 2/3 fail to get a foot in the door and end up taking software engineering roles instead. You are ready to look for a job when you have two high-quality projects that are well-documented, have unique datasets, and are relevant to a specific industry, say banking or insurance. Project Type | Base score | -------------| -----------| Common project | -1 p || Unique project | 10 p | Multiplier Type | Factor -----------------|----------------- Strong documentation | 5x 5000-word article | 5x Kaggle Medal | 10x Employer relevancy | 20x Hireable: 5,250 p Competative: 15,000 p Applied research / research assistant/ residencies: For most companies, the risk of pursuing cutting edge research is often too high, thus only the biggest companies tend to need this skillset. There are smaller research organizations that hire for these positions, but these positions tend to be poorly advertised and have a bias for people in their existing community. Many of these roles don't require a Ph.D., which makes them available to most people with a Bachelor's or Master's degrees, or self-learners with one year of focussed study. Given the status, scarcity, and requirements for these positions, they are the most competitive ML positions. Positions at well-known companies tend to get more than a thousand applicants per position. Daily, these roles require that you understand and can implement SOTA papers, thus that's what they will be looking for in your portfolio. Projects type | Base score --------------| ----------- Common project | -10 p Unique project | 1 p SOTA paper implementation | 20 p Multiplier type | Factor ----------------| --------------- Strong documentation | 5x 5000-word article | 5x SOTA performance | 5x Employer relevancy | 20x Hireable: 52,500 p Competitive: 150,000 p Research Scientist: Research scientist roles require a Ph.D. or equivalent experience. While the former category requires the ability to implement SOTA papers, this category requires you to come up with research ideas. The mainstream research community measure the quality of research ideas by their impact, here is a list of the venues and their impact. To have a competitive portfolio, you need two published papers in the top venues in an area that's relevant to your potential employer. Project type | Base score -------------| ---------------- Common project | -100 p An unpublished paper | 5 p ICML/ICLR/NeurIPS publication | 500p All other publications | 50 p Multiplier type | Factor ------------------| ------------------ First author paper | 10x Employer relevancy | 20x Hireable: 20,000 p Competitive roles and elite PhD positions: 200,000 p Examples: My first portfolio item (after 2 months of learning): Code | Write-up My second portfolio item (after 4 months of learning): Code | Write-up Dylan Djian's first portfolio item: Code | Write-up Dylan Djian's second portfolio item: Code | Write-up Reiichiro Nakano's first portfolio item: Code | Write-up Reiichiro Nakano's second portfolio item: Write-up Most recruiters will spend 10-20 seconds on each of your portfolio items. Unless they can understand the value in that time frame, the value of the project is close to zero. Thus, writing and documentation are key. Here's another thread on how to write about portfolio items. The last key point is relevancy. It's more fun to make a wide range of projects, but if you want to optimize for breaking into the industry, you want to do all projects in one niche, thus making your skillset super relevant for a specific pool of employers. Further Inspiration: FastAI student projects Stanford NLP student projects Stanford CNN student projects Theory 101 [4 months] Learning how to read papers is critical if you want to get into research, and a brilliant asset as an ML engineer. There are three key areas to feel comfortable reading papers: 1) Understanding the details of the most frequent algorithms, gradient descent, linear regression, and MLPs, etc 2) Learning how to translate the most frequent math notations into code 3) Learn the basics of algebra, calculus, statistics, and machine learning For the first week, spend it on 3Blue1Brown's Essence of linear algebra, the Essence of Calculus, and StatQuests' the Basics (of statistics) and Machine Learning. Use a spaced repetition app like Anki and memorize all the key concepts. Use images as much as possible, they are easier to memorize. Spend one month recoding the core concepts in python numpy, including least squares, gradient descent, linear regression, and a vanilla neural network. This will help you reduce a lot of cognitive load down the line. Learning that notations are compact logic and how to translate it into code will make you feel less anxious about the theory. I believe the best deep learning theory curriculum is the Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. I use it as a curriculum, and the use online courses and internet resources to learn the details about each concept. Spend three months on part 1 of the Deep learning book. Use lectures and videos to understand the concepts, Khan academy type exercises to master each concept, and Anki flashcards to remember them long-term. Key Books: Deep Learning Book by Ian Goodfellow and Yoshua Bengio and Aaron Courville. Deep Learning for Coders with fastai and PyTorch: AI Applications Without a PhD by Jeremy Howard and Sylvain. Gugger. Deep Learning with Python by François Chollet. Neural Networks and Deep Learning by Michael Nielsen. Grokking Deep Learning by Andrew W. Trask. Forums FastAI Keras Slack Distill Slack Pytorch Twitter Other good learning strategies: Emil Wallner S. Zayd Enam Catherine Olsson Greg Brockman V2 Greg Brockman V1 Andrew Ng Amid Fish Spinning Up by OpenAI Confession as an AI researcher YC Threads: One and Two If you have suggestions/questions create an issue or ping me on Twitter. UPDATED VERSION: 👉 Check out my 60-page guide, No ML Degree, on how to land a machine learning job without a degree. Language versions: Korean | English

coca
github
LLM Vibe Score0.541
Human Vibe Score0.0750848814969247
phodalMar 21, 2025

coca

Coca - toolbox for system refactoring and analysis !GitHub release (latest SemVer) !GitHub go.mod Go version Coca is a toolbox which is design for legacy system refactoring and analysis, includes call graph, concept analysis, api tree, design patterns suggest. Coca 是一个用于系统重构、系统迁移和系统分析的工具箱。它可以分析代码中的测试坏味道、模块化分析、行数统计、分析调用与依赖、Git 分析以及自动化重构等。 Related Tools: Coco is an effective DevOps analysis and auto-suggest tool. Kotlin version: Chapi Migration Guide (Chinese Version): 《系统重构与迁移指南》 Inspired by: newlee & Tequila Refactoring Modeling: !Refactoring Modeling Languages Support: Java (full features) Features List: Getting started Requirements: graphviz for dot file to image (such as svg, png) The easiest way to get coca is to use one of the pre-built release binaries which are available for OSX, Linux, Windows on the release page. You can also install yourself : Usage Analysis Arch Android Studio Gradle DSL Module (merge header) command: coca arch -x "com.android.tools.idea.gradle.dsl" -H true !Gradle Demo Android Studio Gradle DSL Module Elements Part: command: coca arch -x "com.android.tools.idea.gradle.dsl.parser.elements" !Gradle Demo Find Bad Smells Examples Result: Code Line Count Results: Results to json Cloc by directory results csv: Cloc Top File output to: cocareporter/sortcloc.json and also: Build Deps Tree Examples Results: !Call Demo Identify Spring API !API Demo With Count or multi package: coca api -r com.macro.mall.demo.controller.,com.zheng.cms.admin.,com.phodal.pholedge -c Git Analysis Results: Concept Analyser Results Examples: Count Refs Results: Reverse Call Graph Results: !RCall Demo Auto Refactor support: rename move remove unused import remove unused class Evaluate Arduino Results(Old Version): New Version: Evaluate.json examples Todo results: coca suggest +--------+------------------+--------------------------------+ | CLASS | PATTERN | REASON | +--------+------------------+--------------------------------+ | Insect | factory | too many constructor | | Bee | factory, builder | complex constructor, too | | | | many constructor, too many | | | | parameters | +--------+------------------+--------------------------------+ coca tbs bash +---------------------+---------------------------------------------------------------+------+ | TYPE | FILENAME | LINE | +---------------------+---------------------------------------------------------------+------+ | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 107 | | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 41 | | DuplicateAssertTest | app/test/cc/arduino/i18n/ExternalProcessOutputParserTest.java | 63 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 71 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 72 | | RedundantPrintTest | app/test/cc/arduino/i18n/I18NTest.java | 77 | | DuplicateAssertTest | app/test/cc/arduino/net/PACSupportMethodsTest.java | 19 | | DuplicateAssertTest | app/test/processing/app/macosx/SystemProfilerParserTest.java | 51 | | DuplicateAssertTest | app/test/processing/app/syntax/PdeKeywordsTest.java | 41 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 57 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 83 | | DuplicateAssertTest | app/test/processing/app/tools/ZipDeflaterTest.java | 109 | +---------------------+---------------------------------------------------------------+------+ coca deps -p fixtures/deps/mavensample +---------------------------+----------------------------------------+---------+ | GROUPID | ARTIFACTID | SCOPE | +---------------------------+----------------------------------------+---------+ | org.flywaydb | flyway-core | | | mysql | mysql-connector-java | runtime | | org.springframework.cloud | spring-cloud-starter-contract-verifier | test | +---------------------------+----------------------------------------+---------+ bash brew install go bash export GOROOT=/usr/local/opt/go/libexec export GOPATH=$HOME/.go export PATH=$PATH:$GOROOT/bin:$GOPATH/bin git clone https://github.com/modernizing/coca go get github.com/onsi/ginkgo go get github.com/onsi/gomega `` License Arch based on Tequila Git Analysis inspired by Code Maat Test bad smells inspired by Test Smell Examples @ 2019 A Phodal Huang's Idea. This code is distributed under the MPL license. See LICENSE` in this directory.

aion
github
LLM Vibe Score0.494
Human Vibe Score0.011340905117109681
aionnetworkFeb 28, 2025

aion

Aion Mainstream adoption of blockchains has been limited because of scalability, privacy, and interoperability challenges. Aion is a multi-tier blockchain network designed to address these challenges. Core to our hypothesis is the idea that many blockchains will be created to solve unique business challenges within unique industries. As such, the Aion network is designed to support custom blockchain architectures while providing a trustless mechanism for cross-chain interoperability. The Aion White Papers provides more details regarding our design and project roadmap. This repository contains the main (Java) kernel implementation and releases for the Aion Network. System Requirements Ubuntu 16.04 or a later version Getting Started Blockchain node concept To understand what is blockchain kernel: Node overview Developers If you're interested in building Open Applications, powered by Aion: Visit the Developer site of The Open Application Network : developer.theoan.com If you're interested in making improvements to the Java Implementation of Aion: Refer to the Build Aion kernel from source wiki for information on building this source code to a native binary or Docker image Refer to the Installation wiki for a guide on installing and configuring the kernel. The Owner's Manual wiki will include further instructions and details on working with the kernel. Please refer to the wiki pages for further documentation on mining/validating, using the Web3 API, command line options, etc. Miners/Validators If you're interested in being a validator on the Aion networks, refer to our Validator Docs Users If you're interested in interacting with dApps and using Aion, refer to our Aion Desktop Wallet Docs FAQ Where can I store my Aion? We recommend using the web-based Aion Wallet; more information can be found in “Docs”). Where can I stake my Aion? You can use the original staking interface which has support for staking pool operators, or the web-based Aion Wallet. Where can I check on a transaction on The Open Application Network? You can visit either the web-based Aion Wallet or the Aion Dashboard to view a transaction on the network. Where can I see the current network performance of The Open Application Network? You can visit the Aion Dashboard to see how the Open Application Network is performing. What should I do if the desktop wallet or the web based wallet are not functioning properly? First check in with the community on the community subreddit. If the community is not able to assist then you can submit a ticket through Github. The Open Application Network is currently providing support to help maintain the network; where can I see the funds that The Open Application Network has mined or received as a stake reward? All funds mined or rewarded for staking that the foundation receives are burned to this address: 0x0000000000000000000000000000000000000000000000000000000000000000 users can check the totals burned via the Aion Dashboard here. What is the total circulating supply of Aion? To view the current total circulating supply of Aion you can use the Aion Watch tool located here. Which networks are supported? The Mainnet network is supported. To view the dashboards for this networks use these links: Mainnet How can I export a list of my transactions? If you would like to download a copy of your transaction history you can use https://mainnet.theoan.com and search for your public address. In the bottom right of your screen is a “Download this Account” button which will allow you to select a date range and download a .csv file containing your transactions. Where can I access a copy of The OAN and Aion Brand Guidelines? The OAN and Aion Brand Guidelines can be located here they can be used by the community to create brand aligned content. My Ledger doesn’t seem to be recognized with applications in the Chrome Browser (Staking Interface or Wallet) When using your Ledger hardware wallet with Aion installed to access an account VIA the Chrome browser, users will need to enable the Aion contract on their Ledger device. This can be done by selecting: Aion > Setting > enable Contract. What happened to the Aiwa chrome extension wallet? Aiwa was owned and operated by a third-party organization called BlockX Labs, Aiwa was funded by a community grant during its lifespan. However, BlockX Labs is now reorganizing and will no longer support Aiwa. Usage of Aiwa has decreased significantly with other tools such as the web based wallet now available so the decision was made to deprecate it. I am unable to undelegate my staked Aion In order to undelegate your Aion: – You must have a sufficient Aion balance to perform the undelegation transaction (a minimum of 0.02 Aion is required for the transaction fee) – Your balance will be updated after a lock-up period of 8640 blocks (approximately 24 hours) – Ensure the amount follows this format: 999,999,999.999999999 – If you are using a ledger, please ensure that your firmware is up to date. – If you are using the desktop interface, ensure that you are using the latest version – For more information view this guide What happened to the swap process to convert ERC-20 Aion to the mainnet? As of January 31, 2022 swapping from ERC20 to Aion mainnet is no longer supported. The original Aion token swap from Ethereum to Aion was completed on December 10, 2018. However, in order to support the community members who missed the original swap deadline a manual process was available, this process has now been retired. Community Channels Newsfeed: @AionNewsfeed Info Bot: @AionTGbot Wiki: reddit.com/r/AionNetwork/Wiki Help Desk: https://helpdesk.theoan.com/ Contact To keep up to date and stay connected with current progress and development, reach out to us on the following channels: Aion Telegram Dispatch Alerts Aion on Twitter Aion Blog License Aion is released under the MIT license

Mastering-AI-for-Entrepreneurs-9-Free-Courses
github
LLM Vibe Score0.203
Human Vibe Score0
Softtechhub1Feb 1, 2025

Mastering-AI-for-Entrepreneurs-9-Free-Courses

Mastering-AI-for-Entrepreneurs-9-Free-Courses Introduction: The Entrepreneur's AI RevolutionArtificial Intelligence (AI) is changing the way we do business. It's not just for tech giants anymore. Small businesses and startups are using AI to work smarter, not harder. As an entrepreneur, you need to understand AI to stay ahead.Why AI is a must-have skill for entrepreneursAI is everywhere. It's in the apps we use, the products we buy, and the services we rely on. Businesses that use AI are seeing big improvements:They're making better decisions with data-driven insightsThey're automating routine tasks, freeing up time for creativityThey're personalizing customer experiences, boosting satisfaction and salesIf you're not using AI, you're falling behind. But here's the good news: you don't need to be a tech wizard to harness the power of AI.Breaking the barriers to AI learningThink AI is too complex? Think again. You don't need a computer science degree to understand and use AI in your business. Many AI tools are designed for non-technical users. They're intuitive and user-friendly.The best part? You can learn about AI for free. There are tons of high-quality courses available at no cost. These courses are designed for busy entrepreneurs like you. They cut through the jargon and focus on practical applications.What to expect from this articleWe've handpicked nine free courses that will turn you into an AI-savvy entrepreneur. Each course is unique, offering different perspectives and skills. We'll cover:What makes each course specialWhat you'll learnHow it applies to your businessWho it's best suited forReady to dive in? Let's explore these game-changing courses that will boost your AI knowledge and give your business an edge.1. Google AI Essentials: A Beginner's Guide to Practical AIWhy This Course Is EssentialGoogle AI Essentials is perfect if you're just starting out. It's designed for people who don't have a tech background. The course focuses on how AI can help you in your day-to-day work, not on complex theories.What You'll LearnThis course is all about making AI work for you. You'll discover how to:Use AI to boost your productivity. Generate ideas, create content, and manage tasks more efficiently.Streamline your workflows. Learn how AI can help with everyday tasks like drafting emails and organizing your schedule.Use AI responsibly. Understand the potential biases in AI and how to use it ethically.Key TakeawaysYou'll earn a certificate from Google. This looks great on your resume or LinkedIn profile.You'll learn how to work alongside AI tools to get better results in your business.You'll gain practical skills you can use right away to improve your work.Get StartedEnroll in Google AI Essentials2. Introduction to Generative AI: A Quick Start for EntrepreneursWhy This Course Works for Busy EntrepreneursThis course is short and sweet. In just 30 minutes, you'll get a solid grasp of generative AI. It's perfect if you're short on time but want to understand the basics.What You'll LearnThe fundamentals of generative AI: what it is, how it works, and its limitsHow generative AI differs from other types of AIReal-world applications of generative AI in businessHow It Helps Your BusinessAfter this course, you'll be able to:Make smarter decisions about using AI tools in your businessSpot opportunities where generative AI could solve problems or create valueUnderstand the potential and limitations of this technologyGet StartedEnroll in Introduction to Generative AI3. Generative AI with Large Language Models: Advanced Skills for EntrepreneursWhy This Course Stands OutThis course digs deeper into the technical side of AI. It's ideal if you have some coding experience and want to understand how AI models work under the hood.What You'll LearnYou'll gain key skills for working with Large Language Models (LLMs):How to gather and prepare data for AI modelsChoosing the right model for your needsEvaluating model performance and improving resultsYou'll also learn about:The architecture behind transformer models (the tech powering many AI tools)Techniques for fine-tuning models to your specific business needsWho Should Take This CourseThis course is best for entrepreneurs who:Have basic Python programming skillsUnderstand the fundamentals of machine learningWant to go beyond using AI tools to actually building and customizing themGet StartedEnroll in Generative AI with Large Language Models4. AI for Everyone by Andrew Ng: Simplifying AI for Business LeadersWhy It's Perfect for BeginnersAndrew Ng is a leading figure in AI education. He's known for making complex topics easy to understand. This course is designed for non-technical learners. You don't need any coding or math skills to benefit from it.What You'll LearnHow AI works at a high levelHow to spot problems in your business that AI can solveWays to assess how AI might impact your business processes and strategiesWhy Entrepreneurs Love This CourseIt explains AI concepts in plain English, without technical jargonYou can complete it in just 8 hours, fitting it into your busy scheduleIt focuses on the business value of AI, not just the technologyGet StartedStart with AI for Everyone on Coursera5. Generative AI: Introduction and ApplicationsWhy This Course Is Ideal for EntrepreneursThis course offers a broad view of generative AI applications. You'll learn about AI in text, image, audio, and more. It's packed with hands-on experience using popular AI tools.What You'll LearnThe basics and history of generative AI technologiesHow different industries are using AI, from marketing to creative projectsPractical skills through labs using tools like ChatGPT, DALL-E, and Stable DiffusionHow It Stands OutYou'll hear from real AI practitioners about their experiencesThe course teaches you how to use generative AI to innovate and improve efficiency in your businessGet StartedEnroll in Generative AI: Introduction and Applications6. Generative AI for Everyone by Andrew Ng: Unlocking ProductivityWhy This Course Is a Must-HaveThis course focuses on using generative AI tools for everyday business tasks. It's all about boosting your productivity and efficiency.What You'll LearnHands-on exercises to integrate AI tools into your daily workReal examples of how businesses are using generative AI to save time and moneyTechniques for prompt engineering to get better results from AI toolsHow It Helps EntrepreneursYou'll learn to automate repetitive tasks, freeing up time for strategic thinkingYou'll discover new ways to use AI tools in your business processesYou'll gain confidence in experimenting with AI to solve business challengesGet StartedGo deeper with DeepLearning.AI7. Generative AI for Business Leaders by LinkedIn LearningWhy This Course Focuses on Business ApplicationsThis course is tailored for leaders who want to integrate AI into their business operations. It provides practical insights for improving workflows and decision-making.What You'll LearnStrategies for using AI to optimize your business operationsHow to save time and resources with AI-powered toolsPractical methods for implementing AI in your company, regardless of sizeKey BenefitsThe course is designed for busy professionals, allowing you to learn at your own paceYou'll gain insights you can apply immediately to your businessIt covers both the potential and the limitations of AI in business settingsGet StartedLevel up on LinkedIn Learning8. AI for Beginners by Microsoft: A Structured Learning PathWhy This Course Builds a Strong AI FoundationMicrosoft's AI for Beginners is a comprehensive 12-week program. It covers core AI concepts in a structured, easy-to-follow format. The course combines theoretical knowledge with hands-on practice through quizzes and labs.What You'll LearnThe basics of AI, machine learning, and data scienceStep-by-step guidance to build a strong knowledge basePractical applications of AI in various business contextsHow to Approach This CourseDedicate 2-3 hours per week to complete the curriculumUse the structured format to gradually build your confidence in AI conceptsApply what you learn to real business scenarios as you progressGet StartedBuild foundations with Microsoft9. AI for Business Specialization by UPenn: Strategic Thinking with AIWhy This Course Is Perfect for Business LeadersThis specialization focuses on AI's transformative impact on core business functions. It covers how AI is changing marketing, finance, and operations.What You'll LearnHow to build an AI strategy tailored to your business needsWays to leverage AI to drive innovation across different departmentsTechniques for integrating AI into your business modelHow to Make the Most of This CourseTake detailed notes on how each module applies to your own business challengesUse the specialization to develop a long-term AI vision for your companyNetwork with other business leaders taking the course to share insights and experiencesGet StartedScale up with UPenn's business focusConclusion: Your Path to Becoming an AI-powered EntrepreneurWe've covered nine fantastic free courses that can transform you into an AI-savvy entrepreneur. Let's recap:Google AI Essentials: Perfect for beginners, focusing on practical AI applications.Introduction to Generative AI: A quick start to understand the basics of generative AI.Generative AI with Large Language Models: For those ready to dive into the technical side.AI for Everyone: A non-technical introduction to AI's business impact.Generative AI: Introduction and Applications: A broad look at generative AI across industries.Generative AI for Everyone: Focused on boosting productivity with AI tools.Generative AI for Business Leaders: Tailored for integrating AI into business operations.AI for Beginners: A structured path to build a strong AI foundation.AI for Business Specialization: Strategic thinking about AI in business functions.Remember, you don't need to tackle all these courses at once. Start small and build your knowledge gradually. Pick the course that aligns best with your current needs and business goals.Embracing AI is not just about staying competitive; it's about opening new doors for innovation and growth. These courses will help you see opportunities where AI can solve problems, improve efficiency, and create value for your business.The AI revolution is happening now. The sooner you start learning, the better positioned you'll be to lead in this new era. Each step you take in understanding AI is a step towards future-proofing your business.So, what are you waiting for? Choose a course, dive in, and start your journey to becoming an AI-powered entrepreneur today. The future of your business may depend on it.MORE ARTICLES FOR YOUHumanizzer Fastpass Bundle – OTO1 to OTO4: Get (Humanizzer + All OTOs) Fastpass for Massive 75% Discount Available Limited-Time OneHumanizzer Review: Build Lifelike Human AI Agents That Talk, Listen & Engage Face-To-Face!—In Your Voice, Just Like You!EasyListDetox App Review: A Windows tool with Giveaway Rights for effortlessly cleaning your email lists of duplicates, invalid, and disposable addresses. Simple, efficient, and time-savingAI Copy Kit Review: Google’s Latest AI Tech Tensorflow (Tf) Create Jaw-Dropping And Advanced Ultra HD Videos, Ultra Shorts, 4K Images, Voiceovers, and Any Other GPT 4-Powered Amazing Content In Minutes Without Any Complicated Tools!From Good to Great: 15 Books to Inspire Personal and Business TransformationFTC Affiliate Commission Disclaimer: Some links in this article may earn us a commission if you make a purchase. This doesn't affect our recommendations.

internet-tools-collection
github
LLM Vibe Score0.236
Human Vibe Score0.009333333333333334
bogdanmosicaJan 23, 2025

internet-tools-collection

Internet Tools Collection A collection of tools, website and AI for entrepreneurs, web designers, programmers and for everyone else. Content by category Artificial Intelligence Developers Design Entrepreneur Video Editing Stock videos Stock Photos Stock music Search Engine Optimization Blog Posts Resume Interviews No code website builder No code game builder Side Hustle Browser Extensions Other Students Artificial Intelligence Jasper - The Best AI Writing Assistant [](https://www.jasper.ai/) Create content 5x faster with artificial intelligence. Jasper is the highest quality AI copywriting tool with over 3,000 5-star reviews. Best for writing blog posts, social media content, and marketing copy. AutoDraw [](https://www.autodraw.com/) Fast drawing for everyone. AutoDraw pairs machine learning with drawings from talented artists to help you draw stuff fast. Rytr - Best AI Writer, Content Generator & Writing Assistant [](https://rytr.me/) Rytr is an AI writing assistant that helps you create high-quality content, in just a few seconds, at a fraction of the cost! Neevo - Neevo [](https://www.neevo.ai/) Kinetix Tech [](https://kinetix.tech/) Kinetix is a no-code 3D creation tool powered by Artificial Intelligence. The web-based platform leverages AI motion capture to convert a video into a 3D animation and lets you customize your avatars and environments. We make 3D animation accessible to every creator so they can create engaging stories. LALAL.AI: 100% AI-Powered Vocal and Instrumental Tracks Remover [](https://www.lalal.ai/) Split vocal and instrumental tracks quickly and accurately with LALAL.AI. Upload any audio file and receive high-quality extracted tracks in a few seconds. Copy.ai: Write better marketing copy and content with AI [](https://www.copy.ai/) Get great copy that sells. Copy.ai is an AI-powered copywriter that generates high-quality copy for your business. Get started for free, no credit card required! Marketing simplified! OpenAI [](https://openai.com/) OpenAI is an AI research and deployment company. Our mission is to ensure that artificial general intelligence benefits all of humanity. DALL·E 2 [](https://openai.com/dall-e-2/) DALL·E 2 is a new AI system that can create realistic images and art from a description in natural language. Steve.ai - World’s fastest way to create Videos [](https://www.steve.ai/) Steve.AI is an online Video making software that helps anyone to create Videos and animations in seconds. Octie.ai - Your A.I. ecommerce marketing assistant [](https://octie.ai/) Write emails, product descriptions, and more, with A.I. Created by Octane AI. hypnogram.xyz [](https://hypnogram.xyz/) Generate images from text descriptions using AI FakeYou. Deep Fake Text to Speech. [](https://fakeyou.com/) FakeYou is a text to speech wonderland where all of your dreams come true. Craiyon, formerly DALL-E mini [](https://www.craiyon.com/) Craiyon, formerly DALL-E mini, is an AI model that can draw images from any text prompt! Deck Rocks - Create Pictch Decks [](https://www.deck.rocks/) Writely | Using AI to Improve Your Writing [](https://www.writelyai.com/) Making the art of writing accessible to all Writesonic AI Writer - Best AI Writing Assistant [](https://writesonic.com/) Writesonic is an AI writer that's been trained on top-performing SEO content, high-performing ads, and converting sales copy to help you supercharge your writing and marketing efforts. Smart Copy - AI Copywriting Assistant | Unbounce [](https://unbounce.com/product/smart-copy/) Generate creative AI copy on-the-spot across your favourite tools Synthesia | #1 AI Video Generation Platform [](https://www.synthesia.io/) Create AI videos by simply typing in text. Easy to use, cheap and scalable. Make engaging videos with human presenters — directly from your browser. Free demo. NVIDIA Canvas: Turn Simple Brushstrokes into Realistic Images [](https://www.nvidia.com/en-us/studio/canvas/) Create backgrounds quickly, or speed up your concept exploration so you can spend more time visualizing ideas with the help of NVIDIA Canvas. Hotpot.ai - Hotpot.ai [](https://hotpot.ai/) Hotpot.ai makes graphic design and image editing easy. AI tools allow experts and non-designers to automate tedious tasks while attractive, easy-to-edit templates allow anyone to create device mockups, social media posts, marketing images, app icons, and other work graphics. Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. Search listening tool for market, customer & content research - AnswerThePublic [](https://answerthepublic.com/) Use our free tool to get instant, raw search insights, direct from the minds of your customers. Upgrade to a paid plan to monitor for new ways that people talk & ask questions about your brand, product or topic. Topic Mojo [](https://topicmojo.com/) Discover unique & newest queries around any topic and find what your customers are searching for. Pulling data from 50+ sources to enhance your topic research. AI Image Enlarger | Enlarge Image Without Losing Quality! [](https://imglarger.com/) AI Image Enlarger is a FREE online image enlarger that could upscale and enhance small images automatically. Make jpg/png pictures big without losing quality. Midjourney [](https://www.midjourney.com/app/) Kaedim - AI for turning 2D images to 3D models [](https://www.kaedim3d.com/webapp) AI for turning 2D images, sketches and photos to 3D models in seconds. Overdub: Ultra realistic text to speech voice cloning - Descript [](https://www.descript.com/overdub) Create a text to speech model of your voice. Try a live demo. Getting Started [](https://magenta.tensorflow.org/get-started) Resources to learn about Magenta Photosonic AI Art Generator | Create Unique Images with AI [](https://photosonic.writesonic.com/) Transform your imagination into stunning digital art with Photosonic - the AI art generator. With its creative suggestions, this Writesonic's AI image generator can help unleash your inner artist and share your creations with the world. Image Computer [](https://image.computer/) Most downloaded Instagram Captions App (+more creator tools) [](https://captionplus.app/) Join 3 Million+ Instagram Creators who use CaptionPlus to find Instagram Captions, Hashtags, Feed Planning, Reel Ideas, IG Story Design and more. Writecream - Best AI Writer & Content Generator - Writecream [](https://www.writecream.com/) Sentence Rewriter is a free tool to reword a sentence, paragraph and even entire essays in a short amount of time. Hypotenuse AI: AI Writing Assistant and Text Generator [](https://www.hypotenuse.ai/) Turn a few keywords into original, insightful articles, product descriptions and social media copy with AI copywriting—all in just minutes. Try it free today. Text to Speach Listnr: Generate realistic Text to Speech voiceovers in seconds [](https://www.listnr.tech/) AI Voiceover Generator with over 600+ voiceovers in 80+ languages, go from Text to Voice in seconds. Get started for Free! Free Text to Speech: Online, App, Software, Commercial license with Natural Sounding Voices. [](https://www.naturalreaders.com/) Free text to speech online app with natural voices, convert text to audio and mp3, for personal and commercial use Developers OverAPI.com | Collecting all the cheat sheets [](https://overapi.com/) OverAPI.com is a site collecting all the cheatsheets,all! Search Engine For Devs [](https://you.com/) Spline - Design tool for 3D web browser experiences [](https://spline.design/) Create web-based 3D browser experiences Image to HTML CSS converter. Convert image to HTML CSS with AI: Fronty [](https://fronty.com/) Fronty - Image to HTML CSS code converter. Convert image to HTML powered by AI. Sketchfab - The best 3D viewer on the web [](https://sketchfab.com/) With a community of over one million creators, we are the world’s largest platform to publish, share, and discover 3D content on web, mobile, AR, and VR. Railway [](https://railway.app/) Railway is an infrastructure platform where you can provision infrastructure, develop with that infrastructure locally, and then deploy to the cloud. JSON Crack - Crack your data into pieces [](https://jsoncrack.com/) Simple visualization tool for your JSON data. No forced structure, paste your JSON and view it instantly. Locofy.ai - ship your products 3-4x faster — with low code [](https://www.locofy.ai/) Turn your designs into production-ready frontend code for mobile apps and web. Ship products 3-4x faster with your existing design tools, tech stacks & workflows. Oh Shit, Git!?! [](https://ohshitgit.com/) Carbon | Create and share beautiful images of your source code [](https://carbon.now.sh/) Carbon is the easiest way to create and share beautiful images of your source code. GPRM : GitHub Profile ReadMe Maker [](https://gprm.itsvg.in/) Best Profile Generator, Create your perfect GitHub Profile ReadMe in the best possible way. Lots of features and tools included, all for free ! HubSpot | Software, Tools, and Resources to Help Your Business Grow Better [](https://www.hubspot.com/) HubSpot’s integrated CRM platform contains the marketing, sales, service, operations, and website-building software you need to grow your business. QuickRef.ME - Quick Reference Cheat Sheet [](https://quickref.me/) Share quick reference and cheat sheet for developers massCode | A free and open source code snippets manager for developers [](https://masscode.io/) Code snippets manager for developers, developed using web technologies. Snyk | Developer security | Develop fast. Stay secure. [](https://snyk.io/) Snyk helps software-driven businesses develop fast and stay secure. Continuously find and fix vulnerabilities for npm, Maven, NuGet, RubyGems, PyPI and more. Developer Roadmaps [](https://roadmap.sh/) Community driven roadmaps, articles, guides, quizzes, tips and resources for developers to learn from, identify their career paths, know what they don't know, find out the knowledge gaps, learn and improve. CSS Generators Get Waves – Create SVG waves for your next design [](https://getwaves.io/) A free SVG wave generator to make unique SVG waves for your next web design. Choose a curve, adjust complexity, randomize! Box Shadows [](https://box-shadow.dev/) Tridiv | CSS 3D Editor [](http://tridiv.com/) Tridiv is a web-based editor for creating 3D shapes in CSS Glassmorphism CSS Generator - Glass UI [](https://ui.glass/generator/) Generate CSS and HTML components using the glassmorphism design specifications based on the Glass UI library. Blobmaker - Make organic SVG shapes for your next design [](https://www.blobmaker.app/) Make organic SVG shapes for your next design. Modify the complexity, contrast, and color, to generate unique SVG blobs every time. Keyframes.app [](https://keyframes.app/) cssFilters.co - Custom and Instagram like photo filters for CSS [](https://www.cssfilters.co/) Visual playground for generating CSS for custom and Instagram like photo filters. Experiment with your own uploaded photo or select one from the Unsplash collection. CSS Animations Animista - CSS Animations on Demand [](https://animista.net/) Animista is a CSS animation library and a place where you can play with a collection of ready-made CSS animations and download only those you will use. Build Internal apps Superblocks | Save 100s of developer hours on internal tools [](https://www.superblocks.com/) Superblocks is the fast, easy and secure way for developers to build custom internal tools fast. Connect your databases & APIs. Drag and drop UI components. Extend with Python or Javascript. Deploy in 1-click. Secure and Monitor using your favorite tools Budibase | Build internal tools in minutes, the easy way [](https://budibase.com/) Budibase is a modern, open source low-code platform for building modern internal applications in minutes. Retool | Build internal tools, remarkably fast. [](https://retool.com/) Retool is the fast way to build internal tools. Drag-and-drop our building blocks and connect them to your databases and APIs to build your own tools, instantly. Connects with Postgres, REST APIs, GraphQL, Firebase, Google Sheets, and more. Built by developers, for developers. Trusted by startups and Fortune 500s. Sign up for free. GitHub Repositories GitHub - vasanthk/how-web-works: What happens behind the scenes when we type www.google.com in a browser? [](https://github.com/vasanthk/how-web-works) What happens behind the scenes when we type www.google.com in a browser? - GitHub - vasanthk/how-web-works: What happens behind the scenes when we type www.google.com in a browser? GitHub - kamranahmedse/developer-roadmap: Interactive roadmaps, guides and other educational content to help developers grow in their careers. [](https://github.com/kamranahmedse/developer-roadmap) Interactive roadmaps, guides and other educational content to help developers grow in their careers. - GitHub - kamranahmedse/developer-roadmap: Interactive roadmaps, guides and other educational content to help developers grow in their careers. GitHub - apptension/developer-handbook: An opinionated guide on how to become a professional Web/Mobile App Developer. [](https://github.com/apptension/developer-handbook) An opinionated guide on how to become a professional Web/Mobile App Developer. - GitHub - apptension/developer-handbook: An opinionated guide on how to become a professional Web/Mobile App Developer. ProfileMe.dev | Create an amazing GitHub profile in minutes [](https://www.profileme.dev/) ProfileMe.dev | Create an amazing GitHub profile in minutes GitHub - Kristories/awesome-guidelines: A curated list of high quality coding style conventions and standards. [](https://github.com/Kristories/awesome-guidelines) A curated list of high quality coding style conventions and standards. - GitHub - Kristories/awesome-guidelines: A curated list of high quality coding style conventions and standards. GitHub - tiimgreen/github-cheat-sheet: A list of cool features of Git and GitHub. [](https://github.com/tiimgreen/github-cheat-sheet) A list of cool features of Git and GitHub. Contribute to tiimgreen/github-cheat-sheet development by creating an account on GitHub. GitHub - andreasbm/web-skills: A visual overview of useful skills to learn as a web developer [](https://github.com/andreasbm/web-skills) A visual overview of useful skills to learn as a web developer - GitHub - andreasbm/web-skills: A visual overview of useful skills to learn as a web developer GitHub - Ebazhanov/linkedin-skill-assessments-quizzes: Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers [](https://github.com/Ebazhanov/linkedin-skill-assessments-quizzes) Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers - GitHub - Ebazhanov/linkedin-skill-assessments-quizzes: Full reference of LinkedIn answers 2022 for skill assessments (aws-lambda, rest-api, javascript, react, git, html, jquery, mongodb, java, Go, python, machine-learning, power-point) linkedin excel test lösungen, linkedin machine learning test LinkedIn test questions and answers Blockchain/Crypto Dashboards [](https://dune.com/) Blockchain ecosystem analytics by and for the community. Explore and share data from Ethereum, xDai, Polygon, Optimism, BSC and Solana for free. Introduction - The Anchor Book v0.24.0 [](https://book.anchor-lang.com/introduction/introduction.html) Crypto & Fiat Exchange Super App | Trade, Save & Spend | hi [](https://hi.com/) Buy, Trade, Send and Earn Crypto & Fiat. Deposit Bitcoin, ETH, USDT and other cryptos and start earning. Get the hi Debit Card and Multi-Currency IBAN Account. Moralis Web3 - Enterprise-Grade Web3 APIs [](https://moralis.io/) Bridge the development gap between Web2 and Web3 with Moralis’ powerful Web3 APIs. Mirror [](https://mirror.xyz/) Built on web3 for web3, Mirror’s robust publishing platform pushes the boundaries of writing online—whether it’s the next big white paper or a weekly community update. Makerdao [](https://blog.makerdao.com/) Sholi — software for Investors & Traders / Sholi MetriX [](https://sholi.io/) Sholi — software for Investors & Traders / Sholi MetriX Stock Trading Quiver Quantitative [](https://www.quiverquant.com/) Quiver Quantitative Chart Prime - The only tool you'll need for trading assets across all markets [](https://chartprime.com/) ChartPrime offers a toolkit that will take your trading game to the next level. Visit our site for a full rundown of features and helpful tutorials. Learning Hacker Rank [](https://www.hackerrank.com/) Coderbyte | Code Screening, Challenges, & Interview Prep [](https://coderbyte.com/) Improve your coding skills with our library of 300+ challenges and prepare for coding interviews with content from leading technology companies. Competitive Programming | Participate & Learn | CodeChef [](https://www.codechef.com/) Learn competitive programming with the help of CodeChef's coding competitions. Take part in these online coding contests to level up your skills Learn to Code - for Free | Codecademy [](https://www.codecademy.com/) Learn the technical skills to get the job you want. Join over 50 million people choosing Codecademy to start a new career (or advance in their current one). Free Code Camp [](https://www.freecodecamp.org/) Learn to Code — For Free Sololearn: Learn to Code [](https://www.sololearn.com/home) Join Now to learn the basics or advance your existing skills Mimo: The coding app you need to learn to code! Python, HTML, JavaScript [](https://getmimo.com/) Join more than 17 million learners worldwide. Learn to code for free. Learn Python, JavaScript, CSS, SQL, HTML, and more with our free code learning app. Free for developers [](https://free-for.dev/#/) Your Career in Web Development Starts Here | The Odin Project [](https://www.theodinproject.com/) The Odin Project empowers aspiring web developers to learn together for free Code Learning Games CheckiO - coding games and programming challenges for beginner and advanced [](https://checkio.org/) CheckiO - coding websites and programming games. Improve your coding skills by solving coding challenges and exercises online with your friends in a fun way. Exchanges experience with other users online through fun coding activities Coding for Kids | Game-Based Programming | CodeMonkey [](https://www.codemonkey.com/) CodeMonkey is a leading coding for kids program. Through its award-winning courses, millions of students learn how to code in real programming languages. Coding Games and Programming Challenges to Code Better [](https://www.codingame.com/) CodinGame is a challenge-based training platform for programmers where you can play with the hottest programming topics. Solve games, code AI bots, learn from your peers, have fun. Learn VIM while playing a game - VIM Adventures [](https://vim-adventures.com/) VIM Adventures is an online game based on VIM's keyboard shortcuts. It's the "Zelda meets text editing" game. So come have some fun and learn some VIM! CodeCombat - Coding games to learn Python and JavaScript [](https://codecombat.com/) Learn typed code through a programming game. Learn Python, JavaScript, and HTML as you solve puzzles and learn to make your own coding games and websites. Design Useberry - Codeless prototype analytics [](https://www.useberry.com/) User testing feedback & rich insights in minutes, not months! Figma: the collaborative interface design tool. [](https://www.figma.com/) Build better products as a team. Design, prototype, and gather feedback all in one place with Figma. Dribbble - Discover the World’s Top Designers & Creative Professionals [](https://dribbble.com/) Find Top Designers & Creative Professionals on Dribbble. We are where designers gain inspiration, feedback, community, and jobs. Your best resource to discover and connect with designers worldwide. Photopea | Online Photo Editor [](https://www.photopea.com/) Photopea Online Photo Editor lets you edit photos, apply effects, filters, add text, crop or resize pictures. Do Online Photo Editing in your browser for free! Toools.design – An archive of 1000+ Design Resources [](https://www.toools.design/) A growing archive of over a thousand design resources, weekly updated for the community. Discover highly useful design tools you never thought existed. All Online Tools in One Box | 10015 Tools [](https://10015.io/) All online tools you need in one box for free. Build anything online with “all-in-one toolbox”. All tools are easy-to-use, blazing fast & free. Phase - Digital Design Reinvented| Phase [](https://phase.com/) Design and prototype websites and apps visually and intuitively, in a new powerful product reworked for the digital age. Animated Backgrounds [](https://animatedbackgrounds.me/) A Collection of 30+ animated backgrounds for websites and blogs.With Animated Backgrounds, set a simple, elegant background animations on your websites and blogs. Trianglify.io · Low Poly Pattern Generator [](https://trianglify.io/) Trianglify.io is a tool for generating low poly triangle patterns that can be used as wallpapers and website assets. Cool Backgrounds [](https://coolbackgrounds.io/) Explore a beautifully curated selection of cool backgrounds that you can add to blogs, websites, or as desktop and phone wallpapers. SVG Repo - Free SVG Vectors and Icons [](https://www.svgrepo.com/) Free Vectors and Icons in SVG format. ✅ Download free mono or multi color vectors for commercial use. Search in 300.000+ Free SVG Vectors and Icons. Microcopy - Short copy text for your website. [](https://www.microcopy.me/) Search micro UX copy text: slogans, headlines, notifications, CTA, error messages, email, account preferences, and much more. 3D icons and icon paks - Free3Dicon [](https://free3dicon.com/) All 3D icons you need in one place. This is a collection of free, beautiful, trending 3D icons, that you can use in any project. Love 3D Icon [](https://free3dicons.com/) Downloads free 3D icons GIMP - GNU Image Manipulation Program [](https://www.gimp.org/) GIMP - The GNU Image Manipulation Program: The Free and Open Source Image Editor blender.org - Home of the Blender project - Free and Open 3D Creation Software [](https://www.blender.org/) The Freedom to Create 3D Design Software | 3D Modeling on the Web | SketchUp [](https://www.sketchup.com/) SketchUp is a premier 3D design software that truly makes 3D modeling for everyone, with a simple to learn yet robust toolset that empowers you to create whatever you can imagine. Free Logo Maker - Create a Logo in Seconds - Shopify [](https://www.shopify.com/tools/logo-maker) Free logo maker tool to generate custom design logos in seconds. This logo creator is built for entrepreneurs on the go with hundreds of templates, free vectors, fonts and icons to design your own logo. The easiest way to create business logos online. All your design tools in one place | Renderforest [](https://www.renderforest.com/) Time to get your brand noticed. Create professional videos, logos, mockups, websites, and graphics — all in one place. Get started now! Prompt Hero [](https://prompthero.com/) Type Scale - A Visual Calculator [](https://type-scale.com/) Preview and choose the right type scale for your project. Experiment with font size, scale and different webfonts. DreamFusion: Text-to-3D using 2D Diffusion [](https://dreamfusion3d.github.io/) DreamFusion: Text-to-3D using 2D Diffusion, 2022. The branding style guidelines documents archive [](https://brandingstyleguides.com/) Welcome to the brand design manual documents directory. Search over our worldwide style assets handpicked collection, access to PDF documents for inspiration. Super designer | Create beautiful designs with a few clicks [](https://superdesigner.co/) Create beautiful designs with a few clicks. Simple design tools to generate unique patterns, backgrounds, 3D shapes, colors & images for social media, websites and more Readymag—a design tool to create websites without coding [](https://readymag.com/) Meet the most elegant, simple and powerful web-tool for designing websites, presentations, portfolios and all kinds of digital publications. ffflux: Online SVG Fluid Gradient Background Generator | fffuel [](https://fffuel.co/ffflux/) SVG generator to make fluid gradient backgrounds that feel organic and motion-like. Perfect to add a feeling of motion and fluidity to your web designs. Generate unique SVG design assets | Haikei [](https://haikei.app/) A web-based design tool to generate unique SVG design assets for websites, social media, blog posts, desktop and mobile wallpapers, posters, and more! Our generators let you discover, customize, randomize, and export generative SVG design assets ready to use with your favorite design tools. UI/UX - Inspirational Free Website Builder Software | 10,000+ Free Templates [](https://nicepage.com/) Nicepage is your website builder software breaking limitations common for website builders with revolutionary freehand positioning. 7000+ Free Templates. Easy Drag-n-Drop. No coding. Mobile-friendly. Clean HTML. Super designer | Create beautiful designs with a few clicks [](https://superdesigner.co/) Create beautiful designs with a few clicks. Simple design tools to generate unique patterns, backgrounds, 3D shapes, colors & images for social media, websites and more Pika – Create beautiful mockups from screenshots [](https://pika.style/) Quickly create beautiful website and device mockup from screenshot. Pika lets you capture website screenshots form URL, add device and browser frames, customize background and more LiveTerm [](https://liveterm.vercel.app/) Minimal Gallery – Web design inspiration [](https://minimal.gallery/) For the love of beautiful, clean and functional websites. Awwwards - Website Awards - Best Web Design Trends [](https://www.awwwards.com/) Awwwards are the Website Awards that recognize and promote the talent and effort of the best developers, designers and web agencies in the world. Design Systems For Figma [](https://www.designsystemsforfigma.com/) A collection of Design Systems for Figma from all over the globe. Superside: Design At Scale For Ambitious Brands [](https://www.superside.com/) We are an always-on design company. Get a team of dedicated designers, speedy turnarounds, magical creative collaboration tech and the top 1% of global talent. UXArchive - Made by Waldo [](https://uxarchive.com/) UXArchive the world's largest library of mobile user flows. Be inspired to design the best user experiences. Search by Muzli [](https://search.muz.li/) Search, discover, test and create beautiful color palettes for your projects Siteinspire | Web Design Inspiration [](https://www.siteinspire.com/) SAVEE [](https://savee.it/) The best way to save and share inspiration. A little corner of the internet to find good landing page copywriting examples [](https://greatlandingpagecopy.com/) A little corner of the internet to find great landing page copywriting examples. The Best Landing Page Examples For Design Inspiration - SaaS Landing Page [](https://saaslandingpage.com/) SaaS Landing Page showcases the best landing page examples created by top-class SaaS companies. Get ideas and inspirations for your next design project. Websites Free templates Premium Bootstrap Themes and Templates: Download @ Creative Tim [](https://www.creative-tim.com/) UI Kits, Templates and Dashboards built on top of Bootstrap, Vue.js, React, Angular, Node.js and Laravel. Join over 2,014,387+ creatives to access all our products! Free Bootstrap Themes, Templates, Snippets, and Guides - Start Bootstrap [](https://startbootstrap.com/) Start Bootstrap develops free to download, open source Bootstrap 5 themes, templates, and snippets and creates guides and tutorials to help you learn more about designing and developing with Bootstrap. Free Website Templates [](https://freewebsitetemplates.com/) Get your free website templates here and use them on your website without needing to link back to us. One Page Love - One Page Website Inspiration and Templates [](https://onepagelove.com/) One Page Love is a One Page website design gallery showcasing the best Single Page websites, templates and resources. Free CSS | 3400 Free Website Templates, CSS Templates and Open Source Templates [](https://www.free-css.com/) Free CSS has 3400 free website templates, all templates are free CSS templates, open source templates or creative commons templates. Free Bootstrap Themes and Website Templates | BootstrapMade [](https://bootstrapmade.com/) At BootstrapMade, we create beautiful website templates and bootstrap themes using Bootstrap, the most popular HTML, CSS and JavaScript framework. Free and Premium Bootstrap Themes, Templates by Themesberg [](https://themesberg.com/) Free and Premium Bootstrap themes, templates, admin dashboards and UI kits used by over 38820 web developers and software companies HTML, Vue.js and React templates for startup landing pages - Cruip [](https://cruip.com/) Cruip is a gallery of premium and free HTML, Vue.js and React templates for startups and SaaS. Free Website Templates Download | WordPress Themes - W3Layouts [](https://w3layouts.com/) Want to download free website templates? W3Layouts WordPress themes and website templates are built with responsive web design techniques. Download now! Free HTML Landing Page Templates and UI Kits | UIdeck [](https://uideck.com/) Free HTML Landing Page Templates, Bootstrap Themes, React Templates, HTML Templates, Tailwind Templates, and UI Kits. Create Online Graphics Snappa - Quick & Easy Graphic Design Software [](https://snappa.com/) Snappa makes it easy to create any type of online graphic. Create & publish images for social media, blogs, ads, and more! Canva [](https://www.canva.com/) Polotno Studio - Make graphical designs [](https://studio.polotno.com) Free online design editor. Create images for social media, youtube previews, facebook covers Free Logo Maker: Design Custom Logos | Adobe Express [](https://www.adobe.com/express/create/logo) The Adobe Express logo maker is instant, intuitive, and intelligent. Use it to generate a wide range of possibilities for your own logo. Photo Editor: Fotor – Free Online Photo Editing & Image Editor [](https://www.fotor.com/) Fotor's online photo editor helps you edit photos with free online photo editing tools. Crop photos, resize images, and add effects/filters, text, and graphics in just a few clicks. Photoshop online has never been easier with Fotor's free online photo editor. VistaCreate – Free Graphic Design Software with 70,000+ Free Templates [](https://create.vista.com/) Looking for free graphic design software? Easily create professional designs with VistaCreate, a free design tool with powerful features and 50K+ ready-made templates Draw Freely | Inkscape [](https://inkscape.org/) Inkscape is professional quality vector graphics software which runs on Linux, Mac OS X and Windows desktop computers. Visual & Video Maker Trusted By 11 Million Users - Piktochart [](https://piktochart.com/) With Piktochart, you can create professional-looking infographics, flyers, posters, charts, videos, and more. No design experience needed. Start for free. The Web's Favorite Online Graphic Design Tool | Stencil [](https://getstencil.com/) Stencil is a fantastically easy-to-use online graphic design tool and image editor built for business owners, social media marketers, and bloggers. Pablo by Buffer - Design engaging images for your social media posts in under 30 seconds [](https://pablo.buffer.com/) Buffer makes it super easy to share any page you're reading. Keep your Buffer topped up and we automagically share them for you through the day. Free Online Graphic Design Software | Create stunning designs in seconds. [](https://desygner.com/) Easy drag and drop graphic design tool for anyone to use with 1000's of ready made templates. Create & print professional business cards, flyers, social posts and more. Color Pallet Color Palettes for Designers and Artists - Color Hunt [](https://colorhunt.co/) Discover the newest hand-picked color palettes of Color Hunt. Get color inspiration for your design and art projects. Coolors - The super fast color palettes generator! [](https://coolors.co/) Generate or browse beautiful color combinations for your designs. Get color palette inspiration from nature - colorpalettes.earth [](https://colorpalettes.earth/) Color palettes inspired by beautiful nature photos Color Palette Generator - Create Beautiful Color Schemes [](https://colors.muz.li/) Search, discover, test and create beautiful color palettes for your projects A Most Useful Color Picker | 0to255 [](https://0to255.com/) Find lighter and darker colors based on any color. Discover why over two million people have used 0to255 to choose colors for their website, logo, room interior, and print design projects. Colour Contrast Checker [](https://colourcontrast.cc/) Check the contrast between different colour combinations against WCAG standards Fonts Google Fonts [](https://fonts.google.com/) Making the web more beautiful, fast, and open through great typography Fonts In Use – Type at work in the real world. [](https://fontsinuse.com/) A searchable archive of typographic design, indexed by typeface, format, and topic. Wordmark - Helps you choose fonts! [](https://wordmark.it/) Wordmark helps you choose fonts by quickly displaying your text with your fonts. OH no Type Company [](https://ohnotype.co/) OH no Type Co. Retail and custom typefaces. Life’s a thrill, fonts are chill! Illustrations Illustrations | unDraw [](https://undraw.co/illustrations) The design project with open-source illustrations for any idea you can imagine and create. Create beautiful websites, products and applications with your color, for free. Design Junction [](https://designjunction.xyz/) Design Junction is a one-stop resource library for Designers and Creatives with curated list of best resources handpicked from around the web Humaaans: Mix-&-Match illustration library [](https://www.humaaans.com/) Mix-&-match illustrations of people with a design library for InVIsion Studio and Sketch. Stubborn - Free Illustrations Generator [](https://stubborn.fun/) Free illustrations generator for Figma and Sketch. Get the opportunity to design your characters using symbols and styles. Open Peeps, Hand-Drawn Illustration Library [](https://www.openpeeps.com/) Open Peeps is a hand-drawn illustration library to create scenes of people. You can use them in product illustration, marketing, comics, product states, user flows, personas, storyboarding, quinceañera invitations, or whatever you want! ⠀ Reshot | Free icons & illustrations [](https://www.reshot.com/) Design freely with instant downloads of curated SVG icons and vector illustrations. All free with commercial licensing. No attribution required. Blush: Illustrations for everyone [](https://blush.design/) Blush makes it easy to add free illustrations to your designs. Play with fully customizable graphics made by artists across the globe. Mockups Angle 4 - 5000+ Device Mockups for Figma, Sketch and XD [](https://angle.sh/) Vector mockups for iPhone, iPad, Android and Mac devices, including the new iPhone 13, Pro, Pro Max and Mini. Perfect for presenting your apps. Huge library of components, compositions, wallpapers and plugins made for Figma, Sketch and XD. Make Mockups, Logos, Videos and Designs in Seconds [](https://placeit.net/) Get unlimited downloads on all our 100K templates! You can make a logo, video, mockup, flyer, business card and social media image in seconds right from your browser. Free and premium tools for graphic designers | Lstore Graphics [](https://www.ls.graphics/) Free and premium mockups, UI/UX tools, scene creators for busy designers Logo Design & Brand Identity Platform for Entrepreneurs | Looka [](https://looka.com/) Logojoy is now Looka! Design a Logo, make a website, and create a Brand Identity you’ll love with the power of Artificial Intelligence. 100% free to use. Create stunning product mockups easily and online - Smartmockups [](https://smartmockups.com/) Smartmockups enables you to create stunning high-resolution mockups right inside your browser within one interface across multiple devices. Previewed - Free mockup generator for your app [](https://previewed.app/) Join Previewed to create stunning 3D image shots and animations for your app. Choose from hundreds of ready made mockups, or create your own. Free Design Software - Graphic Online Maker - Glorify [](https://www.glorify.com/) Create professional and high converting social media posts, ads, infographics, presentations, and more with Glorify, a free design software & graphic maker. Other BuiltWith Technology Lookup [](https://builtwith.com/) Web technology information profiler tool. Find out what a website is built with. Compress JPEG Images Online [](https://compressjpeg.com/) Compress JPEG images and photos for displaying on web pages, sharing on social networks or sending by email. PhotoRoom - Remove Background and Create Product Pictures [](https://www.photoroom.com/) Create product and portrait pictures using only your phone. Remove background, change background and showcase products. Magic Eraser - Remove unwanted things from images in seconds [](https://www.magiceraser.io/) Magic Eraser - Use AI to remove unwanted things from images in seconds. Upload an image, mark the bit you need removed, download the fixed up image. Compressor.io - optimize and compress JPEG photos and PNG images [](https://compressor.io/) Optimize and compress JPEG, PNG, SVG, GIF and WEBP images online. Compress, resize and rename your photos for free. Remove Video Background – Unscreen [](https://www.unscreen.com/) Remove the background of any video - 100% automatically, online & free! Goodbye Greenscreen. Hello Unscreen. Noun Project: Free Icons & Stock Photos for Everything [](https://thenounproject.com/) Noun Project features the most diverse collection of icons and stock photos ever. Download SVG and PNG. Browse over 5 million art-quality icons and photos. Design Principles [](https://principles.design/) An Open Source collection of Design Principles and methods Shapefest™ - A massive library of free 3D shapes [](https://www.shapefest.com/) A massive free library of beautifully rendered 3D shapes. 160,000+ high resolution PNG images in one cohesive library. Learning UX Degreeless.design - Everything I Learned in Design School [](https://degreeless.design/) This is a list of everything I've found useful in my journey of learning design, and an ongoing list of things I think you should read. For budding UX, UI, Interaction, or whatever other title designers. UX Tools | Practical UX skills and tools [](https://uxtools.co/) Lessons and resources from two full-time product designers. Built For Mars [](https://builtformars.com/) On a mission to help the world build better user experiences by demystifying UX. Thousands of hours of research packed into UX case studies. Case Study Club – Curated UX Case Study Gallery [](https://www.casestudy.club/) Case Study Club is the biggest curated gallery of the best UI/UX design case studies. Get inspired by industry-leading designers, openly sharing their UX process. The Guide to Design [](https://start.uxdesign.cc/) A self-guided class to help you get started in UX and answer key questions about craft, design, and career Uxcel - Where design careers are built [](https://app.uxcel.com/explore) Available on any device anywhere in the world, Uxcel is the best way to improve and learn UX design online in just 5 minutes per day. UI & UX Design Tips by Jim Raptis. [](https://www.uidesign.tips/) Learn UI & UX Design with practical byte-sized tips and in-depth articles from Jim Raptis. Entrepreneur Instant Username Search [](https://instantusername.com/#/) Instant Username Search checks out if your username is available on more than 100 social media sites. Results appear instantly as you type. Flourish | Data Visualization & Storytelling [](https://flourish.studio/) Beautiful, easy data visualization and storytelling PiPiADS - #1 TikTok Ads Spy Tool [](https://www.pipiads.com/) PiPiADS is the best tiktok ads spy tool .We provide tiktok advertising,advertising on tiktok,tiktok ads examples,tiktok ads library,tiktok ads best practices,so you can understand the tiktok ads cost and master the tiktok ads 2021 and tiktok ads manager. Minea - The best adspy for product search in ecommerce and dropshipping [](https://en.minea.com/) Minea is the ultimate e-commerce product search tool. Minea tracks all ads on all networks. Facebook Ads, influencer product placements, Snapspy, all networks are tracked. Stop paying adspy 149€ for one network and discover Minea. AdSpy [](https://adspy.com/) Google Trends [](https://trends.google.com/) ScoreApp: Advanced Quiz Funnel Marketing | Make a Quiz Today [](https://www.scoreapp.com/) ScoreApp makes quiz funnel marketing easy, so you can attract relevant warm leads, insightful data and increase your sales. Try for free today Mailmodo - Send Interactive Emails That Drive Conversions [](https://www.mailmodo.com/) Use Mailmodo to create and send interactive emails your customers love. Drive conversions and get better email ROI. Sign up for a free trial now. 185 Top E-Commerce Sites Ranked by User Experience Performance – Baymard Institute [](https://baymard.com/ux-benchmark) See the ranked UX performance of the 185 largest e-commerce sites in the US and Europe. The chart summarizes 50,000+ UX performance ratings. Metricool - Analyze, manage and measure your digital content [](https://metricool.com/) Social media scheduling, web analytics, link in bio and reporting. Metricool is free per live for one brand. START HERE Visualping: #1 Website change detection, monitoring and alerts [](https://visualping.io/) More than 1.5 millions users monitor changes in websites with Visualping, the No1 website change detection, website checker, webpage change monitoring and webpage change detection tool. Gumroad – Sell what you know and see what sticks [](https://gumroad.com/) Gumroad is a powerful, but simple, e-commerce platform. We make it easy to earn your first dollar online by selling digital products, memberships and more. Product Hunt – The best new products in tech. [](https://www.producthunt.com/) Product Hunt is a curation of the best new products, every day. Discover the latest mobile apps, websites, and technology products that everyone's talking about. 12ft Ladder [](https://12ft.io/) Show me a 10ft paywall, I’ll show you a 12ft ladder. namecheckr | Social and Domain Name Availability Search For Brand Professionals [](https://www.namecheckr.com/) Social and Domain Name Availability Search For Brand Professionals Excel AI Formula Generator - Excelformulabot.com [](https://excelformulabot.com/) Transform your text instructions into Excel formulas in seconds with the help of AI. Z-Library [](https://z-lib.org/) Global Print On Demand Platform | Gelato [](https://www.gelato.com/) Create and sell custom products online. With local production in 33 countries, easy integration, and 24/7 customer support, Gelato is an all-in-one platform. Freecycle: Front Door [](https://freecycle.org/) Free eBooks | Project Gutenberg [](https://www.gutenberg.org/) Project Gutenberg is a library of free eBooks. Convertio — File Converter [](https://convertio.co/) Convertio - Easy tool to convert files online. More than 309 different document, image, spreadsheet, ebook, archive, presentation, audio and video formats supported. Namechk [](https://namechk.com/) Crazy Egg Website — Optimization | Heatmaps, Recordings, Surveys & A/B Testing [](https://www.crazyegg.com/) Use Crazy Egg to see what's hot and what's not, and to know what your web visitors are doing with tools, such as heatmaps, recordings, surveys, A/B testing & more. Ifttt [](https://ifttt.com/) Also Asked [](https://alsoasked.com/) Business Name Generator - Easily create Brandable Business Names - Namelix [](https://namelix.com/) Namelix uses artificial intelligence to create a short, brandable business name. Search for domain availability, and instantly generate a logo for your new business Merch Informer [](https://merchinformer.com/) Headline Generator [](https://www.title-generator.com/) Title Generator: create 700 headlines with ONE CLICK: Content Ideas + Catchy Headlines + Ad Campaign E-mail Subject Lines + Emotional Titles. Simple - Efficient - One Click Make [](https://www.make.com/en) Create and add calculator widgets to your website | CALCONIC_ [](https://www.calconic.com/) Web calculator builder empowers you to choose from a pre-made templates or build your own calculator widgets from a scratch without any need of programming knowledge Boost Your Views And Subscribers On YouTube - vidIQ [](https://vidiq.com/) vidIQ helps you acquire the tools and knowledge needed to grow your audience faster on YouTube and beyond. Learn More Last Pass [](https://www.lastpass.com/) Starter Story: Learn How People Are Starting Successful Businesses [](https://www.starterstory.com/) Starter Story interviews successful entrepreneurs and shares the stories behind their businesses. In each interview, we ask how they got started, how they grew, and how they run their business today. How To Say No [](https://www.starterstory.com/how-to-say-no) Saying no is hard, but it's also essential for your sanity. Here are some templates for how to say no - so you can take back your life. Think with Google - Discover Marketing Research & Digital Trends [](https://www.thinkwithgoogle.com/) Uncover the latest marketing research and digital trends with data reports, guides, infographics, and articles from Think with Google. ClickUp™ | One app to replace them all [](https://clickup.com/) Our mission is to make the world more productive. To do this, we built one app to replace them all - Tasks, Docs, Goals, and Chat. The Manual [](https://manual.withcompound.com/) Wealth-planning resources for founders and startup employees Software for Amazon FBA Sellers & Walmart Sellers | Helium 10 [](https://www.helium10.com/) If you're looking for the best software for Amazon FBA & Walmart sellers on the market, check out Helium 10's capabilities online today! Buffer: All-you-need social media toolkit for small businesses [](https://buffer.com/) Use Buffer to manage your social media so that you have more time for your business. Join 160,000+ small businesses today. CPGD — The Consumer Packaged Goods Directory [](https://www.cpgd.xyz/) The Consumer Packaged Goods Directory is a platform to discover new brands and resources. We share weekly trends in our newsletter and partner with services to provide vetted, recommended platforms for our Directory brands. Jungle Scout [](https://www.junglescout.com/) BuzzSumo | The World's #1 Content Marketing Platform [](https://buzzsumo.com/) BuzzSumo powers the strategies of 500k+ marketers, with content marketing data on 8b articles, 42m websites, 300t engagements, 500k journalists & 492m questions. Login - Capital [](https://app.capital.xyz/) Raise, hold, spend, and send funds — all in one place. Marketing Pictory – Video Marketing Made Easy - Pictory.ai [](https://pictory.ai/) Pictory's powerful AI enables you to create and edit professional quality videos using text, no technical skills required or software to download. Tolstoy | Communicate with interactive videos [](https://www.gotolstoy.com/) Start having face-to-face conversations with your customers. Create Email Marketing Your Audience Will Love - MailerLite [](https://www.mailerlite.com/) Email marketing tools to grow your audience faster and drive revenue smarter. Get free access to premium features with a 30-day trial! Sign up now! Hypefury - Schedule & Automate Social Media Marketing [](https://hypefury.com/) Save time on social media while creating more value, and growing your audience faster. Schedule & automate your social media experience! Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. Online Email & Lead Scraper | Klean Leads [](https://www.kleanleads.com/) Klean Leads is an online email scraper & email address finder. Use it to book more appointments, get more replies, and close more sales. PhantomBuster [](https://phantombuster.com/) Call to Action Examples - 300+ CTA Phrases [](https://ctaexamples.com/) See the best CTA example in every situation covered by the library of 300+ CTA goals. Use the examples to create your own CTAs in minutes. Creative Center: one-stop creative solution for TikTok [](https://ads.tiktok.com/business/creativecenter/pc/en?from=001010) Come to get your next great idea for TikTok. Here you can find the best performing ads, viral videos, and trending hashtags across regions and verticals. Groove.cm GrooveFunnels, GrooveMail with CRM and Digital Marketing Automation Platform - Groove.cm with GrooveFunnels, GroovePages, GrooveKart [](https://groove.cm/) Groove is a website creator, page builder, sales funnel maker, membership site platform, email autoresponder, blog tool, shopping cart system, ecommerce store solution, affiliate manager, video marketing software and more apps to help build your online business. SurveyMonkey: The World’s Most Popular Free Online Survey Tool [](https://www.surveymonkey.com/) Use SurveyMonkey to drive your business forward by using our free online survey tool to capture the voices and opinions of the people who matter most to you. Video Maker | Create Videos Online | Promo.com [](https://promo.com/) Free customizable video maker to help boost your business. Video creator for ads, social media, product and explainer videos, and for anything else you need! beehiiv — The newsletter platform built for growth [](https://www.beehiiv.com/) Access the best tools available in email, helping your newsletter scale and monetize like never before. GetResponse | Professional Email Marketing for Everyone [](https://www.getresponse.com/) No matter your level of expertise, we have a solution for you. At GetResponse, it's email marketing done right. Start your free account today! Search Email Newsletter Archives : Email Tuna [](https://emailtuna.com/) Explore newsletters without subscribing. Get email design ideas, discount coupon codes and exclusive newsletters deals. Database of email newsletters archived from all over the internet. Other Tools Simplescraper — Scrape Websites and turn them into APIs [](https://simplescraper.io/) Web scraping made easy — a powerful and free Chrome extension for scraping websites in your browser, automated in the cloud, or via API. No code required. Exploding Topics - Discover the hottest new trends. [](https://explodingtopics.com/) See new market opportunities, trending topics, emerging technology, hot startups and more on Exploding Topics. Scribe | Visual step-by-step guides [](https://scribehow.com/) By capturing your process while you work, Scribe automatically generates a visual guide, ready to share with the click of a button. Get It Free – The internet's BEST place to find free stuff! [](https://getitfree.us/) The internet's BEST place to find free stuff! Inflact by Ingramer – Marketing toolkit for Instagram [](https://inflact.com/) Sell on Instagram, build your audience, curate content with the right set of tools. Free Online Form Builder & Form Creator | Jotform [](https://www.jotform.com/) We believe the right form makes all the difference. Go from busywork to less work with powerful forms that use conditional logic, accept payments, generate reports, and automate workflows. Manage Your Team’s Projects From Anywhere | Trello [](https://trello.com/en) Trello is the ultimate project management tool. Start up a board in seconds, automate tedious tasks, and collaborate anywhere, even on mobile. TikTok hashtag generator - tiktokhashtags.com [](https://tiktokhashtags.com/) Find out which are the best hashtags for your TikTok post. Create Infographics, Reports and Maps - Infogram [](https://infogram.com/) Infogram is an easy to use infographic and chart maker. Create and share beautiful infographics, online reports, and interactive maps. Make your own here. Confetto - Create Instagram content in minutes [](https://www.confet.to/) Confetto is an all-in-one social media marketing tool built for SMBs and Social Media Managers. Confetto helps you create high-quality content for your audience that maximizes your reach and engagement on social media. Design, copy-write, plan and schedule content all in one place. Find email addresses in seconds • Hunter (Email Hunter) [](https://hunter.io/) Hunter is the leading solution to find and verify professional email addresses. Start using Hunter and connect with the people that matter for your business. PlayPhrase.me: Site for cinema archaeologists. [](https://playphrase.me/) Travel and explore the world of cinema. Largest collection of video quotes from movies on the web. #1 Free SEO Tools → SEO Review Tools [](https://www.seoreviewtools.com/) SEO Review Tools: 42+ Free Online SEO Tools build with ❤! → Rank checker → Domain Authority Checker → Keyword Tool → Backlink Checker Podcastle: Seamless Podcast Recording & Editing [](https://podcastle.ai/) Podcastle is the simplest way to create professional-quality podcasts. Record, edit, transcribe, and export your content with the power of AI, in an intuitive web-based platform. Save Ads from TikTok & Facebook Ad Library - Foreplay [](https://www.foreplay.co/) The best way to save ads from TikTok Creative Center and Facebook Ad Library, Organize them into boards and share ad inspiration with your team. Supercharge your creative strategy. SiteRight - Automate Your Business [](https://www.siteright.co/) SiteRight combines the abilities of multiple online resources into a single dashboard allowing you to have full control over how you manage your business. Diffchecker - Compare text online to find the difference between two text files [](https://www.diffchecker.com/) Diffchecker will compare text to find the difference between two text files. Just paste your files and click Find Difference! Yout.com [](https://yout.com/) Yout.com allows you to record videos from YouTube, FaceBook, SoundCloud, VK and others too many formats with clipping. Intuitively easy to use, with Yout the Internet DVR, with a bit of extra. AI Content Generation | Competitor Analysis - Predis.ai [](https://predis.ai/) Predis helps brands and influencers communicate better on social media by providing AI-powered content strategy analysis, content and hashtag recommendations. Castr | #1 Live Video Streaming Solution With Video Hosting [](https://castr.io/) Castr is a live video streaming solution platform that delivers enterprise-grade live videos globally with CDN. Live event streaming, video hosting, pre-recorded live, multi stream – all in one place using Castr. Headliner - Promote your podcast, radio show or blog with video [](https://www.headliner.app/) Easily create videos to promote your podcast, radio show or blog. Share to Instagram, Facebook, Twitter, YouTube, Linkedin and anywhere video lives Create Presentations, Infographics, Design & Video | Visme [](https://www.visme.co/) Create professional presentations, interactive infographics, beautiful design and engaging videos, all in one place. Start using Visme today. Designrr - Create eBooks, Kindle books, Leadmagnets, Flipbooks and Blog posts from your content in 2 minutes [](https://designrr.io/) Upload any web page, MS Word, Video, Podcast or YouTube and it will create a stunning ebook and convert it to pdf, epub, Kindle or Flipbook. Quick and Easy to use. Full Training, 24x7 Support and Facebook Group Included. SwipeWell | Swipe File Software [](https://www.swipewell.app/) The only Chrome extension dedicated to helping you save, organize, and reference marketing examples (so you never feel stumped). Tango | Create how-to guides, in seconds [](https://www.tango.us/) Tango takes the pain out of documenting processes by automatically generating how-to guides while you work. Empower your team to do their best work. Ad Creative Bank [](https://www.theadcreativebank.com/) Get inspired by ads from across industries, learn new best practices, and start thinking creatively about your brand’s digital creative. Signature Hound • Free Email Signature and Template Generator [](https://signaturehound.com/) Our email signature generator is free and easy to use. Our customizable templates work with Gmail, Outlook, Office 365, Apple Mail and more. Organize All Of Your Marketing In One Place - CoSchedule [](https://coschedule.com/) Get more done in less time with the only work management software for marketers. B Ok - Books [](https://b-ok.xyz/categories) OmmWriter [](https://ommwriter.com/) Ommwriter Rebrandly | Custom URL Shortener, Branded Link Management, API [](https://www.rebrandly.com/) URL Shortener with custom domains. Shorten, brand and track URLs with the industry-leading link management platform. Free to try. API, Short URL, Custom Domains. Common Tools [](https://www.commontools.org/) Book Bolt [](https://bookbolt.io/) Zazzle [](https://www.zazzle.com/) InspiroBot [](https://inspirobot.me/) Download Free Cheat Sheets or Create Your Own! - Cheatography.com: Cheat Sheets For Every Occasion [](https://cheatography.com/) Find thousands of incredible, original programming cheat sheets, all free to download. No Code Chatbot Platform | Free Chatbot Platform | WotNot [](https://wotnot.io/) WotNot is the best no code chatbot platform to build AI bot easily without coding. Deploy bots and live chat on the Website, Messenger, WhatsApp, and more. SpyFu - Competitor Keyword Research Tools for Google Ads PPC & SEO [](https://www.spyfu.com/) Systeme.io - The only tool you need to launch your online business [](https://systeme.io/) Systeme.io has all the tools you need to grow your online business. Click here to create your FREE account! Productivity Temp Mail [](https://temp-mail.org/en/) The Visual Collaboration Platform for Every Team | Miro [](https://miro.com/) Scalable, secure, cross-device and enterprise-ready team collaboration whiteboard for distributed teams. Join 35M+ users from around the world. Grammarly: Free Online Writing Assistant [](https://www.grammarly.com/) Millions trust Grammarly’s free writing app to make their online writing clear and effective. Getting started is simple — download Grammarly’s extension today. Rize · Maximize Your Productivity [](https://rize.io/) Rize is a smart time tracker that improves your focus and helps you build better work habits. Motion | Manage calendars, meetings, projects & tasks in one app [](https://www.usemotion.com/) Automatically prioritize tasks, schedule meetings, and resolve calendar conflicts. Used by over 10k CEOs and professionals to improve focus, get more done, and streamline workday. Notion – One workspace. Every team. [](https://www.notion.so/) We’re more than a doc. Or a table. Customize Notion to work the way you do. Loom: Async Video Messaging for Work | Loom [](https://www.loom.com/) Record your screen, share your thoughts, and get things done faster with async video. Zapier | Automation that moves you forward [](https://zapier.com/) Workflow automation for everyone. Zapier automates your work across 5,000+ app integrations, so you can focus on what matters. Rows — The spreadsheet with superpowers [](https://rows.com/) Combine the power of a spreadsheet with built-in integrations from your business apps. Automate workflows and build tools that make work simpler. Free Online Form Builder | Tally [](https://tally.so/) Tally is the simplest way to create free forms & surveys. Create any type of form in seconds, without knowing how to code, and for free. Highbrow | Learn Something New Every Day. Join for Free! [](https://gohighbrow.com/) Highbrow helps you learn something new every day with 5-minute lessons delivered to your inbox every morning. Join over 400,000 lifelong learners today! Slick Write | Check your grammar. Proofread online. [](https://www.slickwrite.com/#!home) Slick Write is a powerful, FREE application that makes it easy to check your writing for grammar errors, potential stylistic mistakes, and other features of interest. Whether you're a blogger, novelist, SEO professional, or student writing an essay for school, Slick Write can help take your writing to the next level. Reverso [](https://www.reverso.net) Hemingway Editor [](https://hemingwayapp.com/) Web Apps by 123apps - Edit, Convert, Create [](https://123apps.com/) Splitbee – Your all-in-one analytics and conversion platform [](https://splitbee.io/) Track and optimize your online business with Splitbee. Analytics, Funnels, Automations, A/B Testing and more. PDF Tools Free PDF, Video, Image & Other Online Tools - TinyWow [](https://tinywow.com/) Smallpdf.com - A Free Solution to all your PDF Problems [](https://smallpdf.com/) Smallpdf - the platform that makes it super easy to convert and edit all your PDF files. Solving all your PDF problems in one place - and yes, free. Sejda helps with your PDF tasks [](https://www.sejda.com/) Sejda helps with your PDF tasks. Quick and simple online service, no installation required! Split, merge or convert PDF to images, alternate mix or split scans and many other. iLovePDF | Online PDF tools for PDF lovers [](https://www.ilovepdf.com/) iLovePDF is an online service to work with PDF files completely free and easy to use. Merge PDF, split PDF, compress PDF, office to PDF, PDF to JPG and more! Text rewrite QuillBot [](https://quillbot.com/) Pre Post SEO : Online SEO Tools [](https://www.prepostseo.com/) Free Online SEO Tools: plagiarism checker, grammar checker, image compressor, website seo checker, article rewriter, back link checker Wordtune | Your personal writing assistant & editor [](https://www.wordtune.com/) Wordtune is the ultimate AI writing tool that rewrites, rephrases, and rewords your writing! Trusted by over 1,000,000 users, Wordtune strengthens articles, academic papers, essays, emails and any other online content. Aliexpress alternatives CJdropshipping - Dropshipping from Worldwide to Worldwide! [](https://cjdropshipping.com/) China's reliable eCommerce dropshipping fulfillment supplier, helps small businesses ship worldwide, dropship and fulfillment services that are friendly to start-ups and small businesses, Shopify dropshipping. SaleHoo [](https://www.salehoo.com/) Alibaba.com: Manufacturers, Suppliers, Exporters & Importers from the world's largest online B2B marketplace [](https://www.alibaba.com/) Find quality Manufacturers, Suppliers, Exporters, Importers, Buyers, Wholesalers, Products and Trade Leads from our award-winning International Trade Site. Import & Export on alibaba.com Best Dropshipping Suppliers for US + EU Products | Spocket [](https://www.spocket.co/) Spocket allows you to easily start dropshipping top products from US and EU suppliers. Get started for free and see why Spocket consistently gets 5 stars. Best dropshipping supplier to the US [](https://www.usadrop.com/) THE ONLY AMERICAN-MADE FULFILLMENT CENTER IN CHINA. Our knowledge of the Worldwide dropshipping market and the Chinese Supply-Chain can't be beat! 阿里1688 [](https://www.1688.com/) 阿里巴巴(1688.com)是全球企业间(B2B)电子商务的著名品牌,为数千万网商提供海量商机信息和便捷安全的在线交易市场,也是商人们以商会友、真实互动的社区平台。目前1688.com已覆盖原材料、工业品、服装服饰、家居百货、小商品等12个行业大类,提供从原料--生产--加工--现货等一系列的供应产品和服务 Dropshipping Tools Oberlo | Where Self Made is Made [](https://www.oberlo.com/) Start selling online now with Shopify. All the videos, podcasts, ebooks, and dropshipping tools you'll need to build your online empire. Klaviyo: Marketing Automation Platform for Email & SMS [](https://www.klaviyo.com/) Klaviyo, an ecommerce marketing automation platform for email marketing and sms syncs your tech stack with your website store to scale your business. SMSBump | SMS Marketing E-Commerce App for Shopify [](https://smsbump.com/) SMSBump is an SMS marketing & automation app for Shopify. Segment customers, recover orders, send campaign text messages with a 35%+ click through rate. AfterShip: The #1 Shipment Tracking Platform [](https://www.aftership.com/) Order status lookup, branded tracking page, and multi-carrier tracking API for eCommerce. Supports USPS, FedEx, UPS, and 900+ carriers worldwide. #1 Dropshipping App | Zendrop [](https://zendrop.com/) Start and scale your own dropshipping business with Zendrop. Sell and easily fulfill your orders with the fastest shipping in the industry. Best Dropshipping Suppliers for US + EU Products | Spocket [](https://www.spocket.co/) Spocket allows you to easily start dropshipping top products from US and EU suppliers. Get started for free and see why Spocket consistently gets 5 stars. Video Editing Jitter • The simplest motion design tool on the web. [](https://jitter.video/) Animate your designs easily. Export your creations as videos or GIFs. All in your browser. DaVinci Resolve 18 | Blackmagic Design [](https://www.blackmagicdesign.com/products/davinciresolve) Professional video editing, color correction, visual effects and audio post production all in a single application. Free and paid versions for Mac, Windows and Linux. Online Video Editor | Video Creator | InVideo [](https://invideo.io/) InVideo's Online Video Editor Helps You Make Professional Videos From Premium Templates, Images, And Music. All your video needs in one place | Clipchamp [](https://clipchamp.com/) Fast-forward your creations with our video editing platform. Start with a video template or record your webcam or screen. Get the pro look with filters, transitions, text and more. Then, export in minutes and share in an instant. Descript | All-in-one audio/video editing, as easy as a doc. [](https://www.descript.com/) Record, transcribe, edit, mix, collaborate, and master your audio and video with Descript. Download for free →. Kapwing — Reach more people with your content [](https://www.kapwing.com/) Kapwing is a collaborative, online content creation platform that you can use to edit video and create content. Join over 10 million modern creators who trust Kapwing to create, edit, and grow their content on every channel. Panzoid [](https://panzoid.com/) Powerful, free online apps and community for creating beautiful custom content. Google Web Designer - Home [](https://webdesigner.withgoogle.com/) Kapwing — Reach more people with your content [](https://www.kapwing.com/) Kapwing is a collaborative, online content creation platform that you can use to edit video and create content. Join over 10 million modern creators who trust Kapwing to create, edit, and grow their content on every channel. ClipDrop [](https://clipdrop.co/) Create professional visuals without a photo studio CapCut [](https://www.capcut.com/) CapCut is an all-in-one online video editing software which makes creation, upload & share easier, with frame by frame track editor, cloud drive etc. VEED - Online Video Editor - Video Editing Made Simple [](https://www.veed.io/) Make stunning videos with a single click. Cut, trim, crop, add subtitles and more. Online, no account needed. Try it now, free. VEED Free Video Maker | Create & Edit Your Videos Easily - Animoto [](https://animoto.com/k/welcome) Create, edit, and share videos with our online video maker. Combine your photos, video clips, and music to make quality videos in minutes. Get started free! Runway - Online Video Editor | Everything you need to make content, fast. [](https://runwayml.com/) Discover advanced video editing capabilities to take your creations to the next level. CreatorKit - A.I. video creator for marketers [](https://creatorkit.com/) Create videos with just one click, using our A.I. video editor purpose built for marketers. Create scroll stopping videos, Instagram stories, Ads, Reels, and TikTok videos. Pixar in a Box | Computing | Khan Academy [](https://www.khanacademy.org/computing/pixar) 3D Video Motions Plask - AI Motion Capture and 3D Animation Tool [](https://plask.ai/) Plask is an all-in-one browser-based AI motion capture tool and animation editor that anybody can use, from motion designers to every day content creators. Captions Captions [](https://www.getcaptions.app/) Say hello to Captions, the only camera and editing app that automatically transcribes, captions and clips your talking videos for you. Stock videos Pexels [](https://www.pexels.com/) Pixabay [](https://pixabay.com/) Mixkit - Awesome free assets for your next video project [](https://mixkit.co/) Download Free Stock Video Footage, Stock Music & Premiere Pro Templates for your next video editing project. All assets can be downloaded for free! Free Stock Video Footage HD 4K Download Royalty-Free Clips [](https://www.videvo.net/) Download free stock video footage with over 300,000 video clips in 4K and HD. We also offer a wide selection of music and sound effect files with over 180,000 clips available. Click here to download royalty-free licensing videos, motion graphics, music and sound effects from Videvo today. Free Stock Video Footage HD Royalty-Free Videos Download [](https://mazwai.com/) Download free stock video footage with clips available in HD. Click here to download royalty-free licensing videos from Mazwai now. Royalty Free Stock Video Footage Clips | Vidsplay.com [](https://www.vidsplay.com/) Royalty Free Stock Video Footage Clips Free Stock Video Footage, Royalty Free Videos for Download [](https://coverr.co/) Download royalty free (for personal and commercial use), unique and beautiful video footage for your website or any project. No attribution required. Stock Photos Beautiful Free Images & Pictures | Unsplash [](https://unsplash.com/) Beautiful, free images and photos that you can download and use for any project. Better than any royalty free or stock photos. When we share, everyone wins - Creative Commons [](https://creativecommons.org/) Creative Commons licenses are 20! Honoring 20 years of open sharing using CC licenses, join us in 2022 to celebrate Better Sharing — advancing universal access to knowledge and culture, and fostering creativity, innovation, and collaboration. Help us reach our goal of raising $15 million for a future of Better Sharing.  20 Years of Better … Read More "When we share, everyone wins" Food Pictures • Foodiesfeed • Free Food Photos [](https://www.foodiesfeed.com/) Download 2000+ food pictures ⋆ The best free food photos for commercial use ⋆ CC0 license Free Stock Photos and Images for Websites & Commercial Use [](https://burst.shopify.com/) Browse thousands of beautiful copyright-free images. All our pictures are free to download for personal and commercial use, no attribution required. EyeEm | Authentic Stock Photography and Royalty-Free Images [](https://www.eyeem.com/) Explore high-quality, royalty-free stock photos for commercial use. License individual images or save money with our flexible subscription and image pack plans. picjumbo: Free Stock Photos [](https://picjumbo.com/) Free stock photos and images for your projects and websites.️ Beautiful 100% free high-resolution stock images with no watermark. Free Stock Photos, Images, and Vectors [](https://www.stockvault.net/) 139.738 free stock photos, textures, backgrounds and graphics for your next project. No attribution required. Free Stock Photos, PNGs, Templates & Mockups | rawpixel [](https://www.rawpixel.com/) Free images, PNGs, stickers, backgrounds, wallpapers, graphic templates and PSD mockups. All safe to use with commercial licenses. Free Commercial Stock Photos & Royalty Free Images | PikWizard [](https://pikwizard.com/) Free images, videos & free stock photos. Unlimited downloads ✓ Royalty-free Images ✓Copyright-free for commercial use ✓ No Attribution Required Design Bundles [](https://designbundles.net/) Stock music Royalty Free Music for video creators | Epidemic Sound [](https://www.epidemicsound.com/) Download premium Royalty free Music and SFX! Our free trial gives you access to over 35,000 tracks and 90,000 sound effects for video, streaming and more! Royalty-Free Music & SFX for Video Creators | Artlist [](https://artlist.io/) Explore the ultimate royalty-free music & sound effects catalogs for unlimited use in YouTube videos, social media & films created by inspiring indie artists worldwide. The go-to music licensing choice for all creators Royalty Free Audio Tracks - Envato Elements [](https://elements.envato.com/audio) Download Royalty Free Stock Audio Tracks for your next project from Envato Elements. Premium, High Quality handpicked Audio files ideal for any genre. License popular music for videos • Lickd [](https://lickd.co/) The only place you can license popular music for videos. Access 1M+ mainstream tracks, plus high-quality stock music for content creators NCS (NoCopyrightSounds) - free music for content creators [](https://ncs.io/) NCS is a Record Label dedicated to giving a platform to the next generation of Artists in electronic music, representing genres from house to dubstep via trap, drum & bass, electro pop and more. Search Engine Optimization Keyword Tool For Monthly Search Volume, CPC & Competition [](https://keywordseverywhere.com/) Keywords Everywhere is a browser add-on for Chrome & Firefox that shows search volume, CPC & competition on multiple websites. Semrush - Online Marketing Can Be Easy [](https://www.semrush.com/) Turn the algorithm into a friend. Make your business visible online with 55+ tools for SEO, PPC, content, social media, competitive research, and more. DuckDuckGo — Privacy, simplified. [](https://duckduckgo.com/) The Internet privacy company that empowers you to seamlessly take control of your personal information online, without any tradeoffs. SEO Software for 360° Analysis of Your Website [](https://seranking.com/) Leading SEO software for business owners, agencies, and SEO specialists. Track your rankings, monitor competitors, spot technical errors, and more. Skyrocket your organic traffic with Surfer [](https://surferseo.com/) Use Surfer to research, write, optimize, and audit! Everything you need to create a comprehensive content strategy that yields real results is right here. Ahrefs - SEO Tools & Resources To Grow Your Search Traffic [](https://ahrefs.com/) You don't have to be an SEO pro to rank higher and get more traffic. Join Ahrefs – we're a powerful but easy to learn SEO toolset with a passionate community. Neon Tools [](https://neontools.io/) Google Index Search [](https://lumpysoft.com/) Google Index Search SEO Backlink Checker & Link Building Toolset | Majestic.com [](https://majestic.com/) Develop backlink strategies with our Link Intelligence data, build the strongest SEO backlink campaigns to drive organic traffic and boost your rankings today. PageOptimizer Pro [](https://pageoptimizer.pro/) Plans Services SEO Consulting Learn SEO About Blog POP SEO Community Podcast Support POP On Page Workshops With Kyle Roof POP Chrome Extension Guide Tutorial Videos Frequently Asked Questions Best Practices Login Cancel Anytime Plans Services SEO Consulting Learn SEO About Blog POP SEO Community Podcast Support POP On Page… Keyword Chef - Keywords for Publishers [](https://keywordchef.com/) Rank Insanely Fast for Keywords Your Competition Can’t Find “Every long-tail keyword I find ends up ranking within a day” – Dane Eyerly, Owner at TextGoods.com Keyword Chef automatically finds and filters keywords for you. Real-time SERP analysis lets you find keywords nearly guaranteed to rank. Try for free → Let’s face it, most keyword tools ... Read more Notifier - Social Listening for Social Media and More! [](https://notifier.so/) Track keywords. Market your product for free. Drive the conversation. Easy. Free Trial. No obligation ever. Simple. Fast. Trusted by Top Companies. Free Keyword Research Tool from Wordtracker [](https://www.wordtracker.com/) The best FREE alternative to the Keyword Planner. Use Wordtracker to reveal 1000s of profitable longtail keywords with up to 10,000 results per search Blog Posts The 60 Hottest Front-end Tools of 2021 | CSS-Tricks - CSS-Tricks [](https://css-tricks.com/hottest-front-end-tools-in-2021/) A complete list of the most popular front-end tools in 2021, according to the Web Tools Weekly newsletter. See which resources made the list. Resume ResumeGlow - AI Powered Resume Builder [](https://resumeglow.com/) Get hired fast with a resume that grabs attention. Designed by a team of HR experts and typographers. Customizable templates with more than a million possible Create Your Job-winning Resume - (Free) Resume maker · Resume.io [](https://resume.io/) Free online resume maker, allows you to create a perfect Resume or Cover Letter in 5 minutes. See how easy it is to write a professional resume - apply for jobs today! Rezi - The Leading AI-Powered Free Resume Builder [](https://www.rezi.ai/) Rezi’s award-winning AI-powered resume builder is trusted by hundreds of thousands of job seekers. Create your perfect resume in minutes with Rezi. Create a Perfect Resume | Free Resume Builder | Resumaker.ai [](https://resumaker.ai/) Create your professional resume with this online resume maker. Choose a designer-made template and grab any employer attention in seconds. Trusted AI Resume Maker Helps You Get Hired Fast [](https://skillroads.com/) Reach a 96.4% success rate in the job hunt race with the best resume creator. Our innovative technologies and 24/7 support help you to become a perfect candidate for any job. Do not lose your chance to become the One. Kickresume | Best Online Resume & Cover Letter Builder [](https://www.kickresume.com/) Create your best resume yet. Online resume and cover letter builder used by 1,300,000 job seekers worldwide. Professional templates approved by recruiters. ResumeMaker.Online | Create a Professional Resume for Free [](https://www.resumemaker.online/) Save time with the easiest-to-use Resume Maker Online. Create an effective resume in just minutes and land your dream job. No Sign-up required, start now! Interviews Interview Warmup - Grow with Google [](https://grow.google/certificates/interview-warmup/) A quick way to prepare for your next interview. Practice key questions, get insights about your answers, and get more comfortable interviewing. No code website builder Carrd - Simple, free, fully responsive one-page sites for pretty much anything [](https://carrd.co/) A free platform for building simple, fully responsive one-page sites for pretty much anything. Webflow: Create a custom website | No-code website builder [](https://webflow.com/) Create professional, custom websites in a completely visual canvas with no code. Learn how to create a website by trying Webflow for free! Google Sites: Sign-in [](https://sites.google.com/) FlutterFlow - Build beautiful, modern apps incredibly fast! [](https://flutterflow.io/) FlutterFlow lets you build apps incredibly fast in your browser. Build fully functional apps with Firebase integration, API support, animations, and more. Export your code or even easier deploy directly to the app stores! Free Website Builder: Build a Free Website or Online Store | Weebly [](https://www.weebly.com/) Weebly’s free website builder makes it easy to create a website, blog, or online store. Find customizable templates, domains, and easy-to-use tools for any type of business website. Glide • No Code App Builder • Nocode Application Development [](https://www.glideapps.com/) Create the apps your business needs, without coding, waiting or overpaying. Get started for free and build an app today Adalo - Build Your Own No Code App [](https://www.adalo.com/) Adalo makes creating apps as easy as putting together a slide deck. Turn your idea into a real native app — no code needed! Siter.io - The collaborative web design tool, no-code website builder [](https://siter.io/) Siter.io is a visual website builder for designers. Prototype, design, and create responsive websites in the browser. Work together with your team in one place. Elementor: #1 Free WordPress Website Builder | Elementor.com [](https://elementor.com/) Elementor is the platform web creators choose to build professional WordPress websites, grow their skills, and build their business. Start for free today! No code app builder | Bravo Studio [](https://www.bravostudio.app/) Your no-code mobile app builder for iOS and Android. Create MVP’s, validate ideas and publish on App Store and Google Play Store. Home [](https://typedream.com/) The simplest way to build a website with no-code, as easy as writing on Notion. Try Typedream for free and upgrade for custom domains, collaborators, and unlimited pages. Free Website Builder | Create a Free Website | Wix.com [](https://www.wix.com/) Create a website with Wix’s robust website builder. With 900+ strategically designed templates and advanced SEO and marketing tools, build your brand online today. Free responsive Emails & Landing Pages drag-and-drop Editor | BEE [](https://beefree.io/) Free responsive emails and landing pages editor. With BEE drag-and-drop builders embedded in many software applications you can start designing now! Home [](https://typedream.com/) The simplest way to build a website with no-code, as easy as writing on Notion. Try Typedream for free and upgrade for custom domains, collaborators, and unlimited pages. Ownit Connected Checkout [](https://www.ownit.co/) Ownit Connected Checkout Bookmark.com | No-code Website Builder to Start Your Business [](https://www.bookmark.com/) Our AI powered platform ensures your business is future proof. Try Bookmark for free. The best way to build web apps without code | Bubble [](https://bubble.io/) Bubble introduces a new way to build software. It’s a no-code tool that lets you build SaaS platforms, marketplaces and CRMs without code. Bubble hosts all web apps on its cloud platform. Responsive Web Design | Website Creation | Editor X [](https://www.editorx.com/) Experience the future of website design with responsive layouts, CSS precision and smooth drag and drop. Create a Website for Free. Tilda Website Builder [](https://tilda.cc/) Create a website, online store, landing page with Tilda intuitive website builder. Build your site from hundreds of pre-designed templates and publish it today. No code required. No-code headless commerce and websites | Unstack Inc. [](https://www.unstack.com/) Deploy high performance eCommerce storefronts and websites without the engineering overhead using Unstack's no-code CMS Best Drag-and-Drop Website Builder | Jemi [](https://jemi.so/) The modern website builder for creatives, entrepreneurs, and dreamers. Build a beautiful link in bio site, portfolio, or landing page in minutes. No-code website builder that works like Notion [](https://popsy.co/) Create a beautiful no-code website in minutes. Popsy works just like Notion but is built from the ground up for building websites. Choose a free template. Edit content just like in Notion. Customize styles without code. Free Notion icons and illustrations. Unbounce - The Landing Page Builder & Platform [](https://unbounce.com/) Grow your relevance, leads, and sales with Unbounce. Use Unbounce to easily create and optimize landing pages for your small business and boost conversions with AI insights. Low-code Front-end Design & Development Platform | TeleportHQ [](https://teleporthq.io/) Front-end development platform, with a visual builder and headless content modelling capabilities. Static website creation, and UI development tools. Other tools used in no code website MemberSpace - Turn any part of your website into members-only with just a few clicks [](https://www.memberspace.com/) Create memberships on your website for anything you want like courses, video tutorials, member directories, and more while having 100% control over look & feel. Triggre | The number one true no-code platform to run your business [](https://www.triggre.com/) The best no-code platform to create highly advanced business applications in hours, without programming. Try it now for free! No code game builder Welcome to Buildbox [](https://signup.buildbox.com/) Welcome to Buildbox Flowlab Game Creator - Make games online [](https://flowlab.io/) Flowlab is an online game creator. Make your own games to share with friends. Make 2D Games With GameMaker | Free Video Game Maker [](https://gamemaker.io/) Make a game with GameMaker, the best free video game engine. Perfect for beginners and professionals. Learn to build your own 2D games with our simple tutorials. Side Hustle Side Hustle Stack [](https://sidehustlestack.co/) Side Hustle Stack is a resource for finding platform-based work, ranging from gig work and side hustles to platforms that help you start a small business that can grow. Fiverr [](https://www.fiverr.com/) Remotasks: Work From Home, Online Bootcamp Training [](https://www.remotasks.com/en) Make money doing tasks. Start earning today! Free bootcamp training offered online. Sign up for a free Remotasks account and work from home. Earn up to $200/month. Transcribe Speech to Text | Rev [](https://www.rev.com/) Transcribe Speech to Text with Rev. Reach your audience with clear and accurate captions, transcripts, and subtitles. AI Training Data and other Data Management Services [](https://www.clickworker.com/) AI training data, SEO texts, web research, tagging, surveys and more - Use the crowdsourcing principle with the power of >4.5M Clickworkers. Automate your Busy Work - Byron People-Powered Assistants [](https://www.hibyron.com/) Byron is an on demand US based virtual assistant platform that gives individuals and teams the ability to quickly outsource their non-essential tasks. Jobs Websites - Remote Latest Crypto Jobs, Web3 Jobs and Blockchain Jobs in the leading tech companies. [](https://cryptojobslist.com/) New Cryptocurrency Jobs, Web3 Jobs and Blockchain Jobs on CryptoJobsList — the leading site to find and post jobs. Connect with companies hiring in a few clicks and begin your next experience in the industry. Updated daily. Remote Jobs: Design, Marketing, Programming, Writing & More [](https://justremote.co/) Discover Remote Jobs from around the world. Give up the commute, work remotely and do what you love, daily, from anywhere. Find your perfect remote development, design, sales or marketing job today. Remote Ok [](https://remoteok.com/) Hire Freelancers & Remote Workers For Free [](https://talent.hubstaff.com/) Find and hire the highest quality freelancers from around the world - for free. Choose from thousands of developers, digital marketers, creatives and more. We Work Remotely: Remote jobs in design, programming, marketing and more [](https://weworkremotely.com/) Find the most qualified people in the most unexpected places: Hire remote! We Work Remotely is the best place to find and list remote jobs that aren't restricted by commutes or a particular geographic area. Browse thousands of remote work jobs today. Angel [](https://angel.co/) Remote Work: Jobs, Companies & Virtual Teams - Remote.co [](https://remote.co/) Remote.co is the definitive remote work job board for online job seekers and companies hiring. Start your remote job search here! FlexJobs: Best Remote Jobs, Work from Home Jobs, Online Jobs & More [](https://www.flexjobs.com/) The #1 job search site for hand-screened flexible and remote jobs (work from home jobs) since 2007. Plus get resume, coaching and career help. Join today! Remote jobs remotefront.io [](https://remotefront.io/) All remote jobs at remotefront.io Daily Virtual Events Helping You Grow Professionally [](https://powertofly.com/) PowerToFly is where you receive expert career advice, free video training, coaching and exclusive access to jobs and events at top companies. Best Remote and Work from Home Jobs - Virtual Vocations [](https://www.virtualvocations.com/) Best work from home jobs and remote jobs in over 50 categories for professionals, digital nomads, telecommuting workers and entry level jobseekers. Education, healthcare, medical, customer support and tech job openings. Remote Jobs | Working Nomads [](https://www.workingnomads.com/jobs) Remote jobs for digital working nomads. Start your telecommuting career and work remotely from home or places around the world. Job Search, Companies Hiring Near Me, and Advice | The Muse [](https://www.themuse.com/) Find jobs at the best companies hiring near you and get free career advice. Startupers [](https://www.startupers.com/) NoDesk - Where Everyone Works Remote [](https://nodesk.co/) Browse and apply to the best new remote jobs at leading remote companies and startups for free. Join hundreds of companies that use NoDesk to build their remote teams. Browser Extensions Blackbox - Select. Copy. Paste & Search - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/blackbox-select-copy-past/mcgbeeipkmelnpldkobichboakdfaeon) Fastest Way to Copy Text from Videos & Images Octotree - GitHub code tree - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/octotree-github-code-tree/bkhaagjahfmjljalopjnoealnfndnagc) GitHub on steroids WhatFont - Chrome Web Store [](https://chrome.google.com/webstore/detail/whatfont/jabopobgcpjmedljpbcaablpmlmfcogm?hl=en) The easiest way to identify fonts on web pages. Window Resizer - Chrome Web Store [](https://chrome.google.com/webstore/detail/window-resizer/kkelicaakdanhinjdeammmilcgefonfh?hl=en) Resize the browser window to emulate various screen resolutions. Amino: CSS Editor - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/amino-css-editor/pbcpfbcibpcbfbmddogfhcijfpboeaaf) Live CSS Editor. Write custom CSS for any website and see your changes in real time. Checkbot: SEO, Web Speed & Security Tester 🚀 - Chrome Web Store [](https://chrome.google.com/webstore/detail/checkbot-seo-web-speed-se/dagohlmlhagincbfilmkadjgmdnkjinl?hl=en) Test SEO/speed/security of 100s of pages in a click! Check broken links, HTML/JavaScript/CSS, URL redirects, duplicate titles... Honey: Automatic Coupons & Rewards - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/honey-automatic-coupons-r/bmnlcjabgnpnenekpadlanbbkooimhnj) Save money and earn rewards when you shop online. Tango: screenshots, training, & documentation - Magazinul web Chrome [](https://chrome.google.com/webstore/detail/tango-screenshots-trainin/lggdbpblkekjjbobadliahffoaobaknh) Automatically create beautiful step-by-step guides with screenshots, in seconds. No code browser automation | axiom.ai [](https://axiom.ai/) Build browser bots quickly, without code. Automate website actions and repetitive tasks using just your browser, on any website or web app. No Code Browser extensions builder Bildr - Visual Web Development in your Browser [](https://www.bildr.com/) Visually build SaaS products, Chrome extensions, and web3 dApps Other Repurposing content for social media the easy way » Repurpose.io [](https://repurpose.io/) Repurposing content for social media made easy. Automatically repurpose YouTube, TikTok, Lives, Podcasts, and Zoom calls. Try it for FREE. Smart Serials: Your serial numbers database [](https://smartserials.com/) This is your main source of free serial numbers, unlock keys in a clean environment safe to browse by all ages. Old versions of Windows, Mac and Linux Software, Apps & Abandonware Games - Download at OldVersion.com [](http://www.oldversion.com/) Online Room Planner - Design Your Room [](http://www.planyourroom.com/) Planyourroom.com is a wonderful website to redesign each room in your house by picking out perfect furniture options to fit your unique space. BoredHumans.com - Fun AI Programs You Can Use Online [](https://boredhumans.com/) Fun AI programs you can use online. AI games, fake people, computer generated art, machine learning demos, and more. BNProject | Home [](https://buynothingproject.org/) Open Source Alternatives to Proprietary Software [](https://www.opensourcealternative.to/) Discover 400+ popular open source alternatives to proprietary SaaS. URL Shortener - Short URLs & Custom Free Link Shortener | Bitly [](https://bitly.com/) Bitly’s Connections Platform is more than a free URL shortener, with robust link management software, advanced QR Code features, and a Link-in-bio solution. TinEye Reverse Image Search [](https://tineye.com/) Good Books | Books recommended by successful people [](https://www.goodbooks.io/) Looking for the best books to read in 2022? Discover the best book recommendations from the world's most successful, influential and interesting people. Directory - Website Recommendations [](https://tokapps.com/directory/) 0 TRIED & TESTED WEBSITES LISTED Insanely Useful Websites A combination of useful websites for businesses, freelancers, DIYers, and individuals in a centralised area.All websites have been tried and tested. Filter Websites Audio Business Tools Copywriting Design Entertainment Graphics Guides Health Marketing PC Resources Savings SEO Software Travel Video Apply filter Watch Anime Online, Free Anime Streaming Online on Zoro.to Anime Website [](https://zoro.to/) Zoro is a Free anime streaming website which you can watch English Subbed and Dubbed Anime online with No Account and Daily update. WATCH NOW! Animated Drawings [](https://sketch.metademolab.com/) Bring children's drawings to life, by animating characters to move around! Alternativeto [](https://alternativeto.net/) Chatroulette [](https://chatroulette.com/) Random meetings around the world Tiktok Downloader - Download Video tiktok Without Watermark - SnapTik [](https://snaptik.app/en) TikTok Video Downloader - SnapTik.App is one of the best free Download video Tiktok No Watermark tool available online. You can download TikTok video from any device you have. Imgflip - Create and Share Awesome Images [](https://imgflip.com/) Flip through memes, gifs, and other funny images. Make your own images with our Meme Generator or Animated GIF Maker. Fake Text Message | Make Fake Text Conversation [](https://ifaketextmessage.com/) Fake Text Message is a tool to create a Fake Text Conversation and a Fake iMessage. ✂Templatemaker ︎ [](https://www.templatemaker.nl/en/) Omni Calculator [](https://www.omnicalculator.com/) Omni Calculator solves 2960 problems anywhere from finance and business to health. It’s so fast and easy you won’t want to do the math again! Watch Movies Online Free | Watch Series HD Free [](https://hdtoday.tv/) Free Access to the Biggest library of HD Movies and HD Series online - NO ADS - No Account Required - Fast Free Streaming Students Answers - The Most Trusted Place for Answering Life's Questions [](https://www.answers.com/) Answers is the place to go to get the answers you need and to ask the questions you want Wolfram|Alpha: Computational Intelligence [](https://www.wolframalpha.com/) Compute answers using Wolfram's breakthrough technology & knowledgebase, relied on by millions of students & professionals. For math, science, nutrition, history, geography, engineering, mathematics, linguistics, sports, finance, music… Online Math Tools - Simple, free and easy to use math utilities [](https://onlinemathtools.com/) World's simplest collection of useful mathematics utilities. Generate number sequences, draw fractals, do quick matrix and numerical calculations and more! edX | Free Online Courses by Harvard, MIT, & more | edX [](https://www.edx.org/) Access 2000 free online courses from 140 leading institutions worldwide. Gain new skills and earn a certificate of completion. Join today. Sci-Hub [](https://sci-hub.hkvisa.net/) Sci-Hub,mg.scihub.ltd,sci-hub.tw,The project is supported by user donations. Imagine the world with free access to knowledge for everyone ‐ a world without any paywalls. DigitalDefynd - Find the Best + Free Courses Online [](https://digitaldefynd.com/) 4 Million+ Learners | 96,000+ Courses | 45,000+ Free Courses | 1200+ Free Certificates Learn Anything [](https://learn-anything.xyz/) Search Interactive Mind Maps to learn anything HubSpot Academy - Homepage [](https://academy.hubspot.com/) HubSpot Academy is the worldwide leader in inbound marketing, sales, and customer service/support training.

YT_Emerging_Technologies_Introduction_to_AI
github
LLM Vibe Score0.461
Human Vibe Score0.039054583141409485
zusmaniJan 17, 2025

YT_Emerging_Technologies_Introduction_to_AI

YouTube Channel: Emerging Technologies Playlist: Introduction to AI Instructor: Zeeshan-ul-hassan Usmani Dear Students, I have uploaded all relevant material here for your quick access and learning. I hope you will find it beneficiary Yours Truly, Zeeshan =========================================== Video title: Resources Books to Order: Artificial Intelligence by Zeeshan Usmani - https://gufhtugu.com/artificial-intelligence Artificial Intelligence by Baqir Naqvi - https://gufhtugu.com/masnoi-zahanat/ Recommended Books • Gödel, Escher, Bach : An Eternal Golden Braid by Douglas R. Hofstadter A classic, poetic, philosophical defense of AI. • Machines Who Think by Pamela McCorduck. A good review of early AI history. • Robot: Mere Machine to Transcendent Mind by Hans P. Moravec Somewhat hyped book by a CMU robotics researcher. • Flesh and Machines: How Robots Will Change Us by Rodney Allen Brooks Reasonably decent book by MIT's leading robotics researcher. • Wired for War by Peter Warren Singer Reviews growing use of robots and unmanned vehicles in warfare. • Behind Deep Blue: Building the Computer That Defeated the World Chess Champion by Feng-Hsiung Hsu Autobiographical book on the development of a history making game-playing system. Interesting personal story of the hard engineering work that went into the system, with a few interesting facts on the technical aspects. • The Age of Spiritual Machines : When Computers Exceed Human Intelligence by Ray Kurzweil A recent view by an AI entrepreneur that has content if you ignore all the hype and overly-optimistic trust that Moore's law will magically solve all of the major problems. • Hal's Legacy : 2001's Computer As Dream and Reality An interesting collection of edited articles written to celebrate the fictional birthday of a famous intelligent computer who's true birthday must unfortunately be delayed, pending AI's inevitable progress. • The Sciences of the Artificial by Herbert Simon AI as science by one of its founders. • Models of My Life by Herbert Simon. An autobiography of one of AI's founders who's intellectual contributions also include fundamental contributions to economics (for which he won the Nobel prize), cognitive psychology, and computer science (such as co-inventing the linked list in the 1950's). • Alan Turing: The Enigma by Alan Hodges. A biography of one of the founders of CS and originator of the Turing test. Also a testimony to the tragic implications of homophobia. • The Emperor's New Mind : Concerning Computers, Minds, and the Laws of Physics and Shadows of the Mind : A Search for the Missing Science of Consciousness and The Large, the Small and the Human Mind by Roger Penrose A completely bogus argument against AI by a hopelessly Platonic mathematician. The last book contains an appended article by Stephen Hawking (a colleague of Penrose's) who of course doesn't buy his bogus argument. • The Mind's New Science : A History of the Cognitive Revolution by Howard Gardner A nice history of the development of cognitive science. • How the Mind Works , The Language Instinct , and Words and Rules : The Ingredients of Language by Steven Pinker Fun reading on lots of interesting issues in modern Cognitive Science and Linguistics if you don't take his exaggerated beliefs in nativism and evolutionary psychology too seriously. • Bots : The Origin of New Species by Andrew Leonard A light, somewhat hyped book on on Internet agents, chatterbots, etc. with a few funny stories. • Mathematics: The Loss of Certainty by Morris Kline A very nice book on the failed enterprise of using logic to build a firm foundation for infallible mathematics and the role of Gödel's Incompleteness Theorem in the philosophy of mathematics. • Incompleteness: The Proof and Paradox of Kurt Gödel by Rebecca Goldstein An interesting biography of Kurt Gödel. Too bad he was such a Platonist that, unlike Turing, he did not understand the true implications of his own theorems (interesting author connection: Goldstein is Pinker's wife). Links: • AAAI AI Topics Basic info on AI from the American Association for Artificial Intelligence: http://www.aaai.org/AITopics/html/welcome.html • Loebner Prize for limited Turing test: http://www.loebner.net/Prizef/loebner-prize.html • IBM's Deep Blue Page: http://www.research.ibm.com/deepblue/ • Robocup: Robotic Soccer Competition: http://www.robocup.org/ • NY Times Article on Proof of the Robbins Theorem: http://www.nytimes.com/library/cyber/week/1210math.html • NY Times article on Bayes Nets at Microsoft Research: http://www.nytimes.com/library/tech/00/07/biztech/articles/17lab.html =========================================== Video title: Numbers Infinity Video Link - •https://www.youtube.com/watch?v=hlXHwMgS06c https://www.cbs.com/shows/numb3rs/ http://numb3rs.wolfram.com/ =========================================== Video title: 20 Hours Rule and Assisgnemnt Assignment - https://www.urdufake2020.cicling.org/ =========================================== Video title: Assignments – P1 Mostly Human - https://money.cnn.com/mostly-human =========================================== Video title: Assignments – P2 Assignment – 2 - https://replika.ai/ Assignment – 3 – Teachable Machines https://teachablemachine.withgoogle.com/ Assignment – 4 – Tensor Flow Playground https://playground.tensorflow.org Assignment – 5 – GPT-3 Paper (175B Parameters) https://debuild.co/ Assignment – 6 - Image GPT-3 https://openai.com/blog/image-gpt/ =========================================== Video title: Create your own Deep Fake 1.https://colab.research.google.com/drive/1mGg_fmvhTpvkPkclw2yKkhALVzmawfvT?usp=sharing 2.https://drive.google.com/drive/folders/1wW1bxRV2S7Ce8gc3VDTzMQABE3-WCc_Y?usp=sharing •go into you gdrive > find cloned folder and ensure that this folder must have: vox-adv-cpk.pth.tar & vox-cpk.pth.tar failes •Aliaksandr Siarohin : https://github.com/AliaksandrSiarohin/first-order-model

The 8 AI Skills That Will Separate Winners From Losers in 2025
youtube
LLM Vibe Score0.446
Human Vibe Score0.92
Meet The AI Entrepreneur Who Used LinkedIn To Raise $13.8 Million
youtube
LLM Vibe Score0.436
Human Vibe Score0.64
ForbesApr 19, 2024

Meet The AI Entrepreneur Who Used LinkedIn To Raise $13.8 Million

Benjamin Harvey, the CEO of AI Squared, says he’s added investors including former TIAA CEO Roger Ferguson. Harvey joined Forbes senior writer, Jabari Young, at the Nasdaq MarketSite to discuss the startup’s Series A raise. Read the full story on Forbes: https://www.forbes.com/sites/jabariyoung/2024/04/17/meet-the-ai-entrepreneur-who-used-linkedin-to-raise-138-million/?sh=60958bea5837 0:00 Introduction 2:16 Benjamin Gives Biggest Tip On Learning Profit Loss 5:00 Benjamin Harvey On The State Of AI 8:25 How Will AI Evolve And Change In The Future? 14:04 What Is It Like To Be CEO Of AI Squared? 17:04 How Benjamin's Upbringing And Love Of Cartoons Helps Put Ideas Together In Business 23:02 Benjamin On Getting Investors For AI Squared 25:56 Benjamin's Take On ChatGPT And How Its Used 29:48 Artificial Intelligence: Benjamin's Take On What's Next 34:49 A Good AI Platform vs. A Great One Subscribe to FORBES: https://www.youtube.com/user/Forbes?sub_confirmation=1 Fuel your success with Forbes. Gain unlimited access to premium journalism, including breaking news, groundbreaking in-depth reported stories, daily digests and more. Plus, members get a front-row seat at members-only events with leading thinkers and doers, access to premium video that can help you get ahead, an ad-light experience, early access to select products including NFT drops and more: https://account.forbes.com/membership/?utmsource=youtube&utmmedium=display&utmcampaign=growthnon-subpaidsubscribe_ytdescript Stay Connected Forbes newsletters: https://newsletters.editorial.forbes.com Forbes on Facebook: http://fb.com/forbes Forbes Video on Twitter: http://www.twitter.com/forbes Forbes Video on Instagram: http://instagram.com/forbes More From Forbes: http://forbes.com Forbes covers the intersection of entrepreneurship, wealth, technology, business and lifestyle with a focus on people and success.