VibeBuilders.ai Logo
VibeBuilders.ai

Validation

Explore resources related to validation to help implement AI solutions for your business.

A Structured Approach to Ideation and Validation (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

A Structured Approach to Ideation and Validation (I will not promote)

Hi all, I used to work in VC and wanted to share some startup knowledge and insights from startup founders I know. Recently, I interviewed a friend of mine who built an AI Robotics startup ("Hivebotics") that creates automated toilet-cleaning robots. I can't post the full article because of Reddit's word limit, so I'll be posting it in sections here instead. This first section of the transcript goes through his approach to ideation and validation. Enjoy and let me know what you think! (I will not promote) \\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\ (1) Ideation and Validation Problem-Market-Solution Framework I like to think of startup ideation and validation using this framework: Problem– What exactly are you solving? Observation– How you identify a problem to work on User Research– How you further understand that problem Market– Is there a large enough market for solving this problem? Size– How many people experience this same problem? Demand– How many of those people are willing to pay for the solution? Solution– Your answer to the problem Desirability– Whether people actually want your solution Feasibility– Whether building the solution is practical and realistic Viability– Whether your solution can generate revenue Problem You always need to start problem-first, which is something that was really drilled into me during my time at Stanford. Too often, founders rush to build solutions first—apps or products they find exciting—without confirming whether there's any real demand for it. The first step is always to identify a specific problem, then further understand its scale, urgency and further details by talking to potential users. Observation– To find problems, observation is key. People may not even realise the inefficiencies in their processes until you point them out. That’s why interviews and field research are so important. There are problems all around us, so it's simply a matter of going out, paying attention and being attuned to them as they occur. User Research– To further understand the problem, conducting user research by interviewing potential customers is essential. Personally, I like to use the "Mom Test" when I conduct interviews to avoid biased and generic feedback. Don’t just ask theoretical questions and avoid being too specific—observe how your potential users work, ask about pain points, and use broad, open-ended questions to ensure you aren't leading them to a specific answer. Market Once you've found an actual problem and talked to enough potential users to really understand its specific pain points, the next step is to determine the market size and demand for a solution. Size– Determining the market size is essential because it determines whether or not it's commercially worthwhile to pursue the problem and develop a solution for it. You need to determine if there are enough potential customers out there experiencing this problem to gauge the market size. There's no secret strategy for this; you have to interview as many potential users as possible to confirm that it's a widespread problem in the industry. Demand– Make sure that you're working on a problem that people will gladly pay to have solved. Even if the problem is large enough, you have to make sure it's painful enough to warrant a paid solution. If many people experience the same problem, but aren't willing to pay for a solution, then you don't have a market and should look for a different problem to validate. Another way of looking at it is that your true market size is the number of potential customers actually willing to pay* for the solution to the problem, not the number of people simply experiencing the same problem. Solution When validating a potential solution to the problem, I would look at the 3 factors of desirability, feasibility and viability. Desirability– the degree to which a solution appeals to people and fulfills their wants and needs. Without strong desirability, even the most technically advanced or economically practical product is unlikely to succeed. The best way to test this is to secure financial commitments early on during the proof-of-concept stage. Most people are polite, so they may simply tell you that your startup's product is good even if it's not. However, if they're actually willing to pay for the solution, this is actual evidence of your product's desirability. Don't just ask people if they would pay for it; actually see whether they will pay for it. Feasibility– whether a product can be built using existing technical capabilities. A lack of feasibility makes it challenging or impossible to develop the product, no matter how appealing it might be to users or how promising its financial prospects are. This is just a matter of conducting initial research and actually trying to build a prototype, which will inform you whether the fully-realised product is truly feasible. Viability– the product's ability to generate sustainable financial returns. Without financial viability, the business supporting the product cannot endure, even if the product is highly appealing to users and technically achievable. Here, you need to look at your unit economics, development costs and other expenses to determine the viability of your solution. \\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\ Hope you enjoyed reading this; let me know your honest thoughts in the comments and I'll try to improve how I interview founders based on those!

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀
reddit
LLM Vibe Score0
Human Vibe Score1
Awkward_Ad_9605This week

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀

Hey Members! I’m excited to share an idea for a new application I’m planning to build: Quick Launch . This AI-driven platform is designed to assist solopreneurs or anyone with an idea in launching their Minimum Viable Products (MVPs) by taking on the roles of the entire team needed for the process. Goal: Assistance in quickly moving from Idea to MVP Before I dive into the details, I’d love to hear your thoughts and feedback. Key Features: Product Creation: From Idea to Product Detailing AI-Generated Q&A: Real-time questions generation one-at-a-time to define the product requirements based on their knowledge levels to convert an Idea into a Product. Market Research Reports: In-depth analysis that identifies product-market fit, competitive landscape, and potential marketing strategies. Sentiment Analysis: Evaluation of user feedback and reactions across multiple subreddits to gauge public opinion on ideas. Product Development: Product Detailing to Actual Product User Story Generation: Identification and creation of comprehensive user stories, tasks, and sub-tasks to facilitate development. AI Project Management: AI agents assume roles of project managers and UI/UX designers to streamline product detailing and development. Integration Capabilities: Seamless integration with popular project management tools like Jira, Asana, and Trello for better workflow management. Target Audience: Solopreneurs: Individuals looking to bring their business ideas to life without extensive resources. Indie Hackers: Entrepreneurs focused on building small projects or startups with minimal overhead. Idea Validators: Anyone with a concept seeking initial validation and market feedback before committing significant resources. If you’re interested in learning more, check out our teaser website: Quick Launch Discussion Question: What features would you find most valuable in an application like this? Are there specific pain points you face when launching an MVP? Your insights would be incredibly helpful as I refine this idea! Looking forward to your thoughts! 🙌

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀
reddit
LLM Vibe Score0
Human Vibe Score1
Awkward_Ad_9605This week

Idea Validation Post: Seeking Feedback on My AI-Driven Quick Launch Application! 🚀

Hey Members! I’m excited to share an idea for a new application I’m planning to build: Quick Launch . This AI-driven platform is designed to assist solopreneurs or anyone with an idea in launching their Minimum Viable Products (MVPs) by taking on the roles of the entire team needed for the process. Goal: Assistance in quickly moving from Idea to MVP Before I dive into the details, I’d love to hear your thoughts and feedback. Key Features: Product Creation: From Idea to Product Detailing AI-Generated Q&A: Real-time questions generation one-at-a-time to define the product requirements based on their knowledge levels to convert an Idea into a Product. Market Research Reports: In-depth analysis that identifies product-market fit, competitive landscape, and potential marketing strategies. Sentiment Analysis: Evaluation of user feedback and reactions across multiple subreddits to gauge public opinion on ideas. Product Development: Product Detailing to Actual Product User Story Generation: Identification and creation of comprehensive user stories, tasks, and sub-tasks to facilitate development. AI Project Management: AI agents assume roles of project managers and UI/UX designers to streamline product detailing and development. Integration Capabilities: Seamless integration with popular project management tools like Jira, Asana, and Trello for better workflow management. Target Audience: Solopreneurs: Individuals looking to bring their business ideas to life without extensive resources. Indie Hackers: Entrepreneurs focused on building small projects or startups with minimal overhead. Idea Validators: Anyone with a concept seeking initial validation and market feedback before committing significant resources. If you’re interested in learning more, check out our teaser website: Quick Launch Discussion Question: What features would you find most valuable in an application like this? Are there specific pain points you face when launching an MVP? Your insights would be incredibly helpful as I refine this idea! Looking forward to your thoughts! 🙌

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.
reddit
LLM Vibe Score0
Human Vibe Score1
Alarmed-Argument-605This week

Made 60k mrr for a business by just lead nurturing. Need suggestions and validation.

Apart from the story I need a suggestion and validation here. It's a bit long, skip to tl;dr if you couldn't handle length. A few days ago, I saw a person on Reddit sharing his struggles that, Even after generating a lot of leads from ads of Meta and Google (even with lowest cpc cpa cpl), he was not able to convert them into sales. Out of curiosity I dm'ed him with all fancy services that I offer and expressed that as a agency I would work with him for monthly recurring fee. He suggested for one time consulting fee, I agreed. It was literally a eye opener for me. This guy is in coaching business offering courses for people. His niche was too vague. Courses were on mindset coaching, confidence and public speaking coaching, right attitude coaching, manifestation coaching and all crap shits related to this. At first I thought he was not getting sales because who will pay for all this craps. I openly discussed with him that he has to change what he offers because, if I saw this ad I wouldn't buy this for sure. He then showed me how much money people offering similar service are making . I was literally taken back. He was part of a influencer group (the main guy who encourages these guys to start coaching business, looks like some mlm shit) where people post their succes stories. Literally lot of guys were making above 150k and 200k per month. Even with very basic landing page and average offer They are still winning. Here's where it gets interesting. I tried to clone everything that the top people in this industry are doing in marketing from end to end.( like the same landing page, bonus offers around 50k, exclusive community, free 1 on 1 calls for twice a month).Nothing worked for a month and later surprisingly even the sales started dropping a bit more. I got really confused here. So to do a discovery I went and purchased the competitor course and Man I was literally taken back. Like he has automated everything from end to end. You click the ad, see vsl, you have to fill a form and join a free Skool community where he gives away free stuffs and post success stories of people who took the course. Now every part of this journey you will get a follow up mail and follow up sms. Like after filling the form. after that now if you join and don't purchase the course you will be pampered with email and sms filled with success stories. For sure anybody will be tempted to buy the course. Here is the key take away. He was able to make more sales because he was very successful in nurturing the leads with follow ups after follow ups. Even after you purchased his course he is making passive income from 1 on calls and bonus live webinars. So follow ups will be for 1 on 1 calls and webinars after the course is over. Core point is our guy even after spending 2 to 3k per month on ads was not able to bring huge sales like competitors because he failed the nuture them. Even after making the same offers and the same patterns of marketing as competitors, the sales declined because people thought this is some spam that everyone is doing because the template of the ads was very professional and similar. suprising one is people fall for basic templates thinking it's a underrated one. so what we did here is we integrated a few softwares into one and set up all same webinars, automated email and sms follow ups, ad managers for stats, launched him a free LMS platform where without any additional fees so he can uploaded unlimited courses, skool like community and add product's like Shopify ( he was selling few merchandise with his brand name on) where he can add unlimited products with connection to all payment gateway, integrated with crm with unlimited contacts, workflow and lead nurturing with calender syncing for 1 on 1 calls. But these are a bit old school, what we did was even better. integrated a conversational ai with all of his sales platforms and gave a nocode automation builder with ai for the workflow. we also set him up with a ai voice agent that's automatically calls and markets for his course and also replies for queries when called. we also set up him a dedicated afflitate manager portal with automated commissions. Though he didn't cross 100k Mark, He did a great number after this. He was struggling with 6k sales, now he has reached somewhere mid of 45k to 50k mrr. Max he hit was 61.8k. I see this a big difference.So one small thing, nurturing the lead can bring you immense sales. To set up all of this it costs around 1.2k monthly for me with all the bills. ( I know there are few free for Individual user platforms out there, but It gets very costly when you switch to their premium plans. with heavy volumes you would require more than premium they offer.) I offered him like 3k per month to work as a agency for him who takes care of all these stuffs. He declined and offered for one time set up fee stating that he will pay 1.2k directly. The one time fee was also a bit low, though I agreed since this was a learning for me. what happened next after that is, he referred me to a few other people in the same niche. But the problem is they are not interested in spending 1 to 2 k in bills for software. They requested that if, will I be able to provide the saas alone for less than 500 dollars with one time set up fee. I haven't responded yet since I have to take an enterprise plan for all the software used and pay full advance price for billings. Then to break even that I have to make minimum 50 or odd users for that. let's grantly say 100 users with all other future costs. So here's what I'm planning to do. I'm planning to offer this as saas for let's say 239 dollars per month. with may or may not one time set up fee. ( I checked the entire internet, there is no single person offering at this price point for unlimited. Also one can easily start their marketing agency with this.) The suggestion and validation that I need here is 1.are you going through the same struggles or faced these struggles? would you be interested to buy at 239 dollars per month? let's say you're from a different niche, Did the features I told were okay for you or you need something specific for your industry that you will be interested in buying? please answer in comments and if you will purchase for this price let me know in comments/dms. I will take that into account and if the response rate is above 100 queries, then will integrate this and sell for that price. (ps: If you see this post on similar subs, please bear cause I'm trying to get suggestions from different POV) tl;dr - lead nurturing can massively boost sales *I made a software integration for a client for a 1.2k per month billing and here I want to know if more than 100 people are interested so that I will make this into my own saas and sell it for like a cheap price of 239 dollars per month TIA.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

AI will obsolete most young vertical SAAS startups, I will not promote
reddit
LLM Vibe Score0
Human Vibe Score1
Few_Incident4781This week

AI will obsolete most young vertical SAAS startups, I will not promote

This is an unpopular opinion, but living in New York City and working with a ton of vertical SaaS startups, meaning basically database wrapper startups that engineer workflows for specific industries and specific users, what they built was at one point in time kind of innovative, or their edge was the fact that they built these like very specific workflows. And so a lot of venture capital and seed funding has gone into these types of startups. But with AI, those database wrapper startups are basically obsolete. I personally feel like all of these companies are going to have to shift like quickly to AI or watch all of their edge and what value they bring to the table absolutely evaporate. It's something that I feel like it's not currently being priced in and no one really knows how to price, but it's going to be really interesting to watch as more software becomes generated and workflows get generated. I’m not saying these companies are worth nothing, but their products need to be completely redone EDIT: for people not understanding: The UX is completely different from traditional vertical saas. Also in real world scenarios, AI does not call the same APIs as the front end. The data handling and validation is different. It’s 50% rebuild. Then add in the technical debt, the fact that they might need a different tech stack to build agents correctly, different experience in their engineers. the power struggles that occur inside companies that need a huge change like this could tank the whole thing alone. It can be done, but these companies are vulnerable. The edge they have is working with existing customers to get it right. But they basically blew millions on a tech implementation that’s not as relevant going forwards. Investors maybe better served putting money into a fresh cap table

Technical Co-Founder Seeking Commercial/Marketing Partner for Micro SaaS Projects
reddit
LLM Vibe Score0
Human Vibe Score1
Weekly-Offer-4172This week

Technical Co-Founder Seeking Commercial/Marketing Partner for Micro SaaS Projects

Hi everyone, I’m looking for a commercial or marketing co-founder to join me in developing some Micro SaaS (MSaaS) apps. Here’s a bit about where I’m coming from and what I’m hoping to find: About Me: I’m a full-stack developer with over 15 years of experience, including some work in AI. I’m currently working part-time, which gives me the time to focus on developing MVPs quickly. I’m passionate about creating SaaS solutions and would love to find someone who can help bring these ideas to life. Based in french alps. What I’m Looking For: Role: Non-Technical Co-Founder (Commercial/Marketing) Location: Remote Equity: 50% co-founder stake What I’m Hoping You’ll Bring: Experience: Background in business development, marketing, or similar fields. Vision: An eye for potential in new SaaS ideas and a drive to help make them successful. Commitment: Enthusiasm for building and growing a business together. What’s In It For You: Revenue Potential: Share in the financial rewards of successful products with a 50% equity stake, giving you a direct share of the profits. Fast ROI: Benefit from rapid MVP development, which allows for quicker validation and faster revenue generation. Dynamic Approach: We move quickly—if an app doesn’t gain traction in a few weeks, we pivot to the next idea, keeping our efforts focused on what works. Financial Growth: As we iterate and scale, there are opportunities for significant financial upside based on the success of our products. Shared Success: Be an integral part of a partnership where both of us share equally in the risks and rewards, creating a strong incentive for mutual success. What’s In It For You: Partnership: Equal share in the business (50/50). Opportunity: Work on interesting MSaaS projects with room for creativity. Flexibility: A remote role that fits around your schedule. If you’re interested or would like to learn more, please reach out. I’d be thrilled to discuss how we might work together. Thank you for considering this!

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Why you should consider using small open source fine-tuned models
reddit
LLM Vibe Score0
Human Vibe Score0.929
hamada0001This week

Why you should consider using small open source fine-tuned models

Context I want to start off by giving some context on what fine-tuning is, why it's useful and who it would be useful for: What is fine-tuning? When controlling the output of an LLM there are, broadly, three levels. Prompt engineering, RAG and fine-tuning. Most of you are likely familiar with the first two. Prompt engineering is when you try to optimize the prompt to get the model to do what you want better. RAG (retrieval augmented generation) is when you first do a search on some data (usually stored in a vector database which allows you to search by similarity), then you insert the results into the prompt so that the model can use that context to more accurately answer any questions. It's like letting the LLM access external information right before answering, using that additional context to improve its response Fine-tuning is when you want to fundamentally teach a model something new or teach it to behave in a particular way. You would provide the model with high quality data (i.e. inputs and outputs) which it will train on. Why is it useful? At the moment, many of you use the largest and best LLMs because they give the best results. However, for a lot of use cases you are likely using a sledgehammer for a small nail. Does it do a great job? Damn yeah! Well... why not use a smaller hammer? Because it might miss or hit your finger. The solution shouldn't be to use a sledgehammer, but rather to learn how to use a smaller hammer properly so you never miss! That's exactly what fine-tuning a smaller model is like. Once you fine-tune it on a specific task with good high quality data, it can surpass even the best models at that specific task. It'll be 10x cheaper to run, much faster and, if you use an open source model, you'll own the model (no vendor lock-in!). If you run a SaaS and your biggest expense is AI costs then you should definitely consider fine-tuning. It'll take some time to set up but it'll be well worth it in the medium/long term (a bit like SEO). You can always resort to the best models for more complex tasks. How to fine-tune? I'm going to give you a breakdown of the process from beginning to end. You do need to be (a bit) technical in order to do this. Getting the data Let's suppose we want to fine-tune a model to make high-quality SEO content. At the moment, you might be using a large sophisticated prompt or using multiple large LLMs to write different parts or utilizing RAG. This is all slow and expensive but might be giving you great results. Our goal is to replace this with a fine-tuned model that is great at one thing: writing high-quality SEO content quickly at a much lower cost. The first step is gathering the appropriate data. If you want the model to write 3 or 4 paragraphs based on a prompt that contains the topic and a few keywords, then your data should match that. There are a few way you can do this: You can manually gather high-quality SEO content. You'd write the prompt and the response that the model should give. You can use a larger more powerful LLM to generate the content for you (also known as synthetic data). It'll be expensive but remember that it'll be a larger one-off cost to get the data. If you already have a pipeline that works great then you can use the prompts and the generated content that you already have from that pipeline. You can buy a high-quality dataset or get someone to make it for you. The data is the most important part of this process. Remember, garbage in garbage out. Your data needs to have a good variety and should not contain any bad examples. You should aim for around 1000 examples. The more the better! The actual fine-tuning. At this stage you are now ready to choose a model and setup the fine-tuning. If you are unsure I'd stick to the Llama 3.1 family of models. They are great and reliable. There are three models: 8b, 70b and 405b. Depending on the complexity of the task you should select an appropriate size. However, to really reap the cost saving benefits and the speed you should try to stick with the 8b model or the the 70b model if the 8b is not good enough. For our SEO example, let's use the 8b model. Important note on selecting a model: You might see multiple models with the 8b flag. You might see 4bit-bnb or instruct. The instruct version of the models have basically been trained to be chatbots. So if you want to keep the chatbot-like instruction-following functionality then you should use the instruct version as the base. The non-instruct version simply generates text. It won't 'act' like a chatbot which is better for use cases like creative writing. The 4bit-bnb means that the model has been 'quantized'. Basically it has been made 4x smaller (the original is in 16 bits) so that it is faster to download and faster to fine-tune. This slightly reduces the accuracy of the model but it's usually fine for most use cases :) Fine-tuning should be done on a good GPU. CPU aren't good enough. So you can't spin up a droplet on digital ocean and use that. You'll specifically need to spin up a GPU. One website that I think is great is Runpod .io (I am not affiliated with them). You simply pay for the GPU by the hour. If you want the training to be fast you can use the H100, if you want something cheaper but slower you can use the A40. Although the A40 won't be good enough to run the 70b parameter model. For the 405b model you'll need multiple H100s but let's leave that for more advanced use cases. Once you've spun up your H100 and ssh-ed into it. I would recommend using the unsloth open source library to do the fine-tuning. They have great docs and good boilerplate code. You want to train using a method called QLoRA. This won't train the entire model but only "part of it". I don't want to get into the technical details as t3hat isn't important but essentially it's a very efficient and effective way of fine-tuning models. When fine-tuning you can provide something called a 'validation set'. As your model is training it will be tested against the 'validation set' to see how well it's doing. You'll get an 'eval loss' which basically means how well is your model doing when compared with the unseen validation data. If you have 1000 training examples I'd recommend taking out 100-200 so it can act as the validation set. Your model may start off with an eval loss of 1.1 and by the end of the training (e.g. 3 epochs - the number of epochs is the number of times your model will be trained on the entire dataset. It's like reading a book more than once so you can understand it better. Usually 3-5 epochs is enough) the eval loss would drop to 0.6 or 0.7 which means your model has made great progress in learning your dataset! You don't want it to be too low as that means it is literally memorizing which isn't good. Post fine-tuning You'll want to save the model with the best eval loss. You actually won't have the whole model, just something called the "QLoRA adapters". These are basically like the new neurons that contain the "understanding" of the data you trained the model on. You can combine these with the base model (using unsloth again) to prompt the model. You can also (and I recommend this) convert the model to GGUF format (using unsloth again). This basically packages the QLoRA adapters and model together into an optimized format so you can easily and efficiently run it and prompt it (using unsloth again... lol). I would then recommend running some evaluations on the new model. You can do this by simply prompting the new model and a more powerful model (or using your old pipeline) and then asking a powerful model e.g. Claude to judge which is better. If your model consistently does better then you've hit a winner! You can then use runpod again to deploy the model to their serverless AI endpoint so you only pay when it's actually being inferenced. (Again, I'm not affiliated with them) I hope this was useful and you at least got a good idea of what fine-tuning is and how you might go about doing it. By the way, I've just launched a website where you can easily fine-tune Llama 3.1 models. I'm actually hoping to eventually automate this entire process as I believe small fine-tuned models will be much more common in the future. If you want more info, feel free to DM me :)

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies) (I will not promote)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. (I will not promote) About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

After building an AI Co-founder to solve my startup struggles, I realized we might be onto something bigger. What problems would you want YOUR AI Co-founder to solve?
reddit
LLM Vibe Score0
Human Vibe Score0
Consistent_Yak6765This week

After building an AI Co-founder to solve my startup struggles, I realized we might be onto something bigger. What problems would you want YOUR AI Co-founder to solve?

A few days ago, I shared my entrepreneurial journey and the endless loop of startup struggles I was facing. The response from the community was overwhelming, and it validated something I had stumbled upon while trying to solve my own problems. In just a matter of days, we've built out the core modules I initially used for myself, deep market research capabilities, automated outreach systems, and competitor analysis. It's surreal to see something born out of personal frustration turning into a tool that others might actually find valuable. But here's where it gets interesting (and where I need your help). While we're actively onboarding users for our alpha test, I can't shake the feeling that we're just scratching the surface. We've built what helped me, but what would help YOU? When you're lying awake at 3 AM, stressed about your startup, what tasks do you wish you could delegate to an AI co-founder who actually understands context and can take meaningful action? Of course, it's not a replacement for an actual AI cofounder, but using our prior entrepreneurial experience and conversations with other folks, we understand that OUTREACH and SALES might actually be a big problem statement we can go deeper on as it naturally helps with the following: Idea Validation - Testing your assumptions with real customers before building Pricing strategy - Understanding what the market is willing to pay Product strategy - Getting feedback on features and roadmap Actually revenue - Converting conversations into real paying customers I'm not asking you to imagine some sci-fi scenario, we've already built modules that can: Generate comprehensive 20+ page market analysis reports with actionable insights Handle customer outreach Monitor competitors and target accounts, tracking changes in their strategy Take supervised actions based on the insights gathered (Manual effort is required currently) But what else should it do? What would make you trust an AI co-founder with parts of your business? Or do you think this whole concept is fundamentally flawed? I'm committed to building this the right way, not just another AI tool or an LLM Wrapper, but an agentic system that can understand your unique challenges and work towards overcoming them. Whether you think this is revolutionary or ridiculous, I want to hear your honest thoughts. But more importantly, I want to hear your unfiltered feedback in the comments. What would make this truly valuable for YOU? Edit 1: The AI cofounder will take no equity in your startup.

I started a Tech Startup, and I feel totally STUCK.
reddit
LLM Vibe Score0
Human Vibe Score1
BetAltruistic6556This week

I started a Tech Startup, and I feel totally STUCK.

I made "Visual Love," a Computer Vision/AI-driven matchmaking platform. The idea is that although appearance is one of the biggest factors for starting a relationship, current matchmaking services and dating apps do not have the capability to search for people based on appearance. On Visual Love, you can find your ideal match simply by uploading a picture of your "ideal type." Also, you can connect with someone who thinks of you as their ideal type, simply by uploading your own picture. Or, there might be a perfect (mutually ideal) match. I made this CV/AI algorithm to scan faces, retrieve facial features, and make it possible to find the closest match among millions of others in a second. On average, regular dating app users swipe 8000 times over 8 months until they find their love. On Visual Love, users can find one in a million just in a second. You can try the tech demo on the website if you want to (find the link through my LinkedIn at the bottom of the post; I have to follow the "I will not promote" rule.) I thought this app would have the best chance in Asia, as people care a lot more about appearance in Asia (especially Korea and Japan). Also, my nationality is Korean, and I speak both Korean and Japanese as fluently as I speak English. So I came to Korea, and pitched to a number of VC/AC firms in Korea and Japan, and two of them were typically intersted in making investment. However, they both required me to provide market validation: how much it would cost per user acquisition, how much each user would pay on average, and etc, even after I provided them with a 3-years financial projection including market research based on other dating apps. &#x200B; Everything might be going just as expected, or even better than anticipated, but I'm feeling very stuck now. I am not a business expert, and I don't have much idea on how to proceed from here. The problem is, it wouldn't quite work as expected when there are not many users. If I start with a small group of users, it's not any better than any other dating app. Matching users within a small group doesn't quite reflect the values of Visual Love. So I figured a way around: making a game version of Visual Love targeting 100k to 500k users to work as an initial distribution channel. This version will include finding look-alike celebrities, and solving look-alike face puzzles, and etc. But now, the problem is, I cannot continue this project by myself. I have no social/financial support, and I'm running low on cash. Also, although I'm from Korea, I lived in many different countries. I did my undergraduate in New York (Columbia University) and all my friends are in the US. I don't feel very included here. I can't stop feeling frustrated and distressed :( I'm sure Visual Love can reshape the future of the matchmaking market. But, only if I can continue this project by getting the fund I require. I'm open to any advice, and if you're interested in providing any help or working with me, please contact me through LinkedIn. https://www.linkedin.com/in/don-lee-3853b1264/

I spent 6 months on building a tool, and got 0 zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a tool, and got 0 zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product, Summ, that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

After building an AI Co-founder to solve my startup struggles, I realized we might be onto something bigger. What problems would you want YOUR AI Co-founder to solve?
reddit
LLM Vibe Score0
Human Vibe Score0
Consistent_Yak6765This week

After building an AI Co-founder to solve my startup struggles, I realized we might be onto something bigger. What problems would you want YOUR AI Co-founder to solve?

A few days ago, I shared my entrepreneurial journey and the endless loop of startup struggles I was facing. The response from the community was overwhelming, and it validated something I had stumbled upon while trying to solve my own problems. In just a matter of days, we've built out the core modules I initially used for myself, deep market research capabilities, automated outreach systems, and competitor analysis. It's surreal to see something born out of personal frustration turning into a tool that others might actually find valuable. But here's where it gets interesting (and where I need your help). While we're actively onboarding users for our alpha test, I can't shake the feeling that we're just scratching the surface. We've built what helped me, but what would help YOU? When you're lying awake at 3 AM, stressed about your startup, what tasks do you wish you could delegate to an AI co-founder who actually understands context and can take meaningful action? Of course, it's not a replacement for an actual AI cofounder, but using our prior entrepreneurial experience and conversations with other folks, we understand that OUTREACH and SALES might actually be a big problem statement we can go deeper on as it naturally helps with the following: Idea Validation - Testing your assumptions with real customers before building Pricing strategy - Understanding what the market is willing to pay Product strategy - Getting feedback on features and roadmap Actually revenue - Converting conversations into real paying customers I'm not asking you to imagine some sci-fi scenario, we've already built modules that can: Generate comprehensive 20+ page market analysis reports with actionable insights Handle customer outreach Monitor competitors and target accounts, tracking changes in their strategy Take supervised actions based on the insights gathered (Manual effort is required currently) But what else should it do? What would make you trust an AI co-founder with parts of your business? Or do you think this whole concept is fundamentally flawed? I'm committed to building this the right way, not just another AI tool or an LLM Wrapper, but an agentic system that can understand your unique challenges and work towards overcoming them. Whether you think this is revolutionary or ridiculous, I want to hear your honest thoughts. But more importantly, I want to hear your unfiltered feedback in the comments. What would make this truly valuable for YOU? Edit 1: The AI cofounder will take no equity in your startup.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

I spent 6 months on building a web product, and got zero users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on building a web product, and got zero users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I have stuff to post on Reddit very rarely, but I share how my project is going on, random stuff, and memes on X. Just in case few might want to keep in touch 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2C products beats building B2B products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.
reddit
LLM Vibe Score0
Human Vibe Score0.714
level6-killjoyThis week

GPT Weekly - 19the June Edition - OpenAI's function calling, Meta's free LLM, EU Regulation and more.

This is a recap covering the major news from last week. 🔥Top 3 news - OpenAI’s updates, Meta’s upcoming free LLM and EU Regulation 🗞️Interesting reads include PSA about protecting your keys, The GPT ouroboros, Reddit - OpenAI’s moat, and more.. 🧑‍🎓Learning includes a Step-by-step guide from a non-technical founder who launched his MVP, Chatbot for your Gdrive and more 🔥Top 3 AI news in the past week OpenAI: New Pricing, Models, & Functions OpenAI has been on a roll. Last week we saw the release of OpenAI best practice on using GPT. This week we saw some amazing updates. Three major buckets were: First, the price decreases for both embeddings and GPT-3.5 tokens. Second, new models for gpt-4 and gpt-3.5. A new longer context model for gpt-3.5. Third, a new function calling capability. Why is it important? Previously, the output from OpenAI was all text. So, calling an external API from GPT was quite difficult. You had to parse the text data and things were often incorrect. Langchain created the Agents and Tools feature to tackle this problem. It was still unreliable and prone to issues. Now you get native support to generate a fixed format output. You can use the output to generate functional calls and also pass functions which need to be called. For example, if your app has multiple API endpoints then you can use GPT to generate the API calls with parameters. You can also pass the endpoints as function calls to ensure the correct function is executed. This functionality can further be used to generate structured data (JSON) out of GPT. So, you can generate data from GPT and load it into your backend. What’s next? This functionality allows turning natural language responses into structured data. This can be used to create “intelligent” backends using LLMs. We might see implementations in no-code tools to allow more robust and natural-language tools for non-technical folks. The structured data process goes both ways. You can also feed structured data into GPT for better responses. This feature also has its share of issues. Function calling suffers from the same prompt injection issues. Malicious actors can pass malicious code in function or the responses. For example, creation of queries using functions might contain malicious code to delete data. Without proper user validation this code will be executed automatically and delete data. So, using LLM as the back-end layer needs proper security implementation. Meta's LLM: Commercial Use Ahead Llama has been a boon for the open source community. Many of the open source models rely on Llama. The issue is that Llama is research-only and cannot be used commercially. So, no one can use it to build any product. Meta is now working on the next version of the model. This model will be available for commercial use. This is in stark contrast to both OpenAI and Google. Both safe-guarde their models and make it available through API. Why is it important? Certain industries cannot use LLM APIs because of strict restrictions on data privacy. These companies would want to run their own instance of a foundational model. A commercially available foundational model is also going to help people who want to keep their “API call” costs next to 0. A commercially available free-for-all model will also help push the open source community further. Just like Llama. What’s next? Sam Altman has said OpenAI didn’t release GPT-3 as open-source because they didn’t think people would be able to run it. Now OpenAI is working on an open-source model. This is going to be weaker than GPT-4. Let the battle of LLMs begin. EU's Proposed Legislation and Its Impact on AI Usage The EU parliament voted to move ahead with the E.U. AI Act. This act aims to ensure consumer protection against the dangers of AI. Why is it important? OpenAI and Sam Altman want regulations for models. They have proposed a IAEA-type of agency to stop the proliferation of LLM models. As per OpenAI, all models should be regulated and monitored. The suggestion of a license based regulation has led to significant backlash. Many people have called it “regulatory capture” - with the aim of shutting down competing LLMs. Licensing based regulations might not really be effective. The EU is approaching regulation from a different angle. It doesn’t focus on how models are developed. Rather focuses on how AI will/can be used. They have broken down use cases into 4 categories - unacceptable (prohibited), high, medium and low risk. For example, Building a Pre-Crime software,on%20crimes%20not%20yet%20committed.) to predict crimes? Building a Social credit system? Unacceptable. Using tools to influence elections or recommendation algorithms? High (Highly regulated). Using generative AI tools to create text or images on news sites? Medium (Add label that the content is AI generated) AI providers also need to disclose their training source. To me this sounds like good legislation. What do you guys think? But, OpenAI has warned that EU regulations might force them to pull out completely. What’s next? The disclosure requirements might help various publishing companies. AI and media companies are in talks to pay for training data. Google has been leading the charge. Additionally, OpenAI and Deepmind will open their models for safety and research purposes to the UK government. 🗞️10 AI news highlights and interesting reads PSA: If you are using Repl to write code, you might want to check your OpenAI API keys. If you have left them embedded then people can pirate and steal the keys. LLMs rely on human annotation or human feedback to learn. And one way to generate human annotation is crowdsourcing. But what if the crowdsource human annotators use LLMs? Research shows 33-46% workers used LLMs. So, basically we go from Human -> AI -> Human -> AI. The AI ouroboros. Researchers also say generated data to train models might cause serious issue. All the talks about moats \- Reddit might be OpenAI’s \future\ moat. Given the amount of complaints about how Google search experience has deteriorated during the blackout, this might be true? Doctors are using ChatGPT but not to diagnose.Rather to be more empathetic. We discussed this just a month ago. And guess where the data for this study came from? Reddit AskDocs. Moat FTW?! Beatles to make a comeback…using Generative AI. SnapFusion - Text to Image diffusion on mobile phones. Large context lengths are important for better GPT experience. The secret sauce for 100k context length. There is a lot of bad AI research out there. Some border on snake oil. Most AI “research” should be double checked and challenged. A new research on huggingface said that GPT-4 can ace MIT curriculum. Now someone is replicating the results and say that GPT-4 can’t beat MIT. Are we seeing peak AI? Especially when people from Deepmind and Meta are involved? Mistral AI raised $113 million in seed round with no product. Some might say this funding is for the team and the team is really solid. The issue though is whether the valuation is justified when OpenAI and Google already have a head start. The AI Hype Wall of Shame. \- Collection of articles which mislead people about AI in various aspects. 🧑‍🎓3 Learning Resources Building and Launching a company using GPT-4 with prompts. (The author didn’t know how to code but created and launched the MVP in a month). Chatbot for your Gdrive - https://www.haihai.ai/gpt-gdrive/ Building ChatGPT plugin using Supabase - https://supabase.com/blog/building-chatgpt-plugins-template That’s it folks. Thank you for reading and have a great week ahead. If you are interested in a focused weekly recap delivered to your inbox on Mondays you can subscribe here. It is FREE!

6 principles to data architecture that facilitate innovation
reddit
LLM Vibe Score0
Human Vibe Score1
Competitive_Speech36This week

6 principles to data architecture that facilitate innovation

My team and I have been re-building our company's data architecture. In the process of doing so, I got together six key principles to transforming data architectures and thought I would share them, as a strong data architecture is crucial for businesses looking to stay competitive in the digital landscape, as it improves decision-making, time to market, and data security. When executed with efficiency, a resilient data architecture unleashes unparalleled degrees of agility. Principle 1: Agility and flexibility To quickly adjust to market fluctuations, businesses must create adaptable data infrastructures that can effortlessly manage an ever-growing influx of data. To accomplish this objective, we recommend to our clients to implement Enterprise Service Bus, Enterprise Data Warehouse, and Master Data Management integrated together. &#x200B; I believe the best option is this: \- By centralizing communication, ESB reduces the time and effort required to integrate new systems; \- EDW consolidates data from different sources, resulting in a 50% reduction in software implementation time; \- Finally, MDM ensures consistency and accuracy across the organization, leading to better decision-making and streamlined operations. Implementing these solutions can lead to reduced software implementation time, better ROI, and more manageable data architecture. By fostering a culture of collaboration and adopting modern technologies and practices, businesses can prioritize agility and flexibility in their data architecture to increase the pace of innovation. Principle 2: Modularity and reusability Data architecture that fosters modularity and reusability is essential for accelerating innovation within an organization. By breaking data architecture components into smaller, more manageable pieces, businesses can enable different teams to leverage existing architecture components, reducing redundancy and improving overall efficiency. MDM can promote modularity and reusability by creating a central repository for critical business data. This prevents duplication and errors, improving efficiency and decision-making. MDM enables a single source of truth for data, accessible across multiple systems, which promotes integration and scalability. MDM also provides standardized data models, rules, and governance policies that reduce development time, increase quality, and ensure proper management throughout the data’s lifecycle. Another way to achieve modularity in data architecture is through the use of microservices and scripts for Extract, Transform, and Load (ETL) processes. Adopting a structured methodology and framework can ensure these components are well-organized, making it easier for teams to collaborate and maintain the system. Microservices can also contribute to modularity and reusability in data architecture. These small, independent components can be developed, deployed, and scaled independently of one another. By utilizing microservices, organizations can update or replace individual components without affecting the entire system, improving flexibility and adaptability. Principle 3: Data quality and consistency The efficiency of operations depends on data’s quality, so a meticulously crafted data architecture plays a pivotal role in preserving it, empowering enterprises to make well-informed decisions based on credible information. Here are some key factors to consider that will help your company ensure quality: \- Implementing Master Data Management (MDM) – this way, by consolidating, cleansing, and standardizing data from multiple sources, your IT department will be able to create a single, unified view of the most important data entities (customers, products, and suppliers); \- Assigning data stewardship responsibilities to a small team or an individual specialist; \- Considering implementing data validation, data lineage, and data quality metrics; \- By implementing MDM and adopting a minimal data stewardship approach, organizations can maintain high-quality data that drives innovation and growth. Principle 4: Data governance Data governance is a strategic framework that goes beyond ensuring data quality and consistency. It includes ensuring data security, privacy, accessibility, regulatory compliance, and lifecycle management. Here are some key aspects of data governance: \- Implementing robust measures and controls to protect sensitive data from unauthorized access, breaches, and theft. This is only possible through including encryption, access controls, and intrusion detection systems into your company’s IT architecture; \- Adhering to data privacy regulations and guidelines, such as the General Data Protection Regulation (GDPR) and the California Consumer Privacy Act (CCPA); \- Defining stringent conditions for who has access to specific data assets to maintain control over data and ensure its accessibility only for legitimate purposes. Managing the entire lifecycle of data, from creation and storage to archiving and disposal, including defining policies for data retention, archiving, and deletion in compliance with legal and regulatory requirements. To facilitate effective data governance, organizations can leverage various tools and technologies, such as: \- Data cataloging tools: Solutions like Collibra, Alation, or Informatica Enterprise Data Catalog help organizations discover, understand, and manage their data assets. \- Data lineage tools: Tools like Talend, IBM InfoSphere, or Apache Atlas help track data’s origin, transformation, and usage, providing insights into data quality issues and potential areas for improvement. \- Data quality tools: Solutions like Informatica Data Quality, Trifacta, or SAS Data Quality help organizations maintain high-quality data by identifying and correcting errors, inconsistencies, and inaccuracies. \- Data security and privacy tools: Tools like Varonis, BigID, or Spirion help protect sensitive data and ensure compliance with data privacy regulations. Principle 5: Cloud-first approach A cloud-first approach prioritizes cloud-based solutions over on-premises ones when it comes to data management. Cloud-based data management pros: \- Virtually limitless scalability, so that organizations can grow and adapt to changing data requirements without significant infrastructure investments; \- The pay-as-you-go model of cloud services reduces maintenance costs usually associated with the on-premise choice; \- Greater flexibility for deploying and integrating new technologies and services; \- Cloud can be accessed from anywhere, at any time, turning team collaboration and remote work into a breeze; \- Built-in backup and disaster recovery capabilities, ensuring data safety and minimizing downtime in case of emergencies. Cloud-based data management cons: \- Cloud-first approach raises many data security, privacy, and compliance concerns; \- Transferring large data volumes to and from cloud is often time-consuming and results in increased latency for certain apps; \- Relying on a single cloud provider makes it difficult to switch them or move back to the on-premises option without significant funds and effort. Challenges that organizations that choose a cloud-first approach face: \- Integrating cloud-based systems with on-premises ones can be complex and time-consuming; \- Ensuring data governance and compliance in a multi-cloud or hybrid environment is also another problem reported by my clients. How EDW, ESB, and MDM promote cloud-first approach: A cloud-based EDW centralizes data from multiple sources, enabling a unified view of the organization’s data and simplifying data integration across cloud and on-premises systems. An ESB facilitates communication between disparate cloud and on-premises systems, streamlining data integration and promoting a modular architecture. Cloud-based MDM solutions are used for maintaining data quality and consistency across multiple data sources and environments. Principle 6: Automation and artificial intelligence Incorporating automation tools and AI technologies into data architecture can optimize processes and decision-making. Key Applications: \- Data ingestion and integration: Automation simplifies data schema updates and identifies data quality issues, while AI-assisted development helps create tailored connectors, scripts, and microservices. \- Data quality management: Machine learning algorithms improve data quality and consistency by automatically detecting and correcting inconsistencies and duplicates. \- Predictive analytics: AI and machine learning models analyze historical data to predict trends, identify opportunities, and uncover hidden patterns for better-informed decisions. How No-Code Tools and AI-Assisted Development Work: Business users define data requirements and workflows using no-code tools, enabling AI models to understand their needs. AI models process the information, generating recommendations for connector creation, ETL scripts, and microservices. Developers use AI-generated suggestions to accelerate development and tailor solutions to business needs. By combining automation, AI technologies, and no-code tools, organizations can streamline data architecture processes and bridge the gap between business users and developers, ultimately accelerating innovation. I share more tips on building an agile data architectures in my blog.

My Building Of Trading Order Management System Using AI Agents
reddit
LLM Vibe Score0
Human Vibe Score0
AniketWorkThis week

My Building Of Trading Order Management System Using AI Agents

Practical Guide : Automating Business Transactions with AI-Powered Workflows Full Article | Code https://preview.redd.it/hrkeo00yz4ie1.jpg?width=1911&format=pjpg&auto=webp&s=5bcb6f02c72bbce22fb691e4d8b799c414fed2a7 https://preview.redd.it/1cp0izzxz4ie1.png?width=1899&format=png&auto=webp&s=2598e25e17ab03a95f3009f5333f02b077ce30ca https://preview.redd.it/cjp1640yz4ie1.png?width=1899&format=png&auto=webp&s=13dad0ee8e0b1b22415a60a57b571058f0bdef33 TL;DR A practical implementation of an AI-powered B2B order management system using LangChain and LLM, demonstrating automated order processing, inventory management, and real-time communication between trading partners. https://i.redd.it/kxe4l69105ie1.gif Introduction In today’s fast-paced business environment, efficient order management is crucial for B2B operations. GlobalTrade Nexus AI showcases how artificial intelligence can streamline complex business transactions, reduce errors, and enhance communication between trading partners. What’s This Article About? This article presents a comprehensive B2B trading platform that leverages AI to automate order processing workflows. The system handles everything from order placement to fulfillment, featuring: Real-time inventory verification Automated shipping cost calculations Instant order validation Secure transaction processing Smart order cancellation capabilities State management across the entire order lifecycle The platform demonstrates how modern AI technologies can be integrated into traditional business processes to create a seamless, efficient trading environment. Tech stack Why Read It? As businesses increasingly embrace digital transformation, AI-powered solutions are becoming essential for maintaining competitive advantage. This article provides: A practical example of AI implementation in B2B commerce Insights into modern system architecture for business applications Real-world application of language models in business logic Demonstration of secure and scalable state management Blueprint for building similar AI-enhanced business systems Through our fictional companies’ implementation, readers can understand how AI can transform their business operations and prepare for the future of B2B commerce.

My Building Of Trading Order Management System Using AI Agents
reddit
LLM Vibe Score0
Human Vibe Score0
AniketWorkThis week

My Building Of Trading Order Management System Using AI Agents

Practical Guide : Automating Business Transactions with AI-Powered Workflows Full Article | Code https://preview.redd.it/hrkeo00yz4ie1.jpg?width=1911&format=pjpg&auto=webp&s=5bcb6f02c72bbce22fb691e4d8b799c414fed2a7 https://preview.redd.it/1cp0izzxz4ie1.png?width=1899&format=png&auto=webp&s=2598e25e17ab03a95f3009f5333f02b077ce30ca https://preview.redd.it/cjp1640yz4ie1.png?width=1899&format=png&auto=webp&s=13dad0ee8e0b1b22415a60a57b571058f0bdef33 TL;DR A practical implementation of an AI-powered B2B order management system using LangChain and LLM, demonstrating automated order processing, inventory management, and real-time communication between trading partners. https://i.redd.it/kxe4l69105ie1.gif Introduction In today’s fast-paced business environment, efficient order management is crucial for B2B operations. GlobalTrade Nexus AI showcases how artificial intelligence can streamline complex business transactions, reduce errors, and enhance communication between trading partners. What’s This Article About? This article presents a comprehensive B2B trading platform that leverages AI to automate order processing workflows. The system handles everything from order placement to fulfillment, featuring: Real-time inventory verification Automated shipping cost calculations Instant order validation Secure transaction processing Smart order cancellation capabilities State management across the entire order lifecycle The platform demonstrates how modern AI technologies can be integrated into traditional business processes to create a seamless, efficient trading environment. Tech stack Why Read It? As businesses increasingly embrace digital transformation, AI-powered solutions are becoming essential for maintaining competitive advantage. This article provides: A practical example of AI implementation in B2B commerce Insights into modern system architecture for business applications Real-world application of language models in business logic Demonstration of secure and scalable state management Blueprint for building similar AI-enhanced business systems Through our fictional companies’ implementation, readers can understand how AI can transform their business operations and prepare for the future of B2B commerce.

How I landed an internship in AI
reddit
LLM Vibe Score0
Human Vibe Score1
Any-Reserve-4403This week

How I landed an internship in AI

For motivational purposes only! I see a lot of posts on here from people without “traditional” machine learning, data science, etc.. backgrounds asking how they can break into the field, so I wanted to share my experience. EDIT Learning Resources and Side Project Ideas * My background: I graduated from a decent undergraduate school with a degree in Political Science several years ago. Following school I worked in both a client services role at a market research company and an account management role at a pretty notable fintech start-up. Both of these roles exposed me to ML, AI and more sophisticated software concepts in general, and I didn’t really care for the sales side of things, so I decided to make an attempt at switching careers into something more technical. While working full time I began taking night classes at a local community college, starting with pre calculus all the way up to Calc 2 and eventually more advanced classes like linear algebra and applied probability. I also took some programming courses including DSA. I took these classes for about two years while working, and on the side had been working through various ML books and videos on YouTube. What worked the best for me was Hands-on Machine Learning with Scikit Learn, Keara’s and Tensorflow. I eventually had enough credits where I was able to begin applying to MS in Data Science programs and was fortunate enough to get accepted into one and also get a position in their Robotics Lab doing Computer Vision work. When it came time to apply for internships, it was a BLOODBATH. I must have applied to over 100 roles with my only responses being video interviews and OA’s. Finally I got an interview for an AI Model Validation internship with a large insurance company and after completing the interviews was told I performed well but they were still interviewing several candidates. I ended up getting the offer and accepting the role where I’ll be working on a Computer Vision model and some LLM related tasks this summer and could not be more fortunate / excited. A couple things stood out to them during the interview process. 1, the fact that I was working and taking night classes with the intent to break into the field. It showed a genuine passion as opposed to someone who watched a YouTube video and claims they are now an expert. 2, side projects. I not only had several projects, but I had some that were relevant to the work I’d be doing this summer from the computer vision standpoint. 3, business sense. I emphasized during my interviews how working in a business role prior to beginning my masters would give me a leg up as intern because I would be able to apply the work of a data scientist to solving actual business challenges. For those of you trying to break into the field, keep pushing, keep building, and focus on what makes you unique and able to help a company! Please feel free to contact me if you would like any tips I can share, examples of projects, or anything that would be helpful to your journey.

I am building a free app to fight propaganda online. Need your help
reddit
LLM Vibe Score0
Human Vibe Score-1
Used_Park_1937This week

I am building a free app to fight propaganda online. Need your help

So long story short I was recently bombarded with an insane amounts of right-wing propaganda online and Russian bots so I decided I can somehow simplify the validation of the facts that are posted by people online. I have made a research of tools that are currently available for that and I've found that ChatGpt is quite biased with some topics and it lacks data about latest issues even when search mode is enabled. At least it takes some time for it to update it's memory or something Then I've checked a Google Fact Check tools which were even worse. Bad UI... I had no idea how it validates things because usually proof links that Google provides are quite random. So I decided to build a tool that simplifies the research. The idea was simple - you have a claim - you paste it in app and it goes on web collects articles on topic then processes them with ai and makes a decision based on those articles. The app is running right now. It is completely free and no ads. Also it generates a small article where it tells about how it made a conclusion. Also I need your help - How can I make this app better? Also feel free to test your claims so that I can have more real data to tune this app on. I believe that this is something that is necessary for all of us especially in those crazy times! Here is the link - truthorfake.com https://preview.redd.it/vp5fbzpyfkne1.png?width=854&format=png&auto=webp&s=84df7913382634614814866c8df743d8378d28dd

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

Running and selling multiple side projects alongside a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
leanpreneur1This week

Running and selling multiple side projects alongside a 9-5

My current side project started 56 days ago when I started writing 1,000 words per day. My core businesses are an agency and job board, and I just needed a creative outlet. The likes of Chris Guillebeau and Nathan Barry attribute their progression to writing so I thought I’d see if it might do the same for me. At first I was just vomiting words onto the screen, I made a blog and wrote mainly technical guides related to my skills. Over time I realised I was writing more and more about running a business as a solopreneur, or lean operator. There is tons of content out there giving you the Birds Eye of going from 0 to £10m. Inspiring stuff, but I think there is a void in real content, explaining the nuts and bolts of the how.  What is the day-to-day like for the solopreneurs who make a good living and have plenty of free time? That’s what I’m striving for anyway. I’m not talking about the 7-figure outliers. Or the ones teaching you to make content so you can have a business teaching others how to make content, and so on. I’m also sick of the ‘I made $X in 5 minutes and how you can too’  So, I started chatting to people in my network who run lean businesses and/or side hustles. I ask them a bit about their journey and ask them to teach something - how they operate, or a skill/process/system/tool that other people like you/me will find useful. One of my first chats was with Sam Dickie, who runs multiple side projects so thought I’d share here, see if others find it useful and get some feedback. I’ve removed all links as I’ve never posted on Reddit before so conscious of not being promotional, I’m posting this stuff to a tiny email list of friends with no upsells. Just finding my feet on whether others find it useful or not: — Sam is a serial entrepreneur who builds projects in his spare time whilst working a 9-5. He’s scaled and sold multiple ventures and currently runs one of the best newsletters out there for builders and entrepreneurs. Building audience through newsletters has always been a cornerstone strategy for him, so, along with sharing his advice on solopreneurism, he’s also generously shared his lean newsletter writing process. About Sam Sam is a Senior Product Manager who has spent the last 15 years working in the tech sector after starting his career as a town planner. In addition to his job he spends some of his spare time building side projects. These have included a 3D printing startup, a tech directory, a newsletter, a beta product directory, and consultancy. Sam is the epitome of making a success out of following your interest and curiosity. It’s clear he enjoys his business ventures and builds in a risk-free way.   It’s often touted by business gurus to avoid building around your interests, but Sam bucks the trend successfully. I think he’s someone who has already found his 1,000 true fans.  Descending rabbit holes, Sam’s journey of invention and curation 3D printing Sam’s first foray into launching a startup was with Fiilo, a 3D printing business. This was at the height of the 3D printing craze and he self-admits that he used the launch as an excuse to buy a 3D printer. He ended up with two and launching a product called GrowGo. GrowGo is a sustainable 3D-printed product that turns any bottle into somewhere that you can grow plants and herbs. He eventually sold this business and the printers, making around £10k. Along the way, he was exposed to various business tasks, including building a website in Weebly, the biggest nocode website builder of the time, and built an API that enabled print on demand for his product. NoCode.Tech The experiences of building as someone non-technical led to numerous friends asking how he built all of this tech. Back then, nocode wasn’t popular, and it had almost zero search volume, so Sam created a basic directory. A quick landing page on Weebly with a basic value prop, a short explanation and a list of the tools he had used before. It hit the top spot on Product Hunt, and he landed 2,000 subscribers in the first 48 hours. But, he hadn’t built it at this point, so he set about getting to work. He built the directory and list to 30,000 subs and monetised the site through advertising. At its peak with Sam, it was receiving about £2,000 per month in ad revenue. He was still working his 9-5 at this point, so thought it might be a good time to exit. The site was still growing, but it was becoming anxiety inducing whilst he was still working full-time. So, he ended up selling the site and making friend’s with the buyer. Fast forwarding a bit, Nocode.tech was eventually acquired by Stackr, a nocode app. Sam was working for their competitor at the time and ended up being offered a job by his friend who acquired the site. All of this from a side project in his area of passion. Creator Club After selling the directory, Sam lost his outlet for sharing his tools and learnings.  Being fascinated with curation and loving sifting through for nuggets, he invested more time into his personal website and launched Creator Club newsletter. Sam writes monthly and currently has over 8,000 subs. It’s one of the few newsletters that I let bypass my email filters and land in my main inbox. Life as a Part-Time Multipreneur Side Hustler If it’s not obvious already Sam is a curiosity led business creator. He’s found that the products without a revenue focus or intention have ironically outperformed those created for the sole purpose of creating money. He enjoys working on his side hustles. He could have run the Nocode.Tech for 10 more years and wouldn’t have tired of it as it’s a byproduct of his interest. For this reason, he has also created the Beta Directory, simply because he loves unearthing early-stage products. He admits he gets the fear when he thinks about quitting his 9-5, although he suspects if he devoted the same energy to one of his projects it could replace his income (no doubts from me here). This same fear means that he can run his ventures with less fear. This way, he can experiment with freedom and isn’t risking the ranch with a young family to consider. For example, recently he stopped paid sponsors on his newsletter as it was more stress than the value of the income to him. Sam divides his time on evenings and weekends (unequally) between the following: Creator Club Validation Co Beta directory Consultancy The pure side hustle status magnifies the need to run lean, let’s jump into his process…. Sam’s lean newsletter curation and creation process Starting out publishing his personal newsletter Going against his expertise, Sam originally over-engineered his process.  He curated with Feedly and tried to automate the full writing process with Zapier. The trouble is that there are too many points of failure which can lead the whole  chain to break down, and you spend more time fixing the system. For a 200 subscriber newsletter, he needed to pare things back. His set-up now Sam scaled back and now simple builds automations when he needs them. He keeps the process simple, right down to the design and any welcome automations. Keeping things real We touched on the trend that keeping things raw is better. Content has come full circle with the advent of AI. Everything looks too perfect and consequently, people’s tastes are changing. Sam mentioned watermarks that show content isn’t AI written, and we referenced content such as Greg Isenberg’s sketches, and Chris Donnelly’s image posts. \\Step by Step Process:\\ Using Stoop Inbox to manage sources Curation with Pocket Managing content with Airtable and Zapier Using Bearly to summarise Substack for writing Monitoring content sources Sam uses Stoop Inbox, an RSS curation tool, to manage his content sources. It gives him a dedicated email address for newsletters and he follows an Inbox Zero methodology. He checks in daily in Stoop, and on X, Reddit and IndieHackers. With X, he just uses the standard interface but has been careful to curate his feed, sometimes adding in extra notifications to hear from interesting people. Highlighting content When curating links, Sam uses Arc browser and the Pocket extension to save links. It’s super simple and lightweight. He creates tags which trigger an automation that curates the link to Airtable. If you watch the video, here’s a shoutout to Alice, the AI interface I use which has recently featured on Product Hunt. It’s a fantastic tool with bags of potential to enhance a solopreneur’s life. Ranking and sorting content He sends the links indexed using Pocket to a basic Airtable base via Zapier. From there, he grades the content and sets aside some time to read it in more depth. Pocket pulls through the title, metadata, and URL link. Review Sam does this manually but has used a tool as a shortcut for digesting long form content — Bearly.ai. Bearly.ai was created by Trung Phan and linking back to raw content, Trung is 1/3 of the hosts on the Not Investment Advice podcast. Its irreverent style and thumbnail are an example of a successful podcast that doesn’t over polish. Writing it all up Being a huge Notion fan (check out the free templates on his site), Sam originally used Notion for writing and linked it into Revue. When Elon sunsetted Revue, he switched to Substack. He loves the Substack interface so drafts in Substack based on a duplication of last month’s edition. Before publishing, Sam runs through a 10-point Notion checklist, which he shared with me. Parting Advice Keep your tool stack as lean as possible. Avoid tool switching to the shiny new object. Getting launched quickly is key. Don’t think that you have to be everywhere for distribution, Sam sticks with what he knows on X and LinkedIn. Overall, he advises just keeping things simple and therefore minimising risk. Resources He says they’re cliche, but I don’t agree; they’re timeless. Paul Graham of Y Combinator is someone Sam recommends following. He doesn’t write much, which is great as Sam gets anxiety when someone good often writes and he can’t keep up with the writing. His content is well thought out and distills complex concepts in entrepreneurship and startups. In addition, Sam loves Naval Ravikant’s approach. He mentions checking out the Almanac of Naval Ravikant for collected wisdom. Follow Sam’s Journey Again, not going to link here but you can find Sam’s stuff easily enough if you want to. His personal website is beautiful and contains loads of free downloads. He has also curated personal websites he admires if you need some inspiration. Sam is a super nice guy so reach out to him, I did before I started my personal blog recently, and he gave me some great advice. Also, worth keeping an eye on Validation Co, where he aims to help early-stage makers and creators validate their ideas. He’s building super slow — trying to enjoy the process without unachievable deadlines. Maintaining his stamina and passion. Amazing, I hope he writes more about that soon! -- That’s my second shot at an interview, hope you enjoyed it and found something useful in it. I’m talking to a marketplace founder who spends 2–3 hours per month his project, a multiple job board owner with a 9-5 and a leading book designer next. As this is my side project, should I keep going?

How I Built a $6k/mo Business with Cold Email
reddit
LLM Vibe Score0
Human Vibe Score1
Afraid-Astronomer130This week

How I Built a $6k/mo Business with Cold Email

I scaled my SaaS to a $6k/mo business in under 6 months completely using cold email. However, the biggest takeaway for me is not a business that’s potentially worth 6-figure. It’s having a glance at the power of cold emails in the age of AI. It’s a rapidly evolving yet highly-effective channel, but no one talks about how to do it properly. Below is the what I needed 3 years ago, when I was stuck with 40 free users on my first app. An app I spent 2 years building into the void. Entrepreneurship is lonely. Especially when you are just starting out. Launching a startup feel like shouting into the dark. You pour your heart out. You think you have the next big idea, but no one cares. You write tweets, write blogs, build features, add tests. You talk to some lukewarm leads on Twitter. You do your big launch on Product Hunt. You might even get your first few sales. But after that, crickets... Then, you try every distribution channel out there. SEO Influencers Facebook ads Affiliates Newsletters Social media PPC Tiktok Press releases The reality is, none of them are that effective for early-stage startups. Because, let's face it, when you're just getting started, you have no clue what your customers truly desire. Without understanding their needs, you cannot create a product that resonates with them. It's as simple as that. So what’s the best distribution channel when you are doing a cold start? Cold emails. I know what you're thinking, but give me 10 seconds to change your mind: When I first heard about cold emailing I was like: “Hell no! I’m a developer, ain’t no way I’m talking to strangers.” That all changed on Jan 1st 2024, when I actually started sending cold emails to grow. Over the period of 6 months, I got over 1,700 users to sign up for my SaaS and grew it to a $6k/mo rapidly growing business. All from cold emails. Mastering Cold Emails = Your Superpower I might not recommend cold emails 3 years ago, but in 2024, I'd go all in with it. It used to be an expensive marketing channel bootstrapped startups can’t afford. You need to hire many assistants, build a list, research the leads, find emails, manage the mailboxes, email the leads, reply to emails, do meetings. follow up, get rejected... You had to hire at least 5 people just to get the ball rolling. The problem? Managing people sucks, and it doesn’t scale. That all changed with AI. Today, GPT-4 outperforms most human assistants. You can build an army of intelligent agents to help you complete tasks that’d previously be impossible without human input. Things that’d take a team of 10 assistants a week can now be done in 30 minutes with AI, at far superior quality with less headaches. You can throw 5000 names with website url at this pipeline and you’ll automatically have 5000 personalized emails ready to fire in 30 minutes. How amazing is that? Beyond being extremely accessible to developers who are already proficient in AI, cold email's got 3 superpowers that no other distribution channels can offer. Superpower 1/3 : You start a conversation with every single user. Every. Single. User. Let that sink in. This is incredibly powerful in the early stages, as it helps you establish rapport, bounce ideas off one another, offer 1:1 support, understand their needs, build personal relationships, and ultimately convert users into long-term fans of your product. From talking to 1000 users at the early stage, I had 20 users asking me to get on a call every week. If they are ready to buy, I do a sales call. If they are not sure, I do a user research call. At one point I even had to limit the number of calls I took to avoid burnout. The depth of the understanding of my customers’ needs is unparalleled. Using this insight, I refined the product to precisely cater to their requirements. Superpower 2/3 : You choose exactly who you talk to Unlike other distribution channels where you at best pick what someone's searching for, with cold emails, you have 100% control over who you talk to. Their company Job title Seniority level Number of employees Technology stack Growth rate Funding stage Product offerings Competitive landscape Social activity (Marital status - well, technically you can, but maybe not this one…) You can dial in this targeting to match your ICP exactly. The result is super low CAC and ultra high conversion rate. For example, My competitors are paying $10 per click for the keyword "HARO agency". I pay $0.19 per email sent, and $1.92 per signup At around $500 LTV, you can see how the first means a non-viable business. And the second means a cash-generating engine. Superpower 3/3 : Complete stealth mode Unlike other channels where competitors can easily reverse engineer or even abuse your marketing strategies, cold email operates in complete stealth mode. Every aspect is concealed from end to end: Your target audience Lead generation methods Number of leads targeted Email content Sales funnel This secrecy explains why there isn't much discussion about it online. Everyone is too focused on keeping their strategies close and reaping the rewards. That's precisely why I've chosen to share my insights on leveraging cold email to grow a successful SaaS business. More founders need to harness this channel to its fullest potential. In addition, I've more or less reached every user within my Total Addressable Market (TAM). So, if any competitor is reading this, don't bother trying to replicate it. The majority of potential users for this AI product are already onboard. To recap, the three superpowers of cold emails: You start a conversation with every single user → Accelerate to PMF You choose exactly who you talk to → Super-low CAC Complete stealth mode → Doesn’t attract competition By combining the three superpowers I helped my SaaS reach product-marketing-fit quickly and scale it to $6k per month while staying fully bootstrapped. I don't believe this was a coincidence. It's a replicable strategy for any startup. The blueprint is actually straightforward: Engage with a handful of customers Validate the idea Engage with numerous customers Scale to $5k/mo and beyond More early-stage founders should leverage cold emails for validation, and as their first distribution channel. And what would it do for you? Update: lots of DM asking about more specifics so I wrote about it here. https://coldstartblueprint.com/p/ai-agent-email-list-building

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.
reddit
LLM Vibe Score0
Human Vibe Score0.667
GDbuildsGDThis week

I spent 6 months on a web app as a side project, and got 0 users. Here is my story.

Edit Thank you all so much for your time reading my story. Your support, feedback, criticism, and skepticism; all helped me a lot, and I couldn't appreciate it enough \^\_\^ I very rarely have stuff to post on Reddit, but I share how my project is going on, just random stuff, and memes on X. In case few might want to keep up 👀 TL;DR I spent 6 months on a tool that currently has 0 users. Below is what I learned during my journey, sharing because I believe most mistakes are easily avoidable. Do not overestimate your product and assume it will be an exception to fundamental principles. Principles are there for a reason. Always look for validation before you start. Avoid building products with a low money-to-effort ratio/in very competitive fields. Unless you have the means, you probably won't make it. Pick a problem space, pick your target audience, and talk to them before thinking about a solution. Identify and match their pain points. Only then should you think of a solution. If people are not overly excited or willing to pay in advance for a discounted price, it might be a sign to rethink. Sell one and only one feature at a time. Avoid everything else. If people don't pay for that one core feature, no secondary feature will change their mind. Always spend twice as much time marketing as you do building. You will not get users if they don't know it exists. Define success metrics ("1000 users in 3 months" or "$6000 in the account at the end of 6 months") before you start. If you don't meet them, strongly consider quitting the project. If you can't get enough users to keep going, nothing else matters. VALIDATION, VALIDATION, VALIDATION. Success is not random, but most of our first products will not make a success story. Know when to admit failure, and move on. Even if a product of yours doesn't succeed, what you learned during its journey will turn out to be invaluable for your future. My story So, this is the story of a product that I’ve been working on for the last 6 months. As it's the first product I’ve ever built, after watching you all from the sidelines, I have learned a lot, made many mistakes, and did only a few things right. Just sharing what I’ve learned and some insights from my journey so far. I hope that this post will help you avoid the mistakes I made — most of which I consider easily avoidable — while you enjoy reading it, and get to know me a little bit more 🤓. A slow start after many years Summ isn’t the first product I really wanted to build. Lacking enough dev skills to even get started was a huge blocker for so many years. In fact, the first product I would’ve LOVED to build was a smart personal shopping assistant. I had this idea 4 years ago; but with no GPT, no coding skills, no technical co-founder, I didn’t have the means to make it happen. I still do not know if such a tool exists and is good enough. All I wanted was a tool that could make data-based predictions about when to buy stuff (“buy a new toothpaste every three months”) and suggest physical products that I might need or be strongly interested in. AFAIK, Amazon famously still struggles with the second one. Fast-forward a few years, I learned the very basics of HTML, CSS, and Vanilla JS. Still was not there to build a product; but good enough to code my design portfolio from scratch. Yet, I couldn’t imagine myself building a product using Vanilla JS. I really hated it, I really sucked at it. So, back to tutorial hell, and to learn about this framework I just heard about: React.React introduced so many new concepts to me. “Thinking in React” is a phrase we heard a lot, and with quite good reasons. After some time, I was able to build very basic tutorial apps, both in React, and React Native; but I have to say that I really hated coding for mobile. At this point, I was already a fan of productivity apps, and had a concept for a time management assistant app in my design portfolio. So, why not build one? Surely, it must be easy, since every coding tutorial starts with a todo app. ❌ WRONG! Building a basic todo app is easy enough, but building one good enough for a place in the market was a challenge I took and failed. I wasted one month on that until I abandoned the project for good. Even if I continued working on it, as the productivity landscape is overly competitive, I wouldn’t be able to make enough money to cover costs, assuming I make any. Since I was (and still am) in between jobs, I decided to abandon the project. 👉 What I learned: Do not start projects with a low ratio of money to effort and time. Example: Even if I get 500 monthly users, 200 of which are paid users (unrealistically high number), assuming an average subscription fee of $5/m (such apps are quite cheap, mostly due to the high competition), it would make me around $1000 minus any occurring costs. Any founder with a product that has 500 active users should make more. Even if it was relatively successful, due to the high competition, I wouldn’t make any meaningful money. PS: I use Todoist today. Due to local pricing, I pay less than $2/m. There is no way I could beat this competitive pricing, let alone the app itself. But, somehow, with a project that wasn’t even functional — let alone being an MVP — I made my first Wi-Fi money: Someone decided that the domain I preemptively purchased is worth something. By this point, I had already abandoned the project, certainly wasn’t going to renew the domain, was looking for a FT job, and a new project that I could work on. And out of nowhere, someone hands me some free money — who am I not to take it? Of course, I took it. The domain is still unused, no idea why 🤔. Ngl, I still hate the fact that my first Wi-Fi money came from this. A new idea worth pursuing? Fast-forward some weeks now. Around March, I got this crazy idea of building an email productivity tool. We all use emails, yet we all hate them. So, this must be fixed. Everyone uses emails, in fact everyone HAS TO use emails. So, I just needed to build a tool and wait for people to come. This was all, really. After all, the problem space is huge, there is enough room for another product, everyone uses emails, no need for any further validation, right? ❌ WRONG ONCE AGAIN! We all hear from the greatest in the startup landscape that we must validate our ideas with real people, yet at least some of us (guilty here 🥸) think that our product will be hugely successful and prove them to be an exception. Few might, but most are not. I certainly wasn't. 👉 Lesson learned: Always validate your ideas with real people. Ask them how much they’d pay for such a tool (not if they would). Much better if they are willing to pay upfront for a discount, etc. But even this comes later, keep reading. I think the difference between “How much” and “If” is huge for two reasons: (1) By asking them for “How much”, you force them to think in a more realistic setting. (2) You will have a more realistic idea on your profit margins. Based on my competitive analysis, I already had a solution in my mind to improve our email usage standards and email productivity (huge mistake), but I did my best to learn about their problems regarding those without pushing the idea too hard. The idea is this: Generate concise email summaries with suggested actions, combine them into one email, and send it at their preferred times. Save as much as time the AI you end up with allows. After all, everyone loves to save time. So, what kind of validation did I seek for? Talked with only a few people around me about this crazy, internet-breaking idea. The responses I got were, now I see, mediocre; no one got excited about it, just said things along the lines of “Cool idea, OK”. So, any reasonable person in this situation would think “Okay, not might not be working”, right? Well, I did not. I assumed that they were the wrong audience for this product, and there was this magical land of user segments waiting eagerly for my product, yet unknowingly. To this day, I still have not reached this magical place. Perhaps, it didn’t exist in the first place. If I cannot find it, whether it exists or not doesn’t matter. I am certainly searching for it. 👉 What I should have done: Once I decide on a problem space (time management, email productivity, etc.), I should decide on my potential user segments, people who I plan to sell my product to. Then I should go talk to those people, ask them about their pains, then get to the problem-solving/ideation phase only later. ❗️ VALIDATION COMES FROM THE REALITY OUTSIDE. What validation looks like might change from product to product; but what invalidation looks like is more or less the same for every product. Nico Jeannen told me yesterday “validation = money in the account” on Twitter. This is the ultimate form of validation your product could get. If your product doesn’t make any money, then something is invalidated by reality: Your product, you, your idea, who knows? So, at this point, I knew a little bit of Python from spending some time in tutorial hell a few years ago, some HTML/CSS/JS, barely enough React to build a working app. React could work for this project, but I needed easy-to-implement server interactivity. Luckily, around this time, I got to know about this new gen of indie hackers, and learned (but didn’t truly understand) about their approach to indie hacking, and this library called Nextjs. How good Next.js still blows my mind. So, I was back to tutorial hell once again. But, this time, with a promise to myself: This is the last time I would visit tutorial hell. Time to start building this "ground-breaking idea" Learning the fundamentals of Next.js was easier than learning of React unsurprisingly. Yet, the first time I managed to run server actions on Next.js was one of the rarest moments that completely blew my mind. To this day, I reject the idea that it is something else than pure magic under its hood. Did I absolutely need Nextjs for this project though? I do not think so. Did it save me lots of time? Absolutely. Furthermore, learning Nextjs will certainly be quite helpful for other projects that I will be tackling in the future. Already got a few ideas that might be worth pursuing in the head in case I decide to abandon Summ in the future. Fast-forward few weeks again: So, at this stage, I had a barely working MVP-like product. Since the very beginning, I spent every free hour (and more) on this project as speed is essential. But, I am not so sure it was worth it to overwork in retrospect. Yet, I know I couldn’t help myself. Everything is going kinda smooth, so what’s the worst thing that could ever happen? Well, both Apple and Google announced their AIs (Apple Intelligence and Google Gemini, respectively) will have email summarization features for their products. Summarizing singular emails is no big deal, after all there were already so many similar products in the market. I still think that what truly matters is a frictionless user experience, and this is why I built this product in a certain way: You spend less than a few minutes setting up your account, and you get to enjoy your email summaries, without ever visiting its website again. This is still a very cool concept I really like a lot. So, at this point: I had no other idea that could be pursued, already spent too much time on this project. Do I quit or not? This was the question. Of course not. I just have to launch this product as quickly as possible. So, I did something right, a quite rare occurrence I might say: Re-planned my product, dropped everything secondary to the core feature immediately (save time on reading emails), tried launching it asap. 👉 Insight: Sell only one core feature at one time. Drop anything secondary to this core feature. Well, my primary occupation is product design. So one would expect that a product I build must have stellar design. I considered any considerable time spent on design at this stage would be simply wasted. I still think this is both true and wrong: True, because if your product’s core benefits suck, no one will care about your design. False, because if your design looks amateurish, no one will trust you and your product. So, I always targeted an average level design with it and the way this tool works made it quite easy as I had to design only 2 primary pages: Landing page and user portal (which has only settings and analytics pages). However, even though I knew spending time on design was not worth much of my time, I got a bit “greedy”: In fact, I redesigned those pages three times, and still ended up with a so-so design that I am not proud of. 👉 What I would do differently: Unless absolutely necessary, only one iteration per stage as long as it works. This, in my mind, applies to everything. If your product’s A feature works, then no need to rewrite it from scratch for any reason, or even refactor it. When your product becomes a success, and you absolutely need that part of your codebase to be written, do so, but only then. Ready to launch, now is th etime for some marketing, right? By July 26, I already had a “launchable” product that barely works (I marked this date on a Notion docs, this is how I know). Yet, I had spent almost no time on marketing, sales, whatever. After all, “You build and they will come”. Did I know that I needed marketing? Of course I did, but knowingly didn’t. Why, you might ask. Well, from my perspective, it had to be a dev-heavy product; meaning that you spend most of your time on developing it, mostly coding skills. But, this is simply wrong. As a rule of thumb, as noted by one of the greatests, Marc Louvion, you should spend at least twice of the building time on marketing. ❗️ Time spent on building \* 2 people don’t know your product > they don’t use your product > you don’t get users > you don’t make money Easy as that. Following the same reasoning, a slightly different approach to planning a project is possible. Determine an approximate time to complete the project with a high level project plan. Let’s say 6 months. By the reasoning above, 2 months should go into building, and 4 into marketing. If you need 4 months for building instead of 2, then you need 8 months of marketing, which makes the time to complete the project 12 months. If you don’t have that much time, then quit the project. When does a project count as completed? Well, in reality, never. But, I think we have to define success conditions even before we start for indie projects and startups; so we know when to quit when they are not met. A success condition could look like “Make $6000 in 12 months” or “Have 3000 users in 6 months”. It all depends on the project. But, once you set it, it should be set in stone: You don’t change it unless absolutely necessary. I suspect there are few principles that make a solopreneur successful; and knowing when to quit and when to continue is definitely one of them. Marc Louvion is famously known for his success, but he got there after failing so many projects. To my knowledge, the same applies to Nico Jeannen, Pieter Levels, or almost everyone as well. ❗️ Determining when to continue even before you start will definitely help in the long run. A half-aed launch Time-leap again. Around mid August, I “soft launched” my product. By soft launch, I mean lazy marketing. Just tweeting about it, posting it on free directories. Did I get any traffic? Surely I did. Did I get any users? Nope. Only after this time, it hit me: “Either something is wrong with me, or with this product” Marketing might be a much bigger factor for a project’s success after all. Even though I get some traffic, not convincing enough for people to sign up even for a free trial. The product was still perfect in my eyes at the time (well, still is ^(\_),) so the right people are not finding my product, I thought. Then, a question that I should have been asking at the very first place, one that could prevent all these, comes to my mind: “How do even people search for such tools?” If we are to consider this whole journey of me and my so-far-failed product to be an already destined failure, one metric suffices to show why. Search volume: 30. Even if people have such a pain point, they are not looking for email summaries. So, almost no organic traffic coming from Google. But, as a person who did zero marketing on this or any product, who has zero marketing knowledge, who doesn’t have an audience on social media, there is not much I could do. Finally, it was time to give up. Or not… In my eyes, the most important element that makes a founder (solo or not) successful (this, I am not by any means) is to solve problems. ❗️ So, the problem was this: “People are not finding my product by organic search” How do I make sure I get some organic traffic and gets more visibility? Learn digital marketing and SEO as much as I can within very limited time. Thankfully, without spending much time, I came across Neil Patel's YT channel, and as I said many times, it is an absolute gold mine. I learned a lot, especially about the fundamentals, and surely it will be fruitful; but there is no magic trick that could make people visit your website. SEO certainly helps, but only when people are looking for your keywords. However, it is truly a magical solution to get in touch with REAL people that are in your user segments: 👉 Understand your pains, understand their problems, help them to solve them via building products. I did not do this so far, have to admit. But, in case you would like to have a chat about your email usage, and email productivity, just get in touch; I’d be delighted to hear about them. Getting ready for a ProductHunt launch The date was Sept 1. And I unlocked an impossible achievement: Running out of Supabase’s free plan’s Egres limit while having zero users. I was already considering moving out of their Cloud server and managing a Supabase CLI service on my Hetzner VPS for some time; but never ever suspected that I would have to do this quickly. The cheapest plan Supabase offers is $25/month; yet, at that point, I am in between jobs for such a long time, basically broke, and could barely afford that price. One or two months could be okay, but why pay for it if I will eventually move out of their Cloud service? So, instead of paying $25, I spent two days migrating out of Supabase Cloud. Worth my time? Definitely not. But, when you are broke, you gotta do stupid things. This was the first time that I felt lucky to have zero users: I have no idea how I would manage this migration if I had any. I think this is one of the core tenets of an indie hacker: Controlling their own environment. I can’t remember whose quote this is, but I suspect it was Naval: Entrepreneurs have an almost pathological need to control their own fate. They will take any suffering if they can be in charge of their destiny, and not have it in somebody else’s hands. What’s truly scary is, at least in my case, we make people around us suffer at the expense of our attempting to control our own fates. I know this period has been quite hard on my wife as well, as I neglected her quite a bit, but sadly, I know that this will happen again. It is something that I can barely help with. Still, so sorry. After working the last two weeks on a ProductHunt Launch, I finally launched it this Tuesday. Zero ranking, zero new users, but 36 kind people upvoted my product, and many commented and provided invaluable feedback. I couldn't be more grateful for each one of them 🙏. Considering all these, what lies in the future of Summ though? I have no idea, to be honest. On one hand, I have zero users, have no job, no income. So, I need a way to make money asap. On the other hand, the whole idea of it revolves around one core premise (not an assumption) that I am not so willing to share; and I couldn’t have more trust in it. This might not be the best iteration of it, however I certainly believe that email usage is one of the best problem spaces one could work on. 👉 But, one thing is for certain: I need to get in touch with people, and talk with them about this product I built so far. In fact, this is the only item on my agenda. Nothing else will save my brainchild <3. Below are some other insights and notes that I got during my journey; as they do not 100% fit into this story, I think it is more suitable to list them here. I hope you enjoyed reading this. Give Summ a try, it comes with a generous free trial, no credit card required. Some additional notes and insights: Project planning is one of the most underestimated skills for solopreneurs. It saves you enormous time, and helps you to keep your focus up. Building B2B products beats building B2C products. Businesses are very willing to pay big bucks if your product helps them. On the other hand, spending a few hours per user who would pay $5/m probably is not worth your time. It doesn’t matter how brilliant your product is if no one uses it. If you cannot sell a product in a certain category/niche (or do not know how to sell it), it might be a good idea not to start a project in it. Going after new ideas and ventures is quite risky, especially if you don’t know how to market it. On the other hand, an already established category means that there is already demand. Whether this demand is sufficient or not is another issue. As long as there is enough demand for your product to fit in, any category/niche is good. Some might be better, some might be worse. Unless you are going hardcore B2B, you will need people to find your product by means of organic search. Always conduct thorough keyword research as soon as possible.

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

We've built an AI-powered business building platform, and we're looking for entrepreneurs to try out the MVP!
reddit
LLM Vibe Score0
Human Vibe Score1
UltraIngoThis week

We've built an AI-powered business building platform, and we're looking for entrepreneurs to try out the MVP!

Hey r/sideproject! I'm Felix, co-founder of Buildpad, and we're excited to share our latest project with you. https://reddit.com/link/1eve8n4/video/ahktfda2bgjd1/player Buildpad is an AI-powered (Claude Sonnet 3.5) business-building platform that guides entrepreneurs through every step of creating and growing a business. Here's what makes it unique: Idea validation: Leverage Reddit's API to get real-world data on your ideas through posts, comments and discussions. Structured process: Follow a clear roadmap from idea validation to launch and beyond. Team collaboration: Work with co-founders, all assisted by the same AI. Central context bank: Our AI remembers everything about your project for consistent, informed guidance. We're solving the common problem of entrepreneurs not knowing what to do next, especially during idea generation and validation phases. With Buildpad, you can validate your ideas by searching for relevant keywords across Reddit, helping you understand if people are actually experiencing the problems you're aiming to solve. We're in the MVP stage and looking for early adopters to test the platform and provide feedback. We'd love to hear from you: Does this solution resonate with your entrepreneurial challenges? What features would you find most valuable in a tool like this? Any thoughts or concerns about using AI for startup guidance? If you're interested in trying out Buildpad or have any questions, please comment below or DM me. Thanks for checking it out! buildpad.io

How I Automated Amazon Affiliate Marketing: A Developer's Journey
reddit
LLM Vibe Score0
Human Vibe Score1
siom_cThis week

How I Automated Amazon Affiliate Marketing: A Developer's Journey

From Manual Labor to 1000x Efficiency As a developer who ventured into affiliate marketing, I discovered a significant gap between technical possibilities and current practices. This revelation led me to create AutoPin, a tool that's now helping hundreds of affiliate marketers reclaim their time. The Problem: A Time-Consuming Reality Every affiliate marketer knows this scenario: you spend hours copying and pasting links, checking prices, and updating product information. I found myself dedicating 4-6 hours daily to these repetitive tasks. As a programmer, this felt fundamentally wrong. The typical affiliate marketing workflow looked like this: Find promising products Generate affiliate links one by one Monitor price changes manually Check product availability regularly Update content when things change Repeat this process daily This manual process had several critical issues: Time Waste: 20-30 hours weekly on repetitive tasks Missed Opportunities: Unable to scale beyond 100 products Human Error: Inevitable mistakes in manual updates Delayed Updates: Lost commissions due to outdated information The Solution: Building AutoPin After three months of development and six months of testing, I created a system that could: Generate hundreds of affiliate links in minutes Monitor price changes automatically Update product availability in real-time Export data in multiple formats Scale infinitely without additional effort Real Results, Real Impact The impact was immediate and significant: 📊 Efficiency Metrics: Link generation: From 2 minutes per link to 0.1 seconds Monitoring capacity: From 50 to 5000+ products Update frequency: From daily to real-time Error rate: Reduced by 99.9% 💡 User Success Stories: "Increased my product portfolio by 10x without adding work hours" "Revenue grew 300% in the first month" "Finally able to focus on content creation instead of link management" Technical Insights The system architecture focuses on three core components: Data Extraction Engine Efficient web scraping Rate limiting and proxy management Data validation and cleaning Real-time Monitoring System Websocket connections for instant updates Queue management for large-scale monitoring Smart scheduling based on price volatility Export Framework Multiple format support (CSV, HTML, Markdown) Custom templating engine Batch processing capabilities The Future of Affiliate Marketing Automation We're currently developing AI capabilities to: Generate product descriptions automatically Optimize link placement for conversion Predict price trends and best promotion times Create content variations for different platforms Key Learnings Automation is Essential The future of affiliate marketing lies in automation. Manual processes simply can't compete with automated systems in terms of efficiency and accuracy. Focus on Value Creation When marketers spend less time on repetitive tasks, they can focus on strategy and content quality. Scale Matters With automation, the difference between managing 10 products and 1000 products becomes minimal. Getting Started If you're an affiliate marketer spending hours on manual tasks, it's time to automate. Here's what you can do: Analyze your current workflow Identify repetitive tasks Start with basic automation Scale gradually Monitor and optimize Conclusion The transformation from manual to automated affiliate marketing isn't just about saving time—it's about unlocking potential. When you remove the tedious aspects of affiliate marketing, you create space for creativity, strategy, and growth. Want to experience the difference? Visit AutoPin at autopin.pro and join the automation revolution. Remember: The best time to automate was yesterday. The second best time is now. About the Author: A developer turned affiliate marketer who believes in the power of automation to transform digital marketing. #AffiliateMarketing #Automation #Programming #DigitalMarketing #SaaS #ProductivityTools

Built a side project to help validate... side projects 🤔
reddit
LLM Vibe Score0
Human Vibe Score0.5
ANDYVO_This week

Built a side project to help validate... side projects 🤔

Hey builders! Long-time lurker, first-time poster. Wanted to share something I've been working on after hours. The Problem Like many of you, I was building side projects at night while working full-time. My pattern was painfully consistent: \- Get excited about an idea \- Spend precious evening hours coding \- Launch on Reddit/Twitter \- \crickets\ \- Realize I should've validated first \- Repeat 🥲 The worst part? I was trying to do validation on Reddit, but with limited time after work, I couldn't: \- Properly research different communities \- Keep track of all the responses \- Find people actually interested in what I was building \- Or figure out what features to prioritize in my limited dev time The Solution So I built RediScope (yes, another side project 😅). It automates the Reddit validation process: Quick Loom demo How it works: Tell it what you're trying to validate AI helps write natural-sounding questions for each community Get instant analysis of responses Automatically spot potential early users The goal? Instead of spending your limited free time scrolling through Reddit, get clear validation and potential users in 48 hours. Tech Stack (since we all love this part): \- Frontend: React + Tailwind \- Backend: Node.js \- AI: OpenAI for the smart bits \- DB: PostgreSQL \- Hosting: Digital Ocean Current Status: \- ✅ Core validation engine working \- ✅ Community analysis \- ✅ Response analytics \- ✅ User dashboard \- 🔎 Lead Generator (in progress) \- 📋 API (planned) This is very much a nights-and-weekends project (like everything else we build 😄). Would love feedback from fellow side project builders: \- How do you validate your ideas currently? \- What would make this useful for your next project? \- Any features you'd want to see? Drop your thoughts below! If you want to try it when it launches: https://www.rediscope.app Also huge shoutout to this community - been learning so much from everyone's posts here 🙏

How me and my team made 15+ apps and not made a single sale in 2023
reddit
LLM Vibe Score0
Human Vibe Score0.818
MichaelbetterecycleThis week

How me and my team made 15+ apps and not made a single sale in 2023

Hey, my name is Michael, I am in Auckland NZ. This year was the official beginning of my adult life. I graduated from university and started a full-time job. I’ve also really dug into indiehacking/bootstrapping and started 15 projects (and it will be at least 17 before the year ends). I think I’ve learned a lot but I consciously repeated mistakes. Upto (Nov) Discord Statuses + Your Location + Facebook Poke https://preview.redd.it/4nqt7tp2tf5c1.png?width=572&format=png&auto=webp&s=b0223484bc54b45b5c65e0b1afd0dc52f9c02ad1 This was the end of uni, I often messaged (and got messaged) requests of status and location to (and from my) friends. I thought, what if we make a social app that’s super basic and all it does is show you where your friends are? To differentiate from snap maps and others we wanted something with more privacy where you select the location. However, never finished the codebase or launched it. This is because I slowly started to realize that B2C (especially social networks) are way too hard to make into an actual business and the story with Fistbump would repeat itself. However, this decision not to launch it almost launched a curse on our team. From that point, we permitted ourselves to abandon projects even before launching. Lessons: Don’t do social networks if your goal is 10k MRR ASAP. If you build something to 90% competition ship it or you will think it’s okay to abandon projects Insight Bites (Nov) Youtube Summarizer Extension &#x200B; https://preview.redd.it/h6drqej4tf5c1.jpg?width=800&format=pjpg&auto=webp&s=0f211456c390ac06f4fcb54aa51f9d50b0826658 Right after Upto, we started ideating and conveniently the biggest revolution in the recent history of tech was released → GPT. We instantly began ideating. The first problem we chose to use AI for is to summarize YouTube videos. Comical. Nevertheless, I am convinced we have had the best UX because you could right-click on a video to get a slideshow of insights instead of how everyone else did it. We dropped it because there was too much competition and unit economics didn’t work out (and it was a B2C). PodPigeon (Dec) Podcast → Tweet Threads https://preview.redd.it/0ukge245tf5c1.png?width=2498&format=png&auto=webp&s=23303e1cab330578a3d25cd688fa67aa3b97fb60 Then we thought, to make unit economics work we need to make this worthwhile for podcasters. This is when I got into Twitter and started seeing people summarize podcasts. Then I thought, what if we make something that converts a podcast into tweets? This was probably one of the most important projects because it connected me with Jason and Jonaed, both of whom I regularly stay in contact with and are my go-to experts on ideas related to content creation. Jonaed was even willing to buy Podpigeon and was using it on his own time. However, the unit economics still didn’t work out (and we got excited about other things). Furthermore, we got scared of the competition because I found 1 - 2 other people who did similar things poorly. This was probably the biggest mistake we’ve made. Very similar projects made 10k MRR and more, launching later than we did. We didn’t have a coherent product vision, we didn’t understand the customer well enough, and we had a bad outlook on competition and a myriad of other things. Lessons: I already made another post about the importance of outlook on competition. Do not quit just because there are competitors or just because you can’t be 10x better. Indiehackers and Bootstrappers (or even startups) need to differentiate in the market, which can be via product (UX/UI), distribution, or both. Asking Ace Intro.co + Crowdsharing &#x200B; https://preview.redd.it/0hu2tt16tf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3d397568ef2331e78198d64fafc1a701a3e75999 As I got into Twitter, I wanted to chat with some people I saw there. However, they were really expensive. I thought, what if we made some kind of crowdfunding service for other entrepreneurs to get a private lecture from their idols? It seemed to make a lot of sense on paper. It was solving a problem (validated via the fact that Intro.co is a thing and making things cheaper and accessible is a solid ground to stand on), we understood the market (or so we thought), and it could monetize relatively quickly. However, after 1-2 posts on Reddit and Indiehackers, we quickly learned three things. Firstly, no one cares. Secondly, even if they do, they think they can get the same information for free online. Thirdly, the reasons before are bad because for the first point → we barely talked to people, and for the second people → we barely talked to the wrong people. However, at least we didn’t code anything this time and tried to validate via a landing page. Lessons Don’t give up after 1 Redditor says “I don’t need this” Don’t be scared to choose successful people as your audience. Clarito Journaling with AI analyzer https://preview.redd.it/8ria2wq6tf5c1.jpg?width=1108&format=pjpg&auto=webp&s=586ec28ae75003d9f71b4af2520b748d53dd2854 Clarito is a classic problem all amateur entrepreneurs have. It’s where you lie to yourself that you have a real problem and therefore is validated but when your team asks you how much you would pay you say I guess you will pay, maybe, like 5 bucks a month…? Turns out, you’d have to pay me to use our own product lol. We sent it off to a few friends and posted on some forums, but never really got anything tangible and decided to move away. Honestly, a lot of it is us in our own heads. We say the market is too saturated, it’ll be hard to monetize, it’s B2C, etc. Lessons: You use the Mom Test on other people. You have to do it yourself as well. However, recognizing that the Mom Test requires a lot of creativity in its investigation because knowing what questions to ask can determine the outcome of the validation. I asked myself “Do I journal” but I didn’t ask myself “How often do I want GPT to chyme in on my reflections”. Which was practically never. That being said I think with the right audience and distribution, this product can work. I just don’t know (let alone care) about the audience that much (and I thought I was one of them)/ Horns & Claw Scrapes financial news texts you whether you should buy/sell the stock (news sentiment analysis) &#x200B; https://preview.redd.it/gvfxdgc7tf5c1.jpg?width=1287&format=pjpg&auto=webp&s=63977bbc33fe74147b1f72913cefee4a9ebec9c2 This one we didn’t even bother launching. Probably something internal in the team and also seemed too good to be true (because if this works, doesn’t that just make us ultra-rich fast?). I saw a similar tool making 10k MRR so I guess I was wrong. Lessons: This one was pretty much just us getting into our heads. I declared that without an audience it would be impossible to ship this product and we needed to start a YouTube channel. Lol, and we did. And we couldn’t even film for 1 minute. I made bold statements like “We will commit to this for at least 1 year no matter what”. Learnery Make courses about any subject https://preview.redd.it/1nw6z448tf5c1.jpg?width=1112&format=pjpg&auto=webp&s=f2c73e8af23b0a6c3747a81e785960d4004feb48 This is probably the most “successful” project we’ve made. It grew from a couple of dozen to a couple of hundred users. It has 11 buy events for $9.99 LTD (we couldn’t be bothered connecting Stripe because we thought no one would buy it anyway). However what got us discouraged from seriously pursuing it more is, that this has very low defensibility, “Why wouldn’t someone just use chatGPT?” and it’s B2C so it’s hard to monetize. I used it myself for a month or so but then stopped. I don’t think it’s the app, I think the act of learning a concept from scratch isn’t something you do constantly in the way Learnery delivers it (ie course). I saw a bunch of similar apps that look like Ass make like 10k MRR. Lessons: Don’t do B2C, or if you do, do it properly Don’t just Mixpanel the buy button, connect your Stripe otherwise, it doesn’t feel real and you won’t get momentum. I doubt anyone (even me) will make this mistake again. I live in my GPT bubble where I make assumptions that everyone uses GPT the same way and as much as I do. In reality, the argument that this has low defensibility against GPT is invalid. Platforms that deliver a differentiated UX from ChatGPT to audiences who are not tightly integrated into the habit of using ChatGPT (which is like - everyone except for SOME tech evangelists). CuriosityFM Make podcasts about any subject https://preview.redd.it/zmosrcp8tf5c1.jpg?width=638&format=pjpg&auto=webp&s=d04ddffabef9050050b0d87939273cc96a8637dc This was our attempt at making Learnery more unique and more differentiated from chatGPT. We never really launched it. The unit economics didn’t work out and it was actually pretty boring to listen to, I don’t think I even fully listened to one 15-minute episode. I think this wasn’t that bad, it taught us more about ElevenLabs and voice AI. It took us maybe only 2-3 days to build so I think building to learn a new groundbreaking technology is fine. SleepyTale Make children’s bedtime stories https://preview.redd.it/14ue9nm9tf5c1.jpg?width=807&format=pjpg&auto=webp&s=267e18ec6f9270e6d1d11564b38136fa524966a1 My 8-year-old sister gave me that idea. She was too scared of making tea and I was curious about how she’d react if she heard a bedtime story about that exact scenario with the moral that I wanted her to absorb (which is that you shouldn’t be scared to try new things ie stop asking me to make your tea and do it yourself, it’s not that hard. You could say I went full Goebbels on her). Zane messaged a bunch of parents on Facebook but no one really cared. We showed this to one Lady at the place we worked from at Uni and she was impressed and wanted to show it to her kids but we already turned off our ElevenLabs subscription. Lessons: However, the truth behind this is beyond just “you need to be able to distribute”. It’s that you have to care about the audience. I don’t particularly want to build products for kids and parents. I am far away from that audience because I am neither a kid anymore nor going to be a parent anytime soon, and my sister still asked me to make her tea so the story didn’t work. I think it’s important to ask yourself whether you care about the audience. The way you answer that even when you are in full bias mode is, do you engage with them? Are you interested in what’s happening in their communities? Are you friends with them? Etc. User Survey Analyzer Big User Survey → GPT → Insights Report Me and my coworker were chatting about AI when he asked me to help him analyze a massive survey for him. I thought that was some pretty decent validation. Someone in an actual company asking for help. Lessons Market research is important but moving fast is also important. Ie building momentum. Also don’t revolve around 1 user. This has been a problem in multiple projects. Finding as many users as possible in the beginning to talk to is key. Otherwise, you are just waiting for 1 person to get back to you. AutoI18N Automated Internationalization of the codebase for webapps This one I might still do. It’s hard to find a solid distribution strategy. However, the idea came from me having to do it at my day job. It seems a solid problem. I’d say it’s validated and has some good players already. The key will be differentiation via the simplicity of UX and distribution (which means a slightly different audience). In the backlog for now because I don’t care about the problem or the audience that much. Documate - Part 1 Converts complex PDFs into Excel https://preview.redd.it/8b45k9katf5c1.jpg?width=1344&format=pjpg&auto=webp&s=57324b8720eb22782e28794d2db674b073193995 My mom needed to convert a catalog of furniture into an inventory which took her 3 full days of data entry. I automated it for her and thought this could have a big impact but there was no distribution because there was no ICP. We tried to find the ideal customers by talking to a bunch of different demographics but I flew to Kazakhstan for a holiday and so this kind of fizzled out. I am not writing this blog post linearity, this is my 2nd hour and I am tired and don’t want to finish this later so I don’t even know what lessons I learned. Figmatic Marketplace of high-quality Figma mockups of real apps https://preview.redd.it/h13yv45btf5c1.jpg?width=873&format=pjpg&auto=webp&s=aaa2896aeac2f22e9b7d9eed98c28bb8a2d2cdf1 This was a collab between me and my friend Alex. It was the classic Clarito where we both thought we had this problem and would pay to fix it. In reality, this is a vitamin. Neither I, nor I doubt Alex have thought of this as soon as we bought the domain. We posted it on Gumroad, sent it to a bunch of forums, and called it a day. Same issue as almost all the other ones. No distribution strategy. However, apps like Mobin show us that this concept is indeed profitable but it takes time. It needs SEO. It needs a community. None of those things, me and Alex had or was interested in. However shortly after HTML → Figma came out and it’s the best plugin. Maybe that should’ve been the idea. Podcast → Course Turns Podcaster’s episodes into a course This one I got baited by Jason :P I described to him the idea of repurposing his content for a course. He told me this was epic and he would pay. Then after I sent him the demo, he never checked it out. Anyhow during the development, we realized that doesn’t actually work because A podcast doesn’t have the correct format for the course, the most you can extract are concepts and ideas, seldom explanations. Most creators want video-based courses to be hosted on Kajabi or Udemy Another lesson is that when you pitch something to a user, what you articulate is a platform or a process, they imagine an outcome. However, the end result of your platform can be a very different outcome to what they had in mind and there is even a chance that what they want is not possible. You need to understand really well what the outcome looks like before you design the process. This is a classic problem where we thought of the solution before the problem. Yes, the problem exists. Podcasters want to make courses. However, if you really understand what they want, you can see how repurposing a podcast isn’t the best way to get there. However I only really spoke to 1-2 podcasters about this so making conclusions is dangerous for this can just be another asking ace mistake with the Redditor. Documate Part 2 Same concept as before but now I want to run some ads. We’ll see what happens. https://preview.redd.it/xb3npj0ctf5c1.jpg?width=1456&format=pjpg&auto=webp&s=3cd4884a29fd11d870d010a2677b585551c49193 In conclusion https://preview.redd.it/2zrldc9dtf5c1.jpg?width=1840&format=pjpg&auto=webp&s=2b3105073e752ad41c23f205dbd1ea046c1da7ff It doesn’t actually matter that much whether you choose to do a B2C, or a social network or focus on growing your audience. All of these can make you successful. What’s important is that you choose. If I had to summarize my 2023 in one word it’s indecision. Most of these projects succeeded for other people, nothing was as fundamentally wrong about them as I proclaimed. In reality that itself was an excuse. New ideas seduce, and it is a form of discipline to commit to a single project for a respectful amount of time. https://preview.redd.it/zy9a2vzdtf5c1.jpg?width=1456&format=pjpg&auto=webp&s=901c621227bba0feb4efdb39142f66ab2ebb86fe Distribution is not just posting on Indiehackers and Reddit. It’s an actual strategy and you should think of it as soon as you think of the idea, even before the Figma designs. I like how Denis Shatalin taught me. You have to build a pipeline. That means a reliable way to get leads, launch campaigns at them, close deals, learn from them, and optimize. Whenever I get an idea now I always try to ask myself “Where can I find 1000s leads in one day?” If there is no good answer, this is not a good project to do now. &#x200B; https://preview.redd.it/2boh3fpetf5c1.jpg?width=1456&format=pjpg&auto=webp&s=1c0d5d7b000716fcbbb00cbad495e8b61e25be66 Talk to users before doing anything. Jumping on designing and coding to make your idea a reality is a satisfying activity in the short term. Especially for me, I like to create for the sake of creation. However, it is so important to understand the market, understand the audience, understand the distribution. There are a lot of things to understand before coding. https://preview.redd.it/lv8tt96ftf5c1.jpg?width=1456&format=pjpg&auto=webp&s=6c8735aa6ad795f216ff9ddfa2341712e8277724 Get out of your own head. The real reason we dropped so many projects is that we got into our own heads. We let the negative thoughts creep in and kill all the optimism. I am really good at coming up with excuses to start a project. However, I am equally as good at coming up with reasons to kill a project. And so you have this yin and yang of starting and stopping. Building momentum and not burning out. I can say with certainty my team ran out of juice this year. We lost momentum so many times we got burnt out towards the end. Realizing that the project itself has momentum is important. User feedback and sales bring momentum. Building also creates momentum but unless it is matched with an equal force of impact, it can stomp the project down. That is why so many of our projects died quickly after we launched. The smarter approach is to do things that have a low investment of momentum (like talking to users) but result in high impact (sales or feedback). Yes, that means the project can get invalidated which makes it more short-lived than if we built it first, but it preserves team life energy. At the end of 2023 here is a single sentence I am making about how I think one becomes a successful indiehacker. One becomes a successful Indiehacker when one starts to solve pain-killer problems in the market they understand, for an audience they care about and consistently engage with for a long enough timeframe. Therefore an unsuccessful Indiehacker in a single sentence is An unsuccessful Indiehacker constantly enters new markets they don’t understand to build solutions for people whose problems they don’t care about, in a timeframe that is shorter than than the time they spent thinking about distribution. However, an important note to be made. Life is not just about indiehacking. It’s about learning and having fun. In the human world, the best journey isn’t the one that gets you the fastest to your goals but the one you enjoy the most. I enjoyed making those silly little projects and although I do not regret them, I will not repeat the same mistakes in 2024. But while it’s still 2023, I have 2 more projects I want to do :) EDIT: For Devs, frontend is always react with vite (ts) and backend is either node with express (ts) or python. For DB either Postgres or mongo (usually Prisma for ORM). For deployment all of it is on AWS (S3, EC2). In terms of libraries/APIs Whisper.cpp is best open source for transcription Obviously the gpt apis Eleven labs for voice related stuff And other random stuff here and there

Introducing Novus – an AI-powered QA agent that automates testing for your web apps!
reddit
LLM Vibe Score0
Human Vibe Score1
namish800This week

Introducing Novus – an AI-powered QA agent that automates testing for your web apps!

Hello, I'm excited to introduce a project I've been working on—an AI-powered QA agent designed to streamline and enhance the testing process for web applications. Here's how it works: Key Features: Natural Language Test Definitions: You can define the behavior you want to validate using plain English. Automated Navigation and Validation: The agent autonomously navigates your web app and checks if the specified behavior functions as expected. Comprehensive Reporting: After execution, it provides detailed reports, including step-by-step actions, screenshots, and video recordings.​ How It Works: Define Behavior: Describe the functionality you want to test in simple English.​ Run Test: The agent interprets your description, interacts with your web app accordingly, and validates the outcomes. Review Results: Access detailed reports that include all actions taken, along with visual documentation like screenshots and videos.​ Current Capabilities: Dashboard for Test Management: Create and manage multiple test suites and individual tests through an intuitive interface.​ Visual Regression Analysis: Utilize visual artifacts to perform regression analysis and ensure UI consistency.​ Future Plans: Intelligent Reporting: Implement advanced reporting features to provide deeper insights and analytics. Enhanced Visual Regression: Develop more sophisticated tools for detecting and analyzing visual discrepancies.​ I'm eager to hear your thoughts and feedback. What challenges do you face in QA testing? How do you see AI tools fitting into your workflow? Let's discuss! Here's the demo of what I've built so far https://www.loom.com/share/11b1dd4d18124f9a8032ae81e9cbdab4?sid=56237f10-cffd-4394-b080-0a3fb5ef4b01 Note: This project is currently in development, and I'm actively seeking input to refine and enhance its features.

[D] Here are 17 ways of making PyTorch training faster – what did I miss?
reddit
LLM Vibe Score0
Human Vibe Score1
lorenzkuhnThis week

[D] Here are 17 ways of making PyTorch training faster – what did I miss?

I've been collecting methods to accelerate training in PyTorch – here's what I've found so far. What did I miss? What did I get wrong? The methods – roughly sorted from largest to smallest expected speed-up – are: Consider using a different learning rate schedule. Use multiple workers and pinned memory in DataLoader. Max out the batch size. Use Automatic Mixed Precision (AMP). Consider using a different optimizer. Turn on cudNN benchmarking. Beware of frequently transferring data between CPUs and GPUs. Use gradient/activation checkpointing. Use gradient accumulation. Use DistributedDataParallel for multi-GPU training. Set gradients to None rather than 0. Use .as\_tensor rather than .tensor() Turn off debugging APIs if not needed. Use gradient clipping. Turn off bias before BatchNorm. Turn off gradient computation during validation. Use input and batch normalization. Consider using another learning rate schedule The learning rate (schedule) you choose has a large impact on the speed of convergence as well as the generalization performance of your model. Cyclical Learning Rates and the 1Cycle learning rate schedule are both methods introduced by Leslie N. Smith (here and here), and then popularised by fast.ai's Jeremy Howard and Sylvain Gugger (here and here). Essentially, the 1Cycle learning rate schedule looks something like this: &#x200B; https://preview.redd.it/sc37u5knmxa61.png?width=476&format=png&auto=webp&s=09b309b4dbd67eedb4ab5f86e03e0e83d7b072d1 Sylvain writes: \[1cycle consists of\]  two steps of equal lengths, one going from a lower learning rate to a higher one than go back to the minimum. The maximum should be the value picked with the Learning Rate Finder, and the lower one can be ten times lower. Then, the length of this cycle should be slightly less than the total number of epochs, and, in the last part of training, we should allow the learning rate to decrease more than the minimum, by several orders of magnitude. In the best case this schedule achieves a massive speed-up – what Smith calls Superconvergence – as compared to conventional learning rate schedules. Using the 1Cycle policy he needs \~10x fewer training iterations of a ResNet-56 on ImageNet to match the performance of the original paper, for instance). The schedule seems to perform robustly well across common architectures and optimizers. PyTorch implements both of these methods torch.optim.lrscheduler.CyclicLR and torch.optim.lrscheduler.OneCycleLR, see the documentation. One drawback of these schedulers is that they introduce a number of additional hyperparameters. This post and this repo, offer a nice overview and implementation of how good hyper-parameters can be found including the Learning Rate Finder mentioned above. Why does this work? It doesn't seem entirely clear but one possible explanation might be that regularly increasing the learning rate helps to traverse saddle points in the loss landscape more quickly. Use multiple workers and pinned memory in DataLoader When using torch.utils.data.DataLoader, set numworkers > 0, rather than the default value of 0, and pinmemory=True, rather than the default value of False. Details of this are explained here. Szymon Micacz achieves a 2x speed-up for a single training epoch by using four workers and pinned memory. A rule of thumb that people are using to choose the number of workers is to set it to four times the number of available GPUs with both a larger and smaller number of workers leading to a slow down. Note that increasing num\_workerswill increase your CPU memory consumption. Max out the batch size This is a somewhat contentious point. Generally, however, it seems like using the largest batch size your GPU memory permits will accelerate your training (see NVIDIA's Szymon Migacz, for instance). Note that you will also have to adjust other hyperparameters, such as the learning rate, if you modify the batch size. A rule of thumb here is to double the learning rate as you double the batch size. OpenAI has a nice empirical paper on the number of convergence steps needed for different batch sizes. Daniel Huynh runs some experiments with different batch sizes (also using the 1Cycle policy discussed above) where he achieves a 4x speed-up by going from batch size 64 to 512. One of the downsides of using large batch sizes, however, is that they might lead to solutions that generalize worse than those trained with smaller batches. Use Automatic Mixed Precision (AMP) The release of PyTorch 1.6 included a native implementation of Automatic Mixed Precision training to PyTorch. The main idea here is that certain operations can be run faster and without a loss of accuracy at semi-precision (FP16) rather than in the single-precision (FP32) used elsewhere. AMP, then, automatically decide which operation should be executed in which format. This allows both for faster training and a smaller memory footprint. In the best case, the usage of AMP would look something like this: import torch Creates once at the beginning of training scaler = torch.cuda.amp.GradScaler() for data, label in data_iter: optimizer.zero_grad() Casts operations to mixed precision with torch.cuda.amp.autocast(): loss = model(data) Scales the loss, and calls backward() to create scaled gradients scaler.scale(loss).backward() Unscales gradients and calls or skips optimizer.step() scaler.step(optimizer) Updates the scale for next iteration scaler.update() Benchmarking a number of common language and vision models on NVIDIA V100 GPUs, Huang and colleagues find that using AMP over regular FP32 training yields roughly 2x – but upto 5.5x – training speed-ups. Currently, only CUDA ops can be autocast in this way. See the documentation here for more details on this and other limitations. u/SVPERBlA points out that you can squeeze out some additional performance (\~ 20%) from AMP on NVIDIA Tensor Core GPUs if you convert your tensors to the Channels Last memory format. Refer to this section in the NVIDIA docs for an explanation of the speedup and more about NCHW versus NHWC tensor formats. Consider using another optimizer AdamW is Adam with weight decay (rather than L2-regularization) which was popularized by fast.ai and is now available natively in PyTorch as torch.optim.AdamW. AdamW seems to consistently outperform Adam in terms of both the error achieved and the training time. See this excellent blog post on why using weight decay instead of L2-regularization makes a difference for Adam. Both Adam and AdamW work well with the 1Cycle policy described above. There are also a few not-yet-native optimizers that have received a lot of attention recently, most notably LARS (pip installable implementation) and LAMB. NVIDA's APEX implements fused versions of a number of common optimizers such as Adam. This implementation avoid a number of passes to and from GPU memory as compared to the PyTorch implementation of Adam, yielding speed-ups in the range of 5%. Turn on cudNN benchmarking If your model architecture remains fixed and your input size stays constant, setting torch.backends.cudnn.benchmark = True might be beneficial (docs). This enables the cudNN autotuner which will benchmark a number of different ways of computing convolutions in cudNN and then use the fastest method from then on. For a rough reference on the type of speed-up you can expect from this, Szymon Migacz achieves a speed-up of 70% on a forward pass for a convolution and a 27% speed-up for a forward + backward pass of the same convolution. One caveat here is that this autotuning might become very slow if you max out the batch size as mentioned above. Beware of frequently transferring data between CPUs and GPUs Beware of frequently transferring tensors from a GPU to a CPU using tensor.cpu() and vice versa using tensor.cuda() as these are relatively expensive. The same applies for .item() and .numpy() – use .detach() instead. If you are creating a new tensor, you can also directly assign it to your GPU using the keyword argument device=torch.device('cuda:0'). If you do need to transfer data, using .to(non_blocking=True), might be useful as long as you don't have any synchronization points after the transfer. If you really have to, you might want to give Santosh Gupta's SpeedTorch a try, although it doesn't seem entirely clear when this actually does/doesn't provide speed-ups. Use gradient/activation checkpointing Quoting directly from the documentation: Checkpointing works by trading compute for memory. Rather than storing all intermediate activations of the entire computation graph for computing backward, the checkpointed part does not save intermediate activations, and instead recomputes them in backward pass. It can be applied on any part of a model. Specifically, in the forward pass, function will run in torch.no\grad() manner, i.e., not storing the intermediate activations. Instead, the forward pass saves the inputs tuple and the functionparameter. In the backwards pass, the saved inputs and function is retrieved, and the forward pass is computed on function again, now tracking the intermediate activations, and then the gradients are calculated using these activation values. So while this will might slightly increase your run time for a given batch size, you'll significantly reduce your memory footprint. This in turn will allow you to further increase the batch size you're using allowing for better GPU utilization. While checkpointing is implemented natively as torch.utils.checkpoint(docs), it does seem to take some thought and effort to implement properly. Priya Goyal has a good tutorial demonstrating some of the key aspects of checkpointing. Use gradient accumulation Another approach to increasing the batch size is to accumulate gradients across multiple .backward() passes before calling optimizer.step(). Following a post by Hugging Face's Thomas Wolf, gradient accumulation can be implemented as follows: model.zero_grad() Reset gradients tensors for i, (inputs, labels) in enumerate(training_set): predictions = model(inputs) Forward pass loss = loss_function(predictions, labels) Compute loss function loss = loss / accumulation_steps Normalize our loss (if averaged) loss.backward() Backward pass if (i+1) % accumulation_steps == 0: Wait for several backward steps optimizer.step() Now we can do an optimizer step model.zero_grad() Reset gradients tensors if (i+1) % evaluation_steps == 0: Evaluate the model when we... evaluate_model() ...have no gradients accumulate This method was developed mainly to circumvent GPU memory limitations and I'm not entirely clear on the trade-off between having additional .backward() loops. This discussion on the fastai forum seems to suggest that it can in fact accelerate training, so it's probably worth a try. Use Distributed Data Parallel for multi-GPU training Methods to accelerate distributed training probably warrant their own post but one simple one is to use torch.nn.DistributedDataParallel rather than torch.nn.DataParallel. By doing so, each GPU will be driven by a dedicated CPU core avoiding the GIL issues of DataParallel. In general, I can strongly recommend reading the documentation on distributed training. Set gradients to None rather than 0 Use .zerograd(settonone=True) rather than .zerograd(). Doing so will let the memory allocator handle the gradients rather than actively setting them to 0. This will lead to yield a modest speed-up as they say in the documentation, so don't expect any miracles. Watch out, doing this is not side-effect free! Check the docs for the details on this. Use .as_tensor() rather than .tensor() torch.tensor() always copies data. If you have a numpy array that you want to convert, use torch.astensor() or torch.fromnumpy() to avoid copying the data. Turn on debugging tools only when actually needed PyTorch offers a number of useful debugging tools like the autograd.profiler, autograd.grad\check, and autograd.anomaly\detection. Make sure to use them to better understand when needed but to also turn them off when you don't need them as they will slow down your training. Use gradient clipping Originally used to avoid exploding gradients in RNNs, there is both some empirical evidence as well as some theoretical support that clipping gradients (roughly speaking: gradient = min(gradient, threshold)) accelerates convergence. Hugging Face's Transformer implementation is a really clean example of how to use gradient clipping as well as some of the other methods such as AMP mentioned in this post. In PyTorch this can be done using torch.nn.utils.clipgradnorm(documentation). It's not entirely clear to me which models benefit how much from gradient clipping but it seems to be robustly useful for RNNs, Transformer-based and ResNets architectures and a range of different optimizers. Turn off bias before BatchNorm This is a very simple one: turn off the bias of layers before BatchNormalization layers. For a 2-D convolutional layer, this can be done by setting the bias keyword to False: torch.nn.Conv2d(..., bias=False, ...).  (Here's a reminder why this makes sense.) You will save some parameters, I would however expect the speed-up of this to be relatively small as compared to some of the other methods mentioned here. Turn off gradient computation during validation This one is straightforward: set torch.no_grad() during validation. Use input and batch normalization You're probably already doing this but you might want to double-check: Are you normalizing your input? Are you using batch-normalization? And here's a reminder of why you probably should. Bonus tip from the comments: Use JIT to fuse point-wise operations. If you have adjacent point-wise operations you can use PyTorch JIT to combine them into one FusionGroup which can then be launched on a single kernel rather than multiple kernels as would have been done per default. You'll also save some memory reads and writes. Szymon Migacz shows how you can use the @torch.jit.script decorator to fuse the operations in a GELU, for instance: @torch.jit.script def fused_gelu(x): return x 0.5 (1.0 + torch.erf(x / 1.41421)) In this case, fusing the operations leads to a 5x speed-up for the execution of fused_gelu as compared to the unfused version. See also this post for an example of how Torchscript can be used to accelerate an RNN. Hat tip to u/Patient_Atmosphere45 for the suggestion. Sources and additional resources Many of the tips listed above come from Szymon Migacz' talk and post in the PyTorch docs. PyTorch Lightning's William Falcon has two interesting posts with tips to speed-up training. PyTorch Lightning does already take care of some of the points above per-default. Thomas Wolf at Hugging Face has a number of interesting articles on accelerating deep learning – with a particular focus on language models. The same goes for Sylvain Gugger and Jeremy Howard: they have many interesting posts in particular on learning rates and AdamW. Thanks to Ben Hahn, Kevin Klein and Robin Vaaler for their feedback on a draft of this post! I've also put all of the above into this blog post.

[D] Why is the AI Hype Absolutely Bonkers
reddit
LLM Vibe Score0
Human Vibe Score1
good_riceThis week

[D] Why is the AI Hype Absolutely Bonkers

Edit 2: Both the repo and the post were deleted. Redacting identifying information as the author has appeared to make rectifications, and it’d be pretty damaging if this is what came up when googling their name / GitHub (hopefully they’ve learned a career lesson and can move on). TL;DR: A PhD candidate claimed to have achieved 97% accuracy for coronavirus from chest x-rays. Their post gathered thousands of reactions, and the candidate was quick to recruit branding, marketing, frontend, and backend developers for the project. Heaps of praise all around. He listed himself as a Director of XXXX (redacted), the new name for his project. The accuracy was based on a training dataset of ~30 images of lesion / healthy lungs, sharing of data between test / train / validation, and code to train ResNet50 from a PyTorch tutorial. Nonetheless, thousands of reactions and praise from the “AI | Data Science | Entrepreneur” community. Original Post: I saw this post circulating on LinkedIn: https://www.linkedin.com/posts/activity-6645711949554425856-9Dhm Here, a PhD candidate claims to achieve great performance with “ARTIFICIAL INTELLIGENCE” to predict coronavirus, asks for more help, and garners tens of thousands of views. The repo housing this ARTIFICIAL INTELLIGENCE solution already has a backend, front end, branding, a README translated in 6 languages, and a call to spread the word for this wonderful technology. Surely, I thought, this researcher has some great and novel tech for all of this hype? I mean dear god, we have branding, and the author has listed himself as the founder of an organization based on this project. Anything with this much attention, with dozens of “AI | Data Scientist | Entrepreneur” members of LinkedIn praising it, must have some great merit, right? Lo and behold, we have ResNet50, from torchvision.models import resnet50, with its linear layer replaced. We have a training dataset of 30 images. This should’ve taken at MAX 3 hours to put together - 1 hour for following a tutorial, and 2 for obfuscating the training with unnecessary code. I genuinely don’t know what to think other than this is bonkers. I hope I’m wrong, and there’s some secret model this author is hiding? If so, I’ll delete this post, but I looked through the repo and (REPO link redacted) that’s all I could find. I’m at a loss for thoughts. Can someone explain why this stuff trends on LinkedIn, gets thousands of views and reactions, and gets loads of praise from “expert data scientists”? It’s almost offensive to people who are like ... actually working to treat coronavirus and develop real solutions. It also seriously turns me off from pursuing an MS in CV as opposed to CS. Edit: It turns out there were duplicate images between test / val / training, as if ResNet50 on 30 images wasn’t enough already. He’s also posted an update signed as “Director of XXXX (redacted)”. This seems like a straight up sleazy way to capitalize on the pandemic by advertising himself to be the head of a made up organization, pulling resources away from real biomedical researchers.

DARPA "AI For Critical Minerals Assessment" Competition [D]
reddit
LLM Vibe Score0
Human Vibe Score0
Scherzers_Brown_EyeThis week

DARPA "AI For Critical Minerals Assessment" Competition [D]

DARPA is hosting a competition called “AI for Critical Mineral Assessments,” which is looking for solutions to automatically extract and georeferenced features from scanned or raster maps. The U.S. Geological Survey uses data from these assessments to build reports that can eventually lead to increasing domestic production of critical minerals and reducing U.S. reliance on imports. The competition includes two independent challenges: Map Georeferencing Challenge: Automated map georeferencing is a difficult task as most USGS maps are not digitized, and may be in a multitude of historical coordinate projection systems. Furthermore, the quality of features on scanned maps, critical for the identification of control points for alignment, can vary greatly. Participants will receive a dataset of 1,000 or more maps of various types for training and validation. The goal of this challenge is to accurately geolocate a map of unknown location and coordinate system by fitting coordinate points that can be referenced to known locations in one or more base maps. Register now-Aug. 26. Map Feature Extraction Challenge: Automated map feature extraction is a difficult task because map features (polygons, points, lines, text) often overlap and are sometimes discontinuous. Not only do the features come in all shapes and sizes, but the same feature type can be depicted in different maps using different symbols or patterns. This makes it challenging to create a universal identifier for even a single feature such as a mine location or mineral resource tracts. Participants will be provided a training set consisting of maps with each legend item labeled and characterized (as point, line, or polygon) and a binary pixel map reflecting the feature’s coverage in the map. The goal of the challenge is to identify all features in a map that appear in the map’s legend. Register Sept. 5 - 16. For each of the two challenges, DARPA will award: · $10,000 for the first prize · $3,000 for the second prize · $1,000 for the third prize You can visit criticalminerals.darpa.mil for complete details on how you can compete.

[P] I built an open SotA image tagging model to do what CLIP won't
reddit
LLM Vibe Score0
Human Vibe Score1
fpgaminerThis week

[P] I built an open SotA image tagging model to do what CLIP won't

I'm a hobbyist ML researcher and finally, after a year of work, built a state of the art machine vision model from scratch. It's ViT-B/16 based, 448x448x3 input, 91M parameters, trained for 660M samples, with multi-label classification as the target task, on over 5000 unique tags. All the big foundation vision models today were trained on heavily filtered datasets, greatly limiting the concepts they can represent, in line with arbitrary sets of rules for what is deemed "wholesome" by leading tech companies. Everything from innocuous to spicy is on the chopping block of those filters. And because CLIP pervades the industry, from StableDiffusion to LLaVA, so does OpenAI's sensibilities. My goal was to build a vision model for tagging images, mainly for labelling images for SD finetunes, but which wasn't as heavily filtered and handicapped as CLIP/BLIP/LLaVA. Something more inclusive, diverse, and sex positive. Starting from the wonderful work of SmilingWolf (https://github.com/SmilingWolf/SW-CV-ModelZoo) and the Danbooru2021 dataset, I iterated for a year on the model, training, and manually labeling a thousand images to help the model generalize beyond the danbooru domain. I'm releasing the first version of this model, dubbed JoyTag, today: https://github.com/fpgaminer/joytag It achieves a mean F1 score of 0.578 across all of its over 5000 tags and across both the anime/manga styled images of the original danbooru dataset, but also photographs and other mediums thanks to the auxiliary training data I provided to it. It was quite the struggle getting to this point, and I probably spent more time and money than any sane person should have. I learned a lot about dealing with datasets as large as danbooru2021, training models at scale, and how to keep yourself awake all night so your 8xA100 rental doesn't crash and blow all your money. In my manual testing outside of even the validation set, the model has generalized well to unseen images, so I'm quite happy with the results thus far. There's plenty more work to do expanding its dataset to improve that F1 score further, and roundout its weak points. With inclusivity and diversity being a major goal of this project, I'm disappointed by some of its remaining limitations (as documented in the GitHub README). But I'm already busy manually tagging more images using my model-augmented workflow. I'm happy to answer questions about the project, the training procedure, anything. All the training parameters are documented on GitHub, but there are so many little details that were hard won over the year. Like that damned loss multiplier. Ugh. Github: https://github.com/fpgaminer/joytag Model download: https://huggingface.co/fancyfeast/joytag/tree/main Demo: https://huggingface.co/spaces/fancyfeast/joytag

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

If only someone told me this before my first startup
reddit
LLM Vibe Score0
Human Vibe Score0.625
johnrushxThis week

If only someone told me this before my first startup

If only someone told me this before my first startup: Validate idea first. I wasted a decade building stuff nobody needed. Incubators and VCs served to me as a validation, but I was so wrong. Kill my EGO. It’s not about me, but the user. I must want what the user wants, not what I want. My taste isn't important. The user has expectations, and I must fulfill them. Don’t chaise investors. Chase users, and then investors will be chasing me. I've never had more incoming interest from VC than now when I'm the least interested in them. Never hire managers. Only hire doers until PMF. So many people know how to manage people and so few can actually get sh\*t done barehand. Landing page is the least important thing in a startup. Pick a simple template, edit texts with a no-code website builder in less than an hour and that's it! At the early stage, I win traffic outside of my website, people are already interested, so don't make them search for the signup button among the texts! Focus on conversion optimization only when the traffic is consistent. Keep it to one page. Nobody gonna browse this website. Hire only fullstack devs. There is nothing less productive in this world than a team of developers for an early-stage product. One full stack dev building the whole product. That’s it. Chase global market from day 1. If the product and marketing are good, it will work on the global market too, if it’s bad, it won’t work on the local market too. So better go global from day 1, so that if it works, the upside is 100x bigger. I launched all startups for the Norwegian market, hoping we will scale to international at some point. I wish I launched to international from day 1 as I do now. The size of the market is 10000x bigger. I can validate and grow products in days, not in years as it used to be. Do SEO from day 2. As early as I can. I ignored this for 14 years. It’s my biggest regret. It takes just 5 minutes to get it done on my landing page. I go to Google Keyword Planner, enter a few keywords around my product, sort them by traffic, filter out high competition kws, pick the top 10, and place them natively on my home page and meta tags. Add one blog article every week. Either manually or by paying for an AI blogging tool. Sell features, before building them. Ask existing users if they want this feature. I run DMs with 10-20 users every day, where I chat about all my ideas and features I wanna add. I clearly see what resonates with me most and only go build those. If I don't have followers, try HN, Reddit, or just search on X for posts and ask it in the replies. People are helpful, they will reply if the question is easy to understand. Hire only people I would wanna hug. My cofounder, an old Danish man said this to me in 2015. And it was a big shift. I realized that if I don’t wanna hug the person, it means I dislike them on a chemical/animal level. Even if I can’t say why, but that’s the fact. Sooner or later, we would have a conflict and eventually break up. It takes up to 10 years to build a startup, make sure I do it with people I have this connection with. Invest all money into my startups and friends. Not crypt0, not stockmarket, not properties. I did some math, if I kept investing all my money into all my friends’ startups, that would be about 70 investments. 3 of them turned into unicorns eventually. Even 1 would have made the bank. Since 2022, I have invested all my money into my products, friends, and network. If I don't have friends who do startups, invest it in myself. Post on Twitter daily. I started posting here in March last year. It’s my primary source of new connections and growth. I could have started it earlier, I don't know why I didn't. Don’t work/partner with corporates. Corporations always seem like an amazing opportunity. They’re big and rich, they promise huge stuff, millions of users, etc. But every single time none of this happens. Because I talk to a regular employees there. They waste my time, destroy focus, shift priorities, and eventually bring in no users/money. Don’t get ever distracted by hype e.g. crypt0. I lost 1.5 years of my life this way. I met the worst people along the way. Fricks, scammers, thieves. Some of my close friends turned into thieves along the way, just because it was so common in that space. I wish this didn’t happen to me. I wish I was stronger and stayed on my mission. Don’t build consumer apps. Only b2b. Consumer apps are so hard, like a lottery. It’s just 0.00001% who make it big. The rest don’t. Even if I got many users, then there is a monetization challenge. I’ve spent 4 years in consumer apps and regret it. Don’t hold on bad project for too long, max 1 year. Some projects just don’t work. In most cases, it’s either the idea that’s so wrong that I can’t even pivot it or it’s a team that is good one by one but can’t make it as a team. Don’t drag this out for years. Tech conferences are a waste of time. They cost money, take energy, and time and I never really meet anyone there. Most people there are the “good” employees of corporations who were sent there as a perk for being loyal to the corporation. Very few fellow makers. Scrum is a Scam. For small teams and bootstrapped teams. If I had a team that had to be nagged every morning with questions as if they were children in kindergarten, then things would eventually fail. The only good stuff I managed to do happened with people who were grownups and could manage their stuff on their own. We would just do everything over chat as a sync on goals and plans. Outsource nothing at all until PMF. In a startup, almost everything needs to be done in a slightly different way, more creative, and more integrated into the vision. When outsourcing, the external members get no love and no case for the product. It’s just yet another assignment in their boring job. Instead of coming up with great ideas for my project they will be just focusing on ramping up their skills to get a promotion or a better job offer. Bootstrap. I spent way too much time raising money. I raised more than 10 times, preseed, seeded, and series A. But each time it was a 3-9 month project, meetings every week, and lots of destruction. I could afford to bootstrap, but I still went the VC-funded way, I don’t know why. To be honest, I didn’t know bootstrapping was a thing I could do or anyone does. It may take a decade. When I was 20, I was convinced it takes a few years to build and succeed with a startup. So I kept pushing my plans forward, to do it once I exited. Family, kids. I wish I married earlier. I wish I had kids earlier. No Free Tier. I'd launch a tool with a free tier, and it'd get sign-ups, but very few would convert. I'd treat free sign-ups as KPIs and run on it for years. I'd brag about signups and visitors. I'd even raise VC money with these stats. But eventually, I would fail to reach PMF. Because my main feedback would come from free users and the product turned into a perfect free product. Once I switched to "paid only" until I validated the product, things went really well. Free and paid users often need different products. Don't fall into this trap as I did. Being To Cheap. I always started by checking all competitors and setting the lowest price. I thought this would be one of the key advantages of my product. But no, I was wrong. The audience on $5 and $50 are totally different. $5: pain in the \*ss, never happy, never recommend me to a friend, leave in 4 months. $50: polite, give genuine feedback, happy, share with friends, become my big fan if I solve their request. I will fail. When I started my first startup. I thought if I did everything right, it would work out. But it turned out that almost every startup fails. I wish I knew that and I tried to fail faster, to get to the second iteration, then to the third, and keep going on, until I either find out nothing works or make it work. Use boilerplates. I wasted years of dev time and millions of VC money to pay for basic things. To build yet another sidebar, yet another dashboard, and payment integration... I had too much pride, I couldn't see myself taking someone else code as a basis for my product. I wanted it to be 100% mine, original, from scratch. Because my product seems special to me. Spend more time with Family & Friends. I missed the weddings of all my best friends and family. I was so busy. I thought if I didn't do it on time, the world would end. Looking back today, it was so wrong. I meet my friends and can't share those memories with them, which makes me very sad. I realized now, that spending 10% of my time with family and friends would practically make no negative impact on my startups. Build Products For Audiences I Love. I never thought of this. I'd often build products either for corporates, consumers, or for developers. It turns out I have no love for all 3. But I deeply love indie founders. Because they are risk-takers and partly kids in their hearts. Once I switched the focus to indie makers on my products, my level of joy increased by 100x for me. Ignore Badges and Awards I was chasing those awards just like everyone else. Going to ceremonies, signing up for events and stuff. I've won tons of awards, but none of those were eventually useful to my business. I better focused on my business and users. Write Every Single Day. When I was a kid, I loved writing stories. In school, they would give an assignment, and I'd often write a long story for it, however, the teacher would put an F on it. The reason was simple, I had an issue with the direction of the letters and the sequence of letters in the words. I still have it, it's just the Grammarly app helping me to correct these issues. So the teacher would fail my stories because almost every sentence had a spelling mistake that I couldn't even see. It made me think I'm made at writing. So I stopped, for 15 years. But I kept telling stories all these years. Recently I realized that in any group, the setup ends up turning into me telling stories to everyone. So I tried it all again, here on X 10 months ago. I love it, the process, the feedback from people. I write every day. I wish I had done it all these years. The End. \ this is an updated version of my post on the same topic from 2 months ago. I've edited some of the points and added 9 new ones.* \\ This is not advice, it's my self-reflection that might help you avoid same mistakes if you think those were mistakes

12 months from idea to product - bootstrapping my own mobile app from 0
reddit
LLM Vibe Score0
Human Vibe Score1
MaartinBlack1996This week

12 months from idea to product - bootstrapping my own mobile app from 0

Introduction It has taken 12 months to develop an app that uses a camera to seamlessly detect fridge ingredients and generate recipes—solving the everyday problem I faced while traveling: "What should I cook for dinner today?" Although the end product has evolved from the initial concept, the ingredient detection feature remains one of the key elements that makes this app truly unique. When I started Keto, the biggest challenge I faced was tracking carbs, typically done through barcode scanning or manual searches. While Swifto offers both of these options, we are proud to introduce a feature that allows you to extract net carb values from a single image with just one click. We’ve combined AI with a great user experience to ensure that anyone embarking on their Keto journey can track their progress with ease. My Experience The app is now at a stage where I can truly seek market validation. Yes, this journey took me around 12 months, starting with the idea, creating the website, and developing the app's UI/UX and backend. At this point, many people might wonder: "Did you validate your idea before? Why create such a complex app without first understanding if there's a market need?" While this approach is undoubtedly risky and may not pay off in the future, I had a strong belief that this product could only be validated when people experienced how it works and saw how seamless the UX is compared to other similar apps. Would I Do It Again? Probably not. While developing the mobile app, I learned a lot about how mobile apps are advertised on the Google Play Store and how challenging it is to break into niche markets. You can develop the best application out there, but if no one sees it, it will never reach the top searches, which is crucial for any app's organic reach. I'll need to devise very creative strategies to gain the attention of those who truly matter for this product's validation and then go from there. However, it seems this will require much more effort than I initially anticipated. I'm open to any questions/suggestions.

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a founder built a B2B AI startup to serve with 65+ global brands (including Fortune500 companies)

AI Palette is an AI-driven platform that helps food and beverage companies predict emerging product trends. I had the opportunity recently to sit down with the founder to get his advice on building an AI-first startup, which he'll be going through in this post. About AI Palette: Co-founders: >!2 (Somsubhra GanChoudhuri, Himanshu Upreti)!!100+!!$12.7M USD!!AI-powered predictive analytics for the CPG (Consumer Packaged Goods) industry!!Signed first paying customer in the first year!!65+ global brands, including Cargill, Diageo, Ajinomoto, Symrise, Mondelez, and L’Oréal, use AI Palette!!Every new product launched has secured a paying client within months!!Expanded into Beauty & Personal Care (BPC), onboarding one of India’s largest BPC companies within weeks!!Launched multiple new product lines in the last two years, creating a unified suite for brand innovation!Identify the pain points in your industry for ideas* When I was working in the flavour and fragrance industry, I noticed a major issue CPG companies faced: launching a product took at least one to two years. For instance, if a company decided today to launch a new juice, it wouldn’t hit the market until 2027. This long timeline made it difficult to stay relevant and on top of trends. Another big problem I noticed was that companies relied heavily on market research to determine what products to launch. While this might work for current consumer preferences, it was highly inefficient since the product wouldn’t actually reach the market for several years. By the time the product launched, the consumer trends had already shifted, making that research outdated. That’s where AI can play a crucial role. Instead of looking at what consumers like today, we realised that companies should use AI to predict what they will want next. This allows businesses to create products that are ahead of the curve. Right now, the failure rate for new product launches is alarmingly high, with 8 out of 10 products failing. By leveraging AI, companies can avoid wasting resources on products that won’t succeed, leading to better, more successful launches. Start by talking to as many industry experts as possible to identify the real problems When we first had the idea for AI Palette, it was just a hunch, a gut feeling—we had no idea whether people would actually pay for it. To validate the idea, we reached out to as many people as we could within the industry. Since our focus area was all about consumer insights, we spoke to professionals in the CPG sector, particularly those in the insights departments of CPG companies. Through these early conversations, we began to see a common pattern emerge and identified the exact problem we wanted to solve. Don’t tell people what you’re building—listen to their frustrations and challenges first. Going into these early customer conversations, our goal was to listen and understand their challenges without telling them what we were trying to build. This is crucial as it ensures that you can gather as much data about the problem to truly understand it and that you aren't biasing their answers by showing your solution. This process helped us in two key ways: First, it validated that there was a real problem in the industry through the number of people who spoke about experiencing the same problem. Second, it allowed us to understand the exact scale and depth of the problem—e.g., how much money companies were spending on consumer research, what kind of tools they were currently using, etc. Narrow down your focus to a small, actionable area to solve initially. Once we were certain that there was a clear problem worth solving, we didn’t try to tackle everything at once. As a small team of two people, we started by focusing on a specific area of the problem—something big enough to matter but small enough for us to handle. Then, we approached customers with a potential solution and asked them for feedback. We learnt that our solution seemed promising, but we wanted to validate it further. If customers are willing to pay you for the solution, it’s a strong validation signal for market demand. One of our early customer interviewees even asked us to deliver the solution, which we did manually at first. We used machine learning models to analyse the data and presented the results in a slide deck. They paid us for the work, which was a critical moment. It meant we had something with real potential, and we had customers willing to pay us before we had even built the full product. This was the key validation that we needed. By the time we were ready to build the product, we had already gathered crucial insights from our early customers. We understood the specific information they wanted and how they wanted the results to be presented. This input was invaluable in shaping the development of our final product. Building & Product Development Start with a simple concept/design to validate with customers before building When we realised the problem and solution, we began by designing the product, but not by jumping straight into coding. Instead, we created wireframes and user interfaces using tools like InVision and Figma. This allowed us to visually represent the product without the need for backend or frontend development at first. The goal was to showcase how the product would look and feel, helping potential customers understand its value before we even started building. We showed these designs to potential customers and asked for feedback. Would they want to buy this product? Would they pay for it? We didn’t dive into actual development until we found a customer willing to pay a significant amount for the solution. This approach helped us ensure we were on the right track and didn’t waste time or resources building something customers didn’t actually want. Deliver your solution using a manual consulting approach before developing an automated product Initially, we solved problems for customers in a more "consulting" manner, delivering insights manually. Recall how I mentioned that when one of our early customer interviewees asked us to deliver the solution, we initially did it manually by using machine learning models to analyse the data and presenting the results to them in a slide deck. This works for the initial stages of validating your solution, as you don't want to invest too much time into building a full-blown MVP before understanding the exact features and functionalities that your users want. However, after confirming that customers were willing to pay for what we provided, we moved forward with actual product development. This shift from a manual service to product development was key to scaling in a sustainable manner, as our building was guided by real-world feedback and insights rather than intuition. Let ongoing customer feedback drive iteration and the product roadmap Once we built the first version of the product, it was basic, solving only one problem. But as we worked closely with customers, they requested additional features and functionalities to make it more useful. As a result, we continued to evolve the product to handle more complex use cases, gradually developing new modules based on customer feedback. Product development is a continuous process. Our early customers pushed us to expand features and modules, from solving just 20% of their problems to tackling 50–60% of their needs. These demands shaped our product roadmap and guided the development of new features, ultimately resulting in a more complete solution. Revenue and user numbers are key metrics for assessing product-market fit. However, critical mass varies across industries Product-market fit (PMF) can often be gauged by looking at the size of your revenue and the number of customers you're serving. Once you've reached a certain critical mass of customers, you can usually tell that you're starting to hit product-market fit. However, this critical mass varies by industry and the type of customers you're targeting. For example, if you're building an app for a broad consumer market, you may need thousands of users. But for enterprise software, product-market fit may be reached with just a few dozen key customers. Compare customer engagement and retention with other available solutions on the market for product-market fit Revenue and the number of customers alone isn't always enough to determine if you're reaching product-market fit. The type of customer and the use case for your product also matter. The level of engagement with your product—how much time users are spending on the platform—is also an important metric to track. The more time they spend, the more likely it is that your product is meeting a crucial need. Another way to evaluate product-market fit is by assessing retention, i.e whether users are returning to your platform and relying on it consistently, as compared to other solutions available. That's another key indication that your solution is gaining traction in the market. Business Model & Monetisation Prioritise scalability Initially, we started with a consulting-type model where we tailor-made specific solutions for each customer use-case we encountered and delivered the CPG insights manually, but we soon realized that this wasn't scalable. The problem with consulting is that you need to do the same work repeatedly for every new project, which requires a large team to handle the workload. That is not how you sustain a high-growth startup. To solve this, we focused on building a product that would address the most common problems faced by our customers. Once built, this product could be sold to thousands of customers without significant overheads, making the business scalable. With this in mind, we decided on a SaaS (Software as a Service) business model. The benefit of SaaS is that once you create the software, you can sell it to many customers without adding extra overhead. This results in a business with higher margins, where the same product can serve many customers simultaneously, making it much more efficient than the consulting model. Adopt a predictable, simplistic business model for efficiency. Look to industry practices for guidance When it came to monetisation, we considered the needs of our CPG customers, who I knew from experience were already accustomed to paying annual subscriptions for sales databases and other software services. We decided to adopt the same model and charge our customers an annual upfront fee. This model worked well for our target market, aligning with industry standards and ensuring stable, recurring revenue. Moreover, our target CPG customers were already used to this business model and didn't have to choose from a huge variety of payment options, making closing sales a straightforward and efficient process. Marketing & Sales Educate the market to position yourself as a thought leader When we started, AI was not widely understood, especially in the CPG industry. We had to create awareness around both AI and its potential value. Our strategy focused on educating potential users and customers about AI, its relevance, and why they should invest in it. This education was crucial to the success of our marketing efforts. To establish credibility, we adopted a thought leadership approach. We wrote blogs on the importance of AI and how it could solve problems for CPG companies. We also participated in events and conferences to demonstrate our expertise in applying AI to the industry. This helped us build our brand and reputation as leaders in the AI space for CPG, and word-of-mouth spread as customers recognized us as the go-to company for AI solutions. It’s tempting for startups to offer products for free in the hopes of gaining early traction with customers, but this approach doesn't work in the long run. Free offerings don’t establish the value of your product, and customers may not take them seriously. You should always charge for pilots, even if the fee is minimal, to ensure that the customer is serious about potentially working with you, and that they are committed and engaged with the product. Pilots/POCs/Demos should aim to give a "flavour" of what you can deliver A paid pilot/POC trial also gives you the opportunity to provide a “flavour” of what your product can deliver, helping to build confidence and trust with the client. It allows customers to experience a detailed preview of what your product can do, which builds anticipation and desire for the full functionality. During this phase, ensure your product is built to give them a taste of the value you can provide, which sets the stage for a broader, more impactful adoption down the line. Fundraising & Financial Management Leverage PR to generate inbound interest from VCs When it comes to fundraising, our approach was fairly traditional—we reached out to VCs and used connections from existing investors to make introductions. However, looking back, one thing that really helped us build momentum during our fundraising process was getting featured in Tech in Asia. This wasn’t planned; it just so happened that Tech in Asia was doing a series on AI startups in Southeast Asia and they reached out to us for an article. During the interview, they asked if we were fundraising, and we mentioned that we were. As a result, several VCs we hadn’t yet contacted reached out to us. This inbound interest was incredibly valuable, and we found it far more effective than our outbound efforts. So, if you can, try to generate some PR attention—it can help create inbound interest from VCs, and that interest is typically much stronger and more promising than any outbound strategies because they've gone out of their way to reach out to you. Be well-prepared and deliberate about fundraising. Keep trying and don't lose heart When pitching to VCs, it’s crucial to be thoroughly prepared, as you typically only get one shot at making an impression. If you mess up, it’s unlikely they’ll give you a second chance. You need to have key metrics at your fingertips, especially if you're running a SaaS company. Be ready to answer questions like: What’s your retention rate? What are your projections for the year? How much will you close? What’s your average contract value? These numbers should be at the top of your mind. Additionally, fundraising should be treated as a structured process, not something you do on the side while juggling other tasks. When you start, create a clear plan: identify 20 VCs to reach out to each week. By planning ahead, you’ll maintain momentum and speed up the process. Fundraising can be exhausting and disheartening, especially when you face multiple rejections. Remember, you just need one investor to say yes to make it all worthwhile. When using funds, prioritise profitability and grow only when necessary. Don't rely on funding to survive. In the past, the common advice for startups was to raise money, burn through it quickly, and use it to boost revenue numbers, even if that meant operating at a loss. The idea was that profitability wasn’t the main focus, and the goal was to show rapid growth for the next funding round. However, times have changed, especially with the shift from “funding summer” to “funding winter.” My advice now is to aim for profitability as soon as possible and grow only when it's truly needed. For example, it’s tempting to hire a large team when you have substantial funds in the bank, but ask yourself: Do you really need 10 new hires, or could you get by with just four? Growing too quickly can lead to unnecessary expenses, so focus on reaching profitability as soon as possible, rather than just inflating your team or burn rate. The key takeaway is to spend your funds wisely and only when absolutely necessary to reach profitability. You want to avoid becoming dependent on future VC investments to keep your company afloat. Instead, prioritize reaching break-even as quickly as you can, so you're not reliant on external funding to survive in the long run. Team-Building & Leadership Look for complementary skill sets in co-founders When choosing a co-founder, it’s important to find someone with a complementary skill set, not just someone you’re close to. For example, I come from a business and commercial background, so I needed someone with technical expertise. That’s when I found my co-founder, Himanshu, who had experience in machine learning and AI. He was a great match because his technical knowledge complemented my business skills, and together we formed a strong team. It might seem natural to choose your best friend as your co-founder, but this can often lead to conflict. Chances are, you and your best friend share similar interests, skills, and backgrounds, which doesn’t bring diversity to the table. If both of you come from the same industry or have the same strengths, you may end up butting heads on how things should be done. Having diverse skill sets helps avoid this and fosters a more collaborative working relationship. Himanshu (left) and Somsubhra (right) co-founded AI Palette in 2018 Define roles clearly to prevent co-founder conflict To avoid conflict, it’s essential that your roles as co-founders are clearly defined from the beginning. If your co-founder and you have distinct responsibilities, there is no room for overlap or disagreement. This ensures that both of you can work without stepping on each other's toes, and there’s mutual respect for each other’s expertise. This is another reason as to why it helps to have a co-founder with a complementary skillset to yours. Not only is having similar industry backgrounds and skillsets not particularly useful when building out your startup, it's also more likely to lead to conflicts since you both have similar subject expertise. On the other hand, if your co-founder is an expert in something that you're not, you're less likely to argue with them about their decisions regarding that aspect of the business and vice versa when it comes to your decisions. Look for employees who are driven by your mission, not salary For early-stage startups, the first hires are crucial. These employees need to be highly motivated and excited about the mission. Since the salary will likely be low and the work demanding, they must be driven by something beyond just the paycheck. The right employees are the swash-buckling pirates and romantics, i.e those who are genuinely passionate about the startup’s vision and want to be part of something impactful beyond material gains. When employees are motivated by the mission, they are more likely to stick around and help take the startup to greater heights. A litmus test for hiring: Would you be excited to work with them on a Sunday? One of the most important rounds in the hiring process is the culture fit round. This is where you assess whether a candidate shares the same values as you and your team. A key question to ask yourself is: "Would I be excited to work with this person on a Sunday?" If there’s any doubt about your answer, it’s likely not a good fit. The idea is that you want employees who align with the company's culture and values and who you would enjoy collaborating with even outside of regular work hours. How we structure the team at AI Palette We have three broad functions in our organization. The first two are the big ones: Technical Team – This is the core of our product and technology. This team is responsible for product development and incorporating customer feedback into improving the technology Commercial Team – This includes sales, marketing, customer service, account managers, and so on, handling everything related to business growth and customer relations. General and Administrative Team – This smaller team supports functions like finance, HR, and administration. As with almost all businesses, we have teams that address the two core tasks of building (technical team) and selling (commercial team), but given the size we're at now, having the administrative team helps smoothen operations. Set broad goals but let your teams decide on execution What I've done is recruit highly skilled people who don't need me to micromanage them on a day-to-day basis. They're experts in their roles, and as Steve Jobs said, when you hire the right person, you don't have to tell them what to do—they understand the purpose and tell you what to do. So, my job as the CEO is to set the broader goals for them, review the plans they have to achieve those goals, and periodically check in on progress. For example, if our broad goal is to meet a certain revenue target, I break it down across teams: For the sales team, I’ll look at how they plan to hit that target—how many customers they need to sell to, how many salespeople they need, and what tactics and strategies they plan to use. For the technical team, I’ll evaluate our product offerings—whether they think we need to build new products to attract more customers, and whether they think it's scalable for the number of customers we plan to serve. This way, the entire organization's tasks are cascaded in alignment with our overarching goals, with me setting the direction and leaving the details of execution to the skilled team members that I hire.

Is there any point in building a product with AI anymore?
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

Is there any point in building a product with AI anymore?

Everyone and their grandmother are building AI products. So it begs the question. Is the AI market now too saturated? Have all the AI apps been thought of? Of course not don't be silly, There is still a significant opportunity to create AI products and become profitable. But let's play devils advocate here. Let's say you're a developer and you want to build your first SaaS product. Now, imagine a world where all AI products were already thought of (a scary thought). What would you do? Would you move onto something else? See, when people think of an idea for a SaaS product what often happens is they do a quick Google search and tend to think "oh crap, it already exists, better move on". Maybe that's the right thing to do...but then again maybe it's not. Before you run for the hills, make sure to check the SEO potential for your idea. If your idea has the potential to rank high on Google and there are already hundreds/thousands of people looking for it then you can take that as all the validation you need to start building it. Here are 3 AI ideas that all have good SEO potential. Each idea has keywords that you can target with a difficulty level 500. This means it's easy to rank high in Google for them and they have a high number of people searching for them each month. AI Accounting Software A Saas product that uses AI to analyze bank transactions, invoices, and receipts to automatically match them and reconcile accounts in real-time, reducing manual work and errors. It would also offer predictive insights, suggesting optimal payment times or highlighting potential cash flow issues based on historical data. Could potentially be integrated with popular accounting software like QuickBooks or Xero. SEO Potential Keyword: ai accounting Keyword Difficulty: 9 Average Search Volume 2900 AI Human Resources Software AI Human Resources Software An AI-driven candidate screening and onboarding platform for small to medium-sized businesses. The tool would use AI to automatically filter job applications based on predefined criteria, rank candidates, and even conduct initial interview assessments using natural language processing. It could also manage onboarding tasks by automating the distribution of paperwork, training schedules, and team introductions. SEO Potential Keyword: ai human resources Keyword Difficulty: 17 Average Search Volume 2900 AI Nutrition Tool A Micro SaaS which creates personalized meal planning and nutrition analysis. The platform would use AI to create tailored meal plans based on users' dietary goals, preferences, allergies, and health data (such as activity level or medical conditions). It could analyze food labels, suggest healthier alternatives, and track nutrient intake in real time, helping users maintain balanced diets. SEO Potential Keyword: ai nutrition Keyword Difficulty: 3 Average Search Volume 720 I created a tool (check the first comment) to find ideas like this.

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company
reddit
LLM Vibe Score0
Human Vibe Score0.778
wutangsamThis week

12 months ago, I was unemployed. Last week my side hustle got acquired by a $500m fintech company

I’ve learned so much over the years from this subreddit. I thought I’d return the favour and share some of my own learnings. In November 2020 my best friend and I had an idea. “What if we could find out which stocks the Internet is talking about?” This formed the origins of Ticker Nerd. 9 months later we sold Ticker Nerd to Finder (an Australian fintech company valued at around $500m). In this post, I am going to lay out how we got there. How we came up with the idea First off, like other posts have covered - you don’t NEED a revolutionary or original idea to build a business. There are tonnes of “boring” businesses making over 7 figures a year e.g. law firms, marketing agencies, real estate companies etc. If you’re looking for an exact formula to come up with a great business idea I’m sorry, but it doesn’t exist. Finding new business opportunities is more of an art than a science. Although, there are ways you can make it easier to find inspiration. Below are the same resources I use for inspiration. I rarely ever come up with ideas without first searching one of the resources below for inspiration: Starter Story Twitter Startup Ideas My First Million Trends by the Hustle Trends VC To show how you how messy, random and unpredictable it can be to find an idea - let me explain how my co-founder and I came up with the idea for Ticker Nerd: We discovered a new product on Twitter called Exploding Topics. It was a newsletter that uses a bunch of software and algorithms to find trends that are growing quickly before they hit the mainstream. I had recently listened to a podcast episode from My First Million where they spoke about Motley Fool making hundreds of millions from their investment newsletters. We asked ourselves what if we could build a SaaS platform similar to Exploding Topics but it focused on stocks? We built a quick landing page using Carrd + Gumroad that explained what our new idea will do and included a payment option to get early access for $49. We called it Exploding Stock (lol). We shared it around a bunch of Facebook groups and subreddits. We made $1,000 in pre-sales within a couple days. My co-founder and I can’t code so we had to find a developer to build our idea. We interviewed a bunch of potential candidates. Meanwhile, I was trawling through Wall Street Bets and found a bunch of free tools that did roughly what we wanted to build. Instead of building another SaaS tool that did the same thing as these free tools we decided to pivot from our original idea. Our new idea = a paid newsletter that sends a weekly report that summarises 2 of the best stocks that are growing in interest on the Internet. We emailed everyone who pre-ordered access, telling them about the change and offered a full refund if they wanted. tl;dr: We essentially combined two existing businesses (Exploding Topics and Motley Fool) and made it way better. We validated the idea by finding out if people will actually pay money for it BEFORE we decided to build it. The idea we started out with changed over time. How to work out if your idea will actually make money It’s easy to get hung up on designing the logo or choosing the perfect domain name for your new idea. At this stage none of that matters. The most important thing is working out if people will pay money for it. This is where validation comes in. We usually validate ideas using Carrd. It lets you build a simple one page site without having to code. The Ticker Nerd site was actually built using a Carrd template. Here’s how you can do it yourself (at a high level): Create a Carrd pro account (yes it's a $49 one off payment but you’ll get way more value out of it). Buy a cheap template and send it to your Carrd account. You can build your own template but this will save you a lot of time. Once the template reaches your Carrd account, duplicate it. Leave the original so it can be duplicated for other ideas. Jump onto Canva (free) and create a logo using the free logos provided. Import your logo. Add copy to the page that explains your idea. Use the AIDA formula. Sign up to Gumroad (free) and create a pre-sale campaign. Create a discounted lifetime subscription or version of the product. This will be used pre-sales. Add the copy from the site into the pre-sale campaign on Gumroad. Add a ‘widget’ to Carrd and connect it to Gumroad using the existing easy integration feature. Purchase a domain name. Connect it to Carrd. Test the site works. Share your website Now the site is ready you can start promoting it in various places to see how the market reacts. An easy method is to find relevant subreddits using Anvaka (Github tool) or Subreddit Stats. The Anvaka tool provides a spider map of all the connected subreddits that users are active in. The highlighted ones are most relevant. You can post a thread in these subreddits that offer value or can generate discussion. For example: ‘I’m creating a tool that can write all your copy, would anyone actually use this?’ ‘What does everything think of using AI to get our copy written faster?’ ‘It’s time to scratch my own itch, I’m creating a tool that writes marketing copy using GPT-3. What are the biggest problems you face writing marketing copy? I’ll build a solution for it’ Reddit is pretty brutal these days so make sure the post is genuine and only drop your link in the comments or in the post if it seems natural. If people are interested they’ll ask for the link. Another great place to post is r/entrepreuerridealong and r/business_ideas. These subreddits expect people to share their ideas and you’ll likely make some sales straight off the bat. I also suggest posting in some Facebook groups (related to your idea) as well just for good measure. Assess the results If people are paying you for early access you can assume that it’s worth building your idea. The beauty of posting your idea on Reddit or in Facebook groups is you’ll quickly learn why people love/hate your idea. This can help you decide how to tweak the idea or if you should drop it and move on to the next one. How we got our first 100 customers (for free) By validating Ticker Nerd using subreddits and Facebook groups this gave us our first paying customers. But we knew this wouldn’t be sustainable. We sat down and brainstormed every organic strategy we could use to get traction as quickly as possible. The winner: a Product Hunt launch. A successful Product Hunt launch isn’t easy. You need: Someone that has a solid reputation and audience to “hunt” your product (essentially an endorsement). An aged Product Hunt account - you can’t post any products if your account is less than a week old. To be following relevant Product Hunt members - since they get notified when you launch a new product if they’re following you. Relationships with other builders and makers on Product Hunt that also have a solid reputation and following. Although, if you can pull it off you can get your idea in front of tens of thousands of people actively looking for new products. Over the next few weeks, I worked with my co-founder on connecting with different founders, indie hackers and entrepreneurs mainly via Twitter. We explained to them our plans for the Product Hunt launch and managed to get a small army of people ready to upvote our product on launch day. We were both nervous on the day of the launch. We told ourselves to have zero expectations. The worst that could happen was no one signed up and we were in the same position as we’re in now. Luckily, within a couple of hours Ticker Nerd was on the homepage of Product Hunt and in the top 10. The results were instant. After 24 hours we had around 200 people enter their payment details to sign up for our free trial. These signups were equal to around $5,800 in monthly recurring revenue. \-- I hope this post was useful! Drop any questions you have below and I’ll do my best to respond :)

 I just sold my startup for $200,000 after 11 months. AMA
reddit
LLM Vibe Score0
Human Vibe Score0
jeannenThis week

I just sold my startup for $200,000 after 11 months. AMA

Last August, I was looking for a startup idea I could grow and made a MVP in a week then launched it. I received the $200,000 wire from the buyer a couple of days ago I found tons of useful info online for free, so I hope this can be my way of giving back :) Here is some background: Idea I got the idea when trying to write a tweet using Google Doc's transcription tool, which was terrible. I was pretty sure I wasn't the only one too lazy to type, I made my own solution using AI to transcribe and reformat voice notes into any kind of content. I called it Talknotes, mainly because it was the only domain available lol Validation: My rule is to only reinvest what the project generates. After listing on startup directories and posting on Twitter, I generated $700 in 10 days. It wasn't much, but enough to show interest and keep me motivated. I added user-requested features, but the launch effect wore off, and daily revenues dropped to $0 after a few weeks. I almost gave up, but friends encouraged me to continue. In October, I launched on ProductHunt and it blew up. It became Product of the Day and reached $1500 MRR thanks to media coverage. I initially built everything using vanilla JS/CSS/HTML + Node for backend. But it's pretty limited for apps with lots of interactivity so, I rebuilt the app using Nuxt.js to make it easier to ship new features. Then, I launched ads on Facebook and I implemented a feedback loop: Get new users Learn about them through onboarding Make more ads based on onboarding data This doubled MRR in about 2 months. Burnout and Sale: In May, I had a bad burnout after emergency bug fixes. This made it hard to work on the app after. At this point MRR was around $7000 and total revenues around $70,0000 I listed it on Acquire.com for $200,000, a very good price for the buyer considering revenues and growth. I could've gotten $300,000 with buyer financing or earn-outs, but I wanted cash, $200,000 today is better than $300,000 in a year. Everything was smooth until we tried using Escrow, which almost fucked up the deal (details here). Long story short, had to threaten them to make a sponsored post on Twitter explaining what they did + legal action. They sent the refund the very next day, and we completed the transfer directly. Now, this isn't an overnight success. It's the result of 7 years of grind. I launched over 40 projects since I started, and most of them failed. I often worked 100 hours per week, and I rarely go out or meet many people. It's not for everyone, but I'm fine with it With the profit from the app + sale, and other projects, I have close to 1/3 of a million dollar. I could retire in Asia if I wanted Just mind blowing to think I wrote funny characters in a code editor and sold it for the price of a house lol Edit 1: A few people got confused. I said it's 7 years of grind and most of my projects failed, not that I was not making money. I also said I OFTEN worked 100h/week, not every week :) Since I learned to code 2 years ago I've made close to $400k from my app's profit + exit (this one + another one for $65k last year). And before that I was making money as a marketing freelancer. Also, I dropped after high-school, so, I had to learn everything from scratch, it takes time! Edit 2: Lots of people asked how/where I learned to code in 2 months. I wrote a blog/journal about it back then with links to resources, you can find it here if you're interested

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security
reddit
LLM Vibe Score0
Human Vibe Score1
Royal_Rest8409This week

How a Small Startup in Asia Secured a Contract with the US Department of Homeland Security

Uzair Javaid, a Ph.D. with a passion for data privacy, co-founded Betterdata to tackle one of AI's most pressing challenges: protecting privacy while enabling innovation. Recently, Betterdata secured a lucrative contract with the US Department of Homeland Security, 1 of only 4 companies worldwide to do so and the only one in Asia. Here's how he did it: The Story So what's your story? I grew up in Peshawar, Pakistan, excelling in coding despite studying electrical engineering. Inspired by my professors, I set my sights on studying abroad and eventually earned a Ph.D. scholarship at NUS Singapore, specializing in data security and privacy. During my research, I ethically hacked Ethereum and published 15 papers—three times the requirement. While wrapping up my Ph.D., I explored startup ideas and joined Entrepreneur First, where I met Kevin Yee. With his expertise in generative models and mine in privacy, we founded Betterdata. Now, nearly three years in, we’ve secured a major contract with the U.S. Department of Homeland Security—one of only four companies globally and the only one from Asia. The Startup In a nutshell, what does your startup do? Betterdata is a startup that uses AI and synthetic data generation to address two major challenges: data privacy and the scarcity of high-quality data for training AI models. By leveraging generative models and privacy-enhancing technologies, Betterdata enables businesses, such as banks, to use customer data without breaching privacy regulations. The platform trains AI on real data, learns its patterns, and generates synthetic data that mimics the real thing without containing any personal or sensitive information. This allows companies to innovate and develop AI solutions safely and ethically, all while tackling the growing need for diverse, high-quality data in AI development. How did you conduct ideation and validation for your startup? The initial idea for Betterdata came from personal experience. During my Ph.D., I ethically hacked Ethereum’s blockchain, exposing flaws in encryption-based data sharing. This led me to explore AI-driven deep synthesis technology—similar to deepfakes but for structured data privacy. With GDPR impacting 28M+ businesses, I saw a massive opportunity to help enterprises securely share data while staying compliant. To validate the idea, I spoke to 50 potential customers—a number that strikes the right balance. Some say 100, but that’s impractical for early-stage founders. At 50, patterns emerge: if 3 out of 10 mention the same problem, and this repeats across 50, you have 10–15 strong signals, making it a solid foundation for an MVP. Instead of outbound sales, which I dislike, we used three key methods: Account-Based Marketing (ABM)—targeting technically savvy users with solutions for niche problems, like scaling synthetic data for banks. Targeted Content Marketing—regular customer conversations shaped our thought leadership and outreach. Raising Awareness Through Partnerships—collaborating with NUS, Singapore’s PDPC, and Plug and Play to build credibility and educate the market. These strategies attracted serious customers willing to pay, guiding Betterdata’s product development and market fit. How did you approach the initial building and ongoing product development? In the early stages, we built synthetic data generation algorithms and a basic UI for proof-of-concept, using open-source datasets to engage with banks. We quickly learned that banks wouldn't share actual customer data due to privacy concerns, so we had to conduct on-site installations and gather feedback to refine our MVP. Through continuous consultation with customers, we discovered real enterprise data posed challenges, such as missing values, which led us to adapt our prototype accordingly. This iterative approach of listening to customer feedback and observing their usage allowed us to improve our product, enhance UX, and address unmet needs while building trust and loyalty. Working closely with our customers also gives us a data advantage. Our solution’s effectiveness depends on customer data, which we can't fully access, but bridging this knowledge gap gives us a competitive edge. The more customers we test on, the more our algorithms adapt to diverse use cases, making it harder for competitors to replicate our insights. My approach to iteration is simple: focus solely on customer feedback and ignore external noise like trends or advice. The key question for the team is: which customer is asking for this feature or solution? As long as there's a clear answer, we move forward. External influences, such as AI hype, often bring more confusion than clarity. True long-term success comes from solving real customer problems, not chasing trends. Customers may not always know exactly what they want, but they understand their problems. Our job is to identify these problems and solve them in innovative ways. While customers may suggest specific features, we stay focused on solving the core issue rather than just fulfilling their exact requests. The idea aligns with the quote often attributed to Henry Ford: "If I asked people what they wanted, they would have said faster horses." The key is understanding their problems, not just taking requests at face value. How do you assess product-market fit? To assess product-market fit, we track two key metrics: Customers' Willingness to Pay: We measure both the quantity and quality of meetings with potential customers. A high number of meetings with key decision-makers signals genuine interest. At Betterdata, we focused on getting meetings with people in banks and large enterprises to gauge our product's resonance with the target market. How Much Customers Are Willing to Pay: We monitor the price customers are willing to pay, especially in the early stages. For us, large enterprises, like banks, were willing to pay a premium for our synthetic data platform due to the growing need for privacy tech. This feedback guided our product refinement and scaling strategy. By focusing on these metrics, we refined our product and positioned it for scaling. What is your business model? We employ a structured, phase-driven approach for out business model, as a B2B startup. I initially struggled with focusing on the core value proposition in sales, often becoming overly educational. Eventually, we developed a product roadmap with models that allowed us to match customer needs to specific offerings and justify our pricing. Our pricing structure includes project-based pilots and annual contracts for successful deployments. At Betterdata, our customer engagement unfolds across three phases: Phase 1: Trial and Benchmarking \- We start with outreach and use open-source datasets to showcase results, offering customers a trial period to evaluate the solution. Phase 2: Pilot or PoC \- After positive trial results, we conduct a PoC or pilot using the customer’s private data, with the understanding that successful pilots lead to an annual contract. Phase 3: Multi-Year Contracts \- Following a successful pilot, we transition to long-term commercial contracts, focusing on multi-year agreements to ensure stability and ongoing partnerships. How do you do marketing for your brand? We take a non-conventional approach to marketing, focusing on answering one key question: Which customers are willing to pay, and how much? This drives our messaging to show how our solution meets their needs. Our strategy centers around two main components: Building a network of lead magnets \- These are influential figures like senior advisors, thought leaders, and strategic partners. Engaging with institutions like IMDA, SUTD, and investors like Plug and Play helps us gain access to the right people and foster warm introductions, which shorten our sales cycle and ensure we’re reaching the right audience. Thought leadership \- We build our brand through customer traction, technology evidence, and regulatory guidelines. This helps us establish credibility in the market and position ourselves as trusted leaders in our field. This holistic approach has enabled us to navigate diverse market conditions in Asia and grow our B2B relationships. By focusing on these areas, we drive business growth and establish strong trust with stakeholders. What's your advice for fundraising? Here are my key takeaways for other founders when it comes to fundraising: Fundraise When You Don’t Need To We closed our seed round in April 2023, a time when we weren't actively raising. Founders should always be in fundraising mode, even when they're not immediately in need of capital. Don’t wait until you have only a few months of runway left. Keep the pipeline open and build relationships. When the timing is right, execution becomes much easier. For us, our investment came through a combination of referrals and inbound interest. Even our lead investor initially rejected us, but after re-engaging, things eventually fell into place. It’s crucial to stay humble, treat everyone with respect, and maintain those relationships for when the time is right. Be Mindful of How You Present Information When fundraising, how you present information matters a lot. We created a comprehensive, easily digestible investment memo, hosted on Notion, which included everything an investor might need—problem, solution, market, team, risks, opportunities, and data. The goal was for investors to be able to get the full picture within 30 minutes without chasing down extra details. We also focused on making our financial model clear and meaningful, even though a 5-year forecast might be overkill at the seed stage. The key was clarity and conciseness, and making it as easy as possible for investors to understand the opportunity. I learned that brevity and simplicity are often the best ways to make a memorable impact. For the pitch itself, keep it simple and focus on 4 things: problem, solution, team, and market. If you can summarize each of these clearly and concisely, you’ll have a compelling pitch. Later on, you can expand into market segments, traction, and other metrics, but for seed-stage, focus on those four areas, and make sure you’re strong in at least three of them. If you do, you'll have a compelling case. How do you run things day-to-day? i.e what's your operational workflow and team structure? Here's an overview of our team structure and process: Internally: Our team is divided into two main areas: backend (internal team) and frontend (market-facing team). There's no formal hierarchy within the backend team. We all operate as equals, defining our goals based on what needs to be developed, assigning tasks, and meeting weekly to share updates and review progress. The focus is on full ownership of tasks and accountability for getting things done. I also contribute to product development, identifying challenges and clearing obstacles to help the team move forward. Backend Team: We approach tasks based on the scope defined by customers, with no blame or hierarchy. It's like a sports team—sometimes someone excels, and other times they struggle, but we support each other and move forward together. Everyone has the creative freedom to work in the way that suits them best, but we establish regular meetings and check-ins to ensure alignment and progress. Frontend Team: For the market-facing side, we implement a hierarchy because the market expects this structure. If I present myself as "CEO," it signals authority and credibility. This distinction affects how we communicate with the market and how we build our brand. The frontend team is split into four main areas: Business Product (Software Engineering) Machine Learning Engineering R&D The C-suite sits at the top, followed by team leads, and then the executors. We distill market expectations into actionable tasks, ensuring that everyone is clear on their role and responsibilities. Process: We start by receiving market expectations and defining tasks based on them. Tasks are assigned to relevant teams, and execution happens with no communication barriers between team members. This ensures seamless collaboration and focused execution. The main goal is always effectiveness—getting things done efficiently while maintaining flexibility in how individuals approach their work. In both teams, there's an emphasis on accountability, collaboration, and clear communication, but the structure varies according to the nature of the work and external expectations.

how I built a $6k/mo business with cold email
reddit
LLM Vibe Score0
Human Vibe Score1
Afraid-Astronomer130This week

how I built a $6k/mo business with cold email

I scaled my SaaS to a $6k/mo business in under 6 months completely using cold email. However, the biggest takeaway for me is not a business that’s potentially worth 6-figure. It’s having a glance at the power of cold emails in the age of AI. It’s a rapidly evolving yet highly-effective channel, but no one talks about how to do it properly. Below is the what I needed 3 years ago, when I was stuck with 40 free users on my first app. An app I spent 2 years building into the void. Entrepreneurship is lonely. Especially when you are just starting out. Launching a startup feel like shouting into the dark. You pour your heart out. You think you have the next big idea, but no one cares. You write tweets, write blogs, build features, add tests. You talk to some lukewarm leads on Twitter. You do your big launch on Product Hunt. You might even get your first few sales. But after that, crickets... Then, you try every distribution channel out there. SEO Influencers Facebook ads Affiliates Newsletters Social media PPC Tiktok Press releases The reality is, none of them are that effective for early-stage startups. Because, let's face it, when you're just getting started, you have no clue what your customers truly desire. Without understanding their needs, you cannot create a product that resonates with them. It's as simple as that. So what’s the best distribution channel when you are doing a cold start? Cold emails. I know what you're thinking, but give me 10 seconds to change your mind: When I first heard about cold emailing I was like: “Hell no! I’m a developer, ain’t no way I’m talking to strangers.” That all changed on Jan 1st 2024, when I actually started sending cold emails to grow. Over the period of 6 months, I got over 1,700 users to sign up for my SaaS and grew it to a $6k/mo rapidly growing business. All from cold emails. Mastering Cold Emails = Your Superpower I might not recommend cold emails 3 years ago, but in 2024, I'd go all in with it. It used to be an expensive marketing channel bootstrapped startups can’t afford. You need to hire many assistants, build a list, research the leads, find emails, manage the mailboxes, email the leads, reply to emails, do meetings. follow up, get rejected... You had to hire at least 5 people just to get the ball rolling. The problem? Managing people sucks, and it doesn’t scale. That all changed with AI. Today, GPT-4 outperforms most human assistants. You can build an army of intelligent agents to help you complete tasks that’d previously be impossible without human input. Things that’d take a team of 10 assistants a week can now be done in 30 minutes with AI, at far superior quality with less headaches. You can throw 5000 names with website url at this pipeline and you’ll automatically have 5000 personalized emails ready to fire in 30 minutes. How amazing is that? Beyond being extremely accessible to developers who are already proficient in AI, cold email's got 3 superpowers that no other distribution channels can offer. Superpower 1/3 : You start a conversation with every single user. Every. Single. User. Let that sink in. This is incredibly powerful in the early stages, as it helps you establish rapport, bounce ideas off one another, offer 1:1 support, understand their needs, build personal relationships, and ultimately convert users into long-term fans of your product. From talking to 1000 users at the early stage, I had 20 users asking me to get on a call every week. If they are ready to buy, I do a sales call. If they are not sure, I do a user research call. At one point I even had to limit the number of calls I took to avoid burnout. The depth of the understanding of my customers’ needs is unparalleled. Using this insight, I refined the product to precisely cater to their requirements. Superpower 2/3 : You choose exactly who you talk to Unlike other distribution channels where you at best pick what someone's searching for, with cold emails, you have 100% control over who you talk to. Their company Job title Seniority level Number of employees Technology stack Growth rate Funding stage Product offerings Competitive landscape Social activity (Marital status - well, technically you can, but maybe not this one…) You can dial in this targeting to match your ICP exactly. The result is super low CAC and ultra high conversion rate. For example, My competitors are paying $10 per click for the keyword "HARO agency". I pay $0.19 per email sent, and $1.92 per signup At around $500 LTV, you can see how the first means a non-viable business. And the second means a cash-generating engine. Superpower 3/3 : Complete stealth mode Unlike other channels where competitors can easily reverse engineer or even abuse your marketing strategies, cold email operates in complete stealth mode. Every aspect is concealed from end to end: Your target audience Lead generation methods Number of leads targeted Email content Sales funnel This secrecy explains why there isn't much discussion about it online. Everyone is too focused on keeping their strategies close and reaping the rewards. That's precisely why I've chosen to share my insights on leveraging cold email to grow a successful SaaS business. More founders need to harness this channel to its fullest potential. In addition, I've more or less reached every user within my Total Addressable Market (TAM). So, if any competitor is reading this, don't bother trying to replicate it. The majority of potential users for this AI product are already onboard. To recap, the three superpowers of cold emails: You start a conversation with every single user → Accelerate to PMF You choose exactly who you talk to → Super-low CAC Complete stealth mode → Doesn’t attract competition By combining the three superpowers I helped my SaaS reach product-marketing-fit quickly and scale it to $6k per month while staying fully bootstrapped. I don't believe this was a coincidence. It's a replicable strategy for any startup. The blueprint is actually straightforward: Engage with a handful of customers Validate the idea Engage with numerous customers Scale to $5k/mo and beyond More early-stage founders should leverage cold emails for validation, and as their first distribution channel. And what would it do for you? Update: lots of DM asking about more specifics so I wrote about it here. https://coldstartblueprint.com/p/ai-agent-email-list-building

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

How I went from $27 to $3K as a solopreneur still in a 9-5
reddit
LLM Vibe Score0
Human Vibe Score1
jottrledThis week

How I went from $27 to $3K as a solopreneur still in a 9-5

My journey started back in November 2023. I was scrolling through Twitter and YouTube and saw a word that I had never come across before. Solopreneur. The word caught my eye. Mainly because I was pretty sure I knew what it meant even though it's not a word you'll find in the dictionary. I liked what it was describing. A solo entrepreneur. A one man business. It completely resonated with me. As a software engineer by trade I'm used to working alone, especially since the pandemic hit and we were forced to work remotely. See, I always wanted to ditch the 9-5 thing but thought that was too big and too scary for a single person to do. Surely you would need a lot of money to get started, right? Surely you would need investors? The whole concept seemed impossible to me. That was until I found all the success stories. I became obsessed with the concept of solopreneurship. As I went further down the rabbit hole I found people like Justin Welsh, Kieran Drew and Marc Louvion to name a few. All of whom have one person businesses making huge money every year. So I thought, if they can do it, why can't I? People like this have cleared the pathway for those looking to escape the 9-5 grind. I decided 2024 would be the year I try this out. My main goal for the year? Build a one man business, earn my first $ online and learn a sh\*t ton along the way. My main goal in general? Build my business to $100K per year, quit my 9-5 and live with freedom. From December 2023 to February 2024 I began brainstorming ideas. I was like a lost puppy looking for his ball. How on earth did people find good ideas? I began writing everything and anything that came to mind down in my notes app on my phone. By February I would have approximately 70 ideas. Each as weird and whacky as the other. I was skeptical though. If I went through all the trouble of building a product for one of these ideas how would I know if anyone would even be interested in using it? I got scared and took a break for a week. All these ideas seemed too big and the chance that they would take off into the atmosphere was slim (in my mind anyways). I was learning more and more about solopreneurship as the weeks went on so I decided to build a product centered around everything I was learning about. The idea was simple. Enter a business idea and use AI to give the user details about how to market it, who their target customers were, what to write on their landing page, etc. All for a measly $27 per use. I quickly built it and launched on March 3rd 2024. I posted about it on Indie Hackers, Reddit and Hacker News. I was so excited about the prospect of earning my first internet $! Surely everyone wanted to use my product! Nope...all I got was crickets. I was quickly brought back down to earth. That was until 5 days later. I looked at my phone and had a new Stripe notification! Cha-ching! My first internet $. What a feeling! That was goal number 1 complete. It would be another 6 days before I would get my second sale...and then another 15 days to get my third. It was an emotional rollercoaster. I went from feeling like quitting the 9-5 was actually possible to thinking that maybe the ups and downs aren't worth it. On one hand I had made my first internet dollar so I should my ecstatic, and don't get me wrong, I was but I wanted more. More validation that I could do this long term. By May I was starting to give up on the product. I had learned so much in the past few months about marketing, SEO, building an audience, etc. and I wanted to build something that I thought could have more success so I focused on one critical thing that I had learned about. What was it? Building a product that had SEO potential. A product that I knew hundreds of people were looking for. See this was my thinking - If I could find a keyword that people were searching for on Google hundreds/thousands of times every month and it was easy to rank high on search engines then I would go all in (in SEO land this equates to a Keyword that has a Keyword Difficulty of = 500). I began researching and found that the keyword "micro saas ideas" was being searched for around 600 times each month. Micro Saas was something that really interested me. It was perfect for solopreneurs. Small software products that 1 person could build. What's not to like if you're in the game of software and solopreneurship? Researching keywords like this became like a game for me. I was hooked. I was doing it every day, finding gems that were being searched for hundreds and thousands of times every month that still had potential. That's when I came up with my next product idea. I decided to create a database of Micro Saas Ideas all with this sort of SEO potential. See if you can build a product that you know people are looking for then that's all the validation you need. So I put this theory to the test. I created a database of Micro Saas Ideas with SEO Potential and launched it in June 2024. This time it was different. I made $700 in the first week of launching. A large contrast to my previous failed attempt at becoming the worlds greatest solopreneur. Since launch I have grown the product to $3K and I couldn't be happier. I know what you're saying, $3K isn't a lot. But it's validation. It's validation that I can earn $ online. Validation that I can grow a business and it gives me hope that one day I'll be able to quit that 9-5 grind. My plan is to keep growing the business. I expect there to be a few challenges up ahead but I'll tackle them as I go and learn from the failures and successes. I have a newsletter where I share Micro Saas Ideas with SEO potential every week which I'll leave below in the first comment. Feel free to come along for the ride. If not I hope this post brings you some value If you're thinking about starting as a solopreneur, stop thinking and start doing, you won't regret it.

I’m building a “DesignPickle” for all things Funnels. Would love your feedback...
reddit
LLM Vibe Score0
Human Vibe Score0.846
Gluteous_MaximusThis week

I’m building a “DesignPickle” for all things Funnels. Would love your feedback...

Hey Entrepreneurs, Early next year I’m rolling out a productized service business along the lines of Design Pickle, but instead of design assets, we create on-demand marketing assets: Things like landing pages, lead magnets, email campaigns, etc. This is NOT an agency with client engagements, etc.  It is an on-demand, menu-item style fulfillment platform where we do a few predefined things really, really well, and as much as possible try to reduce the complexity (and required customer inputs) so that creating your next killer Funnel is as easy as ordering dinner on Skip the Dishes. Below I’ve laid out our current thinking (we’re still distilling this into a deck), just so you have the full context.  And at the end, I pose 5 feedback questions. So if this “deck” seems interesting to you, then I’d love to get your feedback at the end 🙂 Thanks! And here goes... \--- The current elevator pitch:  We will research your business, your market and your competitors to develop a killer Lead Magnet, Landing Page, Ad Creatives and a 30-Day Email Drip campaign designed to turn your traffic into a rabid, lifelong buyer tribe (that you can email for years... like having your own, on-demand cash printer).  The overall thesis:  While AI is getting continually better at creating things like one-off graphics, article content, and so on - we do not think it can deeply understand market psychology, what keeps your customers up at night, or the underlying emotions that drive purchase decisions at the individual level, for your specific offer(s). Moreover, it’s also this psychological aspect of marketing where most businesses simply do not have the talent, resources or frankly the experience to create high-performing funnels themselves, regardless of how much "automation" they might have at their fingertips. And that’s because this is where you need to know who your customer really is, and what they’re actually buying (hint: not your features). Few marketers focus on these fundamentals, let alone understand the selling process. This is also why tools like ClickFunnels, HighLevel, LeadPages, etc. while very helpful, can only help with the logistics of selling. It’s still on each business to figure out how to actually tell their story, capture demand, and sell effectively. This is why a productized service that nails market research, competitor analysis & world-class copywriting that can actually turn cold traffic into lifelong customers is going to be a no-brainer for a business that’s currently struggling to actually get a steady flow of online sales. This is not something we see AI replacing effectively, any time soon. Current gaps & unknowns:  At a top level, I’m not overly worried about validation or viability; there are several existing competitors, and obviously the automation platforms have substantial customer bases (ClickFunnels etc). There will be a certain cohort that will want experts to do the actual thinking for them, storytelling, etc. Even if it’s a relatively small cohort, given the CLTV of a service like this, it still makes for a decent sized business. But where I’m less confident is in who our ideal customer actually is... Yes, basically every direct-response internet business needs an effective funnel that can sell. Whether you’re an Enterprise SaaS platform or a solopreneur launching your first $39 ebook, you will benefit from a killer funnel. As a “DesignPickle” type service though, here’s the challenges I see with each core customer category... B2B SaaS: While sales decisions are still emotional, it’s more about account-based considerations; people usually aren’t spending their own money, so it’s more about not looking stupid vs. gaining some benefit. Harder to systemize. Very high stakes. Consumer / SMB SaaS: While I think in general these are ideal customers, there will be resistance to leaning in hard on personality (and personal brand); founders usually want to sell at some point, so if they become the face of the platform, then boosting performance with a high-personality funnel might ironically make it a harder business to sell. SaaS founders are also generally very technical and stereotypically avoid marketing like the plague. Ecommerce: Most DTC brands think of funnels as an extension of their FB ad campaigns; few see their customers as a long-term audience that can become a significant asset. However, certain lifestyle / luxury brands might differ. Online Courses / Coaches: Of all the customer profiles, this group probably has the most appreciation for the effectiveness of marketing psychology, copywriting, etc. and would get the value prop quickly. The problem is that most won’t have the budget or traction to outsource asset creation. This is the “poorest” segment of the market. Service Businesses: Agencies, consultancies, and so on would greatly benefit from having a strong personal brand + storytelling premise (funnel). However, they’re also the worst offenders when it comes to never practicing what they preach / do for others. Client work soaks up all their resources. Local & Brick/Mortar: Generally speaking most local businesses are going to have smaller audiences (email lists under 2K subs), where funnel ops might have limited value long-term due to a lack of scale. And for larger B&M brands with franchises across various locations, you get into stakeholder friction; messaging usually gets watered down to basic corporate-speak as a result. Now, to be clear, I still see a ton of opportunity in each of those main customer categories as well, but I like to be clear-eyed about the overall resistance each niche will have - mainly because this helps to refine messaging to an ideal customer profile within them. In this case though, so far, nothing’s really jumping out at me as a clear “winner” at a category level. So far, what I’m thinking is our ICP might be situational / conditional. For example: A business has a funnel / is invested in the process, but it’s not working yet A business sees their competitor killing it with a funnel, and they’re ultra motivated to do it even better A business has one funnel that’s working awesome, and everything else they try sucks (so they can’t scale / expand) Etc. Basically, our most ideal customer might be ANY type of business who gets it, who’s tried to do this themselves, and now needs the pros to come in and fix things. \--- This is where your feedback would be incredibly valuable... First, if you’ve made it all the way down to this point - thanks for enduring my rambling mess above! But I did think the context might be helpful. Based on our overall biz plan & go-to-market considerations discussed above, if you run a business (or work with one) that might benefit from something like this, I’d love to ask a few questions... What is the nature of your business? (What do you sell)? What do you find hardest about selling to your online audience? Have you built a funnel in the past / are you running one currently? If not, what’s stopping you from building a high-performing funnel? If you had a “magic marketing lamp” where a genie could create ONE amazing marketing asset for you (eg. a killer landing page, video ad, launch strategy, etc), but you could only use it ONCE, what would you have the genie do for you? Please reply below as a comment, or DM me if you’d prefer to keep answers anonymous.  Thanks so much And again, apologies for the novel... Cheers

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

Steep Learning : How I Mapped approximately 10K AI tools to 15K  Replaceable Tasks across 4K professions
reddit
LLM Vibe Score0
Human Vibe Score1
Apprehensive_Form396This week

Steep Learning : How I Mapped approximately 10K AI tools to 15K Replaceable Tasks across 4K professions

Hello Everyone , I would like to share some knowledge today which I went towards countless hours to do . I founded a portal called Seekme.ai, a comprehensive platform that houses over 10,000 AI tools and resources. Today, I'm excited to share with you an insightful and enlightening journey of how I mapped these tools to 15,000 tasks across 4,000 professions. This process, which I've named "Learn by Doing," got me the power of determination, collaboration, and adaptability. The Idea: It all started when I recognized the need for a more efficient and accessible way for professionals to understand which AI tools could help them automate their tasks. The traditional approach of manually researching and testing each AI tool for every profession was time-consuming and inefficient. I envisioned a solution that could streamline this process, making AI adoption easier and more accessible for a broader audience. The Planning: To begin, we needed a clear understanding of the task landscape across various professions. With the help of some Reddit communities , we embarked on an extensive study of common tasks in various industries. We utilized various sources, including government reports, industry surveys, and academic research, to create a comprehensive list of tasks. The result was an impressive list of 15,000 tasks. The Mapping: With the list of tasks in hand, the next step was to identify which AI tools could perform these tasks. I meticulously researched and analyzed each AI tool's capabilities and features. We cross-referenced this information with the tasks I had identified and created a mapping between the two. The process involved a significant amount of collaboration and refinement, as we continually updated and expanded our database of AI tools and tasks. The Challenges: The mapping process was not without its challenges. One of the primary obstacles was ensuring the accuracy and completeness of our data. To address this issue, I implemented a rigorous quality control process that included multiple rounds of checks and validations.I also established partnerships with industry experts and AI vendors to ensure our data was up-to-date and accurate. There is also a challenge that I faced was what is the quality of the tools which is the problem and how do I rank multiple tools if they do the same tasks without user feedback The Results: After months of hard work and dedication, I successfully mapped 10,000 AI tools to 15,000 tasks across 4,000 professions. Our new feature, AI by Profession, was born. This innovative will allow users to quickly and easily identify the AI tools that can automate tasks in their profession, making AI adoption more accessible and efficient than ever before. The Impact: The impact of this project has been significant. By making it easier for professionals to identify AI tools that can automate tasks in their industry, we're helping to drive productivity, efficiency, and innovation. Our users are saving time and resources by not having to manually research and test AI tools. Furthermore, we're contributing to the broader goal of democratizing AI and making it accessible to a broader audience. But there is a still an issue we face of ranking tools who does the similar job. For instance for content creation there 10 tools that can do same video editing so how do we rank it . We are planning to add categories to this to make it more exhaustive Conclusion: The journey to mapping 10,000 AI tools for 15,000 tasks across 4,000 professions was a challenging and rewarding experience. It required a significant amount of planning, determination, and collaboration, but the end result was a powerful tool that's making a difference in the lives of professionals around the world. I don’t know yet how useful it is yet for users So I am inviting you all to see if this feature can help you better equip yourself on the new wave and do things better. I am always up for a chat on anything AI and provide my help if needed. Looking forward to some feedback aswell

Looking for a co-founder for a B2B AI startup. I have a development team and funds for at least a year of operations.
reddit
LLM Vibe Score0
Human Vibe Score0.5
cheech123456This week

Looking for a co-founder for a B2B AI startup. I have a development team and funds for at least a year of operations.

Hello, As the title said I'm looking for a co-founder. I built with my team a few ventures that generate revenues but I don't believe that any of them has a future. I have 15 years of experience in Software Engineering and AI. Worked in various industries, but always in data-driven applications. I spent the last 3 years as an entrepreneur and raised successfully money from VCs. &#x200B; A few preconceptions I have: \- B2C is extremely hard. Very quickly you realize that you need to spend all your resources on marketing. \- B2B is extremely hard - but for different reasons. Sales cycles take months. If you want to reach serious buyers and decision-makers, you need to have an amazing network. Even then, companies will prioritize 90% of the time to do things internally rather than paying for anything. \- I hate when people say that "ideas are garbage", and I think that execution is overhyped. Execution is a matter of finding the right people, and paying them (I am confident to say that I can guarantee good execution). Ideas are not garbage, ideas need validation, and garbage "entrepreneurs" are too lazy to validate anything. &#x200B; Your ideal profile: \- You have a great idea, something that has been brewing for some time but you lack resources or technical experience to execute by yourself. \- You have domain expertise, experience, and a network. If we build an MVP in 3 months, you can get 20 interviews with industry people to validate the solution. Once the MVP is built you can put it in front of another 40 people. \- You are a product person. \- You can do efficient sales calls. (Bonus: You are a sales person) If you are an ideal profile, please reach out.

Founder Pitch: AI Agent for Simplifying Public Cloud Management
reddit
LLM Vibe Score0
Human Vibe Score1
rasvi786This week

Founder Pitch: AI Agent for Simplifying Public Cloud Management

Video to understand : https://youtu.be/9ocUjlUrU\w?si=S0ETDbKSdJqlVDyg Are You Ready to Redefine Cloud Management with AI? Imagine an intelligent AI agent that transforms the complexity of managing public cloud infrastructure into simple, natural language commands. No more navigating through endless configurations or deciphering technical documentation—our AI agent is here to revolutionize the way organizations interact with cloud platforms. About the Project We’re building an AI-powered agent designed to handle public cloud management tasks seamlessly. Whether you’re setting up your organization’s cloud foundation or deploying complex workloads, this AI agent makes it as easy as having a conversation. What Can the AI Agent Do? Cloud Foundation Setup: Example: “Please set up a cloud foundation blueprint for my organization on Google Cloud.”* The AI agent will ask key questions (e.g., organization ID) and guide you through authentication. Once authorized, it sets up the foundation using GCP APIs. Workload Deployment: Example: “Spin up a GKE cluster for me.”* The agent will ask for necessary details (e.g., number of nodes, VPC info), authenticate, and deploy the cluster in minutes. Security and Compliance Validation: Example: “Validate my organization’s cloud setup and check for security vulnerabilities.”* The agent audits your setup, identifies potential risks, and provides actionable insights. Current Progress We’ve developed a working prototype that integrates with major cloud providers like Google Cloud. The AI agent can already: Authenticate with cloud APIs Execute foundational tasks such as setting up organizations and spinning up clusters Perform initial security validations Who I’m Looking For I’m searching for a co-founder with enterprise sales experience and a strategic vision to grow our user base. You will be instrumental in helping us: Build relationships with companies willing to pilot our product Develop go-to-market strategies for enterprise adoption Identify opportunities for partnerships with cloud service providers Your Role As a co-founder, you’ll lead efforts to: Secure Pilot Programs: Identify and onboard enterprises for product trials to gather feedback and refine the solution. Drive Growth: Develop scalable strategies to grow our user base across industries. Market Positioning: Work with me to define our unique value proposition and establish thought leadership in the cloud management space. My Background I bring over a decade of experience in tech, with a strong focus on software engineering and infrastructure. My contributions so far include: Developing the core AI engine and cloud integrations Designing workflows that simplify complex cloud tasks Why Join This Project? Revolutionize Cloud Management: Be part of a project that will redefine how organizations interact with public clouds. Tackle Challenging Problems: Work at the cutting edge of AI and cloud computing. High Growth Potential: Join an industry projected to grow exponentially as enterprises embrace AI-driven automation. Build a Company from Scratch: Shape the product, team, and culture as we grow together. What’s Next? Our immediate priorities include: Expanding the AI agent’s capabilities to support multi-cloud setups. Conducting pilot programs with enterprise clients. Iterating on the product based on real-world feedback. What We Need to Succeed Expertise in enterprise sales and partnerships A deep understanding of enterprise challenges and cloud adoption trends A shared passion for leveraging AI to solve complex problems Let’s work together to build the future of cloud management. If you’re excited about this vision and bring the expertise we need, I’d love to connect and discuss how we can take this project to the next level.

Steep Learning : How I Mapped approximately 10K AI tools to 15K  Replaceable Tasks across 4K professions
reddit
LLM Vibe Score0
Human Vibe Score1
Apprehensive_Form396This week

Steep Learning : How I Mapped approximately 10K AI tools to 15K Replaceable Tasks across 4K professions

Hello Everyone , I would like to share some knowledge today which I went towards countless hours to do . I founded a portal called Seekme.ai, a comprehensive platform that houses over 10,000 AI tools and resources. Today, I'm excited to share with you an insightful and enlightening journey of how I mapped these tools to 15,000 tasks across 4,000 professions. This process, which I've named "Learn by Doing," got me the power of determination, collaboration, and adaptability. The Idea: It all started when I recognized the need for a more efficient and accessible way for professionals to understand which AI tools could help them automate their tasks. The traditional approach of manually researching and testing each AI tool for every profession was time-consuming and inefficient. I envisioned a solution that could streamline this process, making AI adoption easier and more accessible for a broader audience. The Planning: To begin, we needed a clear understanding of the task landscape across various professions. With the help of some Reddit communities , we embarked on an extensive study of common tasks in various industries. We utilized various sources, including government reports, industry surveys, and academic research, to create a comprehensive list of tasks. The result was an impressive list of 15,000 tasks. The Mapping: With the list of tasks in hand, the next step was to identify which AI tools could perform these tasks. I meticulously researched and analyzed each AI tool's capabilities and features. We cross-referenced this information with the tasks I had identified and created a mapping between the two. The process involved a significant amount of collaboration and refinement, as we continually updated and expanded our database of AI tools and tasks. The Challenges: The mapping process was not without its challenges. One of the primary obstacles was ensuring the accuracy and completeness of our data. To address this issue, I implemented a rigorous quality control process that included multiple rounds of checks and validations.I also established partnerships with industry experts and AI vendors to ensure our data was up-to-date and accurate. There is also a challenge that I faced was what is the quality of the tools which is the problem and how do I rank multiple tools if they do the same tasks without user feedback The Results: After months of hard work and dedication, I successfully mapped 10,000 AI tools to 15,000 tasks across 4,000 professions. Our new feature, AI by Profession, was born. This innovative will allow users to quickly and easily identify the AI tools that can automate tasks in their profession, making AI adoption more accessible and efficient than ever before. The Impact: The impact of this project has been significant. By making it easier for professionals to identify AI tools that can automate tasks in their industry, we're helping to drive productivity, efficiency, and innovation. Our users are saving time and resources by not having to manually research and test AI tools. Furthermore, we're contributing to the broader goal of democratizing AI and making it accessible to a broader audience. But there is a still an issue we face of ranking tools who does the similar job. For instance for content creation there 10 tools that can do same video editing so how do we rank it . We are planning to add categories to this to make it more exhaustive Conclusion: The journey to mapping 10,000 AI tools for 15,000 tasks across 4,000 professions was a challenging and rewarding experience. It required a significant amount of planning, determination, and collaboration, but the end result was a powerful tool that's making a difference in the lives of professionals around the world. I don’t know yet how useful it is yet for users So I am inviting you all to see if this feature can help you better equip yourself on the new wave and do things better. I am always up for a chat on anything AI and provide my help if needed. Looking forward to some feedback aswell

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn
youtube
LLM Vibe Score0.369
Human Vibe Score0.26
SimplilearnMar 28, 2025

ChatGPT Full Course For 2025 | ChatGPT Tutorial For Beginnners | ChatGPT Course | Simplilearn

🔥Purdue - Applied Generative AI Specialization - https://www.simplilearn.com/applied-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Professional Certificate Program in Generative AI and Machine Learning - IITG (India Only) - https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube 🔥Advanced Executive Program In Applied Generative AI - https://www.simplilearn.com/applied-generative-ai-course?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube This ChatGPT Full Course 2025 by Simplilearn provides a comprehensive learning journey, starting with an introduction to ChatGPT and Generative AI, followed by insights into AI job opportunities and a comparison between ChatGPT 4.0 and 4.0 Turbo. The tutorial covers prompt engineering techniques, machine learning fundamentals, and running Llama models privately. Learners will explore ChatGPT-powered application development, its role in programming, and Excel automation. The course also dives into blogging, PowerPoint automation, customer support, and finance applications. Advanced topics like RAG vs. Prompt Tuning, prompt injection, and LangChain are included, along with discussions on OpenAI's latest innovations, including Sora and Strawberry. By the end, participants will gain a strong understanding of ChatGPT’s capabilities and monetization strategies. 🚀 Following are the topics covered in the ChatGPT Full Course 2025: 00:00:00 - Introduction to ChatGPT Full Course 2025 00:09:26 - What is ChatGPT 00:10:11 - What is Gen AI 00:26:29 - How to get Job in AI 00:27:06 - ChatGPT 40 vs ChatGPT 4 01:03:14 - Chatgpt analyse 02:13:57 - Prompt Engineering Tutorial 03:10:34 - What is Machine Learning 04:07:06 - Machine Learning Tutorial 04:08:13 - Run Lama Privately 04:23:50 - Search GPT 04:25:31 - Build App Using ChatGPT 06:31:11 - ChatGPT for Programming 06:46:08 - Prompt Formulae Chatgpt 07:58:38 - Automate Excel using Chatgpt 08:00:06 - Blogging with ChatGpt 08:27:25 - Powerpoint using Chatgpt 08:28:31 - Rag Vs Prompt Tuning 09:37:43 - Chatgpt for Customer Support 11:11:06 - ChatGPT for finance 11:17:38 - Prompt injection 11:18:38 - How to Earn Money using ChatGPT 11:41:46 - Open AI Strawberry 11:52:42 - Openai sora 11:54:57 - Langchain 12:22:19 - Open ai chatgpt o1 model ✅ Subscribe to our Channel to learn more about the top Technologies: https://bit.ly/2VT4WtH ⏩ Check out the Artificial Intelligence training videos: https://youtube.com/playlist?list=PLEiEAq2VkUULa5aOQmO_al2VVmhC-eqeI #gpt #chatgpt #chatgptforbeginners #chatgptcourse #genai #generativeai #artificialintelligence #ai #machinelearning #llm #simplilearn #2025 ➡️ About Professional Certificate Program in Generative AI and Machine Learning Dive into the future of AI with our Generative AI & Machine Learning course, in collaboration with E&ICT Academy, IIT Guwahati. Learn tools like ChatGPT, OpenAI, Hugging Face, Python, and more. Join masterclasses led by IITG faculty, engage in hands-on projects, and earn Executive Alumni Status. Key Features: ✅ Program completion certificate from E&ICT Academy, IIT Guwahati ✅ Curriculum delivered in live virtual classes by seasoned industry experts ✅ Exposure to the latest AI advancements, such as generative AI, LLMs, and prompt engineering ✅ Interactive live-virtual masterclasses delivered by esteemed IIT Guwahati faculty ✅ Opportunity to earn an 'Executive Alumni Status' from E&ICT Academy, IIT Guwahati ✅ Eligibility for a campus immersion program organized at IIT Guwahati ✅ Exclusive hackathons and “ask-me-anything” sessions by IBM ✅ Certificates for IBM courses and industry masterclasses by IBM experts ✅ Practical learning through 25+ hands-on projects and 3 industry-oriented capstone projects ✅ Access to a wide array of AI tools such as ChatGPT, Hugging Face, DALL-E 2, Midjourney and more ✅ Simplilearn's JobAssist helps you get noticed by top hiring companies Skills Covered: ✅ Generative AI ✅ Prompt Engineering ✅ Chatbot Development ✅ Supervised and Unsupervised Learning ✅ Model Training and Optimization ✅ Model Evaluation and Validation ✅ Ensemble Methods ✅ Deep Learning ✅ Natural Language Processing ✅ Computer Vision ✅ Reinforcement Learning ✅ Machine Learning Algorithms ✅ Speech Recognition ✅ Statistics Learning Path: ✅ Program Induction ✅ Programming Fundamentals ✅ Python for Data Science (IBM) ✅ Applied Data Science with Python ✅ Machine Learning ✅ Deep Learning with TensorFlow (IBM) ✅ Deep Learning Specialization ✅ Essentials of Generative AI, Prompt Engineering & ChatGPT ✅ Advanced Generative AI ✅ Capstone Electives: ✅ ADL & Computer Vision ✅ NLP and Speech Recognition ✅ Reinforcement Learning ✅ Academic Masterclass ✅ Industry Masterclass 👉 Learn More At: https://www.simplilearn.com/iitg-generative-ai-machine-learning-program?utmcampaign=C4lBsBlloL0&utmmedium=Lives&utm_source=Youtube

GenAI_Agents
github
LLM Vibe Score0.563
Human Vibe Score0.24210481455988786
NirDiamantMar 28, 2025

GenAI_Agents

🌟 Support This Project: Your sponsorship fuels innovation in GenAI agent development. Become a sponsor to help maintain and expand this valuable resource! GenAI Agents: Comprehensive Repository for Development and Implementation 🚀 Welcome to one of the most extensive and dynamic collections of Generative AI (GenAI) agent tutorials and implementations available today. This repository serves as a comprehensive resource for learning, building, and sharing GenAI agents, ranging from simple conversational bots to complex, multi-agent systems. 📫 Stay Updated! 🚀Cutting-edgeUpdates 💡ExpertInsights 🎯Top 0.1%Content Join over 15,000 of AI enthusiasts getting unique cutting-edge insights and free tutorials! Plus, subscribers get exclusive early access and special 33% discounts to my book and the upcoming RAG Techniques course! Introduction Generative AI agents are at the forefront of artificial intelligence, revolutionizing the way we interact with and leverage AI technologies. This repository is designed to guide you through the development journey, from basic agent implementations to advanced, cutting-edge systems. 📚 Learn to Build Your First AI Agent Your First AI Agent: Simpler Than You Think This detailed blog post complements the repository by providing a complete A-Z walkthrough with in-depth explanations of core concepts, step-by-step implementation, and the theory behind AI agents. It's designed to be incredibly simple to follow while covering everything you need to know to build your first working agent from scratch. 💡 Plus: Subscribe to the newsletter for exclusive early access to tutorials and special discounts on upcoming courses and books! Our goal is to provide a valuable resource for everyone - from beginners taking their first steps in AI to seasoned practitioners pushing the boundaries of what's possible. By offering a range of examples from foundational to complex, we aim to facilitate learning, experimentation, and innovation in the rapidly evolving field of GenAI agents. Furthermore, this repository serves as a platform for showcasing innovative agent creations. Whether you've developed a novel agent architecture or found an innovative application for existing techniques, we encourage you to share your work with the community. Related Projects 📚 Dive into my comprehensive guide on RAG techniques to learn about integrating external knowledge into AI systems, enhancing their capabilities with up-to-date and relevant information retrieval. 🖋️ Explore my Prompt Engineering Techniques guide for an extensive collection of prompting strategies, from fundamental concepts to advanced methods, improving your ability to communicate effectively with AI language models. A Community-Driven Knowledge Hub This repository grows stronger with your contributions! Join our vibrant Discord community — the central hub for shaping and advancing this project together 🤝 GenAI Agents Discord Community Whether you're a novice eager to learn or an expert ready to share your knowledge, your insights can shape the future of GenAI agents. Join us to propose ideas, get feedback, and collaborate on innovative implementations. For contribution guidelines, please refer to our CONTRIBUTING.md file. Let's advance GenAI agent technology together! 🔗 For discussions on GenAI, agents, or to explore knowledge-sharing opportunities, feel free to connect on LinkedIn. Key Features 🎓 Learn to build GenAI agents from beginner to advanced levels 🧠 Explore a wide range of agent architectures and applications 📚 Step-by-step tutorials and comprehensive documentation 🛠️ Practical, ready-to-use agent implementations 🌟 Regular updates with the latest advancements in GenAI 🤝 Share your own agent creations with the community GenAI Agent Implementations Explore our extensive list of GenAI agent implementations, sorted by categories: 🌱 Beginner-Friendly Agents Simple Conversational Agent LangChain PydanticAI Overview 🔎 A context-aware conversational AI maintains information across interactions, enabling more natural dialogues. Implementation 🛠️ Integrates a language model, prompt template, and history manager to generate contextual responses and track conversation sessions. Simple Question Answering Agent Overview 🔎 Answering (QA) agent using LangChain and OpenAI's language model understands user queries and provides relevant, concise answers. Implementation 🛠️ Combines OpenAI's GPT model, a prompt template, and an LLMChain to process user questions and generate AI-driven responses in a streamlined manner. Simple Data Analysis Agent LangChain PydanticAI Overview 🔎 An AI-powered data analysis agent interprets and answers questions about datasets using natural language, combining language models with data manipulation tools for intuitive data exploration. Implementation 🛠️ Integrates a language model, data manipulation framework, and agent framework to process natural language queries and perform data analysis on a synthetic dataset, enabling accessible insights for non-technical users. 🔧 Framework Tutorial: LangGraph Introduction to LangGraph: Building Modular AI Workflows Overview 🔎 This tutorial introduces LangGraph, a powerful framework for creating modular, graph-based AI workflows. Learn how to leverage LangGraph to build more complex and flexible AI agents that can handle multi-step processes efficiently. Implementation 🛠️ Step-by-step guide on using LangGraph to create a StateGraph workflow. The tutorial covers key concepts such as state management, node creation, and graph compilation. It demonstrates these principles by constructing a simple text analysis pipeline, serving as a foundation for more advanced agent architectures. Additional Resources 📚 Blog Post 🎓 Educational and Research Agents ATLAS: Academic Task and Learning Agent System Overview 🔎 ATLAS demonstrates how to build an intelligent multi-agent system that transforms academic support through AI-powered assistance. The system leverages LangGraph's workflow framework to coordinate multiple specialized agents that provide personalized academic planning, note-taking, and advisory support. Implementation 🛠️ Implements a state-managed multi-agent architecture using four specialized agents (Coordinator, Planner, Notewriter, and Advisor) working in concert through LangGraph's workflow framework. The system features sophisticated workflows for profile analysis and academic support, with continuous adaptation based on student performance and feedback. Additional Resources 📚 YouTube Explanation Blog Post Scientific Paper Agent - Literature Review Overview 🔎 An intelligent research assistant that helps users navigate, understand, and analyze scientific literature through an orchestrated workflow. The system combines academic APIs with sophisticated paper processing techniques to automate literature review tasks, enabling researchers to efficiently extract insights from academic papers while maintaining research rigor and quality control. Implementation 🛠️ Leverages LangGraph to create a five-node workflow system including decision making, planning, tool execution, and quality validation nodes. The system integrates the CORE API for paper access, PDFplumber for document processing, and advanced language models for analysis. Key features include a retry mechanism for robust paper downloads, structured data handling through Pydantic models, and quality-focused improvement cycles with human-in-the-loop validation options. Additional Resources 📚 YouTube Explanation Blog Post Chiron - A Feynman-Enhanced Learning Agent Overview 🔎 An adaptive learning agent that guides users through educational content using a structured checkpoint system and Feynman-style teaching. The system processes learning materials (either user-provided or web-retrieved), verifies understanding through interactive checkpoints, and provides simplified explanations when needed, creating a personalized learning experience that mimics one-on-one tutoring. Implementation 🛠️ Uses LangGraph to orchestrate a learning workflow that includes checkpoint definition, context building, understanding verification, and Feynman teaching nodes. The system integrates web search for dynamic content retrieval, employs semantic chunking for context processing, and manages embeddings for relevant information retrieval. Key features include a 70% understanding threshold for progression, interactive human-in-the-loop validation, and structured output through Pydantic models for consistent data handling. Additional Resources 📚 YouTube Explanation 💼 Business and Professional Agents Customer Support Agent (LangGraph) Overview 🔎 An intelligent customer support agent using LangGraph categorizes queries, analyzes sentiment, and provides appropriate responses or escalates issues. Implementation 🛠️ Utilizes LangGraph to create a workflow combining state management, query categorization, sentiment analysis, and response generation. Essay Grading Agent (LangGraph) Overview 🔎 An automated essay grading system using LangGraph and an LLM model evaluates essays based on relevance, grammar, structure, and depth of analysis. Implementation 🛠️ Utilizes a state graph to define the grading workflow, incorporating separate grading functions for each criterion. Travel Planning Agent (LangGraph) Overview 🔎 A Travel Planner using LangGraph demonstrates how to build a stateful, multi-step conversational AI application that collects user input and generates personalized travel itineraries. Implementation 🛠️ Utilizes StateGraph to define the application flow, incorporates custom PlannerState for process management. GenAI Career Assistant Agent Overview 🔎 The GenAI Career Assistant demonstrates how to create a multi-agent system that provides personalized guidance for careers in Generative AI. Using LangGraph and Gemini LLM, the system delivers customized learning paths, resume assistance, interview preparation, and job search support. Implementation 🛠️ Leverages a multi-agent architecture using LangGraph to coordinate specialized agents (Learning, Resume, Interview, Job Search) through TypedDict-based state management. The system employs sophisticated query categorization and routing while integrating with external tools like DuckDuckGo for job searches and dynamic content generation. Additional Resources 📚 YouTube Explanation Project Manager Assistant Agent Overview 🔎 An AI agent designed to assist in project management tasks by automating the process of creating actionable tasks from project descriptions, identifying dependencies, scheduling work, and assigning tasks to team members based on expertise. The system includes risk assessment and self-reflection capabilities to optimize project plans through multiple iterations, aiming to minimize overall project risk. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized nodes including task generation, dependency mapping, scheduling, allocation, and risk assessment. Each node uses GPT-4o-mini for structured outputs following Pydantic models. The system implements a feedback loop for self-improvement, where risk scores trigger reflection cycles that generate insights to optimize the project plan. Visualization tools display Gantt charts of the generated schedules across iterations. Additional Resources 📚 YouTube Explanation Contract Analysis Assistant (ClauseAI) Overview 🔎 ClauseAI demonstrates how to build an AI-powered contract analysis system using a multi-agent approach. The system employs specialized AI agents for different aspects of contract review, from clause analysis to compliance checking, and leverages LangGraph for workflow orchestration and Pinecone for efficient clause retrieval and comparison. Implementation 🛠️ Implements a sophisticated state-based workflow using LangGraph to coordinate multiple AI agents through contract analysis stages. The system features Pydantic models for data validation, vector storage with Pinecone for clause comparison, and LLM-based analysis for generating comprehensive contract reports. The implementation includes parallel processing capabilities and customizable report generation based on user requirements. Additional Resources 📚 YouTube Explanation E2E Testing Agent Overview 🔎 The E2E Testing Agent demonstrates how to build an AI-powered system that converts natural language test instructions into executable end-to-end web tests. Using LangGraph for workflow orchestration and Playwright for browser automation, the system enables users to specify test cases in plain English while handling the complexity of test generation and execution. Implementation 🛠️ Implements a structured workflow using LangGraph to coordinate test generation, validation, and execution. The system features TypedDict state management, integration with Playwright for browser automation, and LLM-based code generation for converting natural language instructions into executable test scripts. The implementation includes DOM state analysis, error handling, and comprehensive test reporting. Additional Resources 📚 YouTube Explanation 🎨 Creative and Content Generation Agents GIF Animation Generator Agent (LangGraph) Overview 🔎 A GIF animation generator that integrates LangGraph for workflow management, GPT-4 for text generation, and DALL-E for image creation, producing custom animations from user prompts. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that generates character descriptions, plots, and image prompts using GPT-4, creates images with DALL-E 3, and assembles them into GIFs using PIL. Employs asynchronous programming for efficient parallel processing. TTS Poem Generator Agent (LangGraph) Overview 🔎 An advanced text-to-speech (TTS) agent using LangGraph and OpenAI's APIs classifies input text, processes it based on content type, and generates corresponding speech output. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow that classifies input text using GPT models, applies content-specific processing, and converts the processed text to speech using OpenAI's TTS API. The system adapts its output based on the identified content type (general, poem, news, or joke). Music Compositor Agent (LangGraph) Overview 🔎 An AI Music Compositor using LangGraph and OpenAI's language models generates custom musical compositions based on user input. The system processes the input through specialized components, each contributing to the final musical piece, which is then converted to a playable MIDI file. Implementation 🛠️ LangGraph orchestrates a workflow that transforms user input into a musical composition, using ChatOpenAI (GPT-4) to generate melody, harmony, and rhythm, which are then style-adapted. The final AI-generated composition is converted to a MIDI file using music21 and can be played back using pygame. Content Intelligence: Multi-Platform Content Generation Agent Overview 🔎 Content Intelligence demonstrates how to build an advanced content generation system that transforms input text into platform-optimized content across multiple social media channels. The system employs LangGraph for workflow orchestration to analyze content, conduct research, and generate tailored content while maintaining brand consistency across different platforms. Implementation 🛠️ Implements a sophisticated workflow using LangGraph to coordinate multiple specialized nodes (Summary, Research, Platform-Specific) through the content generation process. The system features TypedDict and Pydantic models for state management, integration with Tavily Search for research enhancement, and platform-specific content generation using GPT-4. The implementation includes parallel processing for multiple platforms and customizable content templates. Additional Resources 📚 YouTube Explanation Business Meme Generator Using LangGraph and Memegen.link Overview 🔎 The Business Meme Generator demonstrates how to create an AI-powered system that generates contextually relevant memes based on company website analysis. Using LangGraph for workflow orchestration, the system combines Groq's Llama model for text analysis and the Memegen.link API to automatically produce brand-aligned memes for digital marketing. Implementation 🛠️ Implements a state-managed workflow using LangGraph to coordinate website content analysis, meme concept generation, and image creation. The system features Pydantic models for data validation, asynchronous processing with aiohttp, and integration with external APIs (Groq, Memegen.link) to create a complete meme generation pipeline with customizable templates. Additional Resources 📚 YouTube Explanation Murder Mystery Game with LLM Agents Overview 🔎 A text-based detective game that utilizes autonomous LLM agents as interactive characters in a procedurally generated murder mystery. Drawing inspiration from the UNBOUNDED paper, the system creates unique scenarios each time, with players taking on the role of Sherlock Holmes to solve the case through character interviews and deductive reasoning. Implementation 🛠️ Leverages two LangGraph workflows - a main game loop for story/character generation and game progression, and a conversation sub-graph for character interactions. The system uses a combination of LLM-powered narrative generation, character AI, and structured game mechanics to create an immersive investigative experience with replayable storylines. Additional Resources 📚 YouTube Explanation 📊 Analysis and Information Processing Agents Memory-Enhanced Conversational Agent Overview 🔎 A memory-enhanced conversational AI agent incorporates short-term and long-term memory systems to maintain context within conversations and across multiple sessions, improving interaction quality and personalization. Implementation 🛠️ Integrates a language model with separate short-term and long-term memory stores, utilizes a prompt template incorporating both memory types, and employs a memory manager for storage and retrieval. The system includes an interaction loop that updates and utilizes memories for each response. Multi-Agent Collaboration System Overview 🔎 A multi-agent collaboration system combining historical research with data analysis, leveraging large language models to simulate specialized agents working together to answer complex historical questions. Implementation 🛠️ Utilizes a base Agent class to create specialized HistoryResearchAgent and DataAnalysisAgent, orchestrated by a HistoryDataCollaborationSystem. The system follows a five-step process: historical context provision, data needs identification, historical data provision, data analysis, and final synthesis. Self-Improving Agent Overview 🔎 A Self-Improving Agent using LangChain engages in conversations, learns from interactions, and continuously improves its performance over time through reflection and adaptation. Implementation 🛠️ Integrates a language model with chat history management, response generation, and a reflection mechanism. The system employs a learning system that incorporates insights from reflection to enhance future performance, creating a continuous improvement loop. Task-Oriented Agent Overview 🔎 A language model application using LangChain that summarizes text and translates the summary to Spanish, combining custom functions, structured tools, and an agent for efficient text processing. Implementation 🛠️ Utilizes custom functions for summarization and translation, wrapped as structured tools. Employs a prompt template to guide the agent, which orchestrates the use of tools. An agent executor manages the process, taking input text and producing both an English summary and its Spanish translation. Internet Search and Summarize Agent Overview 🔎 An intelligent web research assistant that combines web search capabilities with AI-powered summarization, automating the process of gathering information from the internet and distilling it into concise, relevant summaries. Implementation 🛠️ Integrates a web search module using DuckDuckGo's API, a result parser, and a text summarization engine leveraging OpenAI's language models. The system performs site-specific or general searches, extracts relevant content, generates concise summaries, and compiles attributed results for efficient information retrieval and synthesis. Multi agent research team - Autogen Overview 🔎 This technique explores a multi-agent system for collaborative research using the AutoGen library. It employs agents to solve tasks collaboratively, focusing on efficient execution and quality assurance. The system enhances research by distributing tasks among specialized agents. Implementation 🛠️ Agents are configured with specific roles using the GPT-4 model, including admin, developer, planner, executor, and quality assurance. Interaction management ensures orderly communication with defined transitions. Task execution involves collaborative planning, coding, execution, and quality checking, demonstrating a scalable framework for various domains. Additional Resources 📚 comprehensive solution with UI Blogpost Sales Call Analyzer Overview 🔎 An intelligent system that automates the analysis of sales call recordings by combining audio transcription with advanced natural language processing. The analyzer transcribes audio using OpenAI's Whisper, processes the text using NLP techniques, and generates comprehensive reports including sentiment analysis, key phrases, pain points, and actionable recommendations to improve sales performance. Implementation 🛠️ Utilizes multiple components in a structured workflow: OpenAI Whisper for audio transcription, CrewAI for task automation and agent management, and LangChain for orchestrating the analysis pipeline. The system processes audio through a series of steps from transcription to detailed analysis, leveraging custom agents and tasks to generate structured JSON reports containing insights about customer sentiment, sales opportunities, and recommended improvements. Additional Resources 📚 YouTube Explanation Weather Emergency & Response System Overview 🔎 A comprehensive system demonstrating two agent graph implementations for weather emergency response: a real-time graph processing live weather data, and a hybrid graph combining real and simulated data for testing high-severity scenarios. The system handles complete workflow from data gathering through emergency plan generation, with automated notifications and human verification steps. Implementation 🛠️ Utilizes LangGraph for orchestrating complex workflows with state management, integrating OpenWeatherMap API for real-time data, and Gemini for analysis and response generation. The system incorporates email notifications, social media monitoring simulation, and severity-based routing with configurable human verification for low/medium severity events. Additional Resources 📚 YouTube Explanation Self-Healing Codebase System Overview 🔎 An intelligent system that automatically detects, diagnoses, and fixes runtime code errors using LangGraph workflow orchestration and ChromaDB vector storage. The system maintains a memory of encountered bugs and their fixes through vector embeddings, enabling pattern recognition for similar errors across the codebase. Implementation 🛠️ Utilizes a state-based graph workflow that processes function definitions and runtime arguments through specialized nodes for error detection, code analysis, and fix generation. Incorporates ChromaDB for vector-based storage of bug patterns and fixes, with automated search and retrieval capabilities for similar error patterns, while maintaining code execution safety through structured validation steps. Additional Resources 📚 YouTube Explanation DataScribe: AI-Powered Schema Explorer Overview 🔎 An intelligent agent system that enables intuitive exploration and querying of relational databases through natural language interactions. The system utilizes a fleet of specialized agents, coordinated by a stateful Supervisor, to handle schema discovery, query planning, and data analysis tasks while maintaining contextual understanding through vector-based relationship graphs. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-agent workflow including discovery, inference, and planning agents, with NetworkX for relationship graph visualization and management. The system incorporates dynamic state management through TypedDict classes, maintains database context between sessions using a db_graph attribute, and includes safety measures to prevent unauthorized database modifications. Memory-Enhanced Email Agent (LangGraph & LangMem) Overview 🔎 An intelligent email assistant that combines three types of memory (semantic, episodic, and procedural) to create a system that improves over time. The agent can triage incoming emails, draft contextually appropriate responses using stored knowledge, and enhance its performance based on user feedback. Implementation 🛠️ Leverages LangGraph for workflow orchestration and LangMem for sophisticated memory management across multiple memory types. The system implements a triage workflow with memory-enhanced decision making, specialized tools for email composition and calendar management, and a self-improvement mechanism that updates its own prompts based on feedback and past performance. Additional Resources 📚 Blog Post 📰 News and Information Agents News TL;DR using LangGraph Overview 🔎 A news summarization system that generates concise TL;DR summaries of current events based on user queries. The system leverages large language models for decision making and summarization while integrating with news APIs to access up-to-date content, allowing users to quickly catch up on topics of interest through generated bullet-point summaries. Implementation 🛠️ Utilizes LangGraph to orchestrate a workflow combining multiple components: GPT-4o-mini for generating search terms and article summaries, NewsAPI for retrieving article metadata, BeautifulSoup for web scraping article content, and Asyncio for concurrent processing. The system follows a structured pipeline from query processing through article selection and summarization, managing the flow between components to produce relevant TL;DRs of current news articles. Additional Resources 📚 YouTube Explanation Blog Post AInsight: AI/ML Weekly News Reporter Overview 🔎 AInsight demonstrates how to build an intelligent news aggregation and summarization system using a multi-agent architecture. The system employs three specialized agents (NewsSearcher, Summarizer, Publisher) to automatically collect, process and summarize AI/ML news for general audiences through LangGraph-based workflow orchestration. Implementation 🛠️ Implements a state-managed multi-agent system using LangGraph to coordinate the news collection (Tavily API), technical content summarization (GPT-4), and report generation processes. The system features modular architecture with TypedDict-based state management, external API integration, and markdown report generation with customizable templates. Additional Resources 📚 YouTube Explanation Journalism-Focused AI Assistant Overview 🔎 A specialized AI assistant that helps journalists tackle modern journalistic challenges like misinformation, bias, and information overload. The system integrates fact-checking, tone analysis, summarization, and grammar review tools to enhance the accuracy and efficiency of journalistic work while maintaining ethical reporting standards. Implementation 🛠️ Leverages LangGraph to orchestrate a workflow of specialized components including language models for analysis and generation, web search integration via DuckDuckGo's API, document parsing tools like PyMuPDFLoader and WebBaseLoader, text splitting with RecursiveCharacterTextSplitter, and structured JSON outputs. Each component works together through a unified workflow to analyze content, verify facts, detect bias, extract quotes, and generate comprehensive reports. Blog Writer (Open AI Swarm) Overview 🔎 A multi-agent system for collaborative blog post creation using OpenAI's Swarm package. It leverages specialized agents to perform research, planning, writing, and editing tasks efficiently. Implementation 🛠️ Utilizes OpenAI's Swarm Package to manage agent interactions. Includes an admin, researcher, planner, writer, and editor, each with specific roles. The system follows a structured workflow: topic setting, outlining, research, drafting, and editing. This approach enhances content creation through task distribution, specialization, and collaborative problem-solving. Additional Resources 📚 Swarm Repo Podcast Internet Search and Generate Agent 🎙️ Overview 🔎 A two step agent that first searches the internet for a given topic and then generates a podcast on the topic found. The search step uses a search agent and search function to find the most relevant information. The second step uses a podcast generation agent and generation function to create a podcast on the topic found. Implementation 🛠️ Utilizes LangGraph to orchestrate a two-step workflow. The first step involves a search agent and function to gather information from the internet. The second step uses a podcast generation agent and function to create a podcast based on the gathered information. 🛍️ Shopping and Product Analysis Agents ShopGenie - Redefining Online Shopping Customer Experience Overview 🔎 An AI-powered shopping assistant that helps customers make informed purchasing decisions even without domain expertise. The system analyzes product information from multiple sources, compares specifications and reviews, identifies the best option based on user needs, and delivers recommendations through email with supporting video reviews, creating a comprehensive shopping experience. Implementation 🛠️ Uses LangGraph to orchestrate a workflow combining Tavily for web search, Llama-3.1-70B for structured data analysis and product comparison, and YouTube API for review video retrieval. The system processes search results through multiple nodes including schema mapping, product comparison, review identification, and email generation. Key features include structured Pydantic models for consistent data handling, retry mechanisms for robust API interactions, and email delivery through SMTP for sharing recommendations. Additional Resources 📚 YouTube Explanation Car Buyer AI Agent Overview 🔎 The Smart Product Buyer AI Agent demonstrates how to build an intelligent system that assists users in making informed purchasing decisions. Using LangGraph and LLM-based intelligence, the system processes user requirements, scrapes product listings from websites like AutoTrader, and provides detailed analysis and recommendations for car purchases. Implementation 🛠️ Implements a state-based workflow using LangGraph to coordinate user interaction, web scraping, and decision support. The system features TypedDict state management, async web scraping with Playwright, and integrates with external APIs for comprehensive product analysis. The implementation includes a Gradio interface for real-time chat interaction and modular scraper architecture for easy extension to additional product categories. Additional Resources 📚 YouTube Explanation 🎯 Task Management and Productivity Agents Taskifier - Intelligent Task Allocation & Management Overview 🔎 An intelligent task management system that analyzes user work styles and creates personalized task breakdown strategies, born from the observation that procrastination often stems from task ambiguity among students and early-career professionals. The system evaluates historical work patterns, gathers relevant task information through web search, and generates customized step-by-step approaches to optimize productivity and reduce workflow paralysis. Implementation 🛠️ Leverages LangGraph for orchestrating a multi-step workflow including work style analysis, information gathering via Tavily API, and customized plan generation. The system maintains state through the process, integrating historical work pattern data with fresh task research to output detailed, personalized task execution plans aligned with the user's natural working style. Additional Resources 📚 YouTube Explanation Grocery Management Agents System Overview 🔎 A multi-agent system built with CrewAI that automates grocery management tasks including receipt interpretation, expiration date tracking, inventory management, and recipe recommendations. The system uses specialized agents to extract data from receipts, estimate product shelf life, track consumption, and suggest recipes to minimize food waste. Implementation 🛠️ Implements four specialized agents using CrewAI - a Receipt Interpreter that extracts item details from receipts, an Expiration Date Estimator that determines shelf life using online sources, a Grocery Tracker that maintains inventory based on consumption, and a Recipe Recommender that suggests meals using available ingredients. Each agent has specific tools and tasks orchestrated through a crew workflow. Additional Resources 📚 YouTube Explanation 🔍 Quality Assurance and Testing Agents LangGraph-Based Systems Inspector Overview 🔎 A comprehensive testing and validation tool for LangGraph-based applications that automatically analyzes system architecture, generates test cases, and identifies potential vulnerabilities through multi-agent inspection. The inspector employs specialized AI testers to evaluate different aspects of the system, from basic functionality to security concerns and edge cases. Implementation 🛠️ Integrates LangGraph for workflow orchestration, multiple LLM-powered testing agents, and a structured evaluation pipeline that includes static analysis, test case generation, and results verification. The system uses Pydantic for data validation, NetworkX for graph representation, and implements a modular architecture that allows for parallel test execution and comprehensive result analysis. Additional Resources 📚 YouTube Explanation Blog Post EU Green Deal FAQ Bot Overview 🔎 The EU Green Deal FAQ Bot demonstrates how to build a RAG-based AI agent that helps businesses understand EU green deal policies. The system processes complex regulatory documents into manageable chunks and provides instant, accurate answers to common questions about environmental compliance, emissions reporting, and waste management requirements. Implementation 🛠️ Implements a sophisticated RAG pipeline using FAISS vectorstore for document storage, semantic chunking for preprocessing, and multiple specialized agents (Retriever, Summarizer, Evaluator) for query processing. The system features query rephrasing for improved accuracy, cross-reference with gold Q&A datasets for answer validation, and comprehensive evaluation metrics to ensure response quality and relevance. Additional Resources 📚 YouTube Explanation Systematic Review Automation System + Paper Draft Creation Overview 🔎 A comprehensive system for automating academic systematic reviews using a directed graph architecture and LangChain components. The system generates complete, publication-ready systematic review papers, automatically processing everything from literature search through final draft generation with multiple revision cycles. Implementation 🛠️ Utilizes a state-based graph workflow that handles paper search and selection (up to 3 papers), PDF processing, and generates a complete academic paper with all standard sections (abstract, introduction, methods, results, conclusions, references). The system incorporates multiple revision cycles with automated critique and improvement phases, all orchestrated through LangGraph state management. Additional Resources 📚 YouTube Explanation 🌟 Special Advanced Technique 🌟 Sophisticated Controllable Agent for Complex RAG Tasks 🤖 Overview 🔎 An advanced RAG solution designed to tackle complex questions that simple semantic similarity-based retrieval cannot solve. This approach uses a sophisticated deterministic graph as the "brain" 🧠 of a highly controllable autonomous agent, capable of answering non-trivial questions from your own data. Implementation 🛠️ • Implement a multi-step process involving question anonymization, high-level planning, task breakdown, adaptive information retrieval and question answering, continuous re-planning, and rigorous answer verification to ensure grounded and accurate responses. Getting Started To begin exploring and building GenAI agents: Clone this repository: Navigate to the technique you're interested in: Follow the detailed implementation guide in each technique's notebook. Contributing We welcome contributions from the community! If you have a new technique or improvement to suggest: Fork the repository Create your feature branch: git checkout -b feature/AmazingFeature Commit your changes: git commit -m 'Add some AmazingFeature' Push to the branch: git push origin feature/AmazingFeature Open a pull request Contributors License This project is licensed under a custom non-commercial license - see the LICENSE file for details. ⭐️ If you find this repository helpful, please consider giving it a star! Keywords: GenAI, Generative AI, Agents, NLP, AI, Machine Learning, Natural Language Processing, LLM, Conversational AI, Task-Oriented AI

Production-Level-Deep-Learning
github
LLM Vibe Score0.619
Human Vibe Score0.8326638433689385
alirezadirMar 28, 2025

Production-Level-Deep-Learning

:bulb: A Guide to Production Level Deep Learning :clapper: :scroll: :ferry: 🇨🇳 Translation in Chinese.md) :label: NEW: Machine Learning Interviews :label: Note: This repo is under continous development, and all feedback and contribution are very welcome :blush: Deploying deep learning models in production can be challenging, as it is far beyond training models with good performance. Several distinct components need to be designed and developed in order to deploy a production level deep learning system (seen below): This repo aims to be an engineering guideline for building production-level deep learning systems which will be deployed in real world applications. The material presented here is borrowed from Full Stack Deep Learning Bootcamp (by Pieter Abbeel at UC Berkeley, Josh Tobin at OpenAI, and Sergey Karayev at Turnitin), TFX workshop by Robert Crowe, and Pipeline.ai's Advanced KubeFlow Meetup by Chris Fregly. Machine Learning Projects Fun :flushed: fact: 85% of AI projects fail. 1 Potential reasons include: Technically infeasible or poorly scoped Never make the leap to production Unclear success criteria (metrics) Poor team management ML Projects lifecycle Importance of understanding state of the art in your domain: Helps to understand what is possible Helps to know what to try next Mental Model for ML project The two important factors to consider when defining and prioritizing ML projects: High Impact: Complex parts of your pipeline Where "cheap prediction" is valuable Where automating complicated manual process is valuable Low Cost: Cost is driven by: Data availability Performance requirements: costs tend to scale super-linearly in the accuracy requirement Problem difficulty: Some of the hard problems include: unsupervised learning, reinforcement learning, and certain categories of supervised learning Full stack pipeline The following figure represents a high level overview of different components in a production level deep learning system: In the following, we will go through each module and recommend toolsets and frameworks as well as best practices from practitioners that fit each component. Data Management 1.1 Data Sources Supervised deep learning requires a lot of labeled data Labeling own data is costly! Here are some resources for data: Open source data (good to start with, but not an advantage) Data augmentation (a MUST for computer vision, an option for NLP) Synthetic data (almost always worth starting with, esp. in NLP) 1.2 Data Labeling Requires: separate software stack (labeling platforms), temporary labor, and QC Sources of labor for labeling: Crowdsourcing (Mechanical Turk): cheap and scalable, less reliable, needs QC Hiring own annotators: less QC needed, expensive, slow to scale Data labeling service companies: FigureEight Labeling platforms: Diffgram: Training Data Software (Computer Vision) Prodigy: An annotation tool powered by active learning (by developers of Spacy), text and image HIVE: AI as a Service platform for computer vision Supervisely: entire computer vision platform Labelbox: computer vision Scale AI data platform (computer vision & NLP) 1.3. Data Storage Data storage options: Object store: Store binary data (images, sound files, compressed texts) Amazon S3 Ceph Object Store Database: Store metadata (file paths, labels, user activity, etc). Postgres is the right choice for most of applications, with the best-in-class SQL and great support for unstructured JSON. Data Lake: to aggregate features which are not obtainable from database (e.g. logs) Amazon Redshift Feature Store: store, access, and share machine learning features (Feature extraction could be computationally expensive and nearly impossible to scale, hence re-using features by different models and teams is a key to high performance ML teams). FEAST (Google cloud, Open Source) Michelangelo Palette (Uber) Suggestion: At training time, copy data into a local or networked filesystem (NFS). 1 1.4. Data Versioning It's a "MUST" for deployed ML models: Deployed ML models are part code, part data. 1 No data versioning means no model versioning. Data versioning platforms: DVC: Open source version control system for ML projects Pachyderm: version control for data Dolt: a SQL database with Git-like version control for data and schema 1.5. Data Processing Training data for production models may come from different sources, including Stored data in db and object stores, log processing, and outputs of other classifiers*. There are dependencies between tasks, each needs to be kicked off after its dependencies are finished. For example, training on new log data, requires a preprocessing step before training. Makefiles are not scalable. "Workflow manager"s become pretty essential in this regard. Workflow orchestration: Luigi by Spotify Airflow by Airbnb: Dynamic, extensible, elegant, and scalable (the most widely used) DAG workflow Robust conditional execution: retry in case of failure Pusher supports docker images with tensorflow serving Whole workflow in a single .py file Development, Training, and Evaluation 2.1. Software engineering Winner language: Python Editors: Vim Emacs VS Code (Recommended by the author): Built-in git staging and diff, Lint code, open projects remotely through ssh Notebooks: Great as starting point of the projects, hard to scale (fun fact: Netflix’s Notebook-Driven Architecture is an exception, which is entirely based on nteract suites). nteract: a next-gen React-based UI for Jupyter notebooks Papermill: is an nteract library built for parameterizing, executing, and analyzing* Jupyter Notebooks. Commuter: another nteract project which provides a read-only display of notebooks (e.g. from S3 buckets). Streamlit: interactive data science tool with applets Compute recommendations 1: For individuals or startups*: Development: a 4x Turing-architecture PC Training/Evaluation: Use the same 4x GPU PC. When running many experiments, either buy shared servers or use cloud instances. For large companies:* Development: Buy a 4x Turing-architecture PC per ML scientist or let them use V100 instances Training/Evaluation: Use cloud instances with proper provisioning and handling of failures Cloud Providers: GCP: option to connect GPUs to any instance + has TPUs AWS: 2.2. Resource Management Allocating free resources to programs Resource management options: Old school cluster job scheduler ( e.g. Slurm workload manager ) Docker + Kubernetes Kubeflow Polyaxon (paid features) 2.3. DL Frameworks Unless having a good reason not to, use Tensorflow/Keras or PyTorch. 1 The following figure shows a comparison between different frameworks on how they stand for "developement" and "production"*. 2.4. Experiment management Development, training, and evaluation strategy: Always start simple Train a small model on a small batch. Only if it works, scale to larger data and models, and hyperparameter tuning! Experiment management tools: Tensorboard provides the visualization and tooling needed for ML experimentation Losswise (Monitoring for ML) Comet: lets you track code, experiments, and results on ML projects Weights & Biases: Record and visualize every detail of your research with easy collaboration MLFlow Tracking: for logging parameters, code versions, metrics, and output files as well as visualization of the results. Automatic experiment tracking with one line of code in python Side by side comparison of experiments Hyper parameter tuning Supports Kubernetes based jobs 2.5. Hyperparameter Tuning Approaches: Grid search Random search Bayesian Optimization HyperBand and Asynchronous Successive Halving Algorithm (ASHA) Population-based Training Platforms: RayTune: Ray Tune is a Python library for hyperparameter tuning at any scale (with a focus on deep learning and deep reinforcement learning). Supports any machine learning framework, including PyTorch, XGBoost, MXNet, and Keras. Katib: Kubernete's Native System for Hyperparameter Tuning and Neural Architecture Search, inspired by Google vizier and supports multiple ML/DL frameworks (e.g. TensorFlow, MXNet, and PyTorch). Hyperas: a simple wrapper around hyperopt for Keras, with a simple template notation to define hyper-parameter ranges to tune. SIGOPT: a scalable, enterprise-grade optimization platform Sweeps from [Weights & Biases] (https://www.wandb.com/): Parameters are not explicitly specified by a developer. Instead they are approximated and learned by a machine learning model. Keras Tuner: A hyperparameter tuner for Keras, specifically for tf.keras with TensorFlow 2.0. 2.6. Distributed Training Data parallelism: Use it when iteration time is too long (both tensorflow and PyTorch support) Ray Distributed Training Model parallelism: when model does not fit on a single GPU Other solutions: Horovod Troubleshooting [TBD] Testing and Deployment 4.1. Testing and CI/CD Machine Learning production software requires a more diverse set of test suites than traditional software: Unit and Integration Testing: Types of tests: Training system tests: testing training pipeline Validation tests: testing prediction system on validation set Functionality tests: testing prediction system on few important examples Continuous Integration: Running tests after each new code change pushed to the repo SaaS for continuous integration: Argo: Open source Kubernetes native workflow engine for orchestrating parallel jobs (incudes workflows, events, CI and CD). CircleCI: Language-Inclusive Support, Custom Environments, Flexible Resource Allocation, used by instacart, Lyft, and StackShare. Travis CI Buildkite: Fast and stable builds, Open source agent runs on almost any machine and architecture, Freedom to use your own tools and services Jenkins: Old school build system 4.2. Web Deployment Consists of a Prediction System and a Serving System Prediction System: Process input data, make predictions Serving System (Web server): Serve prediction with scale in mind Use REST API to serve prediction HTTP requests Calls the prediction system to respond Serving options: Deploy to VMs, scale by adding instances Deploy as containers, scale via orchestration Containers Docker Container Orchestration: Kubernetes (the most popular now) MESOS Marathon Deploy code as a "serverless function" Deploy via a model serving solution Model serving: Specialized web deployment for ML models Batches request for GPU inference Frameworks: Tensorflow serving MXNet Model server Clipper (Berkeley) SaaS solutions Seldon: serve and scale models built in any framework on Kubernetes Algorithmia Decision making: CPU or GPU? CPU inference: CPU inference is preferable if it meets the requirements. Scale by adding more servers, or going serverless. GPU inference: TF serving or Clipper Adaptive batching is useful (Bonus) Deploying Jupyter Notebooks: Kubeflow Fairing is a hybrid deployment package that let's you deploy your Jupyter notebook* codes! 4.5 Service Mesh and Traffic Routing Transition from monolithic applications towards a distributed microservice architecture could be challenging. A Service mesh (consisting of a network of microservices) reduces the complexity of such deployments, and eases the strain on development teams. Istio: a service mesh to ease creation of a network of deployed services with load balancing, service-to-service authentication, monitoring, with few or no code changes in service code. 4.4. Monitoring: Purpose of monitoring: Alerts for downtime, errors, and distribution shifts Catching service and data regressions Cloud providers solutions are decent Kiali:an observability console for Istio with service mesh configuration capabilities. It answers these questions: How are the microservices connected? How are they performing? Are we done? 4.5. Deploying on Embedded and Mobile Devices Main challenge: memory footprint and compute constraints Solutions: Quantization Reduced model size MobileNets Knowledge Distillation DistillBERT (for NLP) Embedded and Mobile Frameworks: Tensorflow Lite PyTorch Mobile Core ML ML Kit FRITZ OpenVINO Model Conversion: Open Neural Network Exchange (ONNX): open-source format for deep learning models 4.6. All-in-one solutions Tensorflow Extended (TFX) Michelangelo (Uber) Google Cloud AI Platform Amazon SageMaker Neptune FLOYD Paperspace Determined AI Domino data lab Tensorflow Extended (TFX) [TBD] Airflow and KubeFlow ML Pipelines [TBD] Other useful links: Lessons learned from building practical deep learning systems Machine Learning: The High Interest Credit Card of Technical Debt Contributing References: [1]: Full Stack Deep Learning Bootcamp, Nov 2019. [2]: Advanced KubeFlow Workshop by Pipeline.ai, 2019. [3]: TFX: Real World Machine Learning in Production

aima-python
github
LLM Vibe Score0.575
Human Vibe Score0.33114909407186394
aimacodeMar 28, 2025

aima-python

aima-python Python code for the book Artificial Intelligence: A Modern Approach. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Updates for 4th Edition The 4th edition of the book as out now in 2020, and thus we are updating the code. All code here will reflect the 4th edition. Changes include: Move from Python 3.5 to 3.7. More emphasis on Jupyter (Ipython) notebooks. More projects using external packages (tensorflow, etc.). Structure of the Project When complete, this project will have Python implementations for all the pseudocode algorithms in the book, as well as tests and examples of use. For each major topic, such as search, we provide the following files: search.ipynb and search.py: Implementations of all the pseudocode algorithms, and necessary support functions/classes/data. The .py file is generated automatically from the .ipynb file; the idea is that it is easier to read the documentation in the .ipynb file. search_XX.ipynb: Notebooks that show how to use the code, broken out into various topics (the XX). tests/test_search.py: A lightweight test suite, using assert statements, designed for use with py.test, but also usable on their own. Python 3.7 and up The code for the 3rd edition was in Python 3.5; the current 4th edition code is in Python 3.7. It should also run in later versions, but does not run in Python 2. You can install Python or use a browser-based Python interpreter such as repl.it. You can run the code in an IDE, or from the command line with python -i filename.py where the -i option puts you in an interactive loop where you can run Python functions. All notebooks are available in a binder environment. Alternatively, visit jupyter.org for instructions on setting up your own Jupyter notebook environment. Features from Python 3.6 and 3.7 that we will be using for this version of the code: f-strings: all string formatting should be done with f'var = {var}', not with 'var = {}'.format(var) nor 'var = %s' % var. typing module: declare functions with type hints: def successors(state) -> List[State]:; that is, give type declarations, but omit them when it is obvious. I don't need to say state: State, but in another context it would make sense to say s: State. Underscores in numerics: write a million as 1000000 not as 1000000. dataclasses module: replace namedtuple with dataclass. [//]: (There is a sibling [aima-docker]https://github.com/rajatjain1997/aima-docker project that shows you how to use docker containers to run more complex problems in more complex software environments.) Installation Guide To download the repository: git clone https://github.com/aimacode/aima-python.git Then you need to install the basic dependencies to run the project on your system: You also need to fetch the datasets from the aima-data repository: Wait for the datasets to download, it may take a while. Once they are downloaded, you need to install pytest, so that you can run the test suite: pip install pytest Then to run the tests: py.test And you are good to go! Index of Algorithms Here is a table of algorithms, the figure, name of the algorithm in the book and in the repository, and the file where they are implemented in the repository. This chart was made for the third edition of the book and is being updated for the upcoming fourth edition. Empty implementations are a good place for contributors to look for an issue. The aima-pseudocode project describes all the algorithms from the book. An asterisk next to the file name denotes the algorithm is not fully implemented. Another great place for contributors to start is by adding tests and writing on the notebooks. You can see which algorithms have tests and notebook sections below. If the algorithm you want to work on is covered, don't worry! You can still add more tests and provide some examples of use in the notebook! | Figure | Name (in 3rd edition) | Name (in repository) | File | Tests | Notebook |:-------|:----------------------------------|:------------------------------|:--------------------------------|:-----|:---------| | 2 | Random-Vacuum-Agent | RandomVacuumAgent | [agents.py][agents] | Done | Included | | 2 | Model-Based-Vacuum-Agent | ModelBasedVacuumAgent | [agents.py][agents] | Done | Included | | 2.1 | Environment | Environment | [agents.py][agents] | Done | Included | | 2.1 | Agent | Agent | [agents.py][agents] | Done | Included | | 2.3 | Table-Driven-Vacuum-Agent | TableDrivenVacuumAgent | [agents.py][agents] | Done | Included | | 2.7 | Table-Driven-Agent | TableDrivenAgent | [agents.py][agents] | Done | Included | | 2.8 | Reflex-Vacuum-Agent | ReflexVacuumAgent | [agents.py][agents] | Done | Included | | 2.10 | Simple-Reflex-Agent | SimpleReflexAgent | [agents.py][agents] | Done | Included | | 2.12 | Model-Based-Reflex-Agent | ReflexAgentWithState | [agents.py][agents] | Done | Included | | 3 | Problem | Problem | [search.py][search] | Done | Included | | 3 | Node | Node | [search.py][search] | Done | Included | | 3 | Queue | Queue | [utils.py][utils] | Done | No Need | | 3.1 | Simple-Problem-Solving-Agent | SimpleProblemSolvingAgent | [search.py][search] | Done | Included | | 3.2 | Romania | romania | [search.py][search] | Done | Included | | 3.7 | Tree-Search | depth/breadthfirsttree_search | [search.py][search] | Done | Included | | 3.7 | Graph-Search | depth/breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.11 | Breadth-First-Search | breadthfirstgraph_search | [search.py][search] | Done | Included | | 3.14 | Uniform-Cost-Search | uniformcostsearch | [search.py][search] | Done | Included | | 3.17 | Depth-Limited-Search | depthlimitedsearch | [search.py][search] | Done | Included | | 3.18 | Iterative-Deepening-Search | iterativedeepeningsearch | [search.py][search] | Done | Included | | 3.22 | Best-First-Search | bestfirstgraph_search | [search.py][search] | Done | Included | | 3.24 | A\*-Search | astar_search | [search.py][search] | Done | Included | | 3.26 | Recursive-Best-First-Search | recursivebestfirst_search | [search.py][search] | Done | Included | | 4.2 | Hill-Climbing | hill_climbing | [search.py][search] | Done | Included | | 4.5 | Simulated-Annealing | simulated_annealing | [search.py][search] | Done | Included | | 4.8 | Genetic-Algorithm | genetic_algorithm | [search.py][search] | Done | Included | | 4.11 | And-Or-Graph-Search | andorgraph_search | [search.py][search] | Done | Included | | 4.21 | Online-DFS-Agent | onlinedfsagent | [search.py][search] | Done | Included | | 4.24 | LRTA\*-Agent | LRTAStarAgent | [search.py][search] | Done | Included | | 5.3 | Minimax-Decision | minimax_decision | [games.py][games] | Done | Included | | 5.7 | Alpha-Beta-Search | alphabeta_search | [games.py][games] | Done | Included | | 6 | CSP | CSP | [csp.py][csp] | Done | Included | | 6.3 | AC-3 | AC3 | [csp.py][csp] | Done | Included | | 6.5 | Backtracking-Search | backtracking_search | [csp.py][csp] | Done | Included | | 6.8 | Min-Conflicts | min_conflicts | [csp.py][csp] | Done | Included | | 6.11 | Tree-CSP-Solver | treecspsolver | [csp.py][csp] | Done | Included | | 7 | KB | KB | [logic.py][logic] | Done | Included | | 7.1 | KB-Agent | KB_AgentProgram | [logic.py][logic] | Done | Included | | 7.7 | Propositional Logic Sentence | Expr | [utils.py][utils] | Done | Included | | 7.10 | TT-Entails | tt_entails | [logic.py][logic] | Done | Included | | 7.12 | PL-Resolution | pl_resolution | [logic.py][logic] | Done | Included | | 7.14 | Convert to CNF | to_cnf | [logic.py][logic] | Done | Included | | 7.15 | PL-FC-Entails? | plfcentails | [logic.py][logic] | Done | Included | | 7.17 | DPLL-Satisfiable? | dpll_satisfiable | [logic.py][logic] | Done | Included | | 7.18 | WalkSAT | WalkSAT | [logic.py][logic] | Done | Included | | 7.20 | Hybrid-Wumpus-Agent | HybridWumpusAgent | | | | | 7.22 | SATPlan | SAT_plan | [logic.py][logic] | Done | Included | | 9 | Subst | subst | [logic.py][logic] | Done | Included | | 9.1 | Unify | unify | [logic.py][logic] | Done | Included | | 9.3 | FOL-FC-Ask | folfcask | [logic.py][logic] | Done | Included | | 9.6 | FOL-BC-Ask | folbcask | [logic.py][logic] | Done | Included | | 10.1 | Air-Cargo-problem | air_cargo | [planning.py][planning] | Done | Included | | 10.2 | Spare-Tire-Problem | spare_tire | [planning.py][planning] | Done | Included | | 10.3 | Three-Block-Tower | threeblocktower | [planning.py][planning] | Done | Included | | 10.7 | Cake-Problem | havecakeandeatcake_too | [planning.py][planning] | Done | Included | | 10.9 | Graphplan | GraphPlan | [planning.py][planning] | Done | Included | | 10.13 | Partial-Order-Planner | PartialOrderPlanner | [planning.py][planning] | Done | Included | | 11.1 | Job-Shop-Problem-With-Resources | jobshopproblem | [planning.py][planning] | Done | Included | | 11.5 | Hierarchical-Search | hierarchical_search | [planning.py][planning] | Done | Included | | 11.8 | Angelic-Search | angelic_search | [planning.py][planning] | Done | Included | | 11.10 | Doubles-tennis | doubletennisproblem | [planning.py][planning] | Done | Included | | 13 | Discrete Probability Distribution | ProbDist | [probability.py][probability] | Done | Included | | 13.1 | DT-Agent | DTAgent | [probability.py][probability] | Done | Included | | 14.9 | Enumeration-Ask | enumeration_ask | [probability.py][probability] | Done | Included | | 14.11 | Elimination-Ask | elimination_ask | [probability.py][probability] | Done | Included | | 14.13 | Prior-Sample | prior_sample | [probability.py][probability] | Done | Included | | 14.14 | Rejection-Sampling | rejection_sampling | [probability.py][probability] | Done | Included | | 14.15 | Likelihood-Weighting | likelihood_weighting | [probability.py][probability] | Done | Included | | 14.16 | Gibbs-Ask | gibbs_ask | [probability.py][probability] | Done | Included | | 15.4 | Forward-Backward | forward_backward | [probability.py][probability] | Done | Included | | 15.6 | Fixed-Lag-Smoothing | fixedlagsmoothing | [probability.py][probability] | Done | Included | | 15.17 | Particle-Filtering | particle_filtering | [probability.py][probability] | Done | Included | | 16.9 | Information-Gathering-Agent | InformationGatheringAgent | [probability.py][probability] | Done | Included | | 17.4 | Value-Iteration | value_iteration | [mdp.py][mdp] | Done | Included | | 17.7 | Policy-Iteration | policy_iteration | [mdp.py][mdp] | Done | Included | | 17.9 | POMDP-Value-Iteration | pomdpvalueiteration | [mdp.py][mdp] | Done | Included | | 18.5 | Decision-Tree-Learning | DecisionTreeLearner | [learning.py][learning] | Done | Included | | 18.8 | Cross-Validation | cross_validation | [learning.py][learning]\* | | | | 18.11 | Decision-List-Learning | DecisionListLearner | [learning.py][learning]\* | | | | 18.24 | Back-Prop-Learning | BackPropagationLearner | [learning.py][learning] | Done | Included | | 18.34 | AdaBoost | AdaBoost | [learning.py][learning] | Done | Included | | 19.2 | Current-Best-Learning | currentbestlearning | knowledge.py | Done | Included | | 19.3 | Version-Space-Learning | versionspacelearning | knowledge.py | Done | Included | | 19.8 | Minimal-Consistent-Det | minimalconsistentdet | knowledge.py | Done | Included | | 19.12 | FOIL | FOIL_container | knowledge.py | Done | Included | | 21.2 | Passive-ADP-Agent | PassiveADPAgent | [rl.py][rl] | Done | Included | | 21.4 | Passive-TD-Agent | PassiveTDAgent | [rl.py][rl] | Done | Included | | 21.8 | Q-Learning-Agent | QLearningAgent | [rl.py][rl] | Done | Included | | 22.1 | HITS | HITS | [nlp.py][nlp] | Done | Included | | 23 | Chart-Parse | Chart | [nlp.py][nlp] | Done | Included | | 23.5 | CYK-Parse | CYK_parse | [nlp.py][nlp] | Done | Included | | 25.9 | Monte-Carlo-Localization | montecarlolocalization | [probability.py][probability] | Done | Included | Index of data structures Here is a table of the implemented data structures, the figure, name of the implementation in the repository, and the file where they are implemented. | Figure | Name (in repository) | File | |:-------|:--------------------------------|:--------------------------| | 3.2 | romania_map | [search.py][search] | | 4.9 | vacumm_world | [search.py][search] | | 4.23 | onedimstate_space | [search.py][search] | | 6.1 | australia_map | [search.py][search] | | 7.13 | wumpusworldinference | [logic.py][logic] | | 7.16 | hornclausesKB | [logic.py][logic] | | 17.1 | sequentialdecisionenvironment | [mdp.py][mdp] | | 18.2 | waitingdecisiontree | [learning.py][learning] | Acknowledgements Many thanks for contributions over the years. I got bug reports, corrected code, and other support from Darius Bacon, Phil Ruggera, Peng Shao, Amit Patil, Ted Nienstedt, Jim Martin, Ben Catanzariti, and others. Now that the project is on GitHub, you can see the contributors who are doing a great job of actively improving the project. Many thanks to all contributors, especially @darius, @SnShine, @reachtarunhere, @antmarakis, @Chipe1, @ad71 and @MariannaSpyrakou. [agents]:../master/agents.py [csp]:../master/csp.py [games]:../master/games.py [grid]:../master/grid.py [knowledge]:../master/knowledge.py [learning]:../master/learning.py [logic]:../master/logic.py [mdp]:../master/mdp.py [nlp]:../master/nlp.py [planning]:../master/planning.py [probability]:../master/probability.py [rl]:../master/rl.py [search]:../master/search.py [utils]:../master/utils.py [text]:../master/text.py

OpenAI-CLIP
github
LLM Vibe Score0.507
Human Vibe Score0.015912940499642817
moein-shariatniaMar 27, 2025

OpenAI-CLIP

Update (December 2023) I am happy to find out that this code has been used and cited in the following papers: Domino: Discovering Systematic Errors with Cross-Modal Embeddings by Eyuboglu et. al. at ICLR 2022 GSCLIP : A Framework for Explaining Distribution Shifts in Natural Language by Zhu et. al. at ICML 2022 UIC-NLP at SemEval-2022 Task 5: Exploring Contrastive Learning for Multimodal Detection of Misogynistic Memes by Cuervo et. al. at SemEval-2022 cdsBERT - Extending Protein Language Models with Codon Awareness by Hallee et. al. from University of Delaware (Sep 2023) ENIGMA-51: Towards a Fine-Grained Understanding of Human-Object Interactions in Industrial Scenarios by Ragusa et. al. (Nov 2023) You can find the citation info on the right section of this GitHub repo page named: Cite this repository or use the below citation info. Introduction It was in January of 2021 that OpenAI announced two new models: DALL-E and CLIP, both multi-modality models connecting texts and images in some way. In this article we are going to implement CLIP model from scratch in PyTorch. OpenAI has open-sourced some of the code relating to CLIP model but I found it intimidating and it was far from something short and simple. I also came across a good tutorial inspired by CLIP model on Keras code examples and I translated some parts of it into PyTorch to build this tutorial totally with our beloved PyTorch! What does CLIP do? Why is it fun? In Learning Transferable Visual Models From Natural Language Supervision paper, OpenAI introduces their new model which is called CLIP, for Contrastive Language-Image Pre-training. In a nutshell, this model learns the relationship between a whole sentence and the image it describes; in a sense that when the model is trained, given an input sentence it will be able to retrieve the most related images corresponding to that sentence. The important thing here is that it is trained on full sentences instead of single classes like car, dog, etc. The intuition is that when trained on whole sentences, the model can learn a lot more things and finds some pattern between images and texts. They also show that when this model is trained on a huge dataset of images and their corresponding texts, it can also act as a classifier too. I encourage you to study the paper to learn more about this exciting model and their astonishing results on benchmarking datasets . To mention just one, CLIP model trained with this strategy classifies ImageNet better than those SOTA models trained on the ImageNet itself optimized for the only task of classification! As a teaser (!), let's see what the final model that we will build in this article from scratch is capable of: given a query (raw text) like "a boy jumping with skateboard" or "a girl jumping from swing", the model will retrieve the most relevant images: !title_img Let's see some more outputs: Config A note on config and CFG: I wrote the codes with python scripts and then converted it into a Jupyter Notebook. So, in case of python scripts, config is a normal python file where I put all the hyperparameters and in the case of Jupyter Notebook, its a class defined in the beginning of the notebook to keep all the hyperparameters. Utils Dataset As you can see in the tittle image of this article, we need to encode both images and their describing texts. So, the dataset needs to return both images and texts. Of course we are not going to feed raw text to our text encoder! We will use DistilBERT model (which is smaller than BERT but performs nearly as well as BERT) from HuggingFace library as our text encoder; so, we need to tokenize the sentences (captions) with DistilBERT tokenizer and then feed the token ids (input_ids) and the attention masks to DistilBERT. Therefore, the dataset needs to take care of the tokenization as well. Below you can see the dataset's code. Below that I'll explain the most important things that is happening in the code. In the \\init\\ we receive a tokenizer object which is actually a HuggingFace tokinzer; this tokenizer will be loaded when running the model. We are padding and truncating the captions to a specified maxlength. In the \\getitem\\ we will first load an encoded caption which is a dictionary with keys inputids and attention_mask, make tensors out of its values and after that we will load the corresponding image, transform and augment it (if there is any!) and then we make it a tensor and put it in the dictionary with "image" as the key. Finally we put the raw text of the caption with the key "caption" in the dictionary only for visualization purposes. I did not use additional data augmentations but you can add them if you want to improve the model's performance. Image Encoder The image encoder code is straight forward. I'm using PyTorch Image Models library (timm) here which makes a lot of different image models available from ResNets to EfficientNets and many more. Here we will use a ResNet50 as our image encoder. You can easily use torchvision library to use ResNets if you don't want to install a new library. The code encodes each image to a fixed size vector with the size of the model's output channels (in case of ResNet50 the vector size will be 2048). This is the output after the nn.AdaptiveAvgPool2d() layer. Text Encoder As I mentioned before, I'll use DistilBERT as the text encoder. Like its bigger brother BERT, two special tokens will be added to the actual input tokens: CLS and SEP which mark the start and end of a sentence. To grab the whole representation of a sentence (as the related BERT and DistilBERT papers point out) we use the final representations of the CLS token and we hope that this representation captures the overall meaning of the sentence (caption). Thinking it in this way, it is similar to what we did to images and converted them into a fixed size vector. In the case of DistilBERT (and also BERT) the output hidden representation for each token is a vector with size 768. So, the whole caption will be encoded in the CLS token representation whose size is 768. Projection Head I used Keras code example implementation of projection head to write the following in PyTorch. Now that we have encoded both our images and texts into fixed size vectors (2048 for image and 768 for text) we need to bring (project) them into a new world (!) with similar dimensions for both images and texts in order to be able to compare them and push apart the non-relevant image and texts and pull together those that match. So, the following code will bring the 2048 and 768 dimensional vectors into a 256 (projection_dim) dimensional world, where we can compare them. "embeddingdim" is the size of the input vector (2048 for images and 768 for texts) and "projectiondim" is the the size of the output vector which will be 256 for our case. For understanding the details of this part you can refer to the CLIP paper. CLIP This part is where all the fun happens! I'll also talk about the loss function here. I translated some of the code from Keras code examples into PyTorch for writing this part. Take a look at the code and then read the explanation below this code block. Here we will use the previous modules that we built to implement the main model. The \\init\\ function is self-explanatory. In the forward function, we first encode the images and texts separately into fixed size vectors (with different dimensionalities). After that, using separate projection modules we project them to that shared world (space) that I talked about previously. Here the encodings will become of similar shape (256 in our case). After that we will compute the loss. Again I recommend reading CLIP paper to get it better but I'll try my best to explain this part. In Linear Algebra, one common way to measure if two vectors are of similar characteristics (they are like each other) is to calculate their dot product (multiplying the matching entries and take the sum of them); if the final number is big, they are alike and if it is small they are not (relatively speaking)! Okay! What I just said is the most important thing to have in mind to understand this loss function. Let's continue. We talked about two vectors, but, what do we have here? We have imageembeddings, a matrix with shape (batchsize, 256) and textembeddings with shape (batchsize, 256). Easy enough! it means we have two groups of vectors instead of two single vectors. How do we measure how similar two groups of vectors (two matrices) are to each other? Again, with dot product (@ operator in PyTorch does the dot product or matrix multiplication in this case). To be able to multiply these two matrices together, we transpose the second one. Okay, we get a matrix with shape (batchsize, batchsize) which we will call logits. (temperature is equal to 1.0 in our case, so, it does not make a difference. You can play with it and see what difference it makes. Also look at the paper to see why it is here!). I hope you are still with me! If not it's okay, just review the code and check their shapes. Now that we have our logits, we need targets. I need to say that there is a more straight forward way to obtain targets but I had to do this for our case (I'll talk about why in a next paragraph). Let's consider what we hope that this model learns: we want it to learn "similar representations (vectors)" for a given image and the caption describing it. Meaning that either we give it an image or the text describing it, we want it to produce same 256 sized vectors for both. Check the cell below this code block for the continue of the explanations So, in the best case scenario, textembeddings and imageembedding matricies should be the same because they are describing similar things. Let's think now: if this happens, what would the logits matrix be like? Let's see with a simple example! So logits, in the best case, will be a matrix that if we take its softmax, will have 1.0s in the diagonal (An identity matrix to call it with fancy words!). As the loss function's job is to make model's predictions similar to targets (at least in most cases!), we want such a matrix as our target. That's the reason why we are calculating imagessimilarity and textssimilarity matrices in the code block above. Now that we've got our targets matrix, we will use simple cross entropy to calculate the actual loss. I've written the full matrix form of cross entropy as a function which you can see in the bottom of the code block. Okay! We are done! Wasn't it simple?! Alright, you can ignore the next paragraph but if you are curious, there is an important note in that. Here's why I didn't use a simpler approach: I need to admit that there's a simpler way to calculate this loss in PyTorch; by doing this: nn.CrossEntropyLoss()(logits, torch.arange(batch_size)). Why I did not use it here? For 2 reasons. 1- The dataset we are using has multiple captions for a single image; so, there is the possibility that two identical images with their similar captions exist in a batch (it is rare but it can happen). Taking the loss with this easier method will ignore this possibility and the model learns to pull apart two representations (assume them different) that are actually the same. Obviously, we don't want this to happen so I calculated the whole target matrix in a way that takes care of these edge cases. 2- Doing it the way I did, gave me a better understanding of what is happening in this loss function; so, I thought it would give you a better intuition as well! Train Here are some funtions to help us load train and valid dataloaders, our model and then train and evaluate our model on those. There's not much going on here; just simple training loop and utility functions Here's a handy function to train our model. There's not much happening here; just loading the batches, feeding them to the model and stepping the optimizer and lr_scheduler. Running the next cell start training the model. Put the kernel on GPU mode. Every epoch should take about 24 minutes on GPU (even one epoch is enough!). It can take one minute before training actually starts because we are going to encode all the captions once in the train and valid dataset, so please don't stop it! Every thing is working fine. Inference Okay! We are done with training the model. Now, we need to do inference which in our case will be giving the model a piece of text and want it to retrieve the most relevant images from an unseen validation (or test) set. Getting Image Embeddings In this function, we are loading the model that we saved after training, feeding it images in validation set and returning the imageembeddings with shape (validset_size, 256) and the model itself. Finding Matches This function does the final task that we wished our model would be capable of: it gets the model, image_embeddings, and a text query. It will display the most relevant images from the validation set! Isn't it amazing? Let's see how it performs after all! This is how we use this function. Aaaannnndddd the results: Final words I hope you have enjoyed this article. Implementing this paper was a really interesting experience for me. I want to thank Khalid Salama for the great Keras code example he provided which inspired me to write something similar in PyTorch.

aima-java
github
LLM Vibe Score0.521
Human Vibe Score0.06620214044837505
aimacodeMar 25, 2025

aima-java

AIMA3e-Java (JDK 8+) Java implementation of algorithms from Russell and Norvig's Artificial Intelligence - A Modern Approach 3rd Edition. You can use this in conjunction with a course on AI, or for study on your own. We're looking for solid contributors to help. Getting Started Links Overview of Project Interested in Contributing Setting up your own workspace Comments on architecture and design Demo Applications that can be run from your browser (unfortunately not up to date) Javadoc for the aima-core project (outdated) Download the latest official (but outdated) version = 1.9.1 (Dec 18 2016) Latest Maven Information (for integration as a third party library) Index of Implemented Algorithms |Figure|Page|Name (in 3rd edition)|Code | -------- |:--------:| :-----| :----- | |2|34|Environment|Environment| |2.1|35|Agent|Agent| |2.3|36|Table-Driven-Vacuum-Agent|TableDrivenVacuumAgent| |2.7|47|Table-Driven-Agent|TableDrivenAgentProgram| |2.8|48|Reflex-Vacuum-Agent|ReflexVacuumAgent| |2.10|49|Simple-Reflex-Agent|SimpleReflexAgentProgram| |2.12|51|Model-Based-Reflex-Agent|ModelBasedReflexAgentProgram| |3|66|Problem|Problem| |3.1|67|Simple-Problem-Solving-Agent|SimpleProblemSolvingAgent| |3.2|68|Romania|SimplifiedRoadMapOfRomania| |3.7|77|Tree-Search|TreeSearch| |3.7|77|Graph-Search|GraphSearch| |3.10|79|Node|Node| |3.11|82|Breadth-First-Search|BreadthFirstSearch| |3.14|84|Uniform-Cost-Search|UniformCostSearch| |3|85|Depth-first Search|DepthFirstSearch| |3.17|88|Depth-Limited-Search|DepthLimitedSearch| |3.18|89|Iterative-Deepening-Search|IterativeDeepeningSearch| |3|90|Bidirectional search|BidirectionalSearch| |3|92|Best-First search|BestFirstSearch| |3|92|Greedy best-First search|GreedyBestFirstSearch| |3|93|A\* Search|AStarSearch| |3.26|99|Recursive-Best-First-Search |RecursiveBestFirstSearch| |4.2|122|Hill-Climbing|HillClimbingSearch| |4.5|126|Simulated-Annealing|SimulatedAnnealingSearch| |4.8|129|Genetic-Algorithm|GeneticAlgorithm| |4.11|136|And-Or-Graph-Search|AndOrSearch| |4|147|Online search problem|OnlineSearchProblem| |4.21|150|Online-DFS-Agent|OnlineDFSAgent| |4.24|152|LRTA\*-Agent|LRTAStarAgent| |5.3|166|Minimax-Decision|MinimaxSearch| |5.7|170|Alpha-Beta-Search|AlphaBetaSearch| |6|202|CSP|CSP| |6.1|204|Map CSP|MapCSP| |6.3|209|AC-3|AC3Strategy| |6.5|215|Backtracking-Search|AbstractBacktrackingSolver| |6.8|221|Min-Conflicts|MinConflictsSolver| |6.11|224|Tree-CSP-Solver|TreeCspSolver| |7|235|Knowledge Base|KnowledgeBase| |7.1|236|KB-Agent|KBAgent| |7.7|244|Propositional-Logic-Sentence|Sentence| |7.10|248|TT-Entails|TTEntails| |7|253|Convert-to-CNF|ConvertToCNF| |7.12|255|PL-Resolution|PLResolution| |7.15|258|PL-FC-Entails?|PLFCEntails| |7.17|261|DPLL-Satisfiable?|DPLLSatisfiable| |7.18|263|WalkSAT|WalkSAT| |7.20|270|Hybrid-Wumpus-Agent|HybridWumpusAgent| |7.22|272|SATPlan|SATPlan| |9|323|Subst|SubstVisitor| |9.1|328|Unify|Unifier| |9.3|332|FOL-FC-Ask|FOLFCAsk| |9.6|338|FOL-BC-Ask|FOLBCAsk| |9|345|CNF|CNFConverter| |9|347|Resolution|FOLTFMResolution| |9|354|Demodulation|Demodulation| |9|354|Paramodulation|Paramodulation| |9|345|Subsumption|SubsumptionElimination| |10.9|383|Graphplan|GraphPlan| |11.5|409|Hierarchical-Search|HierarchicalSearchAlgorithm| |11.8|414|Angelic-Search|---| |13.1|484|DT-Agent|DT-Agent| |13|484|Probability-Model|ProbabilityModel| |13|487|Probability-Distribution|ProbabilityDistribution| |13|490|Full-Joint-Distribution|FullJointDistributionModel| |14|510|Bayesian Network|BayesianNetwork| |14.9|525|Enumeration-Ask|EnumerationAsk| |14.11|528|Elimination-Ask|EliminationAsk| |14.13|531|Prior-Sample|PriorSample| |14.14|533|Rejection-Sampling|RejectionSampling| |14.15|534|Likelihood-Weighting|LikelihoodWeighting| |14.16|537|GIBBS-Ask|GibbsAsk| |15.4|576|Forward-Backward|ForwardBackward| |15|578|Hidden Markov Model|HiddenMarkovModel| |15.6|580|Fixed-Lag-Smoothing|FixedLagSmoothing| |15|590|Dynamic Bayesian Network|DynamicBayesianNetwork| |15.17|598|Particle-Filtering|ParticleFiltering| |16.9|632|Information-Gathering-Agent|InformationGatheringAgent| |17|647|Markov Decision Process|MarkovDecisionProcess| |17.4|653|Value-Iteration|ValueIteration| |17.7|657|Policy-Iteration|PolicyIteration| |17.9|663|POMDP-Value-Iteration|POMDPValueIteration| |18.5|702|Decision-Tree-Learning|DecisionTreeLearner| |18.8|710|Cross-Validation-Wrapper|CrossValidation| |18.11|717|Decision-List-Learning|DecisionListLearner| |18.24|734|Back-Prop-Learning|BackPropLearning| |18.34|751|AdaBoost|AdaBoostLearner| |19.2|771|Current-Best-Learning|CurrentBestLearning| |19.3|773|Version-Space-Learning|VersionSpaceLearning| |19.8|786|Minimal-Consistent-Det|MinimalConsistentDet| |19.12|793|FOIL|FOIL| |21.2|834|Passive-ADP-Agent|PassiveADPAgent| |21.4|837|Passive-TD-Agent|PassiveTDAgent| |21.8|844|Q-Learning-Agent|QLearningAgent| |22.1|871|HITS|HITS| |23.5|894|CYK-Parse|CYK| |25.9|982|Monte-Carlo-Localization|MonteCarloLocalization| Index of implemented notebooks |Chapter No|Name |Status (in 3rd edition)|Status (in 4th edition) | -------- |:--------:| :-----| :----- | |3| Solving Problems by Searching| In Progress| Not started| |6| Constraint Satisfaction Problems |In Progress|---| |12| Knowledge Representation|Done|---| |13| Quantifying Uncertainty |Done | --- | |14| Probabilistic Reasoning|In Progress| ---| Before starting to work on a new notebook: Open a new issue with the following heading: Notebook: Chapter Name - Version . Check that the issue is not assigned to anyone. Mention a topics list of what you will be implementing in the notebook for that particular chapter. You can iteratively refine the list once you start working. Start a discussion on what can go in that particular notebook. "---" indicates algorithms yet to be implemented. Index of data structures Here is a table of the data structures yet to be implemented. |Fig|Page|Name (in book)|Code| | -------- |:--------:| :-----| :----- | |9.8|341|Append|---| |10.1|369|AIR-CARGO-TRANSPORT-PROBLEM|---| |10.2|370|SPARE-TIRE-PROBLEM|---| |10.3|371|BLOCKS-WORLD |---| |10.7|380|HAVE-CAKE-AND-EAT-CAKE-TOO-PROBLEM|---| |11.1|402|JOB-SHOP-SCHEDULING-PROBLEM|---| |11.4|407|REFINEMENT-HIGH-LEVEL-ACTIONS|---| |23.6|895|SENTENCE-TREE|---| |29.1|1062|POWERS-OF-2|---|

AI-PhD-S24
github
LLM Vibe Score0.472
Human Vibe Score0.0922477795435268
rphilipzhangMar 25, 2025

AI-PhD-S24

Artificial Intelligence for Business Research (Spring 2024) Scribed Lecture Notes Class Recordings (You need to apply for access.) Teaching Team Instructor*: Renyu (Philip) Zhang, Associate Professor, Department of Decisions, Operations and Technology, CUHK Business School, philipzhang@cuhk.edu.hk, @911 Cheng Yu Tung Building. Teaching Assistant*: Leo Cao, Full-time TA, Department of Decisions, Operations and Technology, CUHK Business School, yinglyucao@cuhk.edu.hk. Please be noted that Leo will help with any issues related to the logistics, but not the content, of this course. Tutorial Instructor*: Qiansiqi Hu, MSBA Student, Department of Decisions, Operations and Technology, CUHK Business School, 1155208353@link.cuhk.edu.hk. BS in ECE, Shanghai Jiaotong University Michigan Institute. Basic Information Website: https://github.com/rphilipzhang/AI-PhD-S24 Time: Tuesday, 12:30pm-3:15pm, from Jan 9, 2024 to Apr 16, 2024, except for Feb 13 (Chinese New Year) and Mar 5 (Final Project Discussion) Location: Cheng Yu Tung Building (CYT) LT5 About Welcome to the mono-repo of the PhD course AI for Business Research (DSME 6635) at CUHK Business School in Spring 2024. You may download the Syllabus of this course first. The purpose of this course is to learn the following: Have a basic understanding of the fundamental concepts/methods in machine learning (ML) and artificial intelligence (AI) that are used (or potentially useful) in business research. Understand how business researchers have utilized ML/AI and what managerial questions have been addressed by ML/AI in the recent decade. Nurture a taste of what the state-of-the-art AI/ML technologies can do in the ML/AI community and, potentially, in your own research field. We will meet each Tuesday at 12:30pm in Cheng Yu Tung Building (CYT) LT5 (please pay attention to this room change). Please ask for my approval if you need to join us via the following Zoom links: Zoom link, Meeting ID 996 4239 3764, Passcode 386119. Most of the code in this course will be distributed through the Google CoLab cloud computing environment to avoid the incompatibility and version control issues on your local individual computer. On the other hand, you can always download the Jupyter Notebook from CoLab and run it your own computer. The CoLab files of this course can be found at this folder. The Google Sheet to sign up for groups and group tasks can be found here. The overleaf template for scribing the lecture notes of this course can be found here. If you have any feedback on this course, please directly contact Philip at philipzhang@cuhk.edu.hk and we will try our best to address it. Brief Schedule Subject to modifications. All classes start at 12:30pm and end at 3:15pm. |Session|Date |Topic|Key Words| |:-------:|:-------------:|:----:|:-:| |1|1.09|AI/ML in a Nutshell|Course Intro, ML Models, Model Evaluations| |2|1.16|Intro to DL|DL Intro, Neural Nets, Computational Issues in DL| |3|1.23|Prediction and Traditional NLP|Prediction in Biz Research, Pre-processing| |4|1.30|NLP (II): Traditional NLP|$N$-gram, NLP Performance Evaluations, Naïve Bayes| |5|2.06|NLP (III): Word2Vec|CBOW, Skip Gram| |6|2.20|NLP (IV): RNN|Glove, Language Model Evaluation, RNN| |7|2.27|NLP (V): Seq2Seq|LSTM, Seq2Seq, Attention Mechanism| |7.5|3.05|NLP (V.V): Transformer|The Bitter Lesson, Attention is All You Need| |8|3.12|NLP (VI): Pre-training|Computational Tricks in DL, BERT, GPT| |9|3.19|NLP (VII): LLM|Emergent Abilities, Chain-of-Thought, In-context Learning, GenAI in Business Research| |10|3.26|CV (I): Image Classification|CNN, AlexNet, ResNet, ViT| |11|4.02|CV (II): Image Segmentation and Video Analysis|R-CNN, YOLO, 3D-CNN| |12|4.09|Unsupervised Learning (I): Clustering & Topic Modeling|GMM, EM Algorithm, LDA| |13|4.16|Unsupervised Learning (II): Diffusion Models|VAE, DDPM, LDM, DiT| Important Dates All problem sets are due at 12:30pm right before class. |Date| Time|Event|Note| |:--:|:-:|:---:|:--:| |1.10| 11:59pm|Group Sign-Ups|Each group has at most two students.| |1.12| 7:00pm-9:00pm|Python Tutorial|Given by Qiansiqi Hu, Python Tutorial CoLab| |1.19| 7:00pm-9:00pm|PyTorch Tutorial|Given by Qiansiqi Hu, PyTorch Tutorial CoLab| |3.05|9:00am-6:00pm|Final Project Discussion|Please schedule a meeting with Philip.| |3.12| 12:30pm|Final Project Proposal|1-page maximum| |4.30| 11:59pm|Scribed Lecture Notes|Overleaf link| |5.12|11:59pm|Project Paper, Slides, and Code|Paper page limit: 10| Useful Resources Find more on the Syllabus. Books: ESL, Deep Learning, Dive into Deep Learning, ML Fairness, Applied Causal Inference Powered by ML and AI Courses: ML Intro by Andrew Ng, DL Intro by Andrew Ng, NLP (CS224N) by Chris Manning, CV (CS231N) by Fei-Fei Li, Deep Unsupervised Learning by Pieter Abbeel, DLR by Sergey Levine, DL Theory by Matus Telgarsky, LLM by Danqi Chen, Generative AI by Andrew Ng, Machine Learning and Big Data by Melissa Dell and Matthew Harding, Digital Economics and the Economics of AI by Martin Beraja, Chiara Farronato, Avi Goldfarb, and Catherine Tucker Detailed Schedule The following schedule is tentative and subject to changes. Session 1. Artificial Intelligence and Machine Learning in a Nutshell (Jan/09/2024) Keywords: Course Introduction, Machine Learning Basics, Bias-Variance Trade-off, Cross Validation, $k$-Nearest Neighbors, Decision Tree, Ensemble Methods Slides: Course Introduction, Machine Learning Basics CoLab Notebook Demos: k-Nearest Neighbors, Decision Tree Homework: Problem Set 1: Bias-Variance Trade-Off Online Python Tutorial: Python Tutorial CoLab, 7:00pm-9:00pm, Jan/12/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, https://hastie.su.domains/ElemStatLearn/. Probabilistic Machine Learning: An Introduction, 2022, by Kevin Murphy, https://probml.github.io/pml-book/book1.html. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Athey, Susan, and Guido W. Imbens. 2019. Machine learning methods that economists should know about. Annual Review of Economics 11: 685-725. Hofman, Jake M., et al. 2021. Integrating explanation and prediction in computational social science. Nature 595.7866: 181-188. Bastani, Hamsa, Dennis Zhang, and Heng Zhang. 2022. Applied machine learning in operations management. Innovative Technology at the Interface of Finance and Operations. Springer: 189-222. Kelly, Brian, and Dacheng Xiu. 2023. Financial machine learning, SSRN, https://ssrn.com/abstract=4501707. The Bitter Lesson, by Rich Sutton, which develops so far the most critical insight of AI: "The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most effective, and by a large margin." Session 2. Introduction to Deep Learning (Jan/16/2024) Keywords: Random Forests, eXtreme Gradient Boosting Trees, Deep Learning Basics, Neural Nets Models, Computational Issues of Deep Learning Slides: Machine Learning Basics, Deep Learning Basics CoLab Notebook Demos: Random Forest, Extreme Gradient Boosting Tree, Gradient Descent, Chain Rule Presentation: By Xinyu Li and Qingyu Xu. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing via machine learning. Review of Financial Studies 33: 2223-2273. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets Online PyTorch Tutorial: PyTorch Tutorial CoLab, 7:00pm-9:00pm, Jan/19/2024 (Friday), given by Qiansiqi Hu, 1155208353@link.cuhk.edu.hk. Zoom Link, Meeting ID: 923 4642 4433, Pass code: 178146 References: Deep Learning, 2016, by Ian Goodfellow, Yoshua Bengio and Aaron Courville, https://www.deeplearningbook.org/. Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Probabilistic Machine Learning: Advanced Topics, 2023, by Kevin Murphy, https://probml.github.io/pml-book/book2.html. Deep Learning with PyTorch, 2020, by Eli Stevens, Luca Antiga, and Thomas Viehmann. Gu, Shihao, Brian Kelly, and Dacheng Xiu. 2020. Empirical asset pricing with machine learning. Review of Financial Studies 33: 2223-2273. Session 3. DL Basics, Predictions in Business Research, and Traditonal NLP (Jan/23/2024) Keywords: Optimization and Computational Issues of Deep Learning, Prediction Problems in Business Research, Pre-processing and Word Representations in Traditional Natural Language Processing Slides: Deep Learning Basics, Prediction Problems in Business Research, NLP(I): Pre-processing and Word Representations.pdf) CoLab Notebook Demos: He Initialization, Dropout, Micrograd, NLP Pre-processing Presentation: By Letian Kong and Liheng Tan. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Link to the paper. Homework: Problem Set 2: Implementing Neural Nets, due at 12:30pm, Jan/30/2024 (Tuesday). References: Kleinberg, Jon, Jens Ludwig, Sendhil Mullainathan, and Ziad Obermeyer. 2015. Prediction policy problems. American Economic Review 105(5): 491-495. Mullainathan, Sendhil, and Jann Spiess. 2017. Machine learning: an applied econometric approach. Journal of Economic Perspectives 31(2): 87-106. Kleinberg, Jon, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018. Human decisions and machine predictions. Quarterly Journal of Economics 133(1): 237-293. Bajari, Patrick, Denis Nekipelov, Stephen P. Ryan, and Miaoyu Yang. 2015. Machine learning methods for demand estimation. American Economic Review, 105(5): 481-485. Farias, Vivek F., and Andrew A. Li. 2019. Learning preferences with side information. Management Science 65(7): 3131-3149. Cui, Ruomeng, Santiago Gallino, Antonio Moreno, and Dennis J. Zhang. 2018. The operational value of social media information. Production and Operations Management, 27(10): 1749-1769. Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Chapter 2, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Parameter Initialization and Batch Normalization (in Chinese) GPU Comparisons-vs-NVIDIA-H100-(PCIe)-vs-NVIDIA-RTX-6000-Ada/624vs632vs640) GitHub Repo for Micrograd, by Andrej Karpathy. Hand Written Notes Session 4. Traditonal NLP (Jan/30/2024) Keywords: Pre-processing and Word Representations in NLP, N-Gram, Naïve Bayes, Language Model Evaluation, Traditional NLP Applied to Business/Econ Research Slides: NLP(I): Pre-processing and Word Representations.pdf), NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf) CoLab Notebook Demos: NLP Pre-processing, N-Gram, Naïve Bayes Presentation: By Zhi Li and Boya Peng. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Link to the paper. Homework: Problem Set 3: Implementing Traditional NLP Techniques, due at 12:30pm, Feb/6/2024 (Tuesday). References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Hansen, Stephen, Michael McMahon, and Andrea Prat. 2018. Transparency and deliberation within the FOMC: A computational linguistics approach. Quarterly Journal of Economics, 133(2): 801-870. Chapters 2, 12, & 13, Introduction to Information Retrieval, 2008, Cambridge University Press, by Christopher D. Manning, Prabhakar Raghavan and Hinrich Schutze, https://nlp.stanford.edu/IR-book/information-retrieval-book.html. Chapter 2, 3 & 4, Speech and Language Processing (3rd ed. draft), 2023, by Dan Jurafsky and James H. Martin, https://web.stanford.edu/~jurafsky/slp3/. Natural Language Tool Kit (NLTK) Documentation Hand Written Notes Session 5. Deep-Learning-Based NLP: Word2Vec (Feb/06/2024) Keywords: Traditional NLP Applied to Business/Econ Research, Word2Vec: Continuous Bag of Words and Skip-Gram Slides: NLP(II): N-Gram, Naïve Bayes, and Language Model Evaluation.pdf), NLP(III): Word2Vec.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Xinyu Xu and Shu Zhang. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Link to the paper. Homework: No homework this week. Probably you should think about your final project when enjoying your Lunar New Year Holiday. References: Gentzkow, Matthew, Bryan Kelly, and Matt Taddy. 2019. Text as data. Journal of Economic Literature, 57(3): 535-574. Tetlock, Paul. 2007. Giving content to investor sentiment: The role of media in the stock market. Journal of Finance, 62(3): 1139-1168. Baker, Scott, Nicholas Bloom, and Steven Davis, 2016. Measuring economic policy uncertainty. Quarterly Journal of Economics, 131(4): 1593-1636. Gentzkow, Matthew, and Jesse Shapiro. 2010. What drives media slant? Evidence from US daily newspapers. Econometrica, 78(1): 35-71. Timoshenko, Artem, and John R. Hauser. 2019. Identifying customer needs from user-generated content. Marketing Science, 38(1): 1-20. Mikolov, Tomas, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Efficient estimation of word representations in vector space. ArXiv Preprint, arXiv:1301.3781. Mikolov, Tomas, Ilya Sutskever, Kai Chen, Greg Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in Neural Information Processing Systems (NeurIPS) 26. Parts I - II, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Word Embeddings Trained on Google News Corpus Hand Written Notes Session 6. Deep-Learning-Based NLP: RNN and Seq2Seq (Feb/20/2024) Keywords: Word2Vec: GloVe, Word Embedding and Language Model Evaluations, Word2Vec and RNN Applied to Business/Econ Research, RNN Slides: Guest Lecture Announcement, NLP(III): Word2Vec.pdf), NLP(IV): RNN & Seq2Seq.pdf) CoLab Notebook Demos: Word2Vec: CBOW, Word2Vec: Skip-Gram Presentation: By Qiyu Dai and Yifan Ren. Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. Link to the paper. Link to GitHub Repo. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Ash, Elliot, and Stephen Hansen. 2023. Text algorithms in economics. Annual Review of Economics, 15: 659-688. Associated GitHub with Code Demonstrations. Li, Kai, Feng Mai, Rui Shen, and Xinyan Yan. 2021. Measuring corporate culture using machine learning. Review of Financial Studies, 34(7): 3265-3315. Chen, Fanglin, Xiao Liu, Davide Proserpio, and Isamar Troncoso. 2022. Product2Vec: Leveraging representation learning to model consumer product choice in large assortments. Available at SSRN 3519358. Pennington, Jeffrey, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word representation. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) (pp. 1532-1543). Parts 2 and 5, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9 and 10, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations Hand Written Notes Session 7. Deep-Learning-Based NLP: Attention and Transformer (Feb/27/2024) Keywords: RNN and its Applications to Business/Econ Research, LSTM, Seq2Seq, Attention Mechanism Slides: Final Project, NLP(IV): RNN & Seq2Seq.pdf), NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: RNN & LSTM, Attention Mechanism Presentation: By Qinghe Gui and Chaoyuan Jiang. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Link to the paper. Homework: Problem Set 4 - Word2Vec & LSTM for Sentiment Analysis References: Qi, Meng, Yuanyuan Shi, Yongzhi Qi, Chenxin Ma, Rong Yuan, Di Wu, Zuo-Jun (Max) Shen. 2023. A Practical End-to-End Inventory Management Model with Deep Learning. Management Science, 69(2): 759-773. Sarzynska-Wawer, Justyna, Aleksander Wawer, Aleksandra Pawlak, Julia Szymanowska, Izabela Stefaniak, Michal Jarkiewicz, and Lukasz Okruszek. 2021. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Research, 304, 114135. Hansen, Stephen, Peter J. Lambert, Nicholas Bloom, Steven J. Davis, Raffaella Sadun, and Bledi Taska. 2023. Remote work across jobs, companies, and space (No. w31007). National Bureau of Economic Research. Sutskever, Ilya, Oriol Vinyals, and Quoc V. Le. 2014. Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Parts 5, 6, and 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapters 9, 10, and 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. RNN and LSTM Visualizations PyTorch's Tutorial of Seq2Seq for Machine Translation Illustrated Transformer Transformer from Scratch, with the Code on GitHub Hand Written Notes Session 7.5. Deep-Learning-Based NLP: Attention is All You Need (Mar/05/2024) Keywords: Bitter Lesson: Power of Computation in AI, Attention Mechanism, Transformer Slides: The Bitter Lesson, NLP(V): Attention & Transformer.pdf) CoLab Notebook Demos: Attention Mechanism, Transformer Homework: One-page Proposal for Your Final Project References: The Bitter Lesson, by Rich Sutton Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... and Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30. Part 8, Lecture Notes and Slides for CS224n: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto, https://web.stanford.edu/class/cs224n/. Chapter 11, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola, https://d2l.ai/. Illustrated Transformer Transformer from Scratch, with the Code on GitHub Andrej Karpathy's Lecture to Build Transformers Hand Written Notes Session 8. Deep-Learning-Based NLP: Pretraining (Mar/12/2024) Keywords: Computations in AI, BERT (Bidirectional Encoder Representations from Transformers), GPT (Generative Pretrained Transformers) Slides: Guest Lecture by Dr. Liubo Li on Deep Learning Computation, Pretraining.pdf) CoLab Notebook Demos: Crafting Intelligence: The Art of Deep Learning Modeling, BERT API @ Hugging Face Presentation: By Zhankun Chen and Yiyi Zhao. Noy, Shakked and Whitney Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381: 187-192. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday. References: Devlin, Jacob, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. ArXiv preprint arXiv:1810.04805. GitHub Repo Radford, Alec, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. 2018. Improving language understanding by generative pre-training, (GPT-1) PDF link, GitHub Repo Radford, Alec, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1(8), 9. (GPT-2) PDF Link, GitHub Repo Brown, Tom, et al. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33, 1877-1901. (GPT-3) GitHub Repo Huang, Allen H., Hui Wang, and Yi Yang. 2023. FinBERT: A large language model for extracting information from financial text. Contemporary Accounting Research, 40(2): 806-841. GitHub Repo Part 9, Lecture Notes and Slides for CS 224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N Part 2 & 4, Slides for COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G A Visual Guide to BERT, How GPT-3 Works Andrej Karpathy's Lecture to Build GPT-2 (124M) from Scratch Hand Written Notes Session 9. Deep-Learning-Based NLP: Large Language Models (Mar/19/2024) Keywords: Large Language Models, Generative AI, Emergent Ababilities, Instruction Fine-Tuning (IFT), Reinforcement Learning with Human Feedback (RLHF), In-Context Learning, Chain-of-Thought (CoT) Slides: What's Next, Pretraining.pdf), Large Language Models.pdf) CoLab Notebook Demos: BERT API @ Hugging Face Presentation: By Jia Liu. Liu, Liu, Dzyabura, Daria, Mizik, Natalie. 2020. Visual listening in: Extracting brand image portrayed on social media. Marketing Science, 39(4): 669-686. Link to the Paper Homework: Problem Set 5 - Sentiment Analysis with Hugging Face, due at 12:30pm, March 26, Tuesday (soft-deadline). References: Wei, Jason, et al. 2021. Finetuned language models are zero-shot learners. ArXiv preprint arXiv:2109.01652, link to the paper. Wei, Jason, et al. 2022. Emergent abilities of large language models. ArXiv preprint arXiv:2206.07682, link to the paper. Ouyang, Long, et al. 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730-27744. Wei, Jason, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837. Kaplan, Jared. 2020. Scaling laws for neural language models. ArXiv preprint arXiv:2001.08361, link to the paper. Hoffmann, Jordan, et al. 2022. Training compute-optimal large language models. ArXiv preprint arXiv:2203.15556, link to the paper. Shinn, Noah, et al. 2023. Reflexion: Language agents with verbal reinforcement learning. ArXiv preprint arXiv:2303.11366, link to the paper. Reisenbichler, Martin, Thomas Reutterer, David A. Schweidel, and Daniel Dan. 2022. Frontiers: Supporting content marketing with natural language generation. Marketing Science, 41(3): 441-452. Romera-Paredes, B., Barekatain, M., Novikov, A. et al. 2023. Mathematical discoveries from program search with large language models. Nature, link to the paper. Part 10, Lecture Notes and Slides for CS224N: Natural Language Processing with Deep Learning, by Christopher D. Manning, Diyi Yang, and Tatsunori Hashimoto. Link to CS 224N COS 597G: Understanding Large Language Models, by Danqi Chen. Link to COS 597G Andrej Karpathy's 1-hour Talk on LLM CS224n, Hugging Face Tutorial Session 10. Deep-Learning-Based CV: Image Classification (Mar/26/2024) Keywords: Large Language Models Applications, Convolution Neural Nets (CNN), LeNet, AlexNet, VGG, ResNet, ViT Slides: What's Next, Large Language Models.pdf), Image Classification.pdf) CoLab Notebook Demos: CNN, LeNet, & AlexNet, VGG, ResNet, ViT Presentation: By Yingxin Lin and Zeshen Ye. Netzer, Oded, Alain Lemaire, and Michal Herzenstein. 2019. When words sweat: Identifying signals for loan default in the text of loan applications. Journal of Marketing Research, 56(6): 960-980. Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25. He, Kaiming, Xiangyu Zhang, Shaoqing Ren and Jian Sun. 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770-778. Dosovitskiy, Alexey, et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. ArXiv preprint, arXiv:2010.11929, link to the paper, link to the GitHub repo. Jean, Neal, Marshall Burke, Michael Xie, Matthew W. Davis, David B. Lobell, and Stefand Ermon. 2016. Combining satellite imagery and machine learning to predict poverty. Science, 353(6301), 790-794. Zhang, Mengxia and Lan Luo. 2023. Can consumer-posted photos serve as a leading indicator of restaurant survival? Evidence from Yelp. Management Science 69(1): 25-50. Course Notes (Lectures 5 & 6) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapters 7 and 8, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Fine-Tune ViT for Image Classification with Hugging Face 🤗 Transformers Hugging Face 🤗 ViT CoLab Tutorial Session 11. Deep-Learning-Based CV (II): Object Detection & Video Analysis (Apr/2/2024) Keywords: Image Processing Applications, Localization, R-CNNs, YOLOs, Semantic Segmentation, 3D CNN, Video Analysis Applications Slides: What's Next, Image Classification.pdf), Object Detection and Video Analysis.pdf) CoLab Notebook Demos: Data Augmentation, Faster R-CNN & YOLO v5 Presentation: By Qinlu Hu and Yilin Shi. Yang, Jeremy, Juanjuan Zhang, and Yuhan Zhang. 2023. Engagement that sells: Influencer video advertising on TikTok. Available at SSRN Link to the Paper Homework: Problem Set 6 - AlexNet and ResNet, due at 12:30pm, April 9, Tuesday. References: Girshick, R., Donahue, J., Darrell, T. and Malik, J., 2014. Rich feature hierarchies for accurate object detection and semantic segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 580-587). Redmon, Joseph, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788). Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (pp. 1725-1732). Glaeser, Edward L., Scott D. Kominers, Michael Luca, and Nikhil Naik. 2018. Big data and big cities: The promises and limitations of improved measures of urban life. Economic Inquiry, 56(1): 114-137. Zhang, S., Xu, K. and Srinivasan, K., 2023. Frontiers: Unmasking Social Compliance Behavior During the Pandemic. Marketing Science, 42(3), pp.440-450. Course Notes (Lectures 10 & 11) for CS231n: Deep Learning for Computer Vision, by Fei-Fei Li, Ruohan Gao, & Yunzhu Li. Link to CS231n. Chapter 14, Dive into Deep Learning (2nd Edition), 2023, by Aston Zhang, Zack Lipton, Mu Li, and Alex J. Smola. Link to the book. Hand Written Notes Session 12. Unsupervised Learning: Clustering, Topic Modeling & VAE (Apr/9/2024) Keywords: K-Means, Gaussian Mixture Models, EM-Algorithm, Latent Dirichlet Allocation, Variational Auto-Encoder Slides: What's Next, Clustering, Topic Modeling & VAE.pdf) CoLab Notebook Demos: K-Means, LDA, VAE Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Blei, David M., Ng, Andrew Y., and Jordan, Michael I. 2003. Latent Dirichlet allocation. Journal of Machine Learning Research, 3(Jan): 993-1022. Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Bandiera, O., Prat, A., Hansen, S., & Sadun, R. 2020. CEO behavior and firm performance. Journal of Political Economy, 128(4), 1325-1369. Liu, Jia and Olivier Toubia. 2018. A semantic approach for estimating consumer content preferences from online search queries. Marketing Science, 37(6): 930-952. Mueller, Hannes, and Christopher Rauh. 2018. Reading between the lines: Prediction of political violence using newspaper text. American Political Science Review, 112(2): 358-375. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Chapters 8.5 and 14, The Elements of Statistical Learning (2nd Edition), 2009, by Trevor Hastie, Robert Tibshirani, Jerome Friedman, Link to Book. Course Notes (Lectures 1 & 4) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. Hand Written Notes Session 13. Unsupervised Learning: Diffusion Models (Apr/16/2024) Keywords: VAE, Denoised Diffusion Probabilistic Models, Latent Diffusion Models, CLIP, Imagen, Diffusion Transformers Slides: Clustering, Topic Modeling & VAE.pdf), Diffusion Models.pdf), Course Summary CoLab Notebook Demos: VAE, DDPM, DiT Homework: Problem Set 7 - Unsupervised Learning (EM & LDA), due at 12:30pm, April 23, Tuesday. References: Kingma, D.P. and Welling, M., 2013. Auto-encoding Variational Bayes. arXiv preprint arXiv:1312.6114. Kingma, D.P. and Welling, M., 2019. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4), pp.307-392. Ho, J., Jain, A. and Abbeel, P., 2020. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851. Chan, S.H., 2024. Tutorial on Diffusion Models for Imaging and Vision. arXiv preprint arXiv:2403.18103. Peebles, W. and Xie, S., 2023. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4195-4205. Link to GitHub Repo. Tian, Z., Dew, R. and Iyengar, R., 2023. Mega or Micro? Influencer Selection Using Follower Elasticity. Journal of Marketing Research. Ludwig, J. and Mullainathan, S., 2024. Machine learning as a tool for hypothesis generation. Quarterly Journal of Economics, 139(2), 751-827. Burnap, A., Hauser, J.R. and Timoshenko, A., 2023. Product aesthetic design: A machine learning augmentation. Marketing Science, 42(6), 1029-1056. Course Notes (Lecture 6) for CS294-158-SP24: Deep Unsupervised Learning, taught by Pieter Abbeel, Wilson Yan, Kevin Frans, Philipp Wu. Link to CS294-158-SP24. CVPR 2022 Tutorial: Denoising Diffusion-based Generative Modeling: Foundations and Applications, by Karsten Kreis, Ruiqi Gao, and Arash Vahdat Link to the Tutorial Lilian Weng (OpenAI)'s Blog on Diffusion Models Lilian Weng (OpenAI)'s Blog on Diffusion Models for Video Generation Hugging Face Diffusers 🤗 Library Hand Written Notes

Overmind
github
LLM Vibe Score0.469
Human Vibe Score0.20474237922306593
bencbartlettMar 23, 2025

Overmind

[](https://github.com/bencbartlett/Overmind/releases) [](https://github.com/bencbartlett/Overmind/blob/master/CHANGELOG.md) [](https://bencbartlett.github.io/overmind-docs/) [](https://github.com/bencbartlett/Overmind/wiki) [](https://screeps.slack.com/messages/overmind) [](https://github.com/bencbartlett/Overmind/issues/new) [](https://github.com/bencbartlett/Overmind/issues/new?template=feature_request.md) Current release: Overmind v0.5.2 - Evolution See the changelog for patch notes Documentation is available at the documentation site and the wiki Join the discussion in the #overmind Slack channel! Read blog posts about development Submit an issue here or request a feature here Find me in game here About Overmind What is Screeps? Screeps is an MMO strategy game for programmers. The core objective is to expand your colony, gathering resources and fighting other players along the way. To control your units, you code an AI in JavaScript; everything from moving, mining, building, fighting, and trading is entirely driven by your code. Because Screeps is an MMO, it takes place on a single server that runs 24/7, populated by every other player and their army of creeps. When you log off, your population continues buzzing away with whatever task you set them. Screeps pits your programming prowess head-to-head with other people to see who can think of the most efficient methods of completing tasks or imagine new ways to defeat enemies. What is Overmind? Overmind is my personal codebase that I run on the public server. The structure of the AI is themed loosely around the Zerg's swarm intelligence from Starcraft. Overlords orchestrate Creep actions within each Colony, and the colony Overseer places Directives to adapt to stimuli. Finally, the Assimilator allows all players running Overmind to act as a collective hivemind, sharing creeps and resources and responding jointly to a master ledger of all directives shared by all players. The AI is entirely automated, although it can also run in manual or semiautomatic mode. The latest release should work right out of the box; however, if you find something broken, please submit an issue and I'll try to fix it. Can I use Overmind as my bot? If you're new to Screeps, I would definitely recommend writing your own AI: most of the fun of the game is programming your own bot and watching your little ant farm run! However, I've tried to make the codebase readable and well-documented, so feel free to fork the project or use it as inspiration when writing your AI. If you still want to use Overmind on the public server, that's okay too - there are a number of people already doing this. But please realize that using a mature AI like this gives you a huge advantage over other new players, so don't go out of your way to ruin someone else's fun. In the future, I will be implementing methods for novice players to opt out of excessive aggression by Overmind bots (as long as they don't start a conflict and stay out of its way). Installation Out of the box If you just want to run Overmind without modification, you can copy the compiled main.js file attached to the latest release into your script. While Overmind is fully automated by default, it can be run with varying levels of autonomy; refer to the Overmind wiki for how to configure and operate the bot. Compiling from source To install the full codebase, download or clone the repository. (Please note that while the latest release of Overmind should always be stable, the latest commit may contain unstable features.) Navigate to the Overmind root directory and run . To compile and deploy the codebase, create a screeps.json file from the example file, then do one of the following actions: Compile and deploy to public server: npm run push-main Compile and deploy to private server: npm run push-pserver Compile without deploying: npm run compile Overmind uses rollup to bundle the compiled TypeScript into a single main.js file. The codebase includes functionality to compute checksums for internal validation - if you have a different version of rollup installed globally, different checksums may be computed and some functionality will be disabled. Please ensure the local installation of rollup found in node_modules is used. Setting up the Grafana dashboard Overmind includes a Grafana dashboard (shown below) which tracks detailed operating statistics. To set up the dashboard: Register for grafana service at screepspl.us Setup the ScreepsPlus hosted agent (simpler) or use the NodeJS agent on a free micro instance of Google Compute. Import the dashboard from Overmind.json and change $User to your username. Enjoy your pretty graphs! Design overview Check out the Overmind wiki for in-depth explanations of parts of the design of the AI. (Click the diagram below to see a higher-resolution version.)

bubbln_network-automation
github
LLM Vibe Score0.421
Human Vibe Score0.004537250556463098
olasupoMar 14, 2025

bubbln_network-automation

Bubbln: An AI-driven Network Automation In the world of network engineering, automation has completely transformed the way things work. But, before automation, setting up and managing networks was a tedious job filled with challenges. Engineers had to manually type out configurations, often doing the same tasks repeatedly on different devices. This led to mistakes and wasted time. Then came automation tools like Ansible, Chef, and Puppet, which changed everything. They made network management much easier and allowed for scalability. But there was still a problem: creating automation scripts required a lot of technical know-how and was prone to errors because it relied on human input. And that's why we built Bubbln. It's a game-changer in network engineering, integrating AI into Ansible to take automation to the next level. With Bubbln, we can automatically generate and execute playbooks with incredible accuracy, thereby improving automation efficiency and increasing network engineer’s productivity. It was developed using Python programming language and acts as a bridge between ChatGPT and network systems, making interactions seamless and deployments effortless. Current Capabilities AI-Driven Playbook Generation for OSPF and EIGRP based networks: Bubbln has been rigorously tested to leverage ChatGPT for generation of playbooks for networks based on OSPF and EIGRP networks, with a very high accuracy rate. Auto-creation of Inventory files: Users do not need to prepare the hosts file. Bubbln will auto-generate this file from input provided by the user. Customizable Configurations: Users can input specific router protocols (OSPF or EIGRP), interface configurations, and other network details to tailor the generated playbooks. Documentation: Bubbln automatically creates a report that contains the network configurations, prompts, and generated playbooks for easy reference in future. No expertise required: By auto-generation of the playbooks and inventory file, Bubbln has been able to eliminate a major hurdle to network automation – need for users to learn the automation tools e.g Ansible, Chef. Improved Efficiency: With AI automation, Bubbln speeds up the deployment of network configurations, reducing the time required for manual playbook creation, thereby increasing the productivity of network engineers. Getting Started There are two main approaches to installing Bubbln on your local machine. Docker Container Bubbln has been packaged using docker containers for easy distribution and usage. The following steps can be followed to deploy the Bubbln container on your local machine. Ensure docker is installed on your local machine by entering the below command. This command works for windows and linux OS: The version of docker would be displayed if it is installed. Otherwise, please follow the link below to install docker on your machine: Windows: Docker Desktop for Windows Ubuntu: Docker Engine for Ubuntu CentOS: Docker Engine for CentOS Debian: Docker Engine for Debian Fedora: Docker Engine for Fedora Download the docker image: Create a directory for the project and download Bubbln image using the below command: Run the docker container using the below command: Install nano Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. To do this enter the below command to edit the file: Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key: Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln. Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln by entering the below command: Github Repository Clone You can clone Bubbln’s GitHub repository by following the below steps: Prerequisites Bubbln works well with Python 3.10. You need to ensure python3.10 is installed on your local machine. This can be confirmed by entering the below command: If it is not Installed, then the below command can be utilized to install python 3.10: Build and Prepare the Project Clone the Bubbln repository from GitHub: To clone the repository, first verify you have git installed on your machine by issuing the following commands: If git is installed, the version number would be displayed, otherwise, you can issue the following commands to have git installed on your machine: Navigate or create a directory for the project on your machine and issue the following commands to clone the Bubbln git repository: Create a Virtual Environment for the application Firstly, confirm virtualenv is installed on your machine by inputting the following command: If the output shows something similar to the below, then go to the next step to install virtualenv ` WARNING: Package(s) not found: env, virtual ` Issue the below command to install virtualenv: Create a virtual environment for the project: Activate the virtual environment: Install the dependencies You can then run the below command to install the necessary packages for the app. Update the sshipaddresses.txt file: Update the ssh_addresses.txt file with the SSH IP addresses of the routers you want to configure. Bubbln will utilize this information along with the login credentials (inputted at runtime) to automatically generate a hosts.yml file required by ansible for network configuration. Obtain an OpenAPI API Key: You may follow this guide to sign up and obtain an API key OpenAI Key: OpenAI Key Utilizing a Virtualization machine of choice, setup a network with the following basic configurations: Enable SSH on each of the routers. Configure IP addresses and enable only interfaces required for connectivity by Bubbln Configure static routes to enable Bubbln reach the routers on the network. Ensure all the routers can be reached by ping and SSH from your host machine. Initialize Bubbln While ensuring that python virtual environment is activated as stated in step 5, run the below command to initialize Bubbln How Bubbln Works Bubbln serves as an intermediary between ChatGPT and a network infrastructure, providing logic, control functions, and facilitating network automation. Its operation can be summarized as follows: !image Figure 1Bubbln architecture and interaction with a network of four routers. Initialization: When Bubbln is initialized, it checks the “userconfig.pkl” file to see if Bubbln has ever been initiated. This is indicated by the presence of a welcome message status in the file. If it exists, Bubbln jumps straight to request the user to input the OpenAI key. Otherwise, it displays a welcome message, and updates the userconfig.pkl file accordingly. Upon successful input of the API key, the user is prompted for the SSH credentials of the routers. These parameters are then encrypted and saved in the user_config.pkl file. The SSH credential is later decrypted and parsed as input to dynamically generate a hosts.yml file at runtime. Responsible Code Section: bubbln.py: welcomemessagefeature() !image Figure 2 Bubbln's welcome message. Parameter Input & Validation: In the parameter input stage, Bubbln first checks for the existence of a file called “router_configuration.pkl”. If it exists, the user is prompted to decide whether to load an existing configuration or input a new set of configurations. If the file is empty or non-existent, then users are prompted to input the configuration parameters for each router on the network. These parameters serve as variables that are combined with hardcoded instructions written in natural language to form the prompt sent to ChatGPT. Key parameters include: Router Configurations: OSPF Area OSPF Process ID Number of networks to advertise (OSPF/EIGRP) AS Number (EIGRP) Interface names IP Addresses (in CIDR format) This module also ensures that parameters are keyed in using the correct data type and format e.g. IP addresses are expected in CIDR format and OSPF Area should be of type integer. Upon completion of parameter input, all parameters are saved into a file called “router_configuration.pkl” upon validation of accuracy by the user. Responsible Code Section: parameter_input.py !image Figure 3 Bubbln receiving Network Parameters. Before generating the prompt, a summary of the inputted parameters is displayed for user validation. This step ensures accuracy and minimizes errors. Users are given the option to make corrections if any discrepancies are found. Responsible Code Section: parameterinput.py: validateinputs() !image Figure 4 Bubbln Awaiting Validation of Inputted Network Parameters. Auto-Generation of Prompt: After validation of inputted parameters, Bubbln composes the prompt by combining the inputted parameters with a set of well-engineered hardcoded instructions written in natural language. Responsible Code Section: prompt_generator.py ChatGPT Prompting: The auto-composed prompt is then sent to ChatGPT utilizing gpt-4 chatCompletions model with a temperature parameter of 0.2 and maximum tokens of 1500. The following functions were designed into this process stage Responsible Code Section: chatGPT_prompting.py !image Figure 5 ChatGPT prompting in progress Playbook Generation & Extraction: After ChatGPT processes the prompt from Bubbln, it provides a response which usually contains the generated playbook and explanatory notes. Bubbln then extracts the playbook from the explanatory notes by searching for “---” which usually connotes the start of playbooks and saves each generated playbook uniquely using the nomenclature RouteriPlaybook.yml. Responsible Code Section: playbook_extractor.py !image Figure 6 ChatGPT-generated playbook. Playbook Execution: Bubbln loads the saved “RouteriPlaybook.yml” playbook and dynamically generates the hosts.yml file and parses them to the python library ansiblerunner for further execution on the configured network. Bubbln generates the hosts.yml file at run time by using the pre-inputted SSH credentials in userconfig.pkl file - and decrypts them, as well as IP addresses from the sshipaddresses.txt file, as inputs Responsible Code Section: playbook_execution.py !image Figure 7 Playbook execution in progress Sample result of Executed Playbook Upon successful execution of all playbooks, a query of the routing table on router 4 indicates that router 4 could reach all the prefixes on the network. !image Figure 8 Output of 'sh ip route' executed on R1 File Management and Handling Throughout the execution process, Bubbln manages the creation, saving, and loading of various files to streamline the network automation process. user_config.pkl: This dictionary file dynamically created at run time is used to store encrypted API keys, SSH credentials and initial welcome message information. router_configuration.pkl: It is auto created by Bubbln and used to store network configuration parameters for easy loading during subsequent sessions. hosts.yml: This is a runtime autogenerated file that contains inventory of the network devices. It is auto deleted after the program runs. networkconfigurationreport.pdf: This auto-generated report by Bubbln is a documentation of all the routers configured their parameters, generated playbooks, and prompt for each execution of the Bubbln application. It is created after a successful execution of playbooks and network testing and is meant for auditing and documentation purposes. RouteriPlaybook.yml: After extraction of generated playbooks from ChatGPT’s raw response, Bubbln automatically saves a copy of the generated playbook using unique names for each playbook. !image Figure 9 File structure after successful deployment of a four-router network Providing Feedback We are glad to hear your thoughts and suggestions. Kindly do this through the discussion section of our GitHub - https://github.com/olasupo/bubbln_network-automation/discussions/1#discussion-6487475 We can also be reached on: Olasupo Okunaiya – olasupo.o@gmail.com

ai50
github
LLM Vibe Score0.457
Human Vibe Score0.07953823122984799
nahueespinosaJan 17, 2025

ai50

My work on CS50’s Introduction to AI with Python https://cs50.harvard.edu/ai/ This course explores the concepts and algorithms at the foundation of modern artificial intelligence, diving into the ideas that give rise to technologies like game-playing engines, handwriting recognition, and machine translation. Through hands-on projects, students gain exposure to the theory behind graph search algorithms, classification, optimization, reinforcement learning, and other topics in artificial intelligence and machine learning as they incorporate them into their own Python programs. By course’s end, students emerge with experience in libraries for machine learning as well as knowledge of artificial intelligence principles that enable them to design intelligent systems of their own. Certificate: https://courses.edx.org/certificates/2ec5ff3f06b24bb595c21e3821591538 Notes I've taken some notes on key concepts and algorithms throughout the lectures for future reference. Lecture 0: Search Concepts Agent: entity that perceives its environment and acts upon that environment. State: a configuration of the agent and its environment. Actions: choices that can be made in a state. Transition model: a description of what state results from performing any applicable action in any state. Path cost: numerical cost associated with a given path. Evaluation function: function that estimates the expected utility of the game from a given state. Algorithms DFS (depth first search): search algorithm that always expands the deepest node in the frontier. BFS (breath first search): search algorithm that always expands the shallowest node in the frontier. Greedy best-first search: search algorithm that expands the node that is closest to the goal, as estimated by an heuristic function h(n). A\* search: search algorithm that expands node with lowest value of the "cost to reach node" plus the "estimated goal cost". Minimax: adversarial search algorithm. Projects Degrees Tic-Tac-Toe Lecture 1: Knowledge Concepts Sentence: an assertion about the world in a knowledge representation language. Knowledge base: a set of sentences known by a knowledge-based agent. Entailment: a entails b if in every model in which sentence a is true, sentence b is also true. Inference: the process of deriving new sentences from old ones. Conjunctive normal form: logical sentence that is a conjunction of clauses. First order logic: Propositional logic. Second order logic: Proposition logic with universal and existential quantification. Algorithms Model checking: enumerate all possible models and see if a proposition is true in every one of them. Conversion to CNF and Inference by resolution Projects Knights Minesweeper Lecture 2: Uncertainty Concepts Unconditional probability: degree of belief in a proposition in the absence of any other evidence. Conditional probability: degree of belief in a proposition given some evidence that has already been revealed. Random variable: a variable in probability theory with a domain of possible values it can take on. Independence: the knowledge that one event occurs does not affect the probability of the other event. Bayes' Rule: P(a) P(b|a) = P(b) P(a|b) Bayesian network: data structure that represents the dependencies among random variables. Markov assumption: the assumption that the current state depends on only a finite fixed number of previous states. Markov chain: a sequence of random variables where the distribution of each variable follows the Markov assumption. Hidden Markov Model: a Markov model for a system with hidden states that generate some observed event. Algorithms Inference by enumeration Sampling Likelihood weighting Projects Heredity PageRank Lecture 3: Optimization Concepts Optimization: choosing the best option from a set of options. Algorithms Local Search Hill climbing steepest-ascent: choose the highest-valued neighbor. stochastic: choose randomly from higher-valued neighbors. first-choice: choose the first higher-valued neighbor. random-restart: conduct hill climbing multiple times. local beam search: chooses the k highest-valued neighbors. Simulated annealing: early on, more likely to accept worse-valued neighbors than the current state. Linear programming Simplex Interior-Point Constraint satisfaction problems Arc consistency: to make X arc-consistent with respect to Y, removing elements from X's domain until every choice for X has a possible choice for Y Backtracking search Projects Crossword Lecture 4: Learning Concepts Supervised learning: given a data set of input-output pairs, learn a function to map inputs to outputs. Classification: supervised learning task of learning a function mapping an input point to a discrete category. Regression: supervised learning task of learning a function mapping and input point to a continuous value. Loss function: function that express how poorly our hypothesis performs (L1, L2). Overfitting: when a model fits too closely to a particular data set and therefore may fail to generalize to future data. Regularization: penalizing hypotheses that are more complex to favor simpler, more general hypotheses. Holdout cross-validation: splitting data into a training set and a test set, such that learning happens on the training set and is evaluated on the test set. k-fold cross-validation: splitting data into k sets, and experimenting k times, using each set as a test set once, and using remaining data as training set. Reinforcement learning: given a set of rewards or punishments, learn what actions to take in the future. Unsupervised learning: given input data without any additional feedback, learn patterns. Clustering: organizing a set of objects into groups in such a way that similar objects tend to be in the same group. Algorithms k-nearest-neighbor classification: given an input, chooses the most common class out of the k nearest data points to that input. Support Vector Machines (SVM) Markov decision process: model for decision-making, representing states, actions and their rewards. Q-learning: method for learning a function Q(s, a), estimate of the value of performing action a in state s. Greedy decision-making epsilon-greedy k-means clustering: clustering data based on repeatedly assigning points to clusters and updating those clusters' centers. Projects Shopping Nim Lecture 5: Neural Networks Concepts Artificial neural network: mathematical model for learning inspired by biological neural networks. Multilayer neural network: artificial neural network with an input layer, an output layer, and at least one hidden layer. Deep neural network: neural network with multiple hidden layer. Dropout: temporarily removing units - selected at random - from a neural network to prevent over-reliance on certain units. Image convolution: applying a filter that adds each pixel value of an image to its neighbors, weighted according to a kernel matrix. Pooling: reducing the size of an input by sampling from regions in the input. Convolutional neural network: neural networks that use convolution, usually for analyzing images. Recurrent neural network: neural network that generates output that feeds back into its own inputs. Algorithms Gradient descent: algorithm for minimizing loss when training neural network. Backpropagation: algorithm for training neural networks with hidden layers. Projects Traffic Lecture 6: Language Concepts Natural language processing n-gram: a continuous sequence of n items inside of a text. Tokenization: the task of splitting a sequence of characters into pieces (tokens). Text Categorization Bag-of-words model: represent text as an unordered collection of words. Information retrieval: the task of finding relevant documents in response to a user query. Topic modeling: models for discovering the topics for a set of documents. Term frequency: number of times a term appears in a document. Function words: words that have little meaning on their own, but are used to grammatically connect other words. Content words: words that carry meaning independently. Inverse document frequency: measure of how common or rare a word is across documents. Information extraction: the task of extracting knowledge from documents. WordNet: a lexical database of semantic relations between words. Word representation: looking for a way to represent the meaning of a word for further processing. one-hot: representation of meaning as a vector with a single 1, and with other values as 0. distribution: representation of meaning distributed across multiple values. Algorithms Markov model applied to language: generating the next word based on the previous words and a probability. Naive Bayes: based on the Bayes' Rule to calculate probability of a text being in a certain category, given it contains specific words. Assuming every word is independent of each other. Additive smoothing: adding a value a to each value in our distribution to smooth the data. Laplace smoothing: adding 1 to each value in our distribution (pretending we've seen each value one more time than we actually have). tf-idf: ranking of what words are important in a document by multiplying term frequency (TF) by inverse document frequency (IDF). Automated template generation: giving AI some terms and let it look into a corpus for patterns where those terms show up together. Then it can use those templates to extract new knowledge from the corpus. word2vec: model for generating word vectors. skip-gram architecture: neural network architecture for predicting context words given a target word. Projects Parser Questions